SWE 423: Multimedia Systems Chapter 4: Graphics and Images (4)

Slides:



Advertisements
Similar presentations
Image Analysis Phases Image pre-processing –Noise suppression, linear and non-linear filters, deconvolution, etc. Image segmentation –Detection of objects.
Advertisements

Clustering & image segmentation Goal::Identify groups of pixels that go together Segmentation.
Chapter 8 Content-Based Image Retrieval. Query By Keyword: Some textual attributes (keywords) should be maintained for each image. The image can be indexed.
1 Video Processing Lecture on the image part (8+9) Automatic Perception Volker Krüger Aalborg Media Lab Aalborg University Copenhagen
Image Segmentation Image segmentation (segmentace obrazu) –division or separation of the image into segments (connected regions) of similar properties.
電腦視覺 Computer and Robot Vision I Chapter2: Binary Machine Vision: Thresholding and Segmentation Instructor: Shih-Shinh Huang 1.
Computer Vision Lecture 16: Region Representation
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Chapter 11 Beyond Bag of Words. Question Answering n Providing answers instead of ranked lists of documents n Older QA systems generated answers n Current.
Region labelling Giving a region a name. Image Processing and Computer Vision: 62 Introduction Region detection isolated regions Region description properties.
EE 7730 Image Segmentation.
Thresholding Otsu’s Thresholding Method Threshold Detection Methods Optimal Thresholding Multi-Spectral Thresholding 6.2. Edge-based.
SWE 423: Multimedia Systems
1Ellen L. Walker Segmentation Separating “content” from background Separating image into parts corresponding to “real” objects Complete segmentation Each.
Text Detection in Video Min Cai Background  Video OCR: Text detection, extraction and recognition  Detection Target: Artificial text  Text.
SWE 423: Multimedia Systems
Supervised by Prof. LYU, Rung Tsong Michael Department of Computer Science & Engineering The Chinese University of Hong Kong Prepared by: Chan Pik Wah,
Processing Digital Images. Filtering Analysis –Recognition Transmission.
Segmentation Divide the image into segments. Each segment:
Face Detection: a Survey Speaker: Mine-Quan Jing National Chiao Tung University.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman ICCV 2003 Presented by: Indriyati Atmosukarto.
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
Pores and Ridges: High- Resolution Fingerprint Matching Using Level 3 Features Anil K. Jain Yi Chen Meltem Demirkus.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
Highlights Lecture on the image part (10) Automatic Perception 16
Visual Information Retrieval Chapter 1 Introduction Alberto Del Bimbo Dipartimento di Sistemi e Informatica Universita di Firenze Firenze, Italy.
Tracking Video Objects in Cluttered Background
MULTIPLE MOVING OBJECTS TRACKING FOR VIDEO SURVEILLANCE SYSTEMS.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Content-Based Retrieval in Image Databases. References Ze-Nian Li & Mark S. Drew, Fundamentals of Multimedia, ISBN , Prentice-Hall, 2004.
Thresholding Thresholding is usually the first step in any segmentation approach We have talked about simple single value thresholding already Single value.
Database System Concepts and Architecture
CS654: Digital Image Analysis Lecture 3: Data Structure for Image Analysis.
Edge Linking & Boundary Detection
Digital Image Processing In The Name Of God Digital Image Processing Lecture8: Image Segmentation M. Ghelich Oghli By: M. Ghelich Oghli
Topic 10 - Image Analysis DIGITAL IMAGE PROCESSING Course 3624 Department of Physics and Astronomy Professor Bob Warwick.
S EGMENTATION FOR H ANDWRITTEN D OCUMENTS Omar Alaql Fab. 20, 2014.
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
Digital Image Processing CCS331 Relationships of Pixel 1.
Generalized Hough Transform
Chapter 10 Image Segmentation.
Digital Image Processing (DIP) Lecture # 5 Dr. Abdul Basit Siddiqui Assistant Professor-FURC 1FURC-BCSE7.
Pixel Connectivity Pixel connectivity is a central concept of both edge- and region- based approaches to segmentation The notation of pixel connectivity.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Levels of Image Data Representation 4.2. Traditional Image Data Structures 4.3. Hierarchical Data Structures Chapter 4 – Data structures for.
Metadata Common Vocabulary a journey from a glossary to an ontology of statistical metadata, and back Sérgio Bacelar
Autonomous Robots Vision © Manfred Huber 2014.
1 KMeD: A Knowledge-Based Multimedia Medical Database System Wesley W. Chu Computer Science Department University of California, Los Angeles
Content-Based Image Retrieval QBIC Homepage The State Hermitage Museum db2www/qbicSearch.mac/qbic?selLang=English.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Digital Image Processing
Medical Image Analysis Image Segmentation Figures come from the textbook: Medical Image Analysis, by Atam P. Dhawan, IEEE Press, 2003.
October 1, 2013Computer Vision Lecture 9: From Edges to Contours 1 Canny Edge Detector However, usually there will still be noise in the array E[i, j],
MULTIMEDIA DATA MODELS AND AUTHORING
1 A Methodology for automatic retrieval of similarly shaped machinable components Mark Ascher - Dept of ECE.
Sheng-Fang Huang Chapter 11 part I.  After the image is segmented into regions, how to represent and describe these regions? ◦ In terms of its external.
Managing Data Resources File Organization and databases for business information systems.
Machine Vision ENT 273 Lecture 4 Hema C.R.
Visual Information Retrieval
DIGITAL SIGNAL PROCESSING
Mean Shift Segmentation
Computer Vision Lecture 12: Image Segmentation II
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Presentation transcript:

SWE 423: Multimedia Systems Chapter 4: Graphics and Images (4)

Image Segmentation Assigning a unique number to “object” pixels based on different intensities or colors in the foreground and the background regions of an image –Can be used in the object recognition process, but it is not object recognition on its own Segmentation Methods –Pixel oriented methods –Edge oriented methods –Region oriented methods –....

Pixel-Oriented Segmentation Gray-values of pixels are studied in isolation Looks at the gray-level histogram of an image and finds one or more thresholds in the histogram –Ideally, the histogram has a region without pixels, which is set as the threshold, and hence the image is divided into a foreground and a background based on that (Bimodal Distribution) Major drawback of this approach is that object and background histograms overlap. –Bimodal distribution rarely occurs in nature.

Edge-Oriented Segmentation Segmentation is carried out as follows –Edges of an image are extracted (using Canny operators, e.g.) –Edges are connected to form closed contours around the objects. Hough Transform –Usually very expensive –Works well with regular curves (application in manufactured parts) –May work in presence of noise

Region-Oriented Segmentation A major disadvantage of the previous approaches is the lack of “spatial” relationship considerations of pixels. –Neighboring pixels normally have similar properties The segmentation (region-growing) is carried out as follows –Start with a “seed” pixel. –Pixel’s neighbors are included if they have some similarity to the seed pixel, otherwise they are not. Homogeneity condition Uses an eight-neighborhood (8-nbd) model

Region-Oriented Segmentation Homogeneity criterion: Gray-level mean value of a region is usually used With standard deviation Drawbacks: Computationally expensive.

Water Inflow Segmentation Fill a gray-level image gradually with water. –Gray-levels of pixels are taken as height. –The higher the water rises, the more pixels are flooded Hence, you have lands and waters Lands correspond to “objects”

Object Recognition Layer Features are analyzed to recognize objects and faces in an image database. –Features are matched with object models stored in a knowledge base. –Each template is inspected to find the closest match. –Exact matches are usually impossible and generally computationally expensive. –Occlusion of objects and the existence of spurious features in the image can further diminish the success of matching strategies.

Template Matching Techniques Fixed Template Matching –Useful if object shapes do not change with respect to the viewing angle of the camera. Deformable Template Matching –More suitable for cases where objects in the database may vary due to rigid and non-rigid deformations.

Fixed Template Matching Image Subtraction: –Difference in intensity levels between the image and the template is used in object recognition. –Performs well in restricted environments where imaging conditions (such as image intensity) between the image and the template are the same. Matching by correlation: – utilizes the position of the normalized cross-correlation peak between a template and image. –Generally immune to noise and illumination effects in the image. –Suffers from high computational complexity caused by summations over the entire template.

Deformable Template Matching Template is represented as a bitmap describing the characteristic contour/edges of an object shape. An objective function with transformation parameters which alter the shape of the template is formulated reflecting the cost of such transformations. The objective function is minimized by iteratively updating the transformations parameters to best match the object. Applications include: handwritten character recognition and motion detection of objects in video frames.

Prototype System: KMeD Medical objects belonging only to patients in a small age group are identified automatically in KMeD. –Such objects have high contrast with respect to their background and have relatively simple shapes, large sizes, and little or no overlap with other objects. KMeD resorts to a human-assisted object recognition process otherwise.

Demo gedatabase/demo/cars/ (check car214) gedatabase/demo/cars/

Spatial Modeling and Knowledge Representation Layer (1) Maintain the domain knowledge for representing spatial semantics associated with image databases. At this level, queries are generally descriptive in nature, and focus mostly on semantics and concepts present in image databases. Semantics at this level are based on ``spatial events'' describing the relative locations of multiple objects. –An example involving such semantics is a range query which involves spatial concepts such as close by, in the vicinity, larger than. (e.g. retrieve all images that contain a large tumor in the brain).

Spatial Modeling and Knowledge Representation Layer (2) Identify spatial relationships among objects, once they are recognized and marked by the lower layer using bounding boxes or volumes. Several techniques have been proposed to formally represent spatial knowledge at this layer. –Semantic networks –Mathematical logic –Constraints –Inclusion hierarchies –Frames.

Semantic Networks First introduced to represent the meanings of English sentences in terms of words and relationships between them. Semantic networks are graphs of nodes representing concepts that are linked together by arcs representing relationships between these concepts. Efficiency in semantic networks is gained by representing each concept or object once and using pointers for cross references rather than naming an object explicitly every time it is involved in a relation. Example: Type Abstraction Hierarchies (KMeD)

Brain Lesions Representation

TAH Example

Constraints-based Methodology Domain knowledge is represented using a set of constraints in conjunction with formal expressions such as predicate calculus or graphs. A constraint is a relationship between two or more objects that needs to be satisfied.

Example: PICTION system Its architecture consists of a natural language processing module (NLP), an image understanding module (IU), and a control module. A set of constraints is derived by the NLP module from the picture captions. These constraints (called Visual Semantics by the author) are used with the faces recognized in the picture by the IU module to identify the spatial relationships among people. The control module maintains the constraints generated by the NLP module and acts as a knowledge-base for the IU module to perform face recognition functions.

Mathematical Logic Iconic Indexing by 2D strings: Uses projections of salient objects in a coordinated system. These projections are expressed in the form of 2D strings to form a partial ordering of object projections in 2D. For query processing, 2D subsequence matching is performed to allow similarity-based retrieval. Binary Spatial Relations: Uses Allen's 13 temporal relations to represent spatial relationships.

Inclusion Hierarchies The approach is object-oriented and uses concept classes and attributes to represent domain knowledge. These concepts may represent image features, high-level semantics, semantic operators and conditions.

Frames A frame usually consists of a name and a list of attribute-value pairs. A frame can be associated with a class of objects or with a class of concepts. Frame abstractions allow encapsulation of file names, features, and relevant attributes of image objects.