Download presentation
Presentation is loading. Please wait.
Published byLionel Riley Modified over 8 years ago
1
Ontology-based Automatic Video Annotation Technique in Smart TV Environment Jin-Woo Jeong, Hyun-Ki Hong, and Dong-Ho Lee IEEE Transactions on Consumer Electronics, Vol. 57, No. 4, November 2011 Presented by: You Tithrottanak
2
Contents 1.Introduction 2.Overview of the proposed approach 3.Semantic web technologies for video annotation 4.High-level concept extraction 5.Experiments and analysis
3
Introduction The survey of Korean domestic market for smart TV rate of smart TV in Korea will be 8.9 million households which amounts to 52.6% of all households in Korea by 2022 The reason for this positive prospect comes from the following attractive features of smart TV: ▫Open contents ▫Entertainment & Communication ▫N-screen service ▫Smart multi-tasking ▫Smart advertisement ▫Smart home server
4
Overview of the proposed approach
5
Semantic web technologies for video annotation VideoAnnotation Ontology
6
Semantic web technologies for video annotation There are four important classes: ▫ShotAnno class Describe about information of Objects that exist in the representative frame (key-frame) of each video shot ▫GroupAnno class Describe about information of video groups, dominant concepts in the group and group meta-data ▫SceneAnno class Describe about information of Concepts of each video scene (animal-tracking, interview with tamer) ▫Meta class It presents the meta data of a video sequence (title, createdBy, modifiedBy, length, type, year )
7
Semantic web technologies for video annotation Domain Ontology ▫For providing a sharable and reusable vocabulary ▫LSCOM (Large Scale Concept Ontology for Multimedia) ontology provides a set of standardized concepts for video annotation Full LSCOM : 3,000 high-level concepts They used Light-weight version of the LSCOM : 400 high- level concepts This ontology are used for both a group-level and scene- level annotation ▫Object ontology knowledge-base for a specific domain This type of ontology can be made manually or derived from an existing ontology such as WordNet
8
High-level concept extraction
9
Key Frame Detection and Object Segmentation ▫The detection of a video shot and its corresponding key frame are performed based on the visual similarity ▫They use color structure descriptor (CSD) for representing color feature of each video frame ▫To calculate the visual similarity between video frames, They use Euclidean distance measure
10
High-level concept extraction Semi-concept Mapping ▫the low-level visual features of an object are extracted by MPEG-7 visual descriptors ▫Then mapped to their corresponding semi-concept values 1.For the color feature, they exploit MPEG-7 color structure descriptor (CSD)
11
High-level concept extraction ▫A set of semi-concept values for color feature is defined as {Red-Orange, Yellow-YellowGreen, …, Gray, White} 2.The texture feature of an object is extracted by MPEG-7 Edge Histogram Descriptor (EHD)
12
High-level concept extraction ▫5 types: Vertical Horizontal 45-degree diagonal 135-degree diagonal Non-directional edges
13
High-level concept extraction ▫A shape feature of a region in the image are provide by MPEG-7 (region-based shape descriptor and contour shape descriptor (contour- SD)). ▫Region-based shape descriptor is able to describe the complex objects (company logo and trademark) ▫contour-SD efficiently describes the objects with a single contour(animal objects such as tiger, horse)
14
High-level concept extraction Semantic Inference Rules for High-level Concept Extraction ▫The inference procedure is performed as follows 1.Assume that the extracted semi-concept values are: Color = {“Red_Orange”, “Yellow_Green”, “Black”} Texture = {“Non-directional”, “Horizontal”} 2.Apply rule Cheetah to determine the high-level concept of the object in the video shot
15
High-level concept extraction
16
SVM Learning for High-level Concept Extraction ▫SVM classifier ▫The feature vector of the training set for a particular concept C i is represented as T i = {L i, DC 0-3, DT, DS } L i : a label DC 0-3 : four color semi-concepts, DT : a texture semi-concept DS : a shape semi-concept
17
High-level concept extraction
18
Experiments and analysis They focus on the accuracy of annotation for the video shots The evaluation is performed on the parts of National Geographic videos containing various concepts such as animals, landscape scene. 3 kinds of evaluations for the proposed approaches to investigate: ▫Effectiveness of Semantic Rule-based Approach ▫Effectiveness of Semi-concept based SVM Classifier ▫Relative Effectiveness of the Proposed Approaches
19
Experiments and analysis Effectiveness of Semantic Rule-based Approach
20
Experiments and analysis Effectiveness of Semi-concept based SVM Classifier
21
Experiments and analysis Relative Effectiveness of the Proposed Approaches
22
Thanks
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.