Ontology-based Automatic Video Annotation Technique in Smart TV Environment Jin-Woo Jeong, Hyun-Ki Hong, and Dong-Ho Lee IEEE Transactions on Consumer.

Slides:



Advertisements
Similar presentations
Image Retrieval With Relevant Feedback Hayati Cam & Ozge Cavus IMAGE RETRIEVAL WITH RELEVANCE FEEDBACK Hayati CAM Ozge CAVUS.
Advertisements

Context-based object-class recognition and retrieval by generalized correlograms by J. Amores, N. Sebe and P. Radeva Discussion led by Qi An Duke University.
DONG XU, MEMBER, IEEE, AND SHIH-FU CHANG, FELLOW, IEEE Video Event Recognition Using Kernel Methods with Multilevel Temporal Alignment.
Automatic Video Shot Detection from MPEG Bit Stream Jianping Fan Department of Computer Science University of North Carolina at Charlotte Charlotte, NC.
Automated Shot Boundary Detection in VIRS DJ Park Computer Science Department The University of Iowa.
Image Information Retrieval Shaw-Ming Yang IST 497E 12/05/02.
1 A scheme for racquet sports video analysis with the combination of audio-visual information Visual Communication and Image Processing 2005 Liyuan Xing,
A presentation by Modupe Omueti For CMPT 820:Multimedia Systems
Broadcast News Parsing Using Visual Cues: A Robust Face Detection Approach Yannis Avrithis, Nicolas Tsapatsoulis and Stefanos Kollias Image, Video & Multimedia.
Content-based Video Indexing, Classification & Retrieval Presented by HOI, Chu Hong Nov. 27, 2002.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
ICME 2008 Huiying Liu, Shuqiang Jiang, Qingming Huang, Changsheng Xu.
Chapter 11 Beyond Bag of Words. Question Answering n Providing answers instead of ranked lists of documents n Older QA systems generated answers n Current.
Video retrieval using inference network A.Graves, M. Lalmas In Sig IR 02.
ACM Multimedia th Annual Conference, October , 2004
Expectation Maximization Method Effective Image Retrieval Based on Hidden Concept Discovery in Image Database By Sanket Korgaonkar Masters Computer Science.
SWE 423: Multimedia Systems
Multimedia Search and Retrieval Presented by: Reza Aghaee For Multimedia Course(CMPT820) Simon Fraser University March.2005 Shih-Fu Chang, Qian Huang,
T.Sharon 1 Internet Resources Discovery (IRD) Video IR.
Region-Level Motion- Based Background Modeling and Subtraction Using MRFs Shih-Shinh Huang Li-Chen Fu Pei-Yung Hsiao 2007 IEEE.
LYU 0102 : XML for Interoperable Digital Video Library Recent years, rapid increase in the usage of multimedia information, Recent years, rapid increase.
Visual Information Retrieval Chapter 1 Introduction Alberto Del Bimbo Dipartimento di Sistemi e Informatica Universita di Firenze Firenze, Italy.
Tracking Video Objects in Cluttered Background
A fuzzy video content representation for video summarization and content-based retrieval Anastasios D. Doulamis, Nikolaos D. Doulamis, Stefanos D. Kollias.
DVMM Lab, Columbia UniversityVideo Event Recognition Video Event Recognition: Multilevel Pyramid Matching Dong Xu and Shih-Fu Chang Digital Video and Multimedia.
SIEVE—Search Images Effectively through Visual Elimination Ying Liu, Dengsheng Zhang and Guojun Lu Gippsland School of Info Tech,
Information Retrieval in Practice
1 Image Video & Multimedia Systems Laboratory Multimedia Knowledge Laboratory Informatics and Telematics Institute Exploitation of knowledge in video recordings.
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
Bridge Semantic Gap: A Large Scale Concept Ontology for Multimedia (LSCOM) Guo-Jun Qi Beckman Institute University of Illinois at Urbana-Champaign.
The MPEG-7 Color Descriptors
Exploiting Ontologies for Automatic Image Annotation M. Srikanth, J. Varner, M. Bowden, D. Moldovan Language Computer Corporation
Università degli Studi di Modena and Reggio Emilia Dipartimento di Ingegneria dell’Informazione Prototypes selection with.
Information Systems & Semantic Web University of Koblenz ▪ Landau, Germany Semantic Web - Multimedia Annotation – Steffen Staab
南台科技大學 資訊工程系 Automatic Website Summarization by Image Content: A Case Study with Logo and Trademark Images Evdoxios Baratis, Euripides G.M. Petrakis, Member,
MPEG-7 Interoperability Use Case. Motivation MPEG-7: set of standardized tools for describing multimedia content at different abstraction levels Implemented.
Beauty is Here! Evaluating Aesthetics in Videos Using Multimodal Features and Free Training Data Yanran Wang, Qi Dai, Rui Feng, Yu-Gang Jiang School of.
An MPEG-7 Based Content- aware Album System for Consumer Photographs 2003/12/18 Chen-Hsiu Huang, Chih-Hao Shen, Chun-Hsiang Huang and Ja-Ling Wu Communication.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
IEEE Int'l Symposium on Signal Processing and its Applications 1 An Unsupervised Learning Approach to Content-Based Image Retrieval Yixin Chen & James.
IBM QBIC: Query by Image and Video Content Jianping Fan Department of Computer Science University of North Carolina at Charlotte Charlotte, NC 28223
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Tomohiko TAKAHASHL Masaru SUGANO, Keiichiro HOASHL and Sei NAITO International Conference on Multimedia and Expo 2011 Arbitrary Product Detection from.
A survey of different shape analysis techniques 1 A Survey of Different Shape Analysis Techniques -- Huang Nan.
Case Study 1 Semantic Analysis of Soccer Video Using Dynamic Bayesian Network C.-L Huang, et al. IEEE Transactions on Multimedia, vol. 8, no. 4, 2006 Fuzzy.
1 Applications of video-content analysis and retrieval IEEE Multimedia Magazine 2002 JUL-SEP Reporter: 林浩棟.
Exploiting Ontologies for Automatic Image Annotation Munirathnam Srikanth, Joshua Varner, Mitchell Bowden, Dan Moldovan Language Computer Corporation SIGIR.
MMDB-9 J. Teuhola Standardization: MPEG-7 “Multimedia Content Description Interface” Standard for describing multimedia content (metadata).
2004 謝俊瑋 NTU, CSIE, CMLab 1 A Rule-Based Video Annotation System Andres Dorado, Janko Calic, and Ebroul Izquierdo, Senior Member, IEEE.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Miguel Tavares Coimbra
An MPEG-7 Based Semantic Album for Home Entertainment Presented by Chen-hsiu Huang 2003/08/12 Presented by Chen-hsiu Huang 2003/08/12.
DANIELA KOLAROVA INSTITUTE OF INFORMATION TECHNOLOGIES, BAS Multimedia Semantics and the Semantic Web.
Wonjun Kim and Changick Kim, Member, IEEE
On Using SIFT Descriptors for Image Parameter Evaluation Authors: Patrick M. McInerney 1, Juan M. Banda 1, and Rafal A. Angryk 2 1 Montana State University,
MMM2005The Chinese University of Hong Kong MMM2005 The Chinese University of Hong Kong 1 Video Summarization Using Mutual Reinforcement Principle and Shot.
Relevance Feedback in Image Retrieval System: A Survey Tao Huang Lin Luo Chengcui Zhang.
Automatic Caption Localization in Compressed Video By Yu Zhong, Hongjiang Zhang, and Anil K. Jain, Fellow, IEEE IEEE Transactions on Pattern Analysis and.
SUMMERY 1. VOLUMETRIC FEATURES FOR EVENT DETECTION IN VIDEO correlate spatio-temporal shapes to video clips that have been automatically segmented we.
MPEG 7 &MPEG 21.
MULTIMEDIA SYSTEMS CBIR & CBVR. Schedule Image Annotation (CBIR) Image Annotation (CBIR) Video Annotation (CBVR) Video Annotation (CBVR) Few Project Ideas.
An Ontology framework for Knowledge-Assisted Semantic Video Analysis and Annotation Centre for Research and Technology Hellas/ Informatics and Telematics.
Working meeting of WP4 Task WP4.1
Visual Information Retrieval
Automatic Video Shot Detection from MPEG Bit Stream
Multimedia Content-Based Retrieval
V. Mezaris, I. Kompatsiaris, N. V. Boulgouris, and M. G. Strintzis
Multimedia Information Retrieval
A Novel Smoke Detection Method Using Support Vector Machine
Presentation transcript:

Ontology-based Automatic Video Annotation Technique in Smart TV Environment Jin-Woo Jeong, Hyun-Ki Hong, and Dong-Ho Lee IEEE Transactions on Consumer Electronics, Vol. 57, No. 4, November 2011 Presented by: You Tithrottanak

Contents 1.Introduction 2.Overview of the proposed approach 3.Semantic web technologies for video annotation 4.High-level concept extraction 5.Experiments and analysis

Introduction The survey of Korean domestic market for smart TV rate of smart TV in Korea will be 8.9 million households which amounts to 52.6% of all households in Korea by 2022 The reason for this positive prospect comes from the following attractive features of smart TV: ▫Open contents ▫Entertainment & Communication ▫N-screen service ▫Smart multi-tasking ▫Smart advertisement ▫Smart home server

Overview of the proposed approach

Semantic web technologies for video annotation VideoAnnotation Ontology

Semantic web technologies for video annotation There are four important classes: ▫ShotAnno class  Describe about information of Objects that exist in the representative frame (key-frame) of each video shot ▫GroupAnno class  Describe about information of video groups, dominant concepts in the group and group meta-data ▫SceneAnno class  Describe about information of Concepts of each video scene (animal-tracking, interview with tamer) ▫Meta class  It presents the meta data of a video sequence (title, createdBy, modifiedBy, length, type, year )

Semantic web technologies for video annotation Domain Ontology ▫For providing a sharable and reusable vocabulary ▫LSCOM (Large Scale Concept Ontology for Multimedia) ontology  provides a set of standardized concepts for video annotation  Full LSCOM : 3,000 high-level concepts  They used Light-weight version of the LSCOM : 400 high- level concepts  This ontology are used for both a group-level and scene- level annotation ▫Object ontology  knowledge-base for a specific domain  This type of ontology can be made manually or derived from an existing ontology such as WordNet

High-level concept extraction

Key Frame Detection and Object Segmentation ▫The detection of a video shot and its corresponding key frame are performed based on the visual similarity ▫They use color structure descriptor (CSD) for representing color feature of each video frame ▫To calculate the visual similarity between video frames, They use Euclidean distance measure

High-level concept extraction Semi-concept Mapping ▫the low-level visual features of an object are extracted by MPEG-7 visual descriptors ▫Then mapped to their corresponding semi-concept values 1.For the color feature, they exploit MPEG-7 color structure descriptor (CSD)

High-level concept extraction ▫A set of semi-concept values for color feature is defined as {Red-Orange, Yellow-YellowGreen, …, Gray, White} 2.The texture feature of an object is extracted by MPEG-7 Edge Histogram Descriptor (EHD)

High-level concept extraction ▫5 types:  Vertical  Horizontal  45-degree diagonal  135-degree diagonal  Non-directional edges

High-level concept extraction ▫A shape feature of a region in the image are provide by MPEG-7 (region-based shape descriptor and contour shape descriptor (contour- SD)). ▫Region-based shape descriptor is able to describe the complex objects (company logo and trademark) ▫contour-SD efficiently describes the objects with a single contour(animal objects such as tiger, horse)

High-level concept extraction Semantic Inference Rules for High-level Concept Extraction ▫The inference procedure is performed as follows 1.Assume that the extracted semi-concept values are:  Color = {“Red_Orange”, “Yellow_Green”, “Black”}  Texture = {“Non-directional”, “Horizontal”} 2.Apply rule Cheetah to determine the high-level concept of the object in the video shot

High-level concept extraction

SVM Learning for High-level Concept Extraction ▫SVM classifier ▫The feature vector of the training set for a particular concept C i is represented as T i = {L i, DC 0-3, DT, DS }  L i : a label  DC 0-3 : four color semi-concepts,  DT : a texture semi-concept  DS : a shape semi-concept

High-level concept extraction

Experiments and analysis They focus on the accuracy of annotation for the video shots The evaluation is performed on the parts of National Geographic videos containing various concepts such as animals, landscape scene. 3 kinds of evaluations for the proposed approaches to investigate: ▫Effectiveness of Semantic Rule-based Approach ▫Effectiveness of Semi-concept based SVM Classifier ▫Relative Effectiveness of the Proposed Approaches

Experiments and analysis Effectiveness of Semantic Rule-based Approach

Experiments and analysis Effectiveness of Semi-concept based SVM Classifier

Experiments and analysis Relative Effectiveness of the Proposed Approaches

Thanks