의미정보 해석 - 지식기반 시스템 응용 - 2006.11.21 최보윤 소프트컴퓨팅 연구실 연세대학교.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Kien A. Hua Division of Computer Science University of Central Florida.
Hand Gesture for Taking Self Portrait Shaowei Chu and Jiro Tanaka University of Tsukuba Japan 12th July 15 minutes talk.
Martin Wagner and Gudrun Klinker Augmented Reality Group Institut für Informatik Technische Universität München December 19, 2003.
Face Recognition Face Recognition Using Eigenfaces K.RAMNATH BITS - PILANI.
Page 1 SIXTH SENSE TECHNOLOGY Presented by: KIRTI AGGARWAL 2K7-MRCE-CS-035.
Patch to the Future: Unsupervised Visual Prediction
On the Relationship between Visual Attributes and Convolutional Networks Paper ID - 52.
CS335 Principles of Multimedia Systems Content Based Media Retrieval Hao Jiang Computer Science Department Boston College Dec. 4, 2007.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
University of Athens, Greece Pervasive Computing Research Group Predicting the Location of Mobile Users: A Machine Learning Approach 1 University of Athens,
Location Systems for Ubiquitous Computing Jeffrey Hightower and Gaetano Borriello.
FACE RECOGNITION, EXPERIMENTS WITH RANDOM PROJECTION
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Distributed and Efficient Classifiers for Wireless Audio-Sensor Networks Baljeet Malhotra Ioanis Nikolaidis Mario A. Nascimento University of Alberta Canada.
Oral Defense by Sunny Tang 15 Aug 2003
Overview of Search Engines
Video Surveillance Capturing, Management and Analysis of Security Videos. -Abhinav Goel -Varun Varshney.
Face Recognition Using EigenFaces Presentation by: Zia Ahmed Shaikh (P/IT/2K15/07) Authors: Matthew A. Turk and Alex P. Pentland Vision and Modeling Group,
Smart Environments for Occupancy Sensing and Services Paper by Pirttikangas, Tobe, and Thepvilojanapong Presented by Alan Kelly December 7, 2011.
Knowledge Systems Lab JN 8/24/2015 A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
Research Area B Leif Kobbelt. Communication System Interface Research Area B 2.
Data Mining Techniques
Ambulation : a tool for monitoring mobility over time using mobile phones Computational Science and Engineering, CSE '09. International Conference.
Life Logging Melekam Tsegaye Shaun Bangay Alfredo Terzoli Research area: Wearable, Pervasive and Ubiquitous Computing
Shu Chen,Yan Huang Department of Computer Science & Engineering University of North Texas Denton, TX 76207, USA Recognizing Human Activities from Multi-Modal.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Introduction The large amount of traffic nowadays in Internet comes from social video streams. Internet Service Providers can significantly enhance local.
Autonomous Learning of Object Models on Mobile Robots Xiang Li Ph.D. student supervised by Dr. Mohan Sridharan Stochastic Estimation and Autonomous Robotics.
What’s Making That Sound ?
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Mining Rare and Frequent Events in Multi-camera Surveillance.
Multimedia Specification Design and Production 2013 / Semester 2 / week 8 Lecturer: Dr. Nikos Gazepidis
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Presentation by: K.G.P.Srikanth. CONTENTS  Introduction  Components  Working  Applications.
Image Collection Backend for Cameraphones. Introduction Project Goals Design an integrated system to upload image from a mobile phone to a remote server.
© 2009 Mobile VCE 6 October 2011 Core 4 Secure Applications ‘Instant Knowledge’ Programme Achievements Nigel Jefferies Huawei.
Submitted by:- Vinay kr. Gupta Computer Sci. & Engg. 4 th year.
Recommendation system MOPSI project KAROL WAGA
User Modeling, Recommender Systems & Personalization Pattie Maes MAS 961- week 6.
Images and Sounds: Audio and Video for Education Joe Wise and Michael Hamilton.
Image Classification 영상분류
Students: Anurag Anjaria, Charles Hansen, Jin Bai, Mai Kanchanabal Professors: Dr. Edward J. Delp, Dr. Yung-Hsiang Lu CAM 2 Continuous Analysis of Many.
Research Projects 6v81 Multimedia Database Yohan Jin, T.A.
AUTOMATIC TARGET RECOGNITION OF CIVILIAN TARGETS September 28 th, 2004 Bala Lakshminarayanan.
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
Final Year Project Lego Robot Guided by Wi-Fi (QYA2)
Stable Multi-Target Tracking in Real-Time Surveillance Video
Data and Applications Security Developments and Directions Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #15 Secure Multimedia Data.
Providing User Context for Mobile and Social Networking Applications A. C. Santos et al., Pervasive and Mobile Computing, vol. 6, no. 1, pp , 2010.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
TEMPLATE DESIGN © E-Eye : A Multi Media Based Unauthorized Object Identification and Tracking System Tolgahan Cakaloglu.
Author: Tatsuya Yamazaki National institute of Information and Communications Technology Presenter: Samanvoy Panati.
Face Image-Based Gender Recognition Using Complex-Valued Neural Network Instructor :Dr. Dong-Chul Kim Indrani Gorripati.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Computer Vision Group Department of Computer Science University of Illinois at Urbana-Champaign.
Attila Kiss, Tamás Németh, Szabolcs Sergyán, Zoltán Vámossy, László Csink Budapest Tech Recognition of a Moving Object in a Stereo Environment Using a.
C ONTEXT AWARE SMART PHONE YOGITHA N. & PREETHI G.D. 6 th SEM, B.E.(C.S.E) SIDDAGANGA INSTITUTE OF TECHNOLOGY TUMKUR
TRECVID IES Lab. Intelligent E-commerce Systems Lab. 1 Presented by: Thay Setha 05-Jul-2012.
REAL-TIME DETECTOR FOR UNUSUAL BEHAVIOR
Digital Video Library - Jacky Ma.
Deeply learned face representations are sparse, selective, and robust
Visual Tracking on an Autonomous Self-contained Humanoid Robot
3rd Studierstube Workshop TU Wien
AHED Automatic Human Emotion Detection
Presentation transcript:

의미정보 해석 - 지식기반 시스템 응용 최보윤 소프트컴퓨팅 연구실 연세대학교

Collaborative capturing and interpretation of interactions Y. Sumi, I. Sadanori, T. Matsuguchi, S. Fels, and K. Mase Pervasive 2004 Workshop on Memory and Sharing of Experiences, pp. 1-7, 2004.

Overview Introduction Capturing interactions by multiple sensors Related works Implementation Interpreting interactions Video summary Corpus viewer: Tool for analyzing interaction patterns conclusions

Introduction Interaction corpus –Action highlights Generate diary –Social protocols of human interactions Sensors –Video cameras, microphone and physiological sensors ID tags –LED tag: infrared LED –IR tracker: Infrared signal tracking device Position and identity

Capturing interactions by multiple sensors Recording natural interactions –Multiple presenters and visitors in an exhibition room Sensors & Humanoid robots –Wearable sensors, stationary sensors Monitoring humans Video camera, microphone, IR tracker –Recording robots’ behavior logs and the reactions of the humans which connect the robots Central data server –Getting the data from the sensors and humanoid robots

Related works Smart environment –Supporting humans in a room –The Smart rooms, Intelligent room, AwareHome, Kidsroom and EasyLiving –Recognition of human behavior and understanding of the human’s intention Wearable systems –Collecting personal daily activities –Intelligent recording system Video summary systems –The physical quantity of video data captured by fixed cameras

Exhibition room Implementation

IR tracker & LED tag Implementation

Interpreting interactions Define interaction primitives –Events –Significant intervals or moments of activites IR tracker and LED tag minInterval and maxInterval –minInterval: 5 sec –maxInterval Ubiquitous sensors: 10 sec Wearable sensors: 20 sec

Video summary Assumptions –User, Booth Co-occurences Video summarization

Corpus viewer: Tool for analyzing interaction patterns

conclusions Method to build an interaction corpus using multiple sensors Segment and interpret interactions from huge data Provide a video summary Help social scientists

Using context and similarity for face and location identification M. Davis, M. Smith, F. Stentiford, A. Bambidele, J. Canny, N. Good, S. King and R. Janakiraman Proceedings of the IS&T/SPIE 18th Annual Symposium on Electronic Imaging Science and Technology Internet Imaging VII, 2006.

Overview Introduction System Overview Content Analysis Experimental Data Experimental Design Evaluation Discussion and Results Conclusions & Future Work

Introduction New way for the unsolved image content recognition –Mobile media capture, context-sensing, programmable computation and networking in the form of the nearly ubiquitous cameraphone Cameraphone –Platform for multimedia computing –Combination with the analysis of automatically gathered contextual metadata and media content analysis Contextual metadata –Temporal –Spatial –Social –Face recognition and place recognition Precision of face recognition –PCA 40%, SFA 50% Precision of location recognition –Color histogram 30%, CVA 50%, contextual metadata and CVA 67%

System Overview MMM2 –Gathering data and metadata –Server application: store photo metadata and user profile information –Client application: run the client handset –MMM2 Context Logger University of Helsinki Location information, Bluetooth radio Detect new photos, display interface or web browser, upload MMM2 server –MMM2 website Select a region of a photo and associate a person’s name with this region Creation of Ground-Truth Dataset

Location Recognition Similarity measures –Pattern recognition problem Cognitive Visual Attention –Comparison of two image –Drawn the parts in common –No memory of data Training and Classification –A nearest neighbor classifier –Location classification Visual Sub-cluster Extraction –Many different photos at each location –Location class by several sub-clusters –Adding more exemplars Not guarantee improvements Color Histogram Techniques –Pixel color distributions –Simplest visual sub cluster example corresponding to an exemplar Content Analysis

Face Recognition & GPS PCA –Eigenface principle –Short training time –Best accuracy LDA+PCA –LDA: Multiple images training Bayesian MAP & ML –Maximum a posteriori (MAP), maximum likelihood (ML) –Difference or similarity between two photos SFA (Sparse Factor Analysis) – –Y: a vector of (partially) observed values, X: latent vector representing user preference, m: “model” predicting user behavior, N: noise function GPS Clustering –Suitable format –K-means and farthest first cluster Content Analysis

Experimental Data Face Recognition on Cameraphone Data –NIST FERET dataset Mugshot –Full frontal view –Head-and-shoulders –27,000 cameraphone potos 66user, 10 months Multiple people Real world Photographic Location Data –1209 images Nokia 7610 cameraphones 12 location, 30 cell identities Berkeley Campuss

Experimental Design Training gallery –Hand-labeled with the names –Min of distances between all images in the photo and training gallery image k SFA model –Training Contextual metadata and the face recognizer outputs Contextual metadata only –Evaluation Precision-recall plots for each of the computer vision algorithms –Time Training time: 2 minutes Training for the Bayesian classifiers: 7 hours PCA and LDA classifiers: less than 10 minutes Face recognition for 4 algorithms: less than 1 minute

Evaluation Location by Contextual Metadata –Distribution of metadata: 579 items, 12 location Location by Metadata and Vision

Face Identification Experimental Results Discussion and Results

Location Identification Experimental Results Histogram classifier, the CVA classifier and metadata classifier –Bad performance Metadata –Limit the errors with Cell ID –Specific place at certain times of the day and days of the week Discussion and Results Error Rate Increase Per Feature Removed

Conclusions & Future Work New approach to the automatic identification of human faces and location if mobile images Combination of attributes –Contextual metadata –Image processing Torso-matching Context-aware location recognition research