National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory

Slides:



Advertisements
Similar presentations
GMD German National Research Center for Information Technology Darmstadt University of Technology Perspectives and Priorities for Digital Libraries Research.
Advertisements

5th FP7 Networked Media Concertation Meeting, Brussels 3-4 February 2010 A unIfied framework for multimodal content SEARCH Short Project Overview.
Visual Model-based Software Development EUD-Net Workshop, Pisa, Italy September 23 rd, 2002 University of Paderborn Gregor Engels, Stefan Sauer University.
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Prof. dr. ir. J.H.C. Reiber A Multi-Modal Visualisation Environment for Interactive Analysis of Medical Data MULTI-VIS Lorentz Center, VIEW:
Distributed search for complex heterogeneous media Werner Bailer, José-Manuel López-Cobo, Guillermo Álvaro, Georg Thallinger Search Computing Workshop.
The 20th International Conference on Software Engineering and Knowledge Engineering (SEKE2008) Department of Electrical and Computer Engineering
Manuela Veloso, Anthony Stentz, Alexander Rudnicky Brett Browning, M. Bernardine Dias Faculty Thomas Harris, Brenna Argall, Gil Jones Satanjeev Banerjee.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
DL:Lesson 11 Multimedia Search Luca Dini
Visual Event Detection & Recognition Filiz Bunyak Ersoy, Ph.D. student Smart Engineering Systems Lab.
1 Texmex – November 15 th, 2005 Strategy for the future Global goal “Understand” (= structure…) TV and other MM documents Prepare these documents for applications.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
1 CS 430: Information Discovery Lecture 22 Non-Textual Materials 2.
ICANN workshop, September 14, Athens, Greece Intelligent Affective Interaction technologies and applications.
Introduction to HCC and HCM. Human Centered Computing Philosophical-humanistic position regarding the ethics and aesthetics of a workplace Any system.
Multimodal Interaction. Modalities vs Media Modalities are ways of encoding information e.g. graphics Media are instantiations of modalities e.g. a particular.
Queen Mary, University of London
ICCS-NTUA Contributions to E-teams of MUSCLE WP6 and WP10 Prof. Petros Maragos National Technical University of Athens School of Electrical and Computer.
CS335 Principles of Multimedia Systems Multimedia and Human Computer Interfaces Hao Jiang Computer Science Department Boston College Nov. 20, 2007.
Emotional Intelligence and Agents – Survey and Possible Applications Mirjana Ivanovic, Milos Radovanovic, Zoran Budimac, Dejan Mitrovic, Vladimir Kurbalija,
Smart Learning Services Based on Smart Cloud Computing
HAND GESTURE BASED HUMAN COMPUTER INTERACTION. Hand Gesture Based Applications –Computer Interface A 2D/3D input device (Hand Tracking) Translation of.
Electronic Visualization Laboratory University of Illinois at Chicago Interaction between Real and Virtual Humans: Playing Checkers R. Torre, S. Balcisoy.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
1 Darmstadt, October 02, 2007 Amalia Ortiz Asociación VICOMTech Mikeletegi Pasealekua Donostia - San Sebastián (Gipuzkoa)
Some Thoughts on HPC in Natural Language Engineering Steven Bird University of Melbourne & University of Pennsylvania.
Conversational Applications Workshop Introduction Jim Larson.
NATIONAL TECHNICAL UNIVERSITY OF ATHENS Image, Video And Multimedia Systems Laboratory
Jeanne Corcoran, OTD, OTR/L October 6 th, The mission of Envision Center for Data Perceptualization is to serve, support, and collaborate with faculty,
NATIONAL TECHNICAL UNIVERSITY OF ATHENS Image, Video And Multimedia Systems Laboratory Background
Interactive Discovery and Semantic Labeling of Patterns in Spatial Data Thomas Funkhouser, Adam Finkelstein, David Blei, and Christiane Fellbaum Princeton.
Subtask 1.8 WWW Networked Knowledge Bases August 19, 2003 AcademicsAir force Arvind BansalScott Pollock Cheng Chang Lu (away)Hyatt Rick ParentMark (SAIC)
IST Programme - Key Action III Semantic Web Technologies in IST Key Action III (Multimedia Content and Tools) Hans-Georg Stork CEC DG INFSO/D5
1 CS 430: Information Discovery Lecture 22 Non-Textual Materials: Informedia.
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
I Robot.
ENTERFACE 08 Project 1 “MultiParty Communication with a Tour Guide ECA” Mid-term presentation August 19th, 2008.
Introduction to Dialogue Systems. User Input System Output ?
Towards a Method For Evaluating Naturalness in Conversational Dialog Systems Victor Hung, Miguel Elvir, Avelino Gonzalez & Ronald DeMara Intelligent Systems.
Intelligent Robot Architecture (1-3)  Background of research  Research objectives  By recognizing and analyzing user’s utterances and actions, an intelligent.
Multimodality, universals, natural interaction… and some other stories… Kostas Karpouzis & Stefanos Kollias ICCS/NTUA HUMAINE WP4.
1 Viewing Vision-Language Integration as a Double-Grounding case Katerina Pastra Department of Computer Science, Natural Language Processing Group, University.
M4 / September Integrating multimodal descriptions to index large video collections M4 meeting – Munich Nicolas Moënne-Loccoz, Bruno Janvier,
Semantic Extraction and Semantics-Based Annotation and Retrieval for Video Databases Authors: Yan Liu & Fei Li Department of Computer Science Columbia.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
CONTENT FOCUS FOCUS INTRODUCTION INTRODUCTION COMPONENTS COMPONENTS TYPES OF GESTURES TYPES OF GESTURES ADVANTAGES ADVANTAGES CHALLENGES CHALLENGES REFERENCE.
Electronic visualization laboratory, university of illinois at chicago Towards Lifelike Interfaces That Learn Jason Leigh, Andrew Johnson, Luc Renambot,
1 CS 430 / INFO 430 Information Retrieval Lecture 17 Metadata 4.
Behrooz ChitsazLorrie Apple Johnson Microsoft ResearchU.S. Department of Energy.
Multimedia Analytics Jianping Fan Department of Computer Science University of North Carolina at Charlotte.
Quiz Week 8 Topical. Topical Quiz (Section 2) What is the difference between Computer Vision and Computer Graphics What is the difference between Computer.
Technical Drawings Understanding for the Blind Dr. George Ioannidis
SEESCOASEESCOA SEESCOA Meeting Activities of LUC 9 May 2003.
Slide no 1 Cognitive Systems in FP6 scope and focus Colette Maloney DG Information Society.
Immersive Virtual Characters for Educating Medical Communication Skills J. Hernendez, A. Stevens, D. S. Lind Department of Surgery (College of Medicine)
What is Multimedia Anyway? David Millard and Paul Lewis.
3D Motion Classification Partial Image Retrieval and Download Multimedia Project Multimedia and Network Lab, Department of Computer Science.
Digital Video Library - Jacky Ma.
SAMT 2006.
Next-Generation Search Engines -Perspective and challenges
Pilar Orero, Spain Yoshikazu SEKI, Japan 2018
Towards lifelike Computer Interfaces that learn
Pervasive Computing Happening?
Multimodal Human-Computer Interaction New Interaction Techniques 22. 1
Interactive media.
Motivation It can effectively mine multi-modal knowledge with structured textural and visual relationships from web automatically. We propose BC-DNN method.
Presentation transcript:

National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory A ROADMAP IN MULTIMEDIA KNOWLEDGE TECHNOLOGIES Prof. STEFANOS KOLLIAS Image, Video and Multimedia Systems Laboratory Computer Science Division National Technical University of Athens May 15, 2002

National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory vs I. MULTIMEDIA-BASED KNOWLEDGE MANAGEMENT vs KNOWLEDGE-BASED MULTIMEDIA MANAGEMENT MPEG-7 generated progress in multimedia retrieval and management; focus on semantics is under development in MPEG-21. Knowledge Technologies should specify A/V descriptors, create hooks in Description Languages and update shared knowledge w.r.t. real information (semi or fully automatically). Automatic semantic analysis of A/V content is difficult; semantics are diversely understood and context dependent. The challenge: interweave multimedia analysis and A/V knowledge representation (integrate developments in MPEG-7/MPEG-21 and KT). Semantic Video Segmentation and Retrieval : to be combined with Evaluation in the Context of the video/video-shot description & available visual object representations (involving Tracking /Matching /Recognition).

National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory II. A SPECIFIC FRAMEWORK: ADAPTIVE KNOWLEDGE BASED HUMAN INTERACTION Speech, Face, Gesture and Emotion Recognition are the means to understand users intentions and goals. Virtual humans/agents in virtual/mixed reality environments are the means to personify in a natural way the machine part of the interaction. Integration of multimedia knowledge is required in both analysis of the multimodal systems input and synthesis of its output: inclusion of AV cues / descriptions in knowledge use and update of knowledge for and according to analysis of multimedia inputs, users profiles and context of interaction