TeleMorph & TeleTuras: Bandwidth determined Mobile MultiModal Presentation Student: Anthony J. Solon Supervisors: Prof. Paul Mc Kevitt Kevin Curran School.

Slides:



Advertisements
Similar presentations
CHART or PICTURE INTEGRATING SEMANTIC WEB TO IMPROVE ONLINE Marta Gatius Meritxell González TALP Research Center (UPC) They are friendly and easy to use.
Advertisements

Information Society Technologies Third Call for Proposals Norbert Brinkhoff-Button DG Information Society European Commission Key action III: Multmedia.
TeleMorph & TeleTuras: Bandwidth determined Mobile MultiModal Presentation Student: Anthony J. Solon Supervisors: Prof. Paul Mc Kevitt Kevin Curran School.
Irek Defée Signal Processing for Multimodal Web Irek Defée Department of Signal Processing Tampere University of Technology W3C Web Technology Day.
Empirical and Data-Driven Models of Multimodality Advanced Methods for Multimodal Communication Computational Models of Multimodality Adequate.
MediaHub: An Intelligent Multimedia Distributed Hub Student: Glenn Campbell Supervisors: Dr. Tom Lunney Prof. Paul Mc Kevitt School of Computing and Intelligent.
PGNET, Liverpool JMU, June 2005 MediaHub: An Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney, Paul Mc Kevitt School of Computing.
XISL language XISL= eXtensible Interaction Sheet Language or XISL=eXtensible Interaction Scenario Language.
Languages & The Media, 5 Nov 2004, Berlin 1 New Markets, New Trends The technology side Stelios Piperidis
Network Management Overview IACT 918 July 2004 Gene Awyzio SITACS University of Wollongong.
MULTIMEDIA Development Team.
Presentation Outline  Project Aims  Introduction of Digital Video Library  Introduction of Our Work  Considerations and Approach  Design and Implementation.
Introduction to HCC and HCM. Human Centered Computing Philosophical-humanistic position regarding the ethics and aesthetics of a workplace Any system.
Presentation Outline  Project Aims  Introduction of Digital Video Library  Introduction of Our Work  Considerations and Approach  Design and Implementation.
Professor Michael J. Losacco CIS 1150 – Introduction to Computer Information Systems Communications and Networks Chapter 8.
Multimodal Interaction. Modalities vs Media Modalities are ways of encoding information e.g. graphics Media are instantiations of modalities e.g. a particular.
Architecture & Data Management of XML-Based Digital Video Library System Jacky C.K. Ma Michael R. Lyu.
What is adaptive web technology?  There is an increasingly large demand for software systems which are able to operate effectively in dynamic environments.
Computer Skills Preparatory Year Presented by: L.Obead Alhadreti.
Smart Learning Services Based on Smart Cloud Computing
CPS120: Introduction to Computer Science The World Wide Web Nell Dale John Lewis.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
1 Darmstadt, October 02, 2007 Amalia Ortiz Asociación VICOMTech Mikeletegi Pasealekua Donostia - San Sebastián (Gipuzkoa)
Some Thoughts on HPC in Natural Language Engineering Steven Bird University of Melbourne & University of Pennsylvania.
Introduction to Multimedia Networking (2) Advanced Multimedia University of Palestine University of Palestine Eng. Wisam Zaqoot Eng. Wisam Zaqoot October.
CISC105 General Computer Science Class 1 – 6/5/2006.
MediaHub: An Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney & Paul Mc Kevitt School of Computing and Intelligent Systems Faculty.
© 2007 Tom Beckman Features:  Are autonomous software entities that act as a user’s assistant to perform discrete tasks, simplifying or completely automating.
Spoken dialog for e-learning supported by domain ontologies Dario Bianchi, Monica Mordonini and Agostino Poggi Dipartimento di Ingegneria dell’Informazione.
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Multimodal Information Access Using Speech and Gestures Norbert Reithinger
CMPD273 Multimedia System Prepared by Nazrita Ibrahim © UNITEN2002 Multimedia System Characteristic Reference: F. Fluckiger: “Understanding networked multimedia,
Multimedia is a combination of text, art, sound, animation, and video.
4G-MOBILE COMMUNICATION D.MANOJ KUMAR III. B.Tech, E.C.E N.B.K.R.Institue of Science and technology P.PALLAVI II. B.Tech, E.C.E Kakatiya Institue of Technology.
Apartado Porto Codexwww.inescporto.pt tel (351) fax (351) /April/2005 Research Activities in 4G Networks at INESC Porto.
Challenges in the Migration to 4G Mobile Systems Suk Yu Hui and Kai Hau Yeung City university of Hong Kong Presenter: Shuai Peng
CHAPTER TEN AUTHORING.
Web Services for Satellite Emulation Development Kathy J. LiszkaAllen P. Holtz The University of AkronNASA Glenn Research Center.
FYP: LYU0001 Wireless-based Mobile E-Commerce on the Web Supervisor: Prof. Michael R. Lyu By: Tony, Wat Hong Fai Harris, Yan Wai Keung.
MULTIMEDIA OVER WIRELESS BROADBAND NETWORKS BY: NEELIMA PUNJALA.
卓越發展延續計畫分項三 User-Centric Interactive Media ~ 主 持 人 : 傅立成 共同主持人 : 李琳山,歐陽明,洪一平, 陳祝嵩 水美溫泉會館研討會
Mobile Communication Middleware By:Lekometsa Mokhesi Anisa Ragalo Supervisor: Ken Macgregor.
SoFI – Streaming Music using Song Form Intelligence By Jonathan Doherty Supervisors: Dr. Kevin Curran, Prof. Paul Mc Kevitt School of Computing and Intelligent.
1 Representing New Voice Services and Their Features Ken Turner University of Stirling 11th June 2003.
School of something FACULTY OF OTHER Facing Complexity Using AAC in Human User Interface Design Lisa-Dionne Morris School of Mechanical Engineering
Of 50 E GOV Universal Access Ahmed Gomaa CIMIC Rutgers University.
Toward a Unified Scripting Language 1 Toward a Unified Scripting Language : Lessons Learned from Developing CML and AML Soft computing Laboratory Yonsei.
Audient: An Acoustic Search Engine By Ted Leath Supervisor: Prof. Paul Mc Kevitt School of Computing and Intelligent Systems Faculty of Engineering University.
Österreichisches Forschnungsinstitut für Artificial Intelligence Representational Lego for ECAs Brigitte Krenn.
A MBI L EARN Ambient Intelligent Multimodal Learning Environment for Children 100 day review December 2008 Jennifer Hyndman Supervisors: Dr. Tom Lunney,
Volgograd State Technical University Applied Computational Linguistic Society Undergraduate and post-graduate scientific researches under the direction.
MPEG-4: Multimedia Coding Standard Supporting Mobile Multimedia System Lian Mo, Alan Jiang, Junhua Ding April, 2001.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Intelligent MultiMedia Storytelling System (IMSS) - Automatic Generation of Animation From Natural Language Input By Eunice Ma Supervisor: Prof. Paul Mc.
Multi-Modal Dialogue in Personal Navigation Systems Arthur Chan.
PGNET, Liverpool JMU, June 2005 MediaHub: An Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney, Paul Mc Kevitt School of Computing.
The “Spatial Turing Test” Stephan Winter, Yunhui Wu
Discovering Computers 2010
AIMS’99 Workshop Heidelberg, May 1999 Multimedia Services using IP over Bearer Networks: Quality of Service Aspects based on ACTS Guidelines SII.
Understanding Naturally Conveyed Explanations of Device Behavior Michael Oltmans and Randall Davis MIT Artificial Intelligence Lab.
W3C Multimodal Interaction Activities Deborah A. Dahl August 9, 2006.
WP6 Emotion in Interaction Embodied Conversational Agents WP6 core task: describe an interactive ECA system with capabilities beyond those of present day.
What is Multimedia Anyway? David Millard and Paul Lewis.
A New Approach to Decision-making within an Intelligent MultiMedia Distributed Platform Hub Glenn Campbell, Tom Lunney, Aiden Mc Caughey, Paul Mc Kevitt.
Forschungszentrum Telekommunikation Wien An initiative of the K plus Programme MONA Mobile Multimodal Next Generation Applications Rudolf Pailer
NCP meeting Jan 27-28, 2003, Brussels Colette Maloney Interfaces, Knowledge and Content technologies, Applications & Information Market DG INFSO Multimodal.
Chapter 1: Computer Systems
MediaHub: An Intelligent MultiMedia Distributed Platform Hub
Multimodal Human-Computer Interaction New Interaction Techniques 22. 1
Copyright 2005 Prentice- Hall, Inc.
Presentation transcript:

TeleMorph & TeleTuras: Bandwidth determined Mobile MultiModal Presentation Student: Anthony J. Solon Supervisors: Prof. Paul Mc Kevitt Kevin Curran School of Computing and Intelligent Systems Faculty of Engineering University of Ulster, Magee

Aims of Research To develop an architecture, TeleMorph, that dynamically morphs between output modalities depending on available network bandwidth: Mobile device’s output presentation (unimodal/multimodal) depending on available network bandwidth  network latency and bit error rate  mobile device display, available output abilities, memory, CPU  user modality preferences, cost incurred  user’s cognitive load determined by Cognitive Load Theory (CLT) Utilise Causal Probabilistic Networks (CPNs) for analysing union of constraints giving optimal multimodal output presentation Implement TeleTuras, a tourist information guide for city of Derry

Objectives of Research Receive and interpret questions from user Map questions to multimodal semantic representation Match multimodal representation to knowledge base to retrieve answer Map answers to multimodal semantic representation Monitor user preference or client side choice variations Query bandwidth status Detect client device constraints and limitations Combine affect of all constraints imposed on system using CPNs Generate optimal multimodal presentation based on bandwidth constraint data

Wireless Telecommunications Generations of Mobile networks: 1G - Analog voice service with no data services 2G - Circuit-based, digital networks, capable of data transmission speeds averaging around 9.6K bps 2.5G (GPRS) - Technology upgrades to 2G, boosting data transmission speeds to around 56K bps. Allows packet based “always on” connectivity 3G (UMTS) - digital multimedia, different infrastructure required, data transmission speeds from 144K-384K-2M bps 4G - IP based mobile/wireless networks, Wireless Personal Area Networks (PANs), ‘anywhere and anytime’ ubiquitous services. Speeds up to 100Mbps Network-adaptive multimedia models: Transcoding proxies End-to-end approach Combination approach Mobile/Nomadic computing Active networks

Mobile Intelligent MultiMedia Systems SmartKom (Wahlster, 2003) Mobile, Public, Home/office Saarbr ü cken, Germany Combines speech, gesture and facial expressions on input & output Integrated trip planning, Internet access, communication applications, personal organising VoiceLog (BBN, 2002) BBN technologies in Cambridge, Massachusettes Views/diagrams of military vehicles and direct connection to support Damage identified & ordering of parts using diagrams MUST (Almeida et al., 2002) MUltimodal multilingual information Services for small mobile Terminals EURESCOM, Heidelberg, Germany Future multimodal and multilingual services on mobile networks Please select a parking place from the Map

Intelligent MultiMedia Presentation Flexibly generate various presentations to meet individual requirements of: 1) users, 2) situations, 3) domains Intelligent MultiMedia Presentation can be divided into following processes: determination of communicative intent content selection structuring and ordering allocation to particular media realisation in specific media coordination across media layout design Key research problems: Semantic Representation Fusion, integration & coordination

Semantic representation - represents meaning of media information  Frame-based representations: -CHAMELEON -REA  XML-based representations: -M3L (SmartKom) -MXML (MUST) -SMIL -MPEG-7 Fusion, integration & coordination of modalities  Integrating different media in a consistent and coherent manner  Multimedia coordination leads to effective integrated multiple media in output  Synchronising modalities Time threshold between modalities E.g. Input - “ What building is this? ”, Output - “ This is the Millenium forum ” Not synchronised => side effect can be contradiction  SMIL modality synchronisation and timing elements

Intelligent MultiMedia Presentation Systems Automatically generate coordinated intelligent multimedia presentations User-determined presentation: COMET (Feiner & McKeown, 1991)  COordinated Multimedia Explanation Testbed  Generates instructions for maintenance and repair of military radio receiver-transmitters  Coordinates text and 3D graphics of mechanical devices WIP (Wahlster et al., 1992)  Intelligent multimedia authoring system  presents instructions for assembling/using/maintaining/repairing devices (e.g. espresso machines, lawn mowers, modems) IMPROVISE (Zhou & Feiner, 1998)  Graphics generation system  constructive/parameterised graphics generation approaches  Uses an extensible formalism to represent a visual lexicon for graphics generation

Intelligent MultiMedia Interfaces & Agents Intelligent multimedia interfaces Parse integrated input and generate coordinated output XTRA  Interface to an expert system providing tax form assistance  Generates & interprets natural language text and pointing gestures automatically; relies on pre-stored graphics  Displays relevant tax form and natural language input/output panes Intelligent multimedia agents Embodied Conversational Agents (e.g. MS Agent, REA) Natural human face-face communication - speech, facial expressions, hand gestures & body stance MS Agent  Set of programmable services for interactive presentation  Speech, gesture, audio & text output; speech & haptic input

Project Proposal Research and implement mobile intelligent multimedia presentation architecture called TeleMorph Dynamically generates multimedia presentation determined by bandwidth available; also other constraints: Network latency, bit error rate Mobile device display, available output abilities, memory, CPU user modality preferences, cost incurred Cognitive Load Theory (CLT) Causal Probabilistic Networks (CPNs) for analysing union of constraints giving optimal multimodal output presentation

Implement TeleTuras, a tourist information guide for city of Derry providing testbed for TeleMorph incorporating: route planning, maps, spoken presentations, graphics of points of interest & animations Output modalities used & effectiveness of communication TeleTuras examples: “Where is the Millenium forum?” “How do I get to the GuildHall?” “What buildings are of interest in this area?” “Is there a Chinese restaurant in this area?”

Architecture of TeleMorph

Data flow of TeleMorph Media Analysis : High level :

Comparison of Mobile Intelligent MultiMedia Systems

Comparison of Intelligent MultiMedia Systems

Software Analysis Client output: SMIL media player (InterObject) Java Speech API Markup Language (JSML) Autonomous agent (MSAgent) Client input: Java Speech API Grammar Format (JSGF) J2ME graphics APIs J2ME networking Client device status: SysInfo MIDlet - (type/memory/screen/protocols/input abilities/CPU speed) TeleMorph server tools: SMIL & MPEG-7 HUGIN (CPNs) JATLite/OAA

Project Schedule

Conclusion A Mobile Intelligent MultiModal Presentation Architecture called TeleMorph will be developed Dynamically morphing between output modalities depending on available network bandwidth in conjunction with other relevant constraints CPNs for analysing union of constraints giving optimal multimodal output presentation TeleTuras will be used as testbed for TeleMorph Corpora of questions to test TeleTuras (prospective users/tourists)