Multiple track hypotheses maintained for each possible blob-object mapping Parse driven by blob interaction events Domain-general events are detected (e.g.,

Slides:



Advertisements
Similar presentations
Matching in 2D Readings: Ch. 11: Sec alignment point representations and 2D transformations solving for 2D transformations from point correspondences.
Advertisements

KAIS T The Vision of Autonomic Computing Jeffrey O. Kephart, David M Chess IBM Watson research Center IEEE Computer, Jan 발표자 : 이승학.
1 Approximated tracking of multiple non-rigid objects using adaptive quantization and resampling techniques. J. M. Sotoca 1, F.J. Ferri 1, J. Gutierrez.
2D matching part 2 Review of alignment methods and
Probabilistic Group-Level Motion Analysis and Scenario Recognition Ming-Ching Chang, Nils Krahnstoever, Weina Ge ICCV2011.
CS 355 – Programming Languages
Computational Analysis of Motor Learning. Three paradigms Force field adaptation Visuomotor transformations Sequence learning Does one term (motor learning)
VisualLISA A Quick and Simple Demonstration. The default aspect of an attribute grammar specification on VisualLISA. In this view we can add a new production.
John Hu Nov. 9, 2004 Multimodal Interfaces Oviatt, S. Multimodal interfaces Mankoff, J., Hudson, S.E., & Abowd, G.D. Interaction techniques for ambiguity.
A Bayesian algorithm for tracking multiple moving objects in outdoor surveillance video Department of Electrical Engineering and Computer Science The University.
Experiences with an Architecture for Intelligent Reactive Agents By R. Peter Bonasso, R. James Firby, Erann Gat, David Kortenkamp, David P Miller, Marc.
Zhu, Song-Chun and Mumford, David. A Stochastic Grammar of Images. Foundations and Trends in Computer Graphics and Vision 2(4), (2006) Hemerson.
Hierarchical GUI Test Case Generation Using Automated Planning Atif M. Memon, Student Member, IEEE, Martha E. Pollack, and Mary Lou Soffa, Member, IEEE.
(Hadoop) Pig Dataflow Language B. Ramamurthy Based on Cloudera’s tutorials and Apache’s Pig Manual 6/27/2015.
Tracking Video Objects in Cluttered Background
MVL (Machine Vision Lab) UIC HUMAN MOTION VIDEO DATABASE Jezekiel Ben-Arie ECE Department University Of Illinois at Chicago Scripts, Queries, Recognition.
CSE 341 S. Tanimoto Vis Lang - 1 Programming Languages/Systems A set of possible computations (usually infinite) A set of primitive operations: e.g., arithmetic.
1 Scenario-based Analysis of UML Design Class Models Lijun Yu October 4th, 2010 Oslo, Norway.
1 Video Surveillance systems for Traffic Monitoring Simeon Indupalli.
User Interface Design Principles Gabriel Spitz 1 Lecture # 4.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
User Characteristics & Design Principles Gabriel Spitz 1 Lecture # 11.
[ §6 : 1 ] 6. Basic Methods II Overview 6.1 Models 6.2 Taxonomy 6.3 Finite State Model 6.4 State Transition Model 6.5 Dataflow Model 6.6 User Manual.
Interaction Modeling. Sequence Models  There are two kinds of sequence models: scenarios and sequence diagrams  A scenario is a sequence of events that.
Cookies, Spreadsheets, and Modeling: Dynamic, Interactive, Visual Science and Math Scott A. Sinex Prince George’s Community College Presented at Network.
Activity Diagram. C-S 5462 Activity diagram a diagram that represents a sequence of activities similar to flow chart, used in 1960’s and 1970’s an activity.
A Brief Overview of Computer Vision Jinxiang Chai.
Chapter 10 State Machine Diagrams
Visual Object Tracking Xu Yan Quantitative Imaging Laboratory 1 Xu Yan Advisor: Shishir K. Shah Quantitative Imaging Laboratory Computer Science Department.
Probabilistic Context Free Grammars for Representing Action Song Mao November 14, 2000.
3 April SOA: Services Oriented Architecture MDA: Model Driven Architecture.
A Cognitive Substrate for Natural Language Understanding Nick Cassimatis Arthi Murugesan Magdalena Bugajska.
1 ECE 453 – CS 447 – SE 465 Software Testing & Quality Assurance Instructor Kostas Kontogiannis.
Multiplication Goal 4: Squares through 25.
K. J. O’Hara AMRS: Behavior Recognition and Opponent Modeling Oct Behavior Recognition and Opponent Modeling in Autonomous Multi-Robot Systems.
1 Software Design Reference: Software Engineering, by Ian Sommerville, Ch. 12 & 13, 5 th edition and Ch. 10, 6 th edition.
1 Software Design Overview Reference: Software Engineering, by Ian Sommerville, Ch. 12 & 13.
ISBN Chapter 3 Describing Semantics -Attribute Grammars -Dynamic Semantics.
CS 363 Comparative Programming Languages Semantics.
Tracking People by Learning Their Appearance Deva Ramanan David A. Forsuth Andrew Zisserman.
3.2 Semantics. 2 Semantics Attribute Grammars The Meanings of Programs: Semantics Sebesta Chapter 3.
Xusheng Xiao North Carolina State University CSC 720 Project Presentation 1.
Human Activity Recognition at Mid and Near Range Ram Nevatia University of Southern California Based on work of several collaborators: F. Lv, P. Natarajan,
States.
The Structure of the User Interface Lecture # 8 1 Gabriel Spitz.
Gabriel Spitz 1. Light Switch Issues  Consistency – The switch design is inconsistent with common light switches  Visibility – No visible cue regarding.
Lucent Technologies - Proprietary 1 Interactive Pattern Discovery with Mirage Mirage uses exploratory visualization, intuitive graphical operations to.
Stochastic Grammars: Overview Representation: Stochastic grammar Representation: Stochastic grammar Terminals: object interactions Terminals: object interactions.
STAR Webinars Ontology driven diagram generator for health simulation models Andrew Sutcliffe.
Chapter 7 Part II Structuring System Process Requirements MIS 215 System Analysis and Design.
On Concurrency Idioms and their Effect on Program Analysis Weizmann Institute of Science Guy Katz and David Harel.
1 Minimum Bayes-risk Methods in Automatic Speech Recognition Vaibhava Geol And William Byrne IBM ; Johns Hopkins University 2003 by CRC Press LLC 2005/4/26.
Agenda Preliminaries Motivation and Research questions Exploring GLL
Observer Design Pattern
Leveraging High-Level Expectations for Activity Recognition
Software engineering USER INTERFACE DESIGN.
Above and below the object level
Chapter 1 Database Systems
An Evolutional Model for Operation-driven Visualization Design
States.
Leveraging High-Level Expectations for Activity Recognition
The Vision of Autonomic Computing
States.
Automated Software Integration
Chapter 1 Database Systems
Parsing Unrestricted Text
Detection and tracking of individual SNAP-tagged proteins on the surface of living cells. Detection and tracking of individual SNAP-tagged proteins on.
Med-Fi Prototype Presentation
Getting started – Example 1
Interaction Models (2): Activity Diagrams
Presentation transcript:

Multiple track hypotheses maintained for each possible blob-object mapping Parse driven by blob interaction events Domain-general events are detected (e.g., merge, split, enter, exit) Events transformed into activity-specific terminals of the grammar Abstract model used to detect and prune semantically inconsistent interpretations Likelihood of parse computed from production and symbol probabilities Activity recognized by selecting the most likely, semantically consistent hypothesis Leveraging High-Level Expectations for Activity Recognition David Minnen, Irfan Essa, Thad Starner ToH -> Setup, enter(hand), Solve, exit(hand); Setup -> TowerPlaced, exit(hand); TowerPlaced -> enter(hand, block_A, block_B, block_C), Put_1(block_A, block_B, block_C); Solve -> state(Tower = TowerStart), MakeMoves, state(Tower = TowerGoal); MakeMoves -> Move(block) [0.1] | Move(block), MakeMoves [0.9]; Move -> Move_1-2 | Move_1-3 | Move_2-1 | Move_2-3 | Move_3-1 | Move_3-2; Move_1-2 -> Grab_1, Put_2; Move_1-3 -> Grab_1, Put_3; Move_2-1 -> Grab_2, Put_1; Move_2-3 -> Grab_2, Put_3; Move_3-1 -> Grab_3, Put_1; Move_3-2 -> Grab_3, Put_2; Grab_1 -> touch_1, remove_1(hand, ~) | touch_1(~), remove_last_1(~); Grab_2 -> touch_2, remove_2(hand, ~) | touch_2(~), remove_last_2(~); Grab_3 -> touch_3, remove_3(hand, ~) | touch_3(~), remove_last_3(~); Put_1 -> release_1(~) | touch_1, release_1; Put_2 -> release_2(~) | touch_2, release_2; Put_3 -> release_3(~) | touch_3, release_3; Stochastic Grammar for Towers of Hanoi Task Automatically detect and decompose an activity from video Activity manually specified as a parameterized stochastic grammar Feedback from high-level parse influences interpretation of object identities Experiments use the Towers of Hanoi task to focus on activity recognition rather than visual appearance modeling Overview & Goals Difficult to represent complex temporal constraints such as concurrency and quantitative temporal relationships Inefficient deletion error recovery (exponential growth) Requires activities with strong temporal relationships to disambiguate object identities and actions without relying on appearance models [ (1-29) ToH. ] [ (1-29) Setup, Solve, exit(hand). ] [ (1-4) TowerPlaced, exit(hand), enter(hand). ] [ (1-2) enter(hand, red, green, blue), Put_1(red, green, blue). ] [ (2-2) release_1. ] [ (5-28) state(Tower=TowerStart), MakeMoves, state(Tower=TowerGoal). ] [ (5-28) Move(block), MakeMoves. ] [ (5-7) Move_1-3. ] [ (5-7) Grab_1, Put_3. ] [ (5-6) touch_1, remove_1(hand). ] [ (7-7) release_3. ] [ (8-28) Move(block), MakeMoves. ] [ (8-10) Move_1-2. ] [ (8-10) Grab_1, Put_2. ] [ (8-9) touch_1, remove_1(hand). ] [ (10-10) release_2. ] … Partial Parse of ToH Task Limitations & Future Work Activity Modeling and Recognition Approach Data Flow Recognition of Towers of Hanoi with Occlusion Identifying Objects by Action Rather than Appearance