WEEK 4 PRESENTATION NGOC TA AIDEAN SHARGHI.

Slides:



Advertisements
Similar presentations
Using Closed Captions to Train Activity Recognizers that Improve Video Retrieval Sonal Gupta and Raymond Mooney University of Texas at Austin.
Advertisements

Limin Wang, Yu Qiao, and Xiaoou Tang
Application of light fields in computer vision AMARI LEWIS – REU STUDENT AIDEAN SHARGHI- PH.D STUENT.
SOMM: Self Organizing Markov Map for Gesture Recognition Pattern Recognition 2010 Spring Seung-Hyun Lee G. Caridakis et al., Pattern Recognition, Vol.
Encouraging Students in the Target Language. Target Language Classroom Kit Popsicle Sticks Large / Small Dice Scanned Pictures Video Clips Short Scenario.
Distributed Representations of Sentences and Documents
Recognizing Action at a Distance A.A. Efros, A.C. Berg, G. Mori, J. Malik UC Berkeley.
WEEK 7 Amari Lewis Aidean Sharghi Amari Lewis Aidean Sharghi.
Introduction CSE 1310 – Introduction to Computers and Programming Vassilis Athitsos University of Texas at Arlington 1.
Introduction CSE 1310 – Introduction to Computers and Programming Vassilis Athitsos University of Texas at Arlington 1.
Spatio-temporal constraints for recognizing 3D objects in videos Nicoletta Noceti Università degli Studi di Genova.
Title Slide Title Circle K at [insert University name] [Date]
APPLICATIONS OF LIGHT FIELDS IN COMPUTER VISION WEEK 2 REU STUDENT: AMARI LEWIS P.H.D STUDENT: AIDEAN SHARGHI.
Case Study 1 Semantic Analysis of Soccer Video Using Dynamic Bayesian Network C.-L Huang, et al. IEEE Transactions on Multimedia, vol. 8, no. 4, 2006 Fuzzy.
BPEL Business Process Engineering Language A technology used to build programs in SOA architecture.
Process Model Test In Response to JIRA Issues 31, 55 and 56 8/12/15.
WEEK4 RESEARCH Amari Lewis Aidean Sharghi. PREPARING THE DATASET  Cars – 83 samples  3 images for each sample when x=0  7 images for each sample when.
Unsupervised Salience Learning for Person Re-identification
Using decision trees to build an a framework for multivariate time- series classification 1 Present By Xiayi Kuang.
GROUP NAME THIS LAND IS YOUR LAND BY: Group members’ names here.
AND Gate Inputs Output Input A (Switch) Input B (Switch) Output Y (Lamp) 0 (Open) 0 (OFF) A B Lamp.
Multi-view Synchronization of Human Actions and Dynamic Scenes Emilie Dexter, Patrick Pérez, Ivan Laptev INRIA Rennes - Bretagne Atlantique
Predicting the dropouts rate of online course using LSTM method
NARRATIVE TEXT Susilo Waluyo What do you know about :  F Fables are stories about animals that can talk and act like a man. e.g. the wolf and the house.
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. SHOW.
Human features are those things created by man.
A Hierarchical Deep Temporal Model for Group Activity Recognition
End-To-End Memory Networks
Understanding the Constructs
Tracking parameter optimization
Different Units Ramakrishna Vedantam.
Project Management
Project Management
Project Management
Project Management
week 1 - Introduction Goals
بسم الله الرحمن الرحيم.
REU student -Amari Lewis P.H.D student- Aidean Sharghi June 6th 2014
Master’s Thesis defense Ming Du Advisor: Dr. Yi Shang
Attention-based Caption Description Mun Jonghwan.
إستراتيجيات ونماذج التقويم
Paraphrase Generation Using Deep Learning
Two-Stream Convolutional Networks for Action Recognition in Videos
Video understanding using part based object detection models
دانشگاه شهیدرجایی تهران
تعهدات مشتری در کنوانسیون بیع بین المللی
بسمه تعالی کارگاه ارزشیابی پیشرفت تحصیلی
Session 28 Learning Objectives: - understand how to comment on language features and structure - practise how to answer Paper 1, Q 2 & 3.
Signals and Systems EE235 Leo Lam Leo Lam ©
Pass The Ball.
Recurrent Encoder-Decoder Networks for Time-Varying Dense Predictions
Comparison of EET and Rank Pooling on UCF101 (split 1)
Social Practice of the language: Describe and share information
Reading Tuesday, August 17, 2016.
Learn to Comment Mentor: Mahdi M. Kalayeh
Towards an Unequivocal Representation of Actions
EVENT TITLE Time Date Location Call to Action!
Presented By: Harshul Gupta
Weekly Learning Alex Omar Ruiz Irene.
Describing Objects.
Amari Lewis Aidean Sharghi
Week 3 Presentation Ngoc Ta Aidean Sharghi.
UCF-REU in Computer Vision
Week 8 Presentation Ngoc Ta Aidean Sharghi.
Week 3 Volodymyr Bobyr.
Actor-Object Relation in Videos
Week 7 Presentation Ngoc Ta Aidean Sharghi
Week 6 Presentation Ngoc Ta Aidean Sharghi.
What and How Well You Performed
Presentation transcript:

WEEK 4 PRESENTATION NGOC TA AIDEAN SHARGHI

SST: Single-Stream Temporal Action Proposals Single pass Input: video sentence Output: number of temporal intervals that contain an action Dataset: THUMOS 14 (20min long videos) Performs better at higher tIoU regime Handle very long testing sequences

Dense-Captioning Events in Videos Localize temporal proposal of interest Describe with natural language Dataset: ActivityNet Captions Single pass Using more strides improves recall across all values of IoU’s

Bidirectional Attentive Fusion with Context Gating for Dense Video Captioning Learn the details Modify parameters

Output sentences: a man is seen speaking to the camera and leads into a person holding a ball and a man is standing in a large group of people a man is seen speaking to the camera and leads into a man holding a stick and a man is standing in a black METEOR: 5.2925

Extract features from C3D Model axon-research/c3d-keras facebook/C3D -caffe hx173149/C3D-tensorflow

THANK YOU