CRCV REU 2019 Week 4.

Slides:



Advertisements
Similar presentations
Integers Adding and Subtracting
Advertisements

A) 80 b) 53 c) 13 d) x 2 = : 10 = 3, x 3 = 309.
Content Based Learning Social Learning By Hilda Hernandez
Some of the best books of  &mid= A23D2FC75CD A23D2FC75CD7.
Divide. Evaluate power – 3 = – 3 EXAMPLE – 3 = 3 2 – – 3 = 6 – 3 Multiply. Evaluate expressions Multiply and divide from.
A Robust Luby Transform Encoding Pattern-Aware Symbol Packetization Algorithm for Video Streaming Over Wireless Network Dongju Lee and Hwangjun Song IEEE.
WEEK 1 You have 10 seconds to name…
Communication and Management
Parent Workshop Assessments and Goals Zooming in on the October Progress Report Kindergarten and 1 st Grade.
Haitham Elmarakeby.  Speech recognition
Convolutional LSTM Networks for Subcellular Localization of Proteins
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
Olivier Siohan David Rybach
Unsupervised Learning of Video Representations using LSTMs
Hierarchical Question-Image Co-Attention for Visual Question Answering
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Textual Video Prediction Week 2
Show and Tell: A Neural Image Caption Generator (CVPR 2015)
Week 8 REU Nolan Warner.
Week 6 Cecilia La Place.
VQA: Visual Question Answering
Type Topic in here! Created by Educational Technology Network
Textual Video Prediction
Using Recurrent Neural Networks and Hebbian Learning
Attention Is All You Need
Visual Question Generation
Chapter 13. Mathematics.
Project # 5 Generating Privacy and Security Threat Summary for Internet of Things REU student: Domonique Cox Graduate mentors: Kaiqiang Song Faculty mentor(s):
סדר דין פלילי – חקיקה ומהות ההליך הפלילי
Understanding LSTM Networks

Project 1: Smart Home REU student: Jason Ling Graduate mentors: Safa Bacanli Faculty mentor(s): Damla Turgut Week 9 (July 9 – July ) Accomplishments:
Yi Zhao1, Yanyan Shen*1, Yanmin Zhu1, Junjie Yao2
Project # Investigating the Value versus Privacy Cost of the Internet of Things REU student: Jason Ling Graduate mentors: Safa Bacanli Faculty mentor(s):
Recurrent Encoder-Decoder Networks for Time-Varying Dense Predictions
Project 1: Smart Home REU student: Jason Ling Graduate mentors: Safa Bacanli Faculty mentor(s): Damla Turgut Week 8 (July 2 – July ) Accomplishments:
Nolan Warner Week 12 REU.
Unsupervised Pretraining for Semantic Parsing
Jacob Devlin Ming-Wei Chang Kenton Lee Kristina Toutanova
Please enjoy.
Word Embedding 모든 단어를 vector로 표시 Word vector Word embedding Word
Organization- Unit 1 CDA
You must show all steps of your working out.
Extended Registration KS3
Question 1.
-- Ray Mooney, Association for Computational Linguistics (ACL) 2014
Semi-Supervised Learning and Active Learning
Ask and Answer Questions
Textual Video Prediction
Practice Project Overview
Question Answering System
1MIT CSAIL 2Tsinghua University 3MIT-IBM Watson AI Lab 4DeepMind
Baseline Model CSV Files Pandas DataFrame Sentence Lists
Week 3 Presentation Ngoc Ta Aidean Sharghi.
UCF-REU in Computer Vision
Week 3: Moving Target Detection Using Infrared Sensors
Visual Question Answering Aaron Honculada, Aisha Urooj Khan, Dr
Week 8 Presentation Ngoc Ta Aidean Sharghi.
CRCV REU 2019 Kara Schatz.
CRCV REU 2019 Week 8 Aaron Honculada.
CRCV REU 2019 Week 5.
CRCV REU 2019 Aaron Honculada.
End-to-End Speech-Driven Facial Animation with Temporal GANs
Week 7 Presentation Ngoc Ta Aidean Sharghi
Listen Attend and Spell – a brief introduction
Visual Question Answering
CRCV REU 2019 Aaron Honculada.
Extended Registration KS3
Self-Supervised Cross-View Action Synthesis
The experiment based on hier-attention
Presentation transcript:

CRCV REU 2019 Week 4

Literature Review this week Attention Model/ Transformer Network ELMo OpenAI Transformer Decoder only Word prediction Use Massive Unlabeled Data (books) BERT Conditioned from left and right Encoder only Sequence to Sequence

Literature Review this week Neural-Symbolic Visual Question Answering (NS-VQA)

Baseline Models LSTM BiLSTM

Results Model Used TVQA + S TVQA + V TVQA + IMG TVQA + V + IMG Accuracy (%) Reported 65.15% 45.03% 43.78% N/A Replication 65.74% 45.25% 44.42% 45.52% Q LSTM 42.74% BiLSTM 42.48%

Results Model Used TVQA + S TVQA + V TVQA + IMG TVQA + V + IMG Accuracy (%) Reported 65.15% 45.03% 43.78% N/A Replication 65.74% 45.25% 44.42% 45.52% Q S + Q LSTM 42.74% 42.71% BiLSTM 42.48% 42.67%

Results Model Used TVQA + S TVQA + V TVQA + IMG TVQA + V + IMG Accuracy (%) Reported 65.15% 45.03% 43.78% N/A Replication 65.74% 45.25% 44.42% 45.52% Q S + Q V + Q LSTM 42.74% 42.71% 42.61% BiLSTM 42.48% 42.67%

Results Model Used TVQA + S TVQA + V TVQA + IMG TVQA + V + IMG Accuracy (%) Reported 65.15% 45.03% 43.78% N/A Replication 65.74% 45.25% 44.42% 45.52% Q S + Q V + Q S + V + Q LSTM 42.74% 42.71% 42.61% 42.39% BiLSTM 42.48% 42.67% 42.84%

Next Steps Baseline CNN+LSTM (in Progress) Video Action Transformer Network Neuro-Symbolic Concept Learner (NS-CL)