Report by: 陆纪圆.

Slides:



Advertisements
Similar presentations
Statistical Phrase-Based Translation Authors: Koehn, Och, Marcu Presented by Albert Bertram Titles, charts, graphs, figures and tables were extracted from.
Advertisements

Introduction to Recurrent neural networks (RNN), Long short-term memory (LSTM) Wenjie Pei In this coffee talk, I would like to present you some basic.
Application of RNNs to Language Processing Andrey Malinin, Shixiang Gu CUED Division F Speech Group.
NUDT Machine Translation System for IWSLT2007 Presenter: Boxing Chen Authors: Wen-Han Chao & Zhou-Jun Li National University of Defense Technology, China.
Addressing the Rare Word Problem in Neural Machine Translation
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Haitham Elmarakeby.  Speech recognition
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. SHOW.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
Attention Model in NLP Jichuan ZENG.
Olivier Siohan David Rybach
Fabien Cromieres Chenhui Chu Toshiaki Nakazawa Sadao Kurohashi
Neural Machine Translation
Convolutional Sequence to Sequence Learning
Unsupervised Learning of Video Representations using LSTMs
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
Approaches to Machine Translation
End-To-End Memory Networks
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Deep Learning for Program Repair
Wu et. al., arXiv - sept 2016 Presenter: Lütfi Kerem Şenel
Recurrent Neural Networks for Natural Language Processing
Van-Khanh Tran and Le-Minh Nguyen
Neural Machine Translation by Jointly Learning to Align and Translate
Attention Is All You Need
Show and Tell: A Neural Image Caption Generator (CVPR 2015)
An Overview of Machine Translation
Intro to NLP and Deep Learning
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Neural Machine Translation By Learning to Jointly Align and Translate
Statistical Machine Translation Part III – Phrase-based SMT / Decoding
Attention Is All You Need
Deep Learning based Machine Translation
Grid Long Short-Term Memory
Paraphrase Generation Using Deep Learning
Image Captions With Deep Learning Yulia Kogan & Ron Shiff
Final Presentation: Neural Network Doc Summarization
Understanding LSTM Networks
Word Embedding Word2Vec.
Approaches to Machine Translation
Code Completion with Neural Attention and Pointer Networks
The Big Health Data–Intelligent Machine Paradox
Memory-augmented Chinese-Uyghur Neural Machine Translation
Introduction to Natural Language Processing
Statistical Machine Translation Papers from COLING 2004
Recurrent Encoder-Decoder Networks for Time-Varying Dense Predictions
Machine Translation(MT)
Natural Language to SQL(nl2sql)
RNN Encoder-decoder Architecture
Attention.
Presentation By: Eryk Helenowski PURE Mentor: Vincent Bindschaedler
Please enjoy.
Word embeddings (continued)
Word Embedding 모든 단어를 vector로 표시 Word vector Word embedding Word
Attention for translation
-- Ray Mooney, Association for Computational Linguistics (ACL) 2014
Statistical Machine Translation Part VI – Phrase-based Decoding
Neural Machine Translation - Encoder-Decoder Architecture and Attention Mechanism Anmol Popli CSE 291G.
Neural Machine Translation using CNN
Question Answering System
Neural Machine Translation
Baseline Model CSV Files Pandas DataFrame Sentence Lists
Julia Hockenmaier April 2019
Recurrent Neural Networks
Sequence-to-Sequence Models
CSC 578 Neural Networks and Deep Learning
Neural Machine Translation by Jointly Learning to Align and Translate
Visual Grounding.
CVPR 2019 Poster.
Presentation transcript:

Report by: 陆纪圆

MT(machine translation) gap Example-based, Hybrid “understand”

RBMT(rule-based machine translation) Generally, rule-based methods parse a text, usually creating an intermediary, symbolic representation, from which the text in the target language is generated. The rule-based machine translation paradigm includes transfer-based machine translation, interlingual machine translation and dictionary-based machine translation paradigms. Unlike interlingual MT, transfer-based machine translation depends partially on the language pair involved in the translation.

SMT(statistical machine translation) By Bayes Theorem: The best translation: e: string in the target language f: string in the source language The probability distribution p(e|f) is modeled by: Translation model p(f|e): The probability that the source string f is the translation of the target string e Language model p(e): the probability of seeing the target string e in the target language

ANN(artificial neural networks)

RNN(Recurrent neural networks) FNN(Feedforward neural networks) Recurrent回归的 This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Feedforward前馈 的

Word embedding

NMT(neural machine translation) A bidirectional recurrent neural network (RNN), known as an encoder, is used by the neural network to encode a source sentence for a second RNN, known as a decoder, that is used to predict words in the target language.[8] A sequence of fixed-dimensional vectors

Encoder Decoder E has as many columns as there are words in the source vocabulary and as many rows as you want h0 is all zero

BiRNN(Bidirectional RNN) Future context

Soft attention Softmax normalization Expected vector ci instead of ht for every i

Word alignment Bitext word alignment or simply word alignment is the natural language processing task of identifying translation relationships among the words (or more rarely multiword units) in a bitext, resulting in a bipartite graph between the two sides of the bitext, with an arc between two words if and only if they are translations of one another.

Parse tree

Bidirectional tree encoder Novel: top-down encoder

Bottom-up tree encoder Phrase vector -> sentence vector Dj == Cj,the expected vector Sj is the jth target hidden unit(previous?)

Top-down tree encoder Draw separate pictures h11 up -> h11 down

Incorporate source syntactic tree

Coverage model

Tree coverage model What about top-down coverage?

UNK: out-of-vocabulary words

Model comparison LSTM: long short-term memory

Data run National Institute of Standards and Technology 美国国家标准技术研究院 Linguistic Data Consortium 语言学数据库

BLEU scores (bilingual evaluation understudy) LSTM: long short-term memory GRU: gated recurrent unit

Thanks! Q&A