Report Yang Zhang.

Slides:



Advertisements
Similar presentations
ECG Signal processing (2)
Advertisements

CS590M 2008 Fall: Paper Presentation
Active Learning for Streaming Networked Data Zhilin Yang, Jie Tang, Yutao Zhang Computer Science Department, Tsinghua University.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
Visual Recognition Tutorial
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
CS292 Computational Vision and Language Pattern Recognition and Classification.
An Optimal Learning Approach to Finding an Outbreak of a Disease Warren Scott Warren Powell
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Image Compression Using Neural Networks Vishal Agrawal (Y6541) Nandan Dubey (Y6279)
. Expressive Graphical Models in Variational Approximations: Chain-Graphs and Hidden Variables Tal El-Hay & Nir Friedman School of Computer Science & Engineering.
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Crash Course on Machine Learning
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Machine Learning Seminar: Support Vector Regression Presented by: Heng Ji 10/08/03.
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Model-based Bayesian Reinforcement Learning in Partially Observable Domains by Pascal Poupart and Nikos Vlassis (2008 International Symposium on Artificial.
Visual Information Systems Recognition and Classification.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
BCS547 Neural Decoding.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Dynamic Background Learning through Deep Auto-encoder Networks Pei Xu 1, Mao Ye 1, Xue Li 2, Qihe Liu 1, Yi Yang 2 and Jian Ding 3 1.University of Electronic.
Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Variational Autoencoders Presentation by Yuri Burda CS2523, University of Toronto.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Variational Autoencoders Theory and Extensions
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
CS479/679 Pattern Recognition Dr. George Bebis
Chapter 7. Classification and Prediction
Learning Deep Generative Models by Ruslan Salakhutdinov
Unit 3 Hypothesis.
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Computational Learning Theory
Parameter Estimation 主講人:虞台文.
Statistical Models for Automatic Speech Recognition
Machine Learning Basics
Course: Autonomous Machine Learning
Vision Reading Group, 2nd of May 2018 Martin Rünz
Variational Knowledge Graph Reasoning
Learning with information of features
Learning Hierarchical Features from Generative Models
PixelGAN Autoencoders
9 Tests of Hypotheses for a Single Sample CHAPTER OUTLINE
Bayesian Models in Machine Learning
CS 2750: Machine Learning Support Vector Machines
Collapsed Variational Dirichlet Process Mixture Models
Statistical Models for Automatic Speech Recognition
Support Vector Machines Most of the slides were taken from:
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Hairong Qi, Gonzalez Family Professor
Bayesian Inference for Mixture Language Models
COSC 4335: Other Classification Techniques
An introduction to Graphical Models – Michael Jordan
Support Vector Machines
Support Vector Machines and Kernels
LECTURE 07: BAYESIAN ESTIMATION
Deep learning enhanced Markov State Models (MSMs)
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Background Task Fashion image inpainting Some conceptions
CS249: Neural Language Model
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

Report Yang Zhang

02 03 01 Content VAE My code and Some Experiments Tied VAE for SRE PPT由四个部分组成,分别是项目灵感和简介,系统工作流程,当前工作进度与未来计划。

00 Auto Encoder

01 Variational Auto Encoder

Generative modeling “Generative modeling” is a area of machine learning which deals with models of distributions P(X). For instance, image(pixels)… We can formalize this setup by saying that we get examples X distributed according to some unknown distribution Pgt(X), and our goal is to learn a model P which we can sample from, such that P is as similar as possible to Pgt

How to train it? How to train a generative model is a long-standing problem in machine learning, most approaches have one of three serious drawback: Might require strong assumptions about the structure in the data they might make severe approximations, leading to suboptimal models. they might rely on computationally expensive inference procedures like Markov Chain Monte Carl VAE is a popular frame work to solve those problems

Latent Variable Latent Variable helps if the model first decides which character to generate before it assigns a value to any specific pixel. This kind of decision is formally called a latent variable

Latent Variable

Latent Variable

Network structure

Optimize Computing p(x) is quite difficult. So,the right hand side of Equation 5 is a lower bound to P(x). Our goal is changed to optimize the lower bound to P(x) And, many predecessors look at it in terms of information-theory…

KL: encourages our learned distribution q(z|x) to be similar to the Optimize Reconstruction error KL: encourages our learned distribution q(z|x) to be similar to the true prior distribution p(z)

reparameterization trick Backpropagation can be applied only to the right network.

Visualization of latent space  we're capable of learning smooth latent state representations of the input data

02 some experiments 接下来是第三部分,我们当前完成的工作进度。

VAE for SRE i-vector模型完美地符合PLDA模型的高斯假 设。而对于d-vector模型,无论是句子级别还是说话人级别,其呈现出明显的非高 斯性。显然,d-vector模型不符合PLDA模型的高斯假设,因此其PLDA打分结果 并不理想

VAE for SRE D-vector & VAE encoder Test on SITW dev: Cosine EER | 31.77% PLDA EER | 17.4% LDA_PLDA_EER | 14.52% Test on SITW eval Cosine EER | 32.5% PLDA EER | 18.75% LDA_PLDA_EER | 15.28%

03 Tied VAE for SRE 接下来是最后一个部分,我们的未来计划

Tied VAE This paper generalize the PLDA model to let i-vectors depend non-linearly on the latent factors. Using Tied variational autoencoder as i-vector back-end. VAE framework assume an approximate posterior that belongs a parametric family of distributions. This is an extension of the original VAE formulation, where we added the speaker factor variable yi which is tied across samples of the same speaker

Tied VAE

reference [0] https://www.jeremyjordan.me/autoencoders/ [1] https://www.jeremyjordan.me/variational-autoencoders/ [2] https://arxiv.org/pdf/1606.05908.pdf [3] https://arxiv.org/pdf/1312.6114.pdf …………..

That’s all, thank you! 谢谢老师,我们的介绍结束了!