Carnegie Mellon School of Computer Science Understanding SMT without the “S” (Statistics) Robert Frederking.

Slides:



Advertisements
Similar presentations
What are the S, T, and E in STEM? How are they related?
Advertisements

Simon Fraser University Department of Statistics and Actuarial Sciences Some Random Questions.
PREDICTING BODY COMPOSITION USING SIMPLE MEASUREMENT TECHNIQUES TIM JOHNSON, JAVIER NAVARRO, IDIONG IDONGESIT, MARK WEEKS.
Hazırlayan NEURAL NETWORKS Least Squares Estimation PROF. DR. YUSUF OYSAL.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
Psychology 202b Advanced Psychological Statistics, II February 10, 2011.
SOME SIMPLE MANIPULATIONS OF SOUND USING DIGITAL SIGNAL PROCESSING Richard M. Stern demo August 31, 2004 Department of Electrical and Computer.
The current status of Chinese- English EBMT -where are we now Joy (Ying Zhang) Ralf Brown, Robert Frederking, Erik Peterson Aug 2001.
Modeling the Cost of Misunderstandings in the CMU Communicator System Dan BohusAlex Rudnicky School of Computer Science, Carnegie Mellon University, Pittsburgh,
Temporal Causal Modeling with Graphical Granger Methods
1 A heart fills with loving kindness is a likeable person indeed.
The current status of Chinese-English EBMT research -where are we now Joy, Ralf Brown, Robert Frederking, Erik Peterson Aug 2001.
Visualization of Clusters with a Density-Based Similarity Measure Rebecca Nugent Department of Statistics, Carnegie Mellon University June 9, 2007 Joint.
MT Summit VIII, Language Technologies Institute School of Computer Science Carnegie Mellon University Pre-processing of Bilingual Corpora for Mandarin-English.
Dimensionality Reduction for fMRI Brain Imaging Data Leman Akoglu Carnegie Mellon University, Computer Science Department Abstract Functional Magnetic.
S519: Evaluation of Information Systems Social Statistics Inferential Statistics Chapter 14: linear regression.
SE 501 Software Development Processes Dr. Basit Qureshi College of Computer Science and Information Systems Prince Sultan University Lecture for Week 8.
Students: Nidal Hurani, Ghassan Ibrahim Supervisor: Shai Rozenrauch Industrial Project (234313) Tube Lifetime Predictive Algorithm COMPUTER SCIENCE DEPARTMENT.
Carnegie Mellon School of Computer Science Modeling Website Popularity Competition in the Attention-Activity Marketplace Bruno Ribeiro Christos Faloutsos.
Learning Stable Multivariate Baseline Models for Outbreak Detection Sajid M. Siddiqi, Byron Boots, Geoffrey J. Gordon, Artur W. Dubrawski The Auton Lab.
computer
Regression. Population Covariance and Correlation.
Thomas Knotts. Engineers often: Regress data  Analysis  Fit to theory  Data reduction Use the regression of others  Antoine Equation  DIPPR.
Supplementary Fig. 2. Statistical Classification Analysis Results. Box and whisker plots displaying mean performance metrics returned in the assessment.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Mining Social Network for Personalized Prioritization Language Techonology Institute School of Computer Science Carnegie Mellon University Shinjae.
Structural Equation Models with Directly Observed Variables II James G. Anderson, Ph.D. Purdue University.
Boosting Training Scheme for Acoustic Modeling Rong Zhang and Alexander I. Rudnicky Language Technologies Institute, School of Computer Science Carnegie.
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Characterizing Failure Data in High-Performance-Computing Systems Bianca Schroeder With Garth Gibson and Gary Grider Department of Computer Science Carnegie.
2007 Winter School in Mathematical and Computational Biology Hosted by ARC Centre in Bioinformatics and Institute for Molecular Bioscience The University.
Supercomputing in Social Science Maria Marta Ferreyra Carnegie Mellon University.
How Good is a Model? How much information does AIC give us? –Model 1: 3124 –Model 2: 2932 –Model 3: 2968 –Model 4: 3204 –Model 5: 5436.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Regression Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
CISC Machine Learning for Solving Systems Problems Microarchitecture Design Space Exploration Lecture 4 John Cavazos Dept of Computer & Information.
Fitting normal distribution: ML 1Computer vision: models, learning and inference. ©2011 Simon J.D. Prince.
Copyright © 2003, N. Ahbel Residuals. Copyright © 2003, N. Ahbel Predicted Actual Actual – Predicted = Error Source:
Welcome to Algebra 2! Get out your homework Get out catalogs Get out writing utensils Put bags on the floor Be quiet!!! 3/2/ : Curve Fitting with.
The Idea of the Statistical Test. A statistical test evaluates the "fit" of a hypothesis to a sample.
Lec. 19 – Hypothesis Testing: The Null and Types of Error.
Describing a Score’s Position within a Distribution Lesson 5.
Computacion Inteligente Least-Square Methods for System Identification.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Political Science 30: Political Inquiry. The Magic of the Normal Curve Normal Curves (Essentials, pp ) The family of normal curves The rule of.
 Normal Curves  The family of normal curves  The rule of  The Central Limit Theorem  Confidence Intervals  Around a Mean  Around a Proportion.
Inquiry-Based Learning in Engineering Laboratory Courses Bryan Webler Carnegie Mellon University Keywords: laboratory, project-based learning, active learning.
BIO-PROCESS LAB (B) 2010 KAREN LANCOUR Bio-Process Lab
How Good is a Model? How much information does AIC give us?
CSE 4705 Artificial Intelligence
Pfizer HTS Machine Learning Algorithms: November 2002
Special Topics In Scientific Computing
S519: Evaluation of Information Systems
Finding Lines in Images
ANALYSING DATA.
ريكاوري (بازگشت به حالت اوليه)
Q4 : How does Netflix recommend movies?
Today (2/11/16) Learning objectives (Sections 5.1 and 5.2):
Seminar in Computer Science
Sampling Distribution
Sampling Distribution
Ch. 27: Standard data The reuse of previous times. Advantages
Unitialized, Globally Optimal, Graph-Based Rectilinear Shape Segmentation - The Opposing Metrics Method Computer Science Department – Carnegie Mellon University,
Machine Learning for Actuaries
Data science online training.
Lecture 11 Generalizations of EM.
ENGR 3300 – 505 Advanced Engineering Mathematics
Chapter 7 The Normal Distribution and Its Applications
Scaling collapse. Scaling collapse. Empirical street-covering fractions ⟨C⟩ vs. normalized number of sensor-equipped vehicles NV/⟨B⟩ from the four vehicle-level.
Types of Errors And Error Analysis.
Presentation transcript:

Carnegie Mellon School of Computer Science Understanding SMT without the “S” (Statistics) Robert Frederking

Carnegie Mellon School of Computer Science Statistical modelling Think about statistical modelling as fitting a curve to data points –Start with parameterized function, error metric, and data points –After fitting the function to data using parameters, you can make predictions

Carnegie Mellon School of Computer Science

Carnegie Mellon School of Computer Science

Carnegie Mellon School of Computer Science y = a*x + b Err = sqrt(sum(d i ^2))

Carnegie Mellon School of Computer Science X Y y = a*x + b

Carnegie Mellon School of Computer Science

Carnegie Mellon School of Computer Science

Carnegie Mellon School of Computer Science y = a*x + b Err = sqrt(sum(d i ^2))

Carnegie Mellon School of Computer Science y = a*x + b X Y??

Carnegie Mellon School of Computer Science X Y2 Y1 Err = sqrt(sum(d i ^2)) (Y-y0)^2/a + (X-x0)^2/b = r^2

Carnegie Mellon School of Computer Science Statistical modelling Think about statistical modelling as fitting a curve to data points –Parameterized function, error metric, data points –After fitting parameters, you can make predictions –But you will get some fit for any data set Human researchers need to come up with “good” family of functions, and error metric, for the data you see –Want low error number, good predictions –Tractable, both in training and decoding including data availability, sparseness issues