Funny Factory Mike Cialowicz Zeid Rusan Matt Gamble Keith Harris.

Slides:



Advertisements
Similar presentations
Three Basic Problems Compute the probability of a text: P m (W 1,N ) Compute maximum probability tag sequence: arg max T 1,N P m (T 1,N | W 1,N ) Compute.
Advertisements

School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Machine Learning PoS-Taggers COMP3310 Natural Language Processing Eric.
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING PoS-Tagging theory and terminology COMP3310 Natural Language Processing.
Expectation Maximization Dekang Lin Department of Computing Science University of Alberta.
Three Basic Problems 1.Compute the probability of a text (observation) language modeling – evaluate alternative texts and models P m (W 1,N ) 2.Compute.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
Ling 570 Day 6: HMM POS Taggers 1. Overview Open Questions HMM POS Tagging Review Viterbi algorithm Training and Smoothing HMM Implementation Details.
CPSC 422, Lecture 16Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 16 Feb, 11, 2015.
Albert Gatt Corpora and Statistical Methods – Lecture 7.
Prototype-Driven Grammar Induction Aria Haghighi and Dan Klein Computer Science Division University of California Berkeley.
A random field…
Modeling the Evolution of Product Entities Priya Radhakrishnan 1, Manish Gupta 1,2, Vasudeva Varma 1 1 Search and Information Extraction Lab, IIIT-Hyderabad,
Part of Speech Tagging The DT students NN went VB to P class NN Plays VB NN well ADV NN with P others NN DT Fruit NN flies NN VB NN VB like VB P VB a DT.
Towards Separating Trigram- generated and Real sentences with SVM Jerry Zhu CALD KDD Lab 2001/4/20.
Recognizing Literary Tropes in Plot Synopses Sam Carton April 12, 2013.
Applications of Sequence Learning CMPT 825 Mashaal A. Memon
1 Quasi-Synchronous Grammars  Based on key observations in MT: translated sentences often have some isomorphic syntactic structure, but not usually in.
Learning Bit by Bit Hidden Markov Models. Weighted FSA weather The is outside
Syllabus Text Books Classes Reading Material Assignments Grades Links Forum Text Books עיבוד שפות טבעיות - שיעור חמישי POS Tagging Algorithms עידו.
1 Complementarity of Lexical and Simple Syntactic Features: The SyntaLex Approach to S ENSEVAL -3 Saif Mohammad Ted Pedersen University of Toronto, Toronto.
Maximum Entropy Model LING 572 Fei Xia 02/07-02/09/06.
Maximum Entropy Model LING 572 Fei Xia 02/08/07. Topics in LING 572 Easy: –kNN, Rocchio, DT, DL –Feature selection, binarization, system combination –Bagging.
CS224N Interactive Session Competitive Grammar Writing Chris Manning Sida, Rush, Ankur, Frank, Kai Sheng.
TopicTrend By: Jovian Lin Discover Emerging and Novel Research Topics.
Answering List Questions using Co-occurrence and Clustering Majid Razmara and Leila Kosseim Concordia University
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
CSA2050: Introduction to Computational Linguistics Part of Speech (POS) Tagging II Transformation Based Tagging Brill (1995)
HW7 Extracting Arguments for % Ang Sun March 25, 2012.
Methods for the Automatic Construction of Topic Maps Eric Freese, Senior Consultant ISOGEN International.
Lecture 10 NLTK POS Tagging Part 3 Topics Taggers Rule Based Taggers Probabilistic Taggers Transformation Based Taggers - Brill Supervised learning Readings:
Part-of-Speech Tagging Foundation of Statistical NLP CHAPTER 10.
AQUAINT Workshop – June 2003 Improved Semantic Role Parsing Kadri Hacioglu, Sameer Pradhan, Valerie Krugler, Steven Bethard, Ashley Thornton, Wayne Ward,
Funny Factory Keith Harris Matt Gamble Mike Cialowicz Zeid Rusan.
A Systematic Exploration of the Feature Space for Relation Extraction Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois,
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-16: Probabilistic parsing; computing probability of.
NLP. Introduction to NLP Background –From the early ‘90s –Developed at the University of Pennsylvania –(Marcus, Santorini, and Marcinkiewicz 1993) Size.
Exploiting Reducibility in Unsupervised Dependency Parsing David Mareček and Zdeněk Žabokrtský Institute of Formal and Applied Linguistics Charles University.
Word classes and part of speech tagging. Slide 1 Outline Why part of speech tagging? Word classes Tag sets and problem definition Automatic approaches.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-15: Probabilistic parsing; PCFG (contd.)
NLP. Introduction to NLP #include int main() { int n, reverse = 0; printf("Enter a number to reverse\n"); scanf("%d",&n); while (n != 0) { reverse =
Prototype-Driven Grammar Induction Aria Haghighi and Dan Klein Computer Science Division University of California Berkeley.
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 25– Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 14 th March,
12 x 12 = x 11 = x 10 = x 9 = 108.
CSC 594 Topics in AI – Natural Language Processing
Statistical NLP Winter 2009
Introduction to Machine Learning and Text Mining
Discrete Probability Distributions
DESIGN OF EXPERIMENTS.
A POWERPOINT MADE TO KEEP YOU SAFE ONLINE!
General principles in building a predictive model
SAD: 6º Projecto.
,..,_ -, =.. t :: ;... t---aM 0 d -...c­ " ss ''- !j!j ! )
CSCI 5832 Natural Language Processing
LING/C SC 581: Advanced Computational Linguistics
web1T and deep learning methods
Chapter 23 – Prepositions
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
Modeling in the Time Domain
N-Gram Model Formulas Word sequences Chain rule of probability
Word Embedding Word2Vec.
Reinforcement Learning in MDPs by Lease-Square Policy Iteration
BODY SAFETY RULES KINDERGARTEN.
Brian Nisonger Shauna Eggers Joshua Johanson
Evaluating Logarithms
CPSC-608 Database Systems
Entry Routine Objective: SWBAT evaluate functions, graph functions, and solve equations. Quote of the Day: “Legacy. What is a legacy? It’s planting seeds.
Objective- To graph a relationship in a table.
Descriptive statistics Pearson’s correlation
Part-of-Speech Tagging Using Hidden Markov Models
Example of training and deployment of deep convolutional neural networks. Example of training and deployment of deep convolutional neural networks. During.
Presentation transcript:

Funny Factory Mike Cialowicz Zeid Rusan Matt Gamble Keith Harris

1- To explore strange new worlds. Our Missions: 1- To explore strange new worlds. 2- Given an inputed sentence, output the statistically funniest response based on comedic data. Our Approach: 1- Learn from relationships between words in jokes. 2- Learn from sentence structures of jokes. “On Screen!”

Step 1: Collect data (2.5 MB) Setup 1: “I feel bad going behind Lois' back.” Setup 2: “Don't feel bad Peter.” Zinger!: “Oh I never thought of it like that!” .

Step 2: Tag the jokes (Size = 3.5MB) “I feel bad going behind Lois' back.” “Don't feel bad Peter.” /VB /NN /JJ /NNP “Oh I never thought of it like that!” /UH /PRP /RB /VBD /IN /PRP /IN /DT Attach: /PRP /VBP /JJ /NN /IN /NNP /RB Attach: Attach: “Who tagged that there?”

Step 3a: Zinger word counts (100 MB) I feel bad going behind Lois' back For each word : Count! For word 'feel' : Intuition: Word relations in Zingers should help us construct our own!

Step 3b: Cross sentence counts (## MB) For each adjacent pair in setups : Don't feel bad Peter Count! : Oh I never thought of it like that! For 'feel,bad ' : Intuition: Words in input should help us place a seed word in Zingers we are constructing!

Step 3c: Structure counts (2.2 MB) Oh I never thought of it like that! /UH /PRP /RB /VBD /IN /PRP /IN /DT For each sentence : Count! : Intuition: Using known funny Zinger structures should yield funnier constructed Zingers.

Step 4: Smoothing! Converted dictionary counts to probabilities using: Laplace smoothing (k = 1) Lidstone's law (k = 0.5, 0.05) “Damn that's smooth”

Step 5: Make a sentence! This is an example sense makes sense Input sentence : This is an example sense makes sense /DT makes sense “This makes sense” Get seed word : Highest Prob Generate more words : Highest Prob Get a structure : Highest Prob Complete sentence : Highest Prob

Step 6: DEMO! 5/11/2006 @ 4:13 am in the Linux Lab “YEAH BOYYYYYYYY!”

Step 7: Future Work - Incorporate semantics. - Collect MORE data. (Need a better computer) - Apply weights to cross sentence counts - Evaluate using test subjects (mainly Billy) with different combinations of weight and probability (k = #) parameters. - Do parameters converge along with funny? - Reevaluate using the (better?) parameters.