Markov Logic and Deep Networks Pedro Domingos Dept. of Computer Science & Eng. University of Washington.

Slides:



Advertisements
Similar presentations
Unsupervised Semantic Parsing
Advertisements

Joint Inference in Information Extraction Hoifung Poon Dept. Computer Science & Eng. University of Washington (Joint work with Pedro Domingos)
1 Unsupervised Ontology Induction From Text Hoifung Poon Dept. Computer Science & Eng. University of Washington (Joint work with Pedro Domingos)
1 Gesture recognition Using HMMs and size functions.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Discriminative Structure and Parameter.
Coreference Based Event-Argument Relation Extraction on Biomedical Text Katsumasa Yoshikawa 1), Sebastian Riedel 2), Tsutomu Hirao 3), Masayuki Asahara.
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Markov Logic Networks: Exploring their Application to Social Network Analysis Parag Singla Dept. of Computer Science and Engineering Indian Institute of.
Relational Representations Daniel Lowd University of Oregon April 20, 2015.
Markov Logic Networks Instructor: Pedro Domingos.
1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong.
Review Markov Logic Networks Mathew Richardson Pedro Domingos Xinran(Sean) Luo, u
Efficient Weight Learning for Markov Logic Networks Daniel Lowd University of Washington (Joint work with Pedro Domingos)
Markov Logic: A Unifying Framework for Statistical Relational Learning Pedro Domingos Matthew Richardson
Speaker:Benedict Fehringer Seminar:Probabilistic Models for Information Extraction by Dr. Martin Theobald and Maximilian Dylla Based on Richards, M., and.
CPSC 322, Lecture 23Slide 1 Logic: TD as search, Datalog (variables) Computer Science cpsc322, Lecture 23 (Textbook Chpt 5.2 & some basic concepts from.
CPSC 322, Lecture 19Slide 1 Propositional Logic Intro, Syntax Computer Science cpsc322, Lecture 19 (Textbook Chpt ) February, 23, 2009.
Supervised learning Given training examples of inputs and corresponding outputs, produce the “correct” outputs for new inputs Two main scenarios: –Classification:
The Unreasonable Effectiveness of Data Alon Halevy, Peter Norvig, and Fernando Pereira Kristine Monteith May 1, 2009 CS 652.
CPSC 322, Lecture 23Slide 1 Logic: TD as search, Datalog (variables) Computer Science cpsc322, Lecture 23 (Textbook Chpt 5.2 & some basic concepts from.
Statistical Relational Learning Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Recursive Random Fields Daniel Lowd University of Washington June 29th, 2006 (Joint work with Pedro Domingos)
CSE 574: Artificial Intelligence II Statistical Relational Learning Instructor: Pedro Domingos.
Architecture for Pattern- Base Management Systems Manolis TerrovitisPanos Vassiliadis National Technical Univ. of Athens, Dept. of Electrical and Computer.
Semantics For the Semantic Web: The Implicit, the Formal and The Powerful Amit Sheth, Cartic Ramakrishnan, Christopher Thomas CS751 Spring 2005 Presenter:
Recursive Random Fields Daniel Lowd University of Washington (Joint work with Pedro Domingos)
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Relational Models. CSE 515 in One Slide We will learn to: Put probability distributions on everything Learn them from data Do inference with them.
Markov Logic Networks: A Unified Approach To Language Processing Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with.
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Statistical Relational Learning Pedro Domingos Dept. Computer Science & Eng. University of Washington.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Introduction to Machine Learning Approach Lecture 5.
Named Entity Recognition and the Stanford NER Software Jenny Rose Finkel Stanford University March 9, 2007.
Artificial Intelligence Research Centre Program Systems Institute Russian Academy of Science Pereslavl-Zalessky Russia.
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Pedro Domingos Dept. of Computer Science & Eng.
Some Thoughts to Consider 6 What is the difference between Artificial Intelligence and Computer Science? What is the difference between Artificial Intelligence.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Markov Logic Parag Singla Dept. of Computer Science University of Texas, Austin.
Machine Learning For the Web: A Unified View Pedro Domingos Dept. of Computer Science & Eng. University of Washington Includes joint work with Stanley.
Extracting Semantic Constraint from Description Text for Semantic Web Service Discovery Dengping Wei, Ting Wang, Ji Wang, and Yaodong Chen Reporter: Ting.
L’età della parola Giuseppe Attardi Dipartimento di Informatica Università di Pisa ESA SoBigDataPisa, 24 febbraio 2015.
Annotating Words using WordNet Semantic Glosses Julian Szymański Department of Computer Systems Architecture, Faculty of Electronics, Telecommunications.
Markov Logic And other SRL Approaches
Transfer in Reinforcement Learning via Markov Logic Networks Lisa Torrey, Jude Shavlik, Sriraam Natarajan, Pavan Kuppili, Trevor Walker University of Wisconsin-Madison,
Machine Learning.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Markov Logic Networks Pedro Domingos Dept. Computer Science & Eng. University of Washington (Joint work with Matt Richardson)
Simultaneously Learning and Filtering Juan F. Mancilla-Caceres CS498EA - Fall 2011 Some slides from Connecting Learning and Logic, Eyal Amir 2006.
Modeling Speech Acts and Joint Intentions in Modal Markov Logic Henry Kautz University of Washington.
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
CPSC 322, Lecture 19Slide 1 (finish Planning) Propositional Logic Intro, Syntax Computer Science cpsc322, Lecture 19 (Textbook Chpt – 5.2) Oct,
Pedro Domingos University of Washington. Traditional Programming Machine Learning Computer Data Algorithm Output Computer Data Output Algorithm.
Markov Logic Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Happy Mittal (Joint work with Prasoon Goyal, Parag Singla and Vibhav Gogate) IIT Delhi New Rules for Domain Independent Lifted.
Progress Report ekker. Problem Definition In cases such as object recognition, we can not include all possible objects for training. So transfer learning.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
The Hebrew University of Jerusalem School of Engineering and Computer Science Academic Year: 2011/2012 Instructor: Jeff Rosenschein.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Pedro Domingos University of Washington. Evolution Experience Culture.
An Introduction to Markov Logic Networks in Knowledge Bases
Convolutional Neural Network
CH. 1: Introduction 1.1 What is Machine Learning Example:
Machine Learning Ali Ghodsi Department of Statistics
Unsupervised Learning and Autoencoders
Introduction Task: extracting relational facts from text
Natural Language to SQL(nl2sql)
Artificial Intelligence 2004 Speech & Natural Language Processing
Presentation transcript:

Markov Logic and Deep Networks Pedro Domingos Dept. of Computer Science & Eng. University of Washington

Markov Logic Networks Basic idea: Use first-order logic to compactly specify large non-i.i.d. models MLN = Set of formulas with weights Formula = Feature template (Vars→Objects) E.g., HMM: Weight of formula iNo. of true instances of formula i in x State(+s,t) ^ State(+s',t+1) Obs(+o,t) ^ State(+s,t)

State of the Art in MLNs Many algorithms for learning and inference Inference: Millions of variables, billions of features Learning: Generative, discriminative, max-margin, etc. Best-performing solutions in many application areas Natural language, robot mapping, social networks, computational biology, activity recognition, etc. Open-source software/Web site: Alchemy Book: Domingos & Lowd, Markov Logic, Morgan & Claypool, alchemy.cs.washington.edu

Deep Uses of MLNs Very large scale inference Defining architecture of deep networks Adding knowledge to deep networks Transition to natural language input Learning with many levels of hidden variables

MLNs for Deep Learning Basic idea: Use small amounts of knowledge and large amounts of joint inference to make up for lack of supervision Relational clustering: Cluster objects with similar relations to similar objects Cluster relations that hold between similar sets of objects Coreference resolution: Outperforms supervised approaches on MUC and ACE benchmarks Semantic network extraction: Learns thousands of concepts and relations from millions of Web triples

Example: Unsupervised Semantic Parsing Goal: Read text and answer questions No supervision or annotation Input: Dependency parses of sentences (Nodes→Unary predicates / Edges→Binary preds.) Outputs: Semantic parser and knowledge base Basic idea: Cluster expressions with similar subexpressions Maps syntactic variants to common meaning Discovers its own meaning representation “Part of” lattice of clusters Applied to corpus of biomedical abstracts Three times more correct answers than next best