Effective Reranking for Extracting Protein-protein Interactions from Biomedical Literature Deyu Zhou, Yulan He and Chee Keong Kwoh School of Computer Engineering.

Slides:



Advertisements
Similar presentations
Three Basic Problems Compute the probability of a text: P m (W 1,N ) Compute maximum probability tag sequence: arg max T 1,N P m (T 1,N | W 1,N ) Compute.
Advertisements

Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Three Basic Problems 1.Compute the probability of a text (observation) language modeling – evaluate alternative texts and models P m (W 1,N ) 2.Compute.
C O N T E X T - F R E E LANGUAGES ( use a grammar to describe a language) 1.
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Prof. Carolina Ruiz Computer Science Department Bioinformatics and Computational Biology Program WPI WELCOME TO BCB4003/CS4803 BCB503/CS583 BIOLOGICAL.
Part of Speech Tagging The DT students NN went VB to P class NN Plays VB NN well ADV NN with P others NN DT Fruit NN flies NN VB NN VB like VB P VB a DT.
Contextual Advertising by Combining Relevance with Click Feedback D. Chakrabarti D. Agarwal V. Josifovski.
1 Statistical NLP: Lecture 12 Probabilistic Context Free Grammars.
In Search of a More Probable Parse: Experiments with DOP* and the Penn Chinese Treebank Aaron Meyers Linguistics 490 Winter 2009.
O PTICAL C HARACTER R ECOGNITION USING H IDDEN M ARKOV M ODELS Jan Rupnik.
Re-ranking for NP-Chunking: Maximum-Entropy Framework By: Mona Vajihollahi.
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Parsing with CFG Ling 571 Fei Xia Week 2: 10/4-10/6/05.
Hidden Markov Models. Hidden Markov Model In some Markov processes, we may not be able to observe the states directly.
1 Hidden Markov Model Instructor : Saeed Shiry  CHAPTER 13 ETHEM ALPAYDIN © The MIT Press, 2004.
Presented by Zeehasham Rasheed
Fast Temporal State-Splitting for HMM Model Selection and Learning Sajid Siddiqi Geoffrey Gordon Andrew Moore.
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
Sequence labeling and beam search LING 572 Fei Xia 2/15/07.
Scalable Text Mining with Sparse Generative Models
(Some issues in) Text Ranking. Recall General Framework Crawl – Use XML structure – Follow links to get new pages Retrieve relevant documents – Today.
Introduction to Machine Learning Approach Lecture 5.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Keyphrase Extraction in Scientific Documents Thuy Dung Nguyen and Min-Yen Kan School of Computing National University of Singapore Slides available at.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
1 The Hidden Vector State Language Model Vidura Senevitratne, Steve Young Cambridge University Engineering Department.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
7-Speech Recognition Speech Recognition Concepts
Information Extraction From Medical Records by Alexander Barsky.
Recognizing Names in Biomedical Texts: a Machine Learning Approach GuoDong Zhou 1,*, Jie Zhang 1,2, Jian Su 1, Dan Shen 1,2 and ChewLim Tan 2 1 Institute.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
BioRAT: Extracting Biological Information from Full-length Papers David P.A. Corney, Bernard F. Buxton, William B. Langdon and David T. Jones Bioinformatics.
Dependency Parser for Swedish Project for EDA171 by Jonas Pålsson Marcus Stamborg.
Next Generation Search Engines Ehsun Daroodi 1 Feb, 2003.
Chapter 23: Probabilistic Language Models April 13, 2004.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Mining Logs Files for Data-Driven System Management Advisor.
Statistical Decision-Tree Models for Parsing NLP lab, POSTECH 김 지 협.
Supertagging CMSC Natural Language Processing January 31, 2006.
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
Automatic Grammar Induction and Parsing Free Text - Eric Brill Thur. POSTECH Dept. of Computer Science 심 준 혁.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1Computer Sciences Department. Book: INTRODUCTION TO THE THEORY OF COMPUTATION, SECOND EDITION, by: MICHAEL SIPSER Reference 3Computer Sciences Department.
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. 1 Mining knowledge from natural language texts using fuzzy associated concept mapping Presenter : Wu,
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
NTU & MSRA Ming-Feng Tsai
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Statistical techniques for video analysis and searching chapter Anton Korotygin.
Department of Computer Science The University of Texas at Austin USA Joint Entity and Relation Extraction using Card-Pyramid Parsing Rohit J. Kate Raymond.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Maximum Entropy techniques for exploiting syntactic, semantic and collocational dependencies in Language Modeling Sanjeev Khudanpur, Jun Wu Center for.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Chinese Named Entity Recognition using Lexicalized HMMs.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Natural Language Processing : Probabilistic Context Free Grammars Updated 8/07.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Roadmap Probabilistic CFGs –Handling ambiguity – more likely analyses –Adding probabilities Grammar Parsing: probabilistic CYK Learning probabilities:
N-best list reranking using higher level phonetic, lexical, syntactic and semantic knowledge sources Mithun Balakrishna, Dan Moldovan and Ellis K. Cave.
CSC 594 Topics in AI – Natural Language Processing
CSCI 5832 Natural Language Processing
Hierarchical, Perceptron-like Learning for OBIE
Stance Classification of Ideological Debates
Presentation transcript:

Effective Reranking for Extracting Protein-protein Interactions from Biomedical Literature Deyu Zhou, Yulan He and Chee Keong Kwoh School of Computer Engineering Nanyang Technological University, Singapore 30 August 2007

Outline Protein-protein interactions (PPIs) extraction Hidden Vector State (HVS) model for PPIs extraction Reranking approaches Experimental results Conclusions

Protein Interact Protein Protein-Protein Interactions Extraction Spc97p interacts with Spc98 and Tub4 in the two-hybrid system Spc97p interact Spc98 Spc97p interact Tub4

Existing Approaches Statistics Methods Pattern Matching Parsing- Based Simple to Complicated

An example

Statistics-Based Approaches

Pattern Matching Approaches

Parsing-Based Approaches

Semantic Parser Ĉ = argmax { P(C|W n ) } = argmax { P(C) P(W n |C) } c For each candidate word string W n, need to compute most likely set of embedded concepts semantic model lexical model

We could use a simple finite state tagger … P(W n |C) P(C) … can be robustly trained using EM, but model is too weak to represent embeddings in natural language

Perhaps use some form of hierarchical HMM in which each state is a terminal or a nested HMM … … but when using EM, models rarely converge on good solutions and, in practice, direct maximum-likelihood from “tree-bank” data are needed to train models P(W n |C) P(C)

Hidden Vector State Model

The HVS model is an HMM in which the states correspond to the stack of a push-down automata with a bounded stack size … P(W n |C) … this is a very convenient framework for applying constraints P(C)

HVS model transition constraints: finite stack depth – D push only one non-terminal semantic onto the stack at each step … model defined by three simple probability tables Ĉ = argmax { ∏P(n t |C t-1 ) P(C t [1]|C t [2..D t ]) P(W t |C t ) } c,N t

Parsing with the HVS model P(n t |C t-1 ) 1) POP 1 elements from the previous stack state, n =1 P(C t [1]|C t [2..D t ]) 2) Push 1 pre-terminal semantic concept into stack P(W t |C t ) 3) Generate the next word PROTEIN INTERACT PROTEIN SS … with Spc98 and Tub4 … INTERACT PROTEIN SS DUMMY INERACT PROTEIN SS

Train using EM and apply constraints Abstract semantic annotation PROTEIN ( INTERACT ( PROTEIN) ) CUL-1 was found to interact with SKR-1, SKR-2, SKR-3, and SKR-7 in yeast two-hybrid system Training text Data Constraints EM Parameter Estimation HVS Model Parameters Parse Statistics Limit forward- backward search to only include states which are consistent with the constraints

Reranking Methodology Reranking approaches attempts to improve upon an existing probabilistic parser by reranking the output of the parser. It has benefited applications such as name-entity extraction, semantic parsing and semantic labeling. To rerank parses generated by the HVS model for protein-protein interactions extraction

Architecture

Abstract Annotation need to provide the for each sentence in the training set. An example –Sentences: CUL-1 was found to interact with SKR-1, SKR-2, SKR-3, SKR-7, SKR-8 and SKR-10 in yeast two-hybrid system. Annotation: –PORTEIN NAME(ACTIVATE (PROTEIN NAME)) Results PPI : –CUL-1#SKR-1#ACTIVATE –CUL-1#SKR-2#ACTIVATE –CUL-1#SKR-3#ACTIVATE –CUL-1#SKR-7#ACTIVATE –CUL-1#SKR-8#ACTIVATE –CUL-1#SKR-10#ACTIVATE

Reranking approaches Features for Reranking Suppose sentence S i has its corresponding parse set C i = {C ij, j = 1,.. N} –Parsing Information –Structure Information –Complexity Information

Reranking approaches Score is defined as log-linear regression model Neural Network Support Vector Machines

Experiments Setup –Corpus I comprises of 300 abstracts randomly retrieved from the GENIA corpus GENIA is a collection of research abstracts selected from the search results of MEDLINE database with keyword (MeSH terms) “human, blood cells and transcription factors” split into two parts: –Part I contains 1500 sentences (training data) –Part II consists of 1000 sentences (test data)

Experimental Results Figure 1: F-measure vs number of candidate parses.

Experimental Results (cont ’ d) Experime nts Recall (%) Precision (%) F-Score (%) Baseline SVM NN LLR Table 3: Results based on the interaction category.

Conclusions Three reranking methods for the HVS model in the application of extracting protein-protein interactions from biomedical literature. Experimental results show that 4% relative improvement in F-measure can be obtained through reranking on the semantic parse results Incorporating other semantic or syntactic information might be able to give further gains.