Domain Adaptation for Biomedical Information Extraction Jing Jiang BeeSpace Seminar Oct 17, 2007.

Slides:



Advertisements
Similar presentations
Knowledge Transfer via Multiple Model Local Structure Mapping Jing Gao Wei Fan Jing JiangJiawei Han University of Illinois at Urbana-Champaign IBM T. J.
Advertisements

Knowledge Transfer via Multiple Model Local Structure Mapping Jing Gao, Wei Fan, Jing Jiang, Jiawei Han l Motivate Solution Framework Data Sets Synthetic.
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Rule extraction in neural networks. A survey. Krzysztof Mossakowski Faculty of Mathematics and Information Science Warsaw University of Technology.
Machine learning continued Image source:
Mining External Resources for Biomedical IE Why, How, What Malvina Nissim
Automatic Identification of Cognates, False Friends, and Partial Cognates University of Ottawa, Canada University of Ottawa, Canada.
Multi-Task Transfer Learning for Weakly- Supervised Relation Extraction Jing Jiang Singapore Management University ACL-IJCNLP 2009.
A Two-Stage Approach to Domain Adaptation for Statistical Classifiers Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois.
Frustratingly Easy Domain Adaptation
Reduced Support Vector Machine
An Overview of Text Mining Rebecca Hwa 4/25/2002 References M. Hearst, “Untangling Text Data Mining,” in the Proceedings of the 37 th Annual Meeting of.
Learning from Multiple Outlooks Maayan Harel and Shie Mannor ICML 2011 Presented by Minhua Chen.
Artificial Intelligence Research Centre Program Systems Institute Russian Academy of Science Pereslavl-Zalessky Russia.
Introduction to domain adaptation
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Instance Weighting for Domain Adaptation in NLP Jing Jiang & ChengXiang Zhai University of Illinois at Urbana-Champaign June 25, 2007.
Recognition of Multi-sentence n-ary Subcellular Localization Mentions in Biomedical Abstracts G. Melli, M. Ester, A. Sarkar Dec. 6, 2007
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
BeeSpace Informatics Research: From Information Access to Knowledge Discovery ChengXiang Zhai Nov. 7, 2007.
Domain Adaptation in Natural Language Processing Jing Jiang Department of Computer Science University of Illinois at Urbana-Champaign.
Text Classification, Active/Interactive learning.
Information Extraction from Literature Yue Lu BeeSpace Seminar Oct 24, 2007.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Outline Quick review of GS Current problems with GS Our solutions Future work Discussion …
The Necessity of Combining Adaptation Methods Cognitive Computation Group, University of Illinois Experimental Results Title Ming-Wei Chang, Michael Connor.
Exploiting Domain Structure for Named Entity Recognition Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois at Urbana-Champaign.
Xiaoxiao Shi, Qi Liu, Wei Fan, Philip S. Yu, and Ruixin Zhu
Automatically Generating Gene Summaries from Biomedical Literature (To appear in Proceedings of PSB 2006) X. LING, J. JIANG, X. He, Q.~Z. MEI, C.~X. ZHAI,
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
Transfer Learning with Applications to Text Classification Jing Peng Computer Science Department.
Transfer Learning Motivation and Types Functional Transfer Learning Representational Transfer Learning References.
BioSumm A novel summarizer oriented to biological information Elena Baralis, Alessandro Fiori, Lorenzo Montrucchio Politecnico di Torino Introduction text.
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
University of Illinois at Urbana-Champaign BeeSpace Navigator v4.0 and Gene Summarizer beespace.uiuc.edu `
Pattern Recognition April 19, 2007 Suggested Reading: Horn Chapter 14.
Entity Set Expansion in Opinion Documents Lei Zhang Bing Liu University of Illinois at Chicago.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Artificial Intelligence Research Center Pereslavl-Zalessky, Russia Program Systems Institute, RAS.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
BeeSpace Informatics Research: From Information Access to Knowledge Discovery ChengXiang Zhai Nov. 14, 2007.
HAITHAM BOU AMMAR MAASTRICHT UNIVERSITY Transfer for Supervised Learning Tasks.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Copyright Paula Matuszek Kinds of Machine Learning.
Text Categorization by Boosting Automatically Extracted Concepts Lijuan Cai and Tommas Hofmann Department of Computer Science, Brown University SIGIR 2003.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Artificial Intelligence: Research and Collaborative Possibilities a presentation by: Dr. Ernest L. McDuffie, Assistant Professor Department of Computer.
BeeSpace Informatics Research ChengXiang (“Cheng”) Zhai Department of Computer Science Institute for Genomic Biology Statistics Graduate School of Library.
ID Identification in Online Communities Yufei Pan Rutgers University.
Machine Learning Lecture 1: Intro + Decision Trees Moshe Koppel Slides adapted from Tom Mitchell and from Dan Roth.
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Data Mining and Text Mining. The Standard Data Mining process.
BeeSpace Informatics Research
Big data classification using neural network
Sentiment analysis algorithms and applications: A survey
School of Computer Science & Engineering
Done Done Course Overview What is AI? What are the Major Challenges?
Cross Domain Distribution Adaptation via Kernel Mapping
Introductory Seminar on Research: Fall 2017
Introduction Task: extracting relational facts from text
Ying Dai Faculty of software and information science,
Knowledge Transfer via Multiple Model Local Structure Mapping
Topic 5: Cluster Analysis
Extracting Information from Diverse and Noisy Scanned Document Images
Presentation transcript:

Domain Adaptation for Biomedical Information Extraction Jing Jiang BeeSpace Seminar Oct 17, 2007

10/17/072 Outline Why do we need domain adaptation? Solutions:  Intelligent learning methods  Knowledge bases  Expert supervision Connections with BeeSpace V4

10/17/073 Why do we need domain adaptation? Many biomedical information extraction problems are solved by supervised machine learning methods such as support vector machines (SVMs).  Entity recognition  Relation extraction  Sentence categorization In supervised machine learning, it is assumed that the training data and the test data have the same distribution.

10/17/074 Why do we need domain adaptation? Existing labeled training data is often limited to certain domains.  GENIA corpus  human, blood cells, transcription factors  PennBioIE  Genetic variation in malignancy, Cytochrome P450 inhibition  Training data for sentence categorization in gene summarizer  fly Even when the training data is diverse (containing multiple domains), it would still be nice to customize the classifier for the particular target domain that we are working on.

10/17/075 Why do we need domain adaptation? NER TaskTrain → TestF1 to find PER, LOC, ORG from news text NYT → NYT0.855 Reuters → NYT0.641 to find gene/protein from biomedical literature mouse → mouse0.541 fly → mouse0.281

10/17/076 Solutions to domain adaptation Intelligent learning methods  Instance weighting  Feature selection Knowledge bases Expert supervision thesis research future work discussion

10/17/077 Domain adaptive learning methods Two-stage approach Two frameworks  Instance weighting  Feature selection Use of unlabeled data

10/17/078 Intuition Source Domain Target Domain

10/17/079 Goal Target Domain Source Domain

10/17/0710 Start from the source domain Source Domain Target Domain

10/17/0711 Focus on the common part Source Domain Target Domain

10/17/0712 Pick up some part from the target domain Source Domain Target Domain

10/17/0713 Formal formulation? Source Domain Target Domain How to formally formulate these ideas?

10/17/0714 Instance weighting Source Domain Target Domain instance space (each point represents an example) to assign different weights to different instances in the objective function

10/17/0715 Instance weighting Observation source domain target domain

10/17/0716 Instance weighting Observation source domain target domain

10/17/0717 Instance weighting Analysis of domain difference p(x, y) p(x)p(y | x) p s (y | x) ≠ p t (y | x) p s (x) ≠ p t (x) labeling difference instance difference labeling adaptation instance adaptation ?

10/17/0718 Instance weighting Three sets of instances DsDs D t, l D t, u X  D s + D t,l + D t,u ?

10/17/0719 Instance weighting Framework a flexible setup covering both standard methods and new domain adaptive methods labeled source data labeled target data unlabeled target data

10/17/0720 Feature selection Source Domain Target Domain feature space (each point represents a feature) to identify features that behave similarly across domains

10/17/0721 Feature selection Observation Domain-specific features wingless daughterless eyeless apexless … “suffix -less” weighted high in the model trained from fly data Useful for other organisms? in general NO ! May cause generalizable features to be downweighted fly genes

10/17/0722 Feature selection Observation Generalizable features: generalize well in all domains …decapentaplegic and wingless are expressed in analogous patterns in each … …that CD38 is expressed by both neurons and glial cells…that PABPC5 is expressed in fetal brain and in a range of adult tissues. flymouse

10/17/0723 Feature selection Observation Generalizable features: generalize well in all domains …decapentaplegic and wingless are expressed in analogous patterns in each … …that CD38 is expressed by both neurons and glial cells…that PABPC5 is expressed in fetal brain and in a range of adult tissues. flymouse “w i+2 = expressed” is generalizable

10/17/0724 Feature selection Intuition for identification of generalizable features … source domains … -less … expressed … expressed … -less … expressed … -less … expressed … -less … … expressed … -less … flymouseD3D3 DKDK

10/17/0725 Feature selection Framework Matrix A is for feature selection

10/17/0726 Feature selection results on gene/protein recognition

10/17/0727 New directions to explore Knowledge bases Expert supervision

10/17/0728 Knowledge bases – entity recognition Well-documented nomenclatures  Fly, Mouse, Rat FlyMouseRat  Help filter out false positives?  Help select features? Dictionaries of entities  “Dictionary features”  Automatic summarization of nomenclatures?  Automatic identification of good features?

10/17/0729 Knowledge bases – sentence categorization in gene summarizer For fly, the training sentences are automatically extracted from FlyBase. For other organisms, do we have similar resources?

10/17/0730 Expert supervision – entity recognition Computer system selects ambiguous examples for human experts to judge. Computer system asks human experts other questions.  Similar organisms?  Typical surface features? (e.g. cis-regulatory elements, “-RE”) Computer system summarizes possible features from pseudo labeled data, and asks human experts for confirmation.

10/17/0731 Connections to BeeSpace V4 A major challenge in BeeSpace V4 is extraction of new types of entities and relations. Exploiting knowledge bases and expert supervision is especially important. For new types, no labeled data is available even from other domains. Use of bootstrapping methods should be explored.

10/17/0732 New entity types Recognition of many new types will be dictionary based: organism, anatomy, biological process, etc. Recognition of some new types will need some NER techniques: chemical, regulatory element

10/17/0733 New relation types Bootstrapping (?)  Seed patterns from knowledge bases or human experts  Human inspection of newly discovered patterns?

10/17/0734 The end