Deep Learning for the Soft Cutoff Problem

Slides:



Advertisements
Similar presentations
Text Classification: An Implementation Project Prerak Sanghvi Computer Science and Engineering Department State University of New York at Buffalo.
Advertisements

Sufficient Dimensionality Reduction with Irrelevance Statistics Amir Globerson 1 Gal Chechik 2 Naftali Tishby 1 1 Center for Neural Computation and School.
Important Task in Patents Retrieval Recall is an Important Factor Given Query Patent -> the Task is to Search all Related Patents Patents have Complex.
EVALUATION David Kauchak CS 451 – Fall Admin Assignment 3 - change constructor to take zero parameters - instead, in the train method, call getFeatureIndices()
A Neural Network Approach to Predicting Stock Performance John Piefer ECE/CS 539 Project Presentation.
IMSS005 Computer Science Seminar
윤언근 DataMining lab.  The Web has grown exponentially in size but this growth has not been isolated to good-quality pages.  spamming and.
Artificial Neural Networks Approach to Stock Prediction Presented by Justin Jaeck.
Principals of Research Writing. What is Research Writing? Process of communicating your research  Before the fact  Research proposal  After the fact.
1 What Makes a Query Difficult? David Carmel, Elad YomTov, Adam Darlow, Dan Pelleg IBM Haifa Research Labs SIGIR 2006.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
1 Learning to Impress in Sponsored Search Xin Supervisors: Prof. King and Prof. Lyu.
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Harnessing the Deep Web : Present and Future -Tushar Mhaskar Jayant Madhavan, Loredana Afanasiev, Lyublena Antova, Alon Halevy January 7,
Sparse Coding: A Deep Learning using Unlabeled Data for High - Level Representation Dr.G.M.Nasira R. Vidya R. P. Jaia Priyankka.
When deep learning meets object detection: Introduction to two technologies: SSD and YOLO Wenchi Ma.
Big data classification using neural network
TensorFlow– A system for large-scale machine learning
Deep Learning for Dual-Energy X-Ray
Maximum Entropy Models and Feature Engineering CSCI-GA.2591
Quantum Simulation Neural Networks
Computer Science and Engineering, Seoul National University
Wenhan Xiong, Thien Hoang, William Wang Department of Computer Science
Pick samples from task t
Multimodal Learning with Deep Boltzmann Machines
Intro to NLP and Deep Learning
Natural Language Processing of Knee MRI Reports
Chapter 15 QUERY EXECUTION.
MLP Based Feedback System for Gas Valve Control in a Madison Symmetric Torus Andrew Seltzman Dec 14, 2010.
Using Tensorflow to Detect Objects in an Image
Master’s Thesis defense Ming Du Advisor: Dr. Yi Shang
Lecture 23: Feature Selection
Bird-species Recognition Using Convolutional Neural Network
Distributed Learning of Multilingual DNN Feature Extractors using GPUs
Hyperparameters, bias-variance tradeoff, validation
Research Interests.
Final Presentation: Neural Network Doc Summarization
Exploring Matching Networks for Cross Language Information Retrieval
Goodfellow: Chapter 14 Autoencoders
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Interpret the execution mode of SQL query in F1 Query paper
MATERIAL Resources for Cross-Lingual Information Retrieval
Neural Networks Geoff Hulten.
Deep Robust Unsupervised Multi-Modal Network
Resource Recommendation for AAN
Semantic Similarity Detection
Dialogue State Tracking & Dialogue Corpus Survey
Unsupervised Pretraining for Semantic Parsing
Natural Language to SQL(nl2sql)
Using Multilingual Neural Re-ranking Models for Low Resource Target Languages in Cross-lingual Document Detection Using Multilingual Neural Re-ranking.
Dennis Zhao,1 Dragomir Radev PhD1 LILY Lab
Sequential Question Answering Using Neural Coreference Resolution
Stylistic Author Attribution Methods for Identifying Parody Trump Tweets Isaac Pena Department of Computer Science, Yale University, New Haven, CT LILY.
Visual Recognition of American Sign Language Using Hidden Markov Models 문현구 문현구.
Department of Computer Science Ben-Gurion University of the Negev
Department of Computer Science Ben-Gurion University of the Negev
Lab 2: Information Retrieval
Automatic Handwriting Generation
The Updated experiment based on LSTM
Introduction to Neural Networks
Object Detection Implementations
Single Parameter Tuning
LHC beam mode classification
Cross-lingual Information Retrieval (CLIR) Johns Hopkins University
Austin Karingada, Jacob Handy, Adviser : Dr
Goodfellow: Chapter 14 Autoencoders
Machine Learning for Cyber
Evaluation David Kauchak CS 158 – Fall 2019.
Presentation transcript:

Deep Learning for the Soft Cutoff Problem Miles Saffran Department of Computer Science, Yale University, New Haven, CT LILY Lab Introduction Results The MATERIAL project, commissioned by IARPA, seeks to create a search engine capable of retrieving content in low-resource languages using English queries. An important component of this project is discerning which results are relevant enough to be included. Though there might be 100 documents associated with a certain query, users are presumably not interested in every particular document. The metric being used to assess the quality of a cutoff system, AQWV, penalizes both for missing relevant documents and for returning irrelevant documents. The training data provided to us comes with a means of calculating the cutoff that maximizes AQWV; however, this information is undisclosed on the private cases the contractor will run. My project attempts to predict the optimal cutoff score based on AQWV by using a multi-layered neural network to learn certain characteristics about query/document pairs and successfully predict the best cutoff. Figures 3 and 4 demonstrate the loss of the model with various hyper-parameters tested, while Figure 5 shows the test results of training on ANALYSIS-EN and testing on a set from DEV-EN. Generally speaking, the results had high levels of variance depending on the training and testing data as well as the parameters of the model. The overall AQWV score of the network when trained on one English dataset and tested on another was 0.1, which was close to the optimal score of .1408. However, when performing cross-lingual training (training on one language and testing on another), the model underperformed with an overall AQWV score of .151 compared to an optimal .3575. More specifically, the performance of the model improved drastically when I added two features. First, using word embedding of queries (sourcing a script written by Caitlin Westerfield) as a feature helped improve model accuracy. Second, adding another hidden layer allowed the model to perform better on cross-lingual tasks than it previously had, indicating that the model was actually capable of picking up on generalizable features related to the input. Figure 3. French loss in the simple model Figure 1. Training loss over epochs Materials and Methods The data – query and document pairs – were training examples provided to the MATERIAL group at Yale. To format the information in a structure useful for a neural network, I wrote code to extract certain features for input and used a program written by Javid Dadashkarimi to calculate the ideal cutoff point. These features include document length, relevant Indri score for each document, and query embedding in GloVe. I then used the TensorFlow library to create a neural network with a hidden layer and softmax activation function. Once this was set up, I iterated through different hyperparameters like learning rate and hidden layer size to optimize the model. Upon further experimentation, I added another hidden layer to the network to compress the input and avoid overfitting. For each language I was training on, I saved the best performing results(those with lowest loss) and used it to predict cutoff points on test data. Conclusion Ultimately, though the network in its current iteration is not consistently successful in finding the optimal cutoff, initial results indicate room for improvement. Adding dropout and regularization might help improve model performance while adding more training and test data could amplify the learning process. In addition, collecting other features to use as input (Duet scores, etc) might yield a more robust model capable of generalized learning both within and across languages. Figure 2. English loss with different learning rates Figure 4. English loss with the complex model Acknowledgement Figure 5. Output of the ANALYSIS-EN set tested on DEV-EN Thanks to Prof. Radev for advising, Javid and Jungo for helping me navigate the amount of material in Material and reviewing my code, and Caitlin for her word embedding script.