Recommendation for English multiple-choice cloze questions based on expected test scores 2011, International Journal of Knowledge-Based and Intelligent.

Slides:



Advertisements
Similar presentations
Answering Approximate Queries over Autonomous Web Databases Xiangfu Meng, Z. M. Ma, and Li Yan College of Information Science and Engineering, Northeastern.
Advertisements

Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Fast Algorithms For Hierarchical Range Histogram Constructions
Biointelligence Laboratory, Seoul National University
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
What is Science anyway.
Fast Bayesian Matching Pursuit Presenter: Changchun Zhang ECE / CMR Tennessee Technological University November 12, 2010 Reading Group (Authors: Philip.
Linear Model Incorporating Feature Ranking for Chinese Documents Readability Gang Sun, Zhiwei Jiang, Qing Gu and Daoxu Chen State Key Laboratory for Novel.
Optimal Design Laboratory | University of Michigan, Ann Arbor 2011 Design Preference Elicitation Using Efficient Global Optimization Yi Ren Panos Y. Papalambros.
Prediction of fault-proneness at early phase in object-oriented development Toshihiro Kamiya †, Shinji Kusumoto † and Katsuro Inoue †‡ † Osaka University.
Active Learning and Collaborative Filtering
Visual Recognition Tutorial
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
Computer Science Department Jeff Johns Autonomous Learning Laboratory A Dynamic Mixture Model to Detect Student Motivation and Proficiency Beverly Woolf.
Carnegie Mellon 1 Maximum Likelihood Estimation for Information Thresholding Yi Zhang & Jamie Callan Carnegie Mellon University
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Project  Now it is time to think about the project  It is a team work Each team will consist of 2 people  It is better to consider a project of your.
Expectation Maximization Algorithm
Visual Recognition Tutorial
Student Evaluation System
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
CONTENT-BASED BOOK RECOMMENDING USING LEARNING FOR TEXT CATEGORIZATION TRIVIKRAM BHAT UNIVERSITY OF TEXAS AT ARLINGTON DATA MINING CSE6362 BASED ON PAPER.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Multi-Style Language Model for Web Scale Information Retrieval Kuansan Wang, Xiaolong Li and Jianfeng Gao SIGIR 2010 Min-Hsuan Lai Department of Computer.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Computational Methods to Vocalize Arabic Texts H. Safadi*, O. Al Dakkak** & N. Ghneim**
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Focused Matrix Factorization for Audience Selection in Display Advertising BHARGAV KANAGAL, AMR AHMED, SANDEEP PANDEY, VANJA JOSIFOVSKI, LLUIS GARCIA-PUEYO,
1 Logistic Regression Adapted from: Tom Mitchell’s Machine Learning Book Evan Wei Xiang and Qiang Yang.
Probabilistic Question Recommendation for Question Answering Communities Mingcheng Qu, Guang Qiu, Xiaofei He, Cheng Zhang, Hao Wu, Jiajun Bu, Chun Chen.
Segmental Hidden Markov Models with Random Effects for Waveform Modeling Author: Seyoung Kim & Padhraic Smyth Presentor: Lu Ren.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Regression Approaches to Voice Quality Control Based on One-to-Many Eigenvoice Conversion Kumi Ohta, Yamato Ohtani, Tomoki Toda, Hiroshi Saruwatari, and.
Department of Electrical Engineering, Southern Taiwan University Robotic Interaction Learning Lab 1 The optimization of the application of fuzzy ant colony.
Universit at Dortmund, LS VIII
DISCRIMINATIVE TRAINING OF LANGUAGE MODELS FOR SPEECH RECOGNITION Hong-Kwang Jeff Kuo, Eric Fosler-Lussier, Hui Jiang, Chin-Hui Lee ICASSP 2002 Min-Hsuan.
ETeacher: Providing personalized assistance to e-learning students Schiaffino, S., Garcia, P. & Amandi, A. (2008). eTeacher: Providing personalized assistance.
You Are What You Tag Yi-Ching Huang and Chia-Chuan Hung and Jane Yung-jen Hsu Department of Computer Science and Information Engineering Graduate Institute.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
1 Learning styles and formative assessment strategy: enhancing student achievement in Web-based learning Authors: K.H. Wang, T.H. Wang, W.L. Wang, S.C.
Detecting Group Differences: Mining Contrast Sets Author: Stephen D. Bay Advisor: Dr. Hsu Graduate: Yan-Cheng Lin.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Pairwise Preference Regression for Cold-start Recommendation Speaker: Yuanshuai Sun
A Classification-based Approach to Question Answering in Discussion Boards Liangjie Hong, Brian D. Davison Lehigh University (SIGIR ’ 09) Speaker: Cho,
Spectrum Sensing In Cognitive Radio Networks
NTU & MSRA Ming-Feng Tsai
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Chance Constrained Robust Energy Efficiency in Cognitive Radio Networks with Channel Uncertainty Yongjun Xu and Xiaohui Zhao College of Communication Engineering,
Reinforcement Learning for Mapping Instructions to Actions S.R.K. Branavan, Harr Chen, Luke S. Zettlemoyer, Regina Barzilay Computer Science and Artificial.
Computacion Inteligente Least-Square Methods for System Identification.
An unsupervised conditional random fields approach for clustering gene expression time series Chang-Tsun Li, Yinyin Yuan and Roland Wilson Bioinformatics,
Dependency Networks for Inference, Collaborative filtering, and Data Visualization Heckerman et al. Microsoft Research J. of Machine Learning Research.
Shadow Detection in Remotely Sensed Images Based on Self-Adaptive Feature Selection Jiahang Liu, Tao Fang, and Deren Li IEEE TRANSACTIONS ON GEOSCIENCE.
Bayesian Optimization. Problem Formulation Goal  Discover the X that maximizes Y  Global optimization Active experimentation  We can choose which values.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Yun, Hyuk Jin. Theory A.Nonuniformity Model where at location x, v is the measured signal, u is the true signal emitted by the tissue, is an unknown.
Deep Feedforward Networks
Statistical Models for Automatic Speech Recognition
A Modified Naïve Possibilistic Classifier for Numerical Data
Ying shen Sse, tongji university Sep. 2016
The Alpha-Beta Procedure
Test Drop Rules: If not:
Linear Discrimination
Chapter 12 Analyzing Semistructured Decision Support Systems
Implementation of Learning Systems
Stochastic Methods.
Presentation transcript:

Recommendation for English multiple-choice cloze questions based on expected test scores 2011, International Journal of Knowledge-Based and Intelligent Engineering Systems Tomoharu Iwata 1, Tomoko Kojiri 2, Takeshi Yamada 1, Toyohide Watanabe 2 1 NTT Communication Science Laboratories, Japan 2 Graduate School of Information Science, Nagoya University, Japan Presenter: Chi-Shan Yeh

Introduction English multiple-choice cloze questions are one popular learning material in eLearning. In e-Learning systems, providing suitable learning materials that teach unacquired knowledge is crucial. In the e-Learning of English multiple-choice cloze questions, students tend to repeatedly tackle the same type of questions to heuristically acquire the knowledge asked in the questions.

Introduction This research assumes students who are studying to acquire English knowledge within the given tests for English multiple-choice cloze questions. Students take English multiple-choice cloze tests that contain a specific number of questions and study using the same type of questions to acquire the knowledge of the incorrectly answered questions. The objective of our research is to provide effective questions in the studying phase to increase the scores of subsequent tests taken after the studying phase.

Introduction Students are given many questions to deeply understand the unacquired knowledge and giving a minimum number of questions so that they can quickly pass the test. This research focuses on increasing test scores with great efficiency by giving a small set of questions that are adequate for gaining good test scores.

Introduction Although many personalized learning material recommendation algorithms have been proposed, they do not directly support learning efficiency. This research focuses on situations where students take the same test before and after studying to quantify the expected improvement in test scores. In the research field of recommendation systems, since the cognitive load on users to assign explicit ratings is heavy, gathering enough appropriate ratings is difficult.

Introduction Therefore, in our approach, test scores are used to determine the relations between questions and student explicit ratings are not applied. Our method recommends questions by which students can provide correct answers to questions that they answered incorrectly before studying. Effective questions for incorrectly answered questions may be identical for most students.

Introduction If particular questions are studied between tests and student test scores are increased after such studying, they may provide knowledge about questions whose answers were changed and became correct. The expected improvementof test scores based on questions is calculated by logistic regression models. By training the logistic regression models using student test scores and their learning log data, namely, the studied questions, questions that improve test scores can be automatically extracted.

Related Work Traditional intelligent tutoring systems, metadata about materials are attached beforehand, including the difficulty, topic, prerequisites, and the relationship between materials. It may be appropriate if the number of learning materials is not so large and their relations and metadata can be defined uniquely. Our research focuses on situations where plenty of questions exist whose relations are not defined beforehand and the unacquired knowledge of students is not uniquely determined by the metadata.

Related Work Chen et al. (2005) proposed Personalized E-Learning system based on Item Response Theory automatically adjusts the difficulty of learning materials according to the student’s level of understanding based on item response theory to provide appropriate learning materials. However, since PEL-IRT does not consider questionsthat students need to tackle in the test, it does not necessarily provide learning materials that can improve test scores.

Related Work Li and Chang (2005) proposed a personalized e-learning system which can automatically adapt to the interests and levels of learners. A feedback extractor with fusion capability is proposed to combine multiple feedback measures to infer user preferences. User profile, which stores user preferences and levels of expertise, is collected by user profiler to deliver personalized information using the collaborative filtering algorithm. However, if students only use their favorite materials their test scores may not improve.

Related Work Chen et al. (2008) proposed a personalized system that recommends materials considering their difficulty and incorrectly answere questions. However, this method does not consider learning efficiency, students need to cope with many questions before understanding the questions that they answered incorrectly. Our research increases specific test scores previously tackled by students with a limited number of questions.

Approach This research assumes that the student, who took a couple of tests in the past, is studying a set of multiple choice cloze questions in the studying phase to take a new test after the studying phase, where each test consists of a number of questions. The aim of this research is not to give a large volume of questions. Instead, it provides a limited number of suitable questions that can help increase test scores.

Approach Effective questions for maximizing test scores have the same important knowledge as incorrectly answered questions. When two questions share important knowledge and students cannot answer one of the questions in the tests, they may answer correctly after studying with other questions. Therefore, studied questions are detected as effective for test questions whose answers are changed from incorrect to correct answer by many students.

Framework

Method Problem setting The goal of our method is to select questions to be used during the studying phase from a set of questions to enhance learning efficiency, which is quantified by the expected improvement in test scores. Let z j be a variable that represents whether question j is recommended in the studying phase as follows: The studied questions are represented by vector z = (z j ) j ∈ M, where M represents a set of all candidate questions to study.

Method Problem setting The results of the set of test questions V before and after the studying phase are represented by vectors x = (x i ) i ∈ V and y = (y i ) i ∈ V, respectively.

Method Problem setting The improvement in the test scores expected from recommended questions z is written as follows: If the test result before studying phase x is given, the expected improvement in the test score can be simplified as follows:

Method Recommendation algorithm Our method sequentially selects a question that maximizes the expected improvement from questions that have not yet been recommended as follows:

Method Recommendation algorithm where z = (z j ) j ∈ M represents the currently studied questions and z +j represents the studied questions when question j is newly recommended. Examples of end conditions include those where the number of studied questions, the expected improvement, or the time period of the studying phase exceeds a certain threshold.

Method Improvement model When recommending questions,our method requires improvement model P (y ni = 1|x ni = −1, z n ), which is the probability of the improvement of answering question i with study questions z.

Method Improvement model Unknown parameters Θ = {µ i, θ i } i ∈ V can be estimated by maximizing the following log likelihood, which consists of the learning log data and the test results for set of students N:

Method Improvement model In the experiments, we used a quasiNewton method for the optimization and Gaussian priors with zero means for the unknown parameters to avoid overitting. The Hessian of the log likelihood function with Gaussian priors with respect to the parameters is positive definite. Therefore, the log likelihood is a convex function, and the global optimum solution is guaranteed.

Experiment Setting

Experiment Evaluation of improvement models We constructed and evaluated improvement models using the log data of 52 students, such as |N| = 52, with random material recommendations. To evaluate the effect of the relationships, we compared the proposed model with a model that does not consider relationships among questions or the Bernoulli model as follows: Parameter φi can be estimated based on the maximum likelihood as follows:

Experiment Evaluation of improvement models For the evaluation measurement, we used the AUC of the problem to predict whether questions that were incorrectly answered in the pre-test are correctly answered in the post-test.

Experiment Analysis of improvement models

Experiment Evaluation of recommendation algorithms by simulations We compared our algorithm with the following three algorithms: Random, Mistakable, and Level. Random randomly recommends a question.

Experiment Evaluation of recommendation algorithms by simulations Recommendation question is determined as follows: Level recommends a question whose difficulty is appropriate to the student’s understanding level by selecting recommendation question as follows: Where is the difficulty of question j and is the level of student n.

Experiment Evaluation of recommendation algorithms by simulations We evaluated the recommendation algorithm by the expected improvement rate, which is the expected number of correctly answered questions after studying to the number of incorrectly answered questions before studying as follows: ps. Here, we assumed that questions correctly answered before studying were also correct after studying.

Experiment Evaluation of recommendation algorithms by students Because the evaluation of recommendation with many actual students costs much, we compared the proposed method with the most basic random method. 38 students studied questions recommended randomly, and 49 studied questions recommended by our method. We evaluated the recommendation algorithm by the improvement rate, which is the correctly answered questions after studying to the incorrectly answered questions before studying as follows:

Experiment Evaluation of recommendation algorithms by students

Conclusion and Future Direction We proposed a method that maximizes learning efficiency based on the expected improvement in test scores. The experimental results suggest that our question recommendation approach is promising. Although we modeled the expected improvement in test scores with simple logistic regression using only the learning log data and test results to simplify our framework’s novelty, we could also use more information about the questions and student attributes for the modeling.

Conclusion and Future Direction Our recommendation method is not only specific to multiple-choice questions, but also be applied to math or physics. In addition, learning materials tackled by students in the studying phase can also be other types of materials. Our approach does not depend on questions; it can be applied to other types of materials. We plan further verification of our proposed method by comparing other methods with actual students.