Bayes Net Toolbox for Student Modeling (BNT-SM)

Slides:



Advertisements
Similar presentations
Introducing Bayesian Networks in Neuropsychological Measures Presented at 2006 APS Annual Conference Toshi Yumoto University of Maryland, Abt Associate.
Advertisements

Bayesian network for gene regulatory network construction
Author: Carlos Pacheco, Shuvendu K. Lahiri, Michael D. Ernst, Thomas Ball MIT CSAIL.
Foundational Certified Code in a Metalogical Framework Karl Crary and Susmit Sarkar Carnegie Mellon University.
Dynamic Causal Modelling for ERP/ERFs Valentina Doria Georg Kaegi Methods for Dummies 19/03/2008.
Bridgette Parsons Megan Tarter Eva Millan, Tomasz Loboda, Jose Luis Perez-de-la-Cruz Bayesian Networks for Student Model Engineering.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
What do you need to know about DCM for ERPs/ERFs to be able to use it?
Weka & Rapid Miner Tutorial By Chibuike Muoh. WEKA:: Introduction A collection of open source ML algorithms – pre-processing – classifiers – clustering.
Educational data mining overview & Introduction to Exploratory Data Analysis Ken Koedinger CMU Director of PSLC Professor of Human-Computer Interaction.
Navigating the parameter space of Bayesian Knowledge Tracing models Visualizations of the convergence of the Expectation Maximization algorithm Zachary.
Knowledge Inference: Advanced BKT Week 4 Video 5.
Bayesian Knowledge Tracing and Other Predictive Models in Educational Data Mining Zachary A. Pardos PSLC Summer School 2011 Bayesian Knowledge Tracing.
Effective Skill Assessment Using Expectation Maximization in a Multi Network Temporal Bayesian Network By Zach Pardos, Advisors: Neil Heffernan, Carolina.
S5-1 © 2001 Carnegie Mellon University OCTAVE SM Process 5 Identify Key Components Software Engineering Institute Carnegie Mellon University Pittsburgh,
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
Mining Data from Randomized Within-Subject Experiments in an Automated Reading Tutor Joseph E. Beck and Jack Mostow Project LISTEN (
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
MML, inverse learning and medical data-sets Pritika Sanghi Supervisors: A./Prof. D. L. Dowe Dr P. E. Tischer.
MML, inverse learning and medical data-sets Pritika Sanghi Supervisors: A./Prof. D. L. Dowe Dr P. E. Tischer.
+ Doing More with Less : Student Modeling and Performance Prediction with Reduced Content Models Yun Huang, University of Pittsburgh Yanbo Xu, Carnegie.
CSE 590ST Statistical Methods in Computer Science Instructor: Pedro Domingos.
Educational Data Mining Overview John Stamper PSLC Summer School /25/2011 1PSLC Summer School 2011.
Educational Data Mining Ryan S.J.d. Baker PSLC/HCII Carnegie Mellon University Richard Scheines Professor of Statistics, Machine Learning, and Human-Computer.
Determining the Significance of Item Order In Randomized Problem Sets Zachary A. Pardos, Neil T. Heffernan Worcester Polytechnic Institute Department of.
Open Source Workshop1 IBM Software Group Working with Apache Tuscany A Hands-On Workshop Luciano Resende Haleh.
Case Study – San Pedro Week 1, Video 6. Case Study of Classification  San Pedro, M.O.Z., Baker, R.S.J.d., Bowers, A.J., Heffernan, N.T. (2013) Predicting.
Genetic Regulatory Network Inference Russell Schwartz Department of Biological Sciences Carnegie Mellon University.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Leslie Luyt Supervisor: Dr. Karen Bradshaw 2 November 2009.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 April 2, 2012.
R2WinBUGS: Using R for Bayesian Analysis Matthew Russell Rongxia Li 2 November Northeastern Mensurationists Meeting.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
M Machine Learning F# and Accord.net. Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher.
©Ian Sommerville 2004 Software Engineering. Chapter 21Slide 1 Chapter 21 Software Evolution.
Instance Construction via Likelihood- Based Data Squashing Madigan D., Madigan D., et. al. (Ch 12, Instance selection and Construction for Data Mining.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Estimating Component Availability by Dempster-Shafer Belief Networks Estimating Component Availability by Dempster-Shafer Belief Networks Lan Guo Lane.
Automatic Generation of Programming Feedback: A Data-Driven Approach Kelly Rivers and Ken Koedinger 1.
GeoProMT Purpose of today’s meeting – Present some research ideas Identify people willing to make a commitment to the project – Development could be part.
Assessment embedded in step- based tutors (SBTs) CPI 494 Feb 12, 2009 Kurt VanLehn ASU.
Carnegie Mellon How does the amount of context in which words are practiced affect fluency growth? Experimental results Jack Mostow, Jessica Nelson, Martin.
CrossCheckSimulation Results Conclusions References Model Instrumentation Modeling with CUTS Property Specification SPRUCE Challenge Problem Checking Model.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 January 25, 2012.
Bayesian Evolutionary Analysis by Sampling Trees (BEAST) LEE KIM-SUNG Environmental Health Institute National Environment Agency.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 6, 2012.
Oct 29th, 2001Copyright © 2001, Andrew W. Moore Bayes Net Structure Learning Andrew W. Moore Associate Professor School of Computer Science Carnegie Mellon.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Dependency Networks for Inference, Collaborative filtering, and Data Visualization Heckerman et al. Microsoft Research J. of Machine Learning Research.
Center for Assessment and Improvement of Learning
Core Methods in Educational Data Mining
Development Environment
EPANET-MATLAB Toolkit An Open-Source Software for Interfacing EPANET with MATLAB™ Demetrios ELIADES, Marios KYRIAKOU, Stelios VRACHIMIS and Marios POLYCARPOU.
Using Bayesian Networks to Predict Test Scores
Core Methods in Educational Data Mining
CAREER: Intelligent Generation of Text and Information Graphics*
the goal of Bayesian divergence time estimation
Addressing the Assessing Challenge with the ASSISTment System
Educational Data Mining Success Stories
Deep Knowledge Tracing
Core Methods in Educational Data Mining
This material is based upon work supported by the National Science Foundation under Grant #XXXXXX. Any opinions, findings, and conclusions or recommendations.
Multiple DAGs Learning with Non-negative Matrix Factorization
Human-centered Machine Learning
Tractography Algorithm: Toolbox:
Presentation transcript:

Bayes Net Toolbox for Student Modeling (BNT-SM) 21/10/10 Bayes Net Toolbox for Student Modeling (BNT-SM) Kai-min Chang & Yanbo Xu EDM2015 SMART Tutorial June 26, 2015 This work was supported by the National Science Foundation under Cyberlearning Grant IIS1124240. The opinions expressed are those of the authors and do not necessarily represent the views of the Institute, or the National Science Foundation.

Download BNT-SM Matlab Octave http://www.cs.cmu.edu/~listen/BNT-SM Check your university’s license Octave http://www.gnu.org/software/octave

Bayes Net Toolkit for Student Modeling (BNT-SM) Estimates of model parameters BayesNet.xml BNT-SM BNT Observations of student performance Estimates of student knowledge

Model Knowledge Tracing with Some Bayes Nets Packages Popular Bayes Nets packages (at the time…) BNT, BUGS, GMTK Complicated coding effort to use the packages Create toolkit to hide the complexity from ITS researchers. Extend Bayes Nets Toolkit (BNT; Murphy, 1998)

Revision History First developed 9 years ago (Chang et al., 2006) Incorporated LR-DBN (Xu & Mostow, 2011) Updated for the tutorial (SMART, 2015) To run on Octave (used a workaround since xmlread.m is not implemented in Octave). Attempted to use the most updated BNT (unsuccessful).

Knowledge Tracing (KT) Corbett & Anderson (1995) Infers a student’s proficiency on a skill from observation of applying the skill. Maintain 4 skill-specific parameters: Learning parameters: Already know & learn Performance parameters: guess & slip Knowledge of the skill “cat” 0.40 0.32 0.21 0.29 0.45 0.38 Correct  

Bayesian Knowledge Tracing (BKT) Affect future attempt Student Knowledge (K0) Student Knowledge (Kn) Already know learn, forget Affect current attempt guess, slip Student Performance (C0) Student Performance (Cn)

Bayes Net Toolkit for Student Modeling (BNT-SM) Step 4 Step 1 Step 3 Estimates of model parameters BayesNet.xml BNT-SM Step 2 BNT Step 5 Observations of student performance Estimates of student knowledge

1a. Specify Network Structure <nodes> <node> <name>StudentKnowledge</name> <values>2 </values> <type> discrete </type> <observed> latent</observed> <within> StudentPerformance </within> <between>StudentKnowledge </between> </node> …… Student Knowledge (K0) Kn Student Performance (C0) Cn

1b. Initialize Model Parameters Already know Student Knowledge (K0) Kn Learn XML Guess, Slip Student Performance (C0) Cn

2. Provide Observations of Student Performance Ordered attempts within a skill Comprehensive observational data Skill User Performance Time Student Knowledge Student Performance my Kevin 2002-08-14 15:30:21 ?  cat 2002-08-14 15:30:22 …

3. Run BNT-SM Estimates of model parameters BayesNet.xml BNT-SM BNT Observations of student performance Estimates of student knowledge I will describe the output of BNT-SM more in the evaluation section.

4. Examine Model Parameters Parameter values can be used to infer which skill is more difficult Skill # Users # Cases LL L0 Guess Slip Learn Forget My 14 23 -5.15 0.46 0.73 0.04 0.20 0.00 Cat 46 219 -94.07 0.44 0.70 0.11 0.24 Average log likelihood can be used to compare across different models

5. Examine Model Inferences User Skill Knowledge Correct Alice My 0.463045 NULL Bob Cathy … John Cat 0.440226 Kevin 0.620417 Estimates the progression of student knowledge that can be used by the tutor

Matlab / Octave >> cd BNT-SM/src >> setup >> cd ../model/kt >> [property evidence hash_bnet] = RunBnet('property.xml'); Sample Bayes net models An XML file that specifies the Bayes net we are constructing.

Examine Some Files >> vim property.xml (Matlab) >> vim load_property.m (Octave) >> vim evidence.train.xls >> vim evidence.test.xls >> vim param_table.xls >> vim inference_result_header.xls >> vim inference_result.xls Input Output

Model 1: KT Corbett & Anderson (1995) … … l0 l f l f K(1) K(t) K(t+1) g s g s g s C(1) C(t) C(t+1)

Model 1: KT  ? ? ?   Corbett & Anderson (1995) … … l0 l f l f g s

Model 2: LR-DBN An Extension of BNT-SM to trace multiple subskills (Xu & Mostow, 2012) S1(t+1) S1(t) S2(t+1) S2(t) Sm(t) Sm(t+1) … … K(1) l0 le fe K(t) le fe K(t+1) … … g s g s g s C(1) C(t) C(t+1) 19

Model 3: EEG-KT Insert a binary EEG measure into KT (Yuan et al., 2014) l0 le fe le fe K(1) K(t) K(t+1) … … E(1) E(t) E(t+1) ge se ge se ge se ge se C(1) C(t) C(t+1)

Model 3: EEG-KT Insert a binary EEG measure into KT (Yuan et al., 2014) le fe le fe l0 ? ? ? … … ge se ge se ge se ge se    Idea Model Evaluate Discuss

References Chang, K.M., Beck, J.E., Mostow, J., & Corbett, A. (2006). A Bayes net toolkit for student modeling in intelligent tutoring systems. Proceedings of the 8th International Conference on Intelligent Tutoring Systems, Jhongli, Taiwan, 104-113. Xu, Y., & Mostow, J. (2011). Using Logistic Regression to Trace Multiple Subskills in a Dynamic Bayes Net. Proceedings of the 4th International Conference on Educational Data Mining Eindhoven, Netherlands, 241-245. Yuan, Y., Chang, K.M., Xu, Y., & Mostow, J. (2014). A Public Toolkit and ITS Dataset for EEG. Proceedings of the 12th International Conference on Intelligent Tutoring Systems Workshop on Utilizing EEG Input in Intelligent Tutoring Systems, Honolulu, Hawaii.

Questions? Kai-min Kevin Chang Yanbo Xu Language Technologies Institute, Carnegie Mellon University E-mail: kkchang@cs.cmu.edu Yanbo Xu Robotics Institute, Carnegie Mellon University E-mail: yanbox@cs.cmu.edu