Core Methods in Educational Data Mining HUDK4050 Fall 2014.

Slides:



Advertisements
Similar presentations
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 February 18, 2013.
Advertisements

Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Comparison of Data Mining Algorithms on Bioinformatics Dataset Melissa K. Carroll Advisor: Sung-Hyuk Cha March 4, 2003.
Knowledge Inference: Advanced BKT Week 4 Video 5.
Indian Statistical Institute Kolkata
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 27, 2012.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 7, 2013.
Three kinds of learning
ML ALGORITHMS. Algorithm Types Classification (supervised) Given -> A set of classified examples “instances” Produce -> A way of classifying new examples.
Educational Data Mining Overview John Stamper PSLC Summer School /25/2011 1PSLC Summer School 2011.
Feature Engineering Studio February 23, Let’s start by discussing the HW.
Classifiers, Part 3 Week 1, Video 5 Classification  There is something you want to predict (“the label”)  The thing you want to predict is categorical.
Diagnostic Metrics, Part 1 Week 2 Video 2. Different Methods, Different Measures  Today we’ll focus on metrics for classifiers  Later this week we’ll.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 6, 2012.
Classifiers, Part 1 Week 1, video 3:. Prediction  Develop a model which can infer a single aspect of the data (predicted variable) from some combination.
Case Study – San Pedro Week 1, Video 6. Case Study of Classification  San Pedro, M.O.Z., Baker, R.S.J.d., Bowers, A.J., Heffernan, N.T. (2013) Predicting.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 13, 2012.
ENSEMBLE LEARNING David Kauchak CS451 – Fall 2013.
Categorical data. Decision Tree Classification Which feature to split on? Try to classify as many as possible with each split (This is a good split)
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Educational Data Mining: Discovery with Models Ryan S.J.d. Baker PSLC/HCII Carnegie Mellon University Ken Koedinger CMU Director of PSLC Professor of Human-Computer.
Feature Engineering Studio September 23, Let’s start by discussing the HW.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
Learning Analytics: Process & Theory March 24, 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 13, 2013.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Validation methods.
Feature Engineering Studio September 9, Welcome to Feature Engineering Studio Design studio-style course teaching how to distill and engineer features.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 6, 2013.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Feature Engineering Studio Special Session September 25, 2013.
Machine Learning in Practice Lecture 10 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 January 25, 2012.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Machine Learning in Practice Lecture 2 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining
Introduction to Machine Learning
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Special Topics in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Feature Engineering Studio Special Session
Big Data, Education, and Society
Gerd Kortemeyer, William F. Punch
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Presentation transcript:

Core Methods in Educational Data Mining HUDK4050 Fall 2014

Any administrative questions?

The Homework Let’s go over the homework

Q1 Build a decision tree (using operator W-J48 from the Weka Extension Pack) on the entire data set. What is the non-cross-validated kappa? What was process? What was result?

Q2 That’s the correct answer, but let’s think about it. The kappa value you just obtained is artificially high – the model is over-fitting to which student it is. What is the non-cross-validated kappa, if you build the model (using the same operator), excluding student? How did you modify the model to remove the student term? – Were there multiple ways to accomplish this?

Q2 That’s the correct answer, but let’s think about it. Why was the kappa value artificially high?

Q2 That’s the correct answer, but let’s think about it. Why was the kappa value artificially high? How do we know that we were over-fitting to the student?

Q2 How did you remove student from the model? There were multiple ways to accomplish this

Q2 What is the non-cross-validated kappa, if you build the model (using the same operator), excluding student? Did the number go up? Go down? Stay the same? – What does this mean?

Q3 Some other features in the data set may make your model overly specific to the current data set. Which data features would not apply outside of the population sampled in the current data set? Answers?

Q4 What is the non-cross-validated kappa, if you build the W-J48 decision tree model (using the same operator), excluding student and the variables from Question 3? Recall that you decided to eliminate School, Class, and Coder, as well as STUDENTID. Answers?

Q4 What is the non-cross-validated kappa, if you build the W-J48 decision tree model (using the same operator), excluding student and the variables from Question 3? Recall that you decided to eliminate School, Class, and Coder, as well as STUDENTID. Answers? Was this algorithm successful?

Q5 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use Naïve Bayes? Answers?

Q6 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use W-JRip? Answers?

Q7 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use Logistic Regression? (Hint: You will need to transform some variables to make this work; RapidMiner will tell you what to do) How did you do the variable transform? – Why did you need to do the variable transform? Answers?

Q8 Wow, that was a lot of waiting for nothing. What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use Step Regression (called Linear Regression)? Answers?

Q9 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use k-NN instead of W-J48? (We’ll discuss the results of this test later). Answers? Why did you get that result?

Q10 What is the kappa, if you delete School, Class, Coder, and STUDENTID, use W-J48, and conduct 10-fold stratified-sample cross- validation? How did you set this up? Answers?

Q11 Why is the kappa lower for question 11 (cross- validation) than question 4 (no cross- validation?)

Q12 What is the kappa, for the same set of variables you used for question 4, if you use k- NN, and conduct 10-fold stratified-sample cross-validation?

Q13 k-NN and W-J48 got almost the same Kappa when compared using cross-validation. But the kappa for k-NN was much higher (1.000) when cross-validation wasn't used. Why is that?

Questions? Comments? Concerns?

How did you like RapidMiner?

Other RapidMiner questions?

What is the difference between a classifier and a regressor?

What are some things you might use a classifier for? Bonus points for examples other than those in the BDE videos

Any questions about any algorithms?

Do folks feel like they understood logistic regression? Any questions?

Anyone willing to come up and do a couple examples?

Logistic Regression m = 0.5A - B + C ABCMP(M) 000

Logistic Regression m = 0.5A - B + C ABCMP(M) 001

Logistic Regression m = 0.5A - B + C ABCMP(M) 011

Logistic Regression m = 0.5A - B + C ABCMP(M) 411

Logistic Regression m = 0.5A - B + C ABCMP(M)

Why would someone Use a decision tree rather than, say, logistic regression?

Has anyone Used any classification algorithms outside the set discussed/recommended in the videos? Say more?

Other questions, comments, concerns about lectures?

Did anyone read Hand article? Thoughts?

Did anyone read Pardos article? Thoughts?

Creative HW 1

Questions about Creative HW 1?

Questions? Concerns?

Other questions or comments?

Next Class Tuesday, September 15 Behavior Detection Baker, R.S. (2014) Big Data and Education. Ch.1, V5. Ch. 3, V1, V2. Baker, R.S.J.d., Corbett, A.T., Roll, I., Koedinger, K.R. (2008) Developing a Generalizable Detector of When Students Game the System. User Modeling and User-Adapted Interaction, 18, 3, Sao Pedro, M.A., Baker, R.S.J.d., Gobert, J., Montalvo, O. Nakama, A. (2013) Leveraging Machine-Learned Detectors of Systematic Inquiry Behavior to Estimate and Predict Transfer of Inquiry Skill. User Modeling and User-Adapted Interaction, 23 (1), Creative HW 1 due

The End