Core Methods in Educational Data Mining

Slides:



Advertisements
Similar presentations
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 February 18, 2013.
Advertisements

Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Comparison of Data Mining Algorithms on Bioinformatics Dataset Melissa K. Carroll Advisor: Sung-Hyuk Cha March 4, 2003.
Knowledge Inference: Advanced BKT Week 4 Video 5.
Indian Statistical Institute Kolkata
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 27, 2012.
Machine Learning in Practice Lecture 3 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Feature Engineering Studio February 23, Let’s start by discussing the HW.
Classifiers, Part 3 Week 1, Video 5 Classification  There is something you want to predict (“the label”)  The thing you want to predict is categorical.
Classifiers, Part 1 Week 1, video 3:. Prediction  Develop a model which can infer a single aspect of the data (predicted variable) from some combination.
Case Study – San Pedro Week 1, Video 6. Case Study of Classification  San Pedro, M.O.Z., Baker, R.S.J.d., Bowers, A.J., Heffernan, N.T. (2013) Predicting.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 13, 2012.
Feature Engineering Studio September 23, Let’s start by discussing the HW.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 13, 2013.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Feature Engineering Studio September 9, Welcome to Feature Engineering Studio Design studio-style course teaching how to distill and engineer features.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 6, 2013.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Feature Engineering Studio Special Session September 25, 2013.
Machine Learning in Practice Lecture 10 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 January 25, 2012.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Oct 29th, 2001Copyright © 2001, Andrew W. Moore Bayes Net Structure Learning Andrew W. Moore Associate Professor School of Computer Science Carnegie Mellon.
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Data Transformation: Normalization
Chapter 7. Classification and Prediction
EXPERIMENTAL RESEARCH
Core Methods in Educational Data Mining
Part III – Gathering Data
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Special Topics in Educational Data Mining
Wed June 12 Goals of today’s lecture. Learning Mechanisms
Core Methods in Educational Data Mining
Big Data, Education, and Society
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Feature Engineering Studio Special Session
Big Data, Education, and Society
TOP DM 10 Algorithms C4.5 C 4.5 Research Issue:
Core Methods in Educational Data Mining
Machine Learning with Weka
Lesson 4 – Social Skill: Giving Feedback.
Core Methods in Educational Data Mining
CSCI N317 Computation for Scientific Applications Unit Weka
Machine Learning in Practice Lecture 23
Machine Learning in Practice Lecture 7
Support Vector Machine _ 2 (SVM)
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
CS639: Data Management for Data Science
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Reminder for next week CUELT Conference.
Presentation transcript:

Core Methods in Educational Data Mining EDUC545 Spring 2017

The Homework Let’s go over basic homework 2

Q1 Build a decision tree (using operator W-J48 from the Weka Extension Pack) on the entire data set. What is the non-cross-validated kappa? What was process? What was result?

Q2 That’s the correct answer, but let’s think about it. The kappa value you just obtained is artificially high – the model is over-fitting to which student it is. What is the non-cross-validated kappa, if you build the model (using the same operator), excluding student? How did you modify the model to remove the student term? Were there multiple ways to accomplish this?

Q2 That’s the correct answer, but let’s think about it. Why was the kappa value artificially high?

Q2 That’s the correct answer, but let’s think about it. Why was the kappa value artificially high? How do we know that we were over-fitting to the student?

Q2 How did you remove student from the model? There were multiple ways to accomplish this

Q2 What is the non-cross-validated kappa, if you build the model (using the same operator), excluding student? Did the number go up? Go down? Stay the same? What does this mean?

Q3 Some other features in the data set may make your model overly specific to the current data set. Which data features would not apply outside of the population sampled in the current data set? Answers?

Q4 What is the non-cross-validated kappa, if you build the W-J48 decision tree model (using the same operator), excluding student and the variables from Question 3? Recall that you decided to eliminate School, Class, and Coder, as well as STUDENTID. Answers?

Q4 What is the non-cross-validated kappa, if you build the W-J48 decision tree model (using the same operator), excluding student and the variables from Question 3? Recall that you decided to eliminate School, Class, and Coder, as well as STUDENTID. Answers? Was this algorithm successful?

Q5 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use Naïve Bayes? Answers?

Q6 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use W-JRip? Answers?

Q7 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use Logistic Regression? (Hint: You will need to transform some variables to make this work; RapidMiner will tell you what to do) How did you do the variable transform? Why did you need to do the variable transform? Answers?

Q8 Wow, that was a lot of waiting for nothing. What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use Step Regression (called Linear Regression)? Answers?

Q9 What is the non-cross-validated kappa, for the same set of variables you used for question 4, if you use k-NN instead of W-J48? (We’ll discuss the results of this test later). Answers? Why did you get that result?

Q10 What is the kappa, if you delete School, Class, Coder, and STUDENTID, use W-J48, and conduct 10-fold stratified-sample cross-validation? How did you set this up? Answers?

Q11 Why is the kappa lower for question 11 (cross-validation) than question 4 (no cross-validation?)

Q12 What is the kappa, for the same set of variables you used for question 4, if you use k-NN, and conduct 10-fold stratified-sample cross-validation?

Q13 k-NN and W-J48 got almost the same Kappa when compared using cross-validation. But the kappa for k-NN was much higher (1.000) when cross-validation wasn't used. Why is that?

Questions? Comments? Concerns?

How are you liking RapidMiner?

Other RapidMiner questions?

What is the difference between a classifier and a regressor?

What are some things you might use a classifier for? Bonus points for examples other than those in the BDE videos

Any questions about any algorithms?

Do folks feel like they understood logistic regression? Any questions?

Anyone willing to come up and do a couple examples?

Logistic Regression m = 0.5A - B + C A B C M P(M)

Logistic Regression m = 0.5A - B + C A B C M P(M) 1

Logistic Regression m = 0.5A - B + C A B C M P(M) 1

Logistic Regression m = 0.5A - B + C A B C M P(M) 4 1

Logistic Regression m = 0.5A - B + C A B C M P(M) 100 -100

Why would someone Use a decision tree rather than, say, logistic regression?

Has anyone Used any classification algorithms outside the set discussed/recommended in the videos? Say more?

Other questions, comments, concerns about lectures?

Did anyone read Hand article? Thoughts?

Did anyone read Pardos article last week? Thoughts?

Creative HW 1

Questions about Creative HW 1?

Questions? Concerns?

Back to clustering lecture Slides 26-50

Other questions or comments?

Next Class Wednesday, February15 Behavior Detection Baker, R.S. (2015) Big Data and Education. Ch.1, V5. Ch. 3, V1, V2. Sao Pedro, M.A., Baker, R.S.J.d., Gobert, J., Montalvo, O. Nakama, A. (2013) Leveraging Machine-Learned Detectors of Systematic Inquiry Behavior to Estimate and Predict Transfer of Inquiry Skill. User Modeling and User-Adapted Interaction, 23 (1), 1-39.  Kai, S., Paquette, L., Baker, R.S., Bosch, N., D'Mello, S., Ocumpaugh, J., Shute, V., Ventura, M. (2015) A Comparison of Face-based and Interaction-based Affect Detectors in Physics Playground. Proceedings of the 8th International Conference on Educational Data Mining, 77-84.  Creative HW 1 due

The End