By: Diane Marie Lee Ma. Mercedes Rodrigo Ryan Baker Jessica Sugay

Slides:



Advertisements
Similar presentations
Author: Carlos Pacheco, Shuvendu K. Lahiri, Michael D. Ernst, Thomas Ball MIT CSAIL.
Advertisements

Describing Quantitative Variables
Course Web Site – Also linked from Blackboard Course Materials – Excel Tutorials – Access Tutorials – PPT.
Active Learning for Streaming Networked Data Zhilin Yang, Jie Tang, Yutao Zhang Computer Science Department, Tsinghua University.
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
Reliability and Validity checks S-005. Checking on reliability of the data we collect  Compare over time (test-retest)  Item analysis  Internal consistency.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
-Arthur Lewis (alew525). Contents Background and Introduction Overview of Popular Environments Empirical Studies Borland Delphi v/s SimplifIDE Gild v/s.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
Bangpeng Yao Li Fei-Fei Computer Science Department, Stanford University, USA.
Statistics Micro Mini Threats to Your Experiment!
An Experimental Evaluation on Reliability Features of N-Version Programming Xia Cai, Michael R. Lyu and Mladen A. Vouk ISSRE’2005.
Affective Computing and Intelligent Interaction Ma. Mercedes T. Rodrigo Ateneo Laboratory for the Learning Sciences Department of Information Systems and.
Development of an Affect-Sensitive Agent for an Intelligent Tutor for Algebra Thor Collin S. Andallaza August 4, 2012.
Software Development Unit 6.
Measures of Central Tendency
Jared A. Rowland, M.S., Michael M. Knepp, M.S., Sheri L. Towe, M.S., Chris S. Immel, M.S., Ryoichi J.P. Noguchi, M.S., Chad L. Stephens, M.S. & David W.
Case Study – San Pedro Week 1, Video 6. Case Study of Classification  San Pedro, M.O.Z., Baker, R.S.J.d., Bowers, A.J., Heffernan, N.T. (2013) Predicting.
by B. Zadrozny and C. Elkan
Chapter 8 Introduction to Hypothesis Testing
Topic 3 Z-Scores Unit 5 Topic 3. Explore Lindsay’s class wrote three diploma examinations. The results are shown in the table below. Relative to the other.
Mining and Analysis of Control Structure Variant Clones Guo Qiao.
Designing and Debugging Batch and Interactive COBOL Programs Chapter 5.
Leena Razzaq Office: 310BWVH Office hours: Monday 11am-1pm or by appointment jys.
Chartese Jones - MCIS Department - Mississippi Valley State University Mentor: Dr. Raymond Williams, Mississippi Valley State University What effect does.
Expert Systems with Applications 34 (2008) 459–468 Multi-level fuzzy mining with multiple minimum supports Yeong-Chyi Lee, Tzung-Pei Hong, Tien-Chin Wang.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Intel Confidential – Internal Only Co-clustering of biological networks and gene expression data Hanisch et al. This paper appears in: bioinformatics 2002.
Data Structures and Algorithms Lecture 1 Instructor: Quratulain Date: 1 st Sep, 2009.
The Affective and Learning Profiles of Students when Using an Intelligent Tutoring System for Algebra by: Maria Carminda V. Lagud Ma. Mercedes T. Rodrigo.
Exploring the Relationship Between Novice Programmer Confusion and Achievement By: Diane Marie Lee Ma. Mercedes Rodrigo Ryan Baker Jessica Sugay Andrei.
Learning Analytics: Process & Theory March 24, 2014.
Finding Near-Duplicate Web Pages: A Large-Scale Evaluation of Algorithms Author: Monika Henzinger Presenter: Chao Yan.
Supervisor Ebtsam AbdelHakam Department of Computer Science Najran University 24/2/2014 Ebtsam Abdelhakam 1.
McGraw-Hill/Irwin Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. Using Nonexperimental Research.
Unsupervised Mining of Statistical Temporal Structures in Video Liu ze yuan May 15,2011.
® IBM Software Group © 2009 IBM Corporation Essentials of Modeling with the IBM Rational Software Architect, V7.5 Module 15: Traceability and Static Analysis.
Investigating semantic similarity measures across the Gene Ontology: the relationship between sequence and annotation Bioinformatics, July 2003 P.W.Load,
 Software Development Life Cycle  Software Development Tools  High Level Programming:  Structures  Algorithms  Iteration  Pseudocode  Order of.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 6, 2013.
CHAPTER Windows Server Management. Chapter Objectives Give an overview of the Server Manager Provide details of accessing the Server Manager Explain the.
Dan Thompson Oklahoma State University Center for Health Science Evaluating Assessments: Utilizing ExamSoft’s item-analysis to better understand student.
1 Estimation Chapter Introduction Statistical inference is the process by which we acquire information about populations from samples. There are.
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Iceberg: A Tool for Static Analysis of Java Critical Sections Michael D. Shah, Samuel Z. Guyer Tufts University Soap Workshop Session II, 10:50-11:10am,
Experience Report: System Log Analysis for Anomaly Detection
Learning to Detect and Classify Malicious Executables in the Wild by J
Constructing a Predictor to Identify Drug and Adverse Event Pairs
Different Types of Data
Outline Introduction State-of-the-art solutions
Selecting the Best Measure for Your Study
Coding Manual and Process
Software Design and Architecture
Statistical Data Analysis
Longitudinal Designs.
SAD: 6º Projecto.
Big Data, Education, and Society
Algorithms Chapter 3 With Question/Answer Animations
ID1050– Quantitative & Qualitative Reasoning
Predict Failures with Developer Networks and Social Network Analysis
The Behavior of Tutoring Systems
Discriminative Frequent Pattern Analysis for Effective Classification
Section Ii: statistics
Designing a Method and Test Program
Systems Update
False discovery rate estimation
The Essentials of 2-Level Design of Experiments Part I: The Essentials of Full Factorial Designs Developed by Don Edwards, John Grego and James Lynch.
Xiao-Yu Zhang, Shupeng Wang, Xiaochun Yun
Essentials of Statistics 4th Edition
Presentation transcript:

Exploring the Relationship Between Novice Programmer Confusion and Achievement By: Diane Marie Lee Ma. Mercedes Rodrigo Ryan Baker Jessica Sugay Andrei Coronel

Affective States and Achievement Recent studies have illustrated the relationships between affective states and achievement Negative affective states have negative impact on student’s achievement (Craig et al, 2006; Rodrigo, 2009; Lagud, 2010) Craig = Autotutor; boredom has negative correlation with learning gains Rodrigo = boredom and confusion = lower achievement in the programming course Lagud = Aplusix = highest levels of boredom and confusion among low-achieving students

Confusion Double-edged/ Dual Nature (D’Mello 2009) Harmful Helpful

Goal Discovery-with-models approach to finding the relationship between novice programmer confusion and achievement

Data Collection 149 students enrolled in CS21a – Introduction to Computing I Four lab sessions BlueJ IDE BlueJ Plug-in (Jadud and Henriksen, 2009) BlueJ IDE plug-in by Jadud Connected to a SQLite database server

Data Collection Compilation logs include Compilation logs = all submissions made to the compiler Compilation logs include Computer number Timestamp Code Error message (if any) And many more!

Data Collection Total of 340 student-lab sessions Total of 13,528 compilation logs collected 13,000++ of compilation logs are to be too hard to label Can’t see confusion in just one compilation = must look at group of compilations

Data Labeling Sorted the compilations by student and by Java class name Grouped the compilations into clips Clips = 8 compilations Total: 2,386 clips Raters were asked to label a sample of 664 clips Wanted to have a representative sample, so we’ll get 2 clips per student However, some students are so amazing that they had less than 16 compilations during the whole lab session

Data Labeling Used low-fidelity text replays Labels Maintains good inter-rater reliability and efficient in aiding coders to label student disengagement (Baker et al. 2006) Labels Confused Not Confused Bad Clip Cohen’s Kappa between raters: 0.77 The same error appeared in the same general vicinity within the code for several consecutive compilations. The coders inferred that the student did not know what was causing the error and how to fix it. An assortment of errors appeared in consecutive compilations and remained unresolved. The coders inferred that the student was experimenting solutions, changing the actual error message but not addressing the real source of the error. Code malformations that showed a poor understanding of Java constructs,e.g. “return outside method”. The coders inferred that the student did not grasp even the basics of program construction, despite the availability of written aids such as Java code samples and explanatory slides.

Data Labeling Filter out “bad clips” Remove clips where raters disagreed on the label Left with 418 clips for model construction

Model Construction Used RapidMiner version 5.1 Used J48 Decision Trees Features were mined from the clips J48 Decision Trees with 10 fold batch cross validation at the student level

Model Construction Feature set used: Average time between compilations Maximum time between compilations Average time between compilations w/ errors Maximum time between compilations w/ errors Number of compilations w/ errors Number of pairs consecutive compilations ending w/ the same error Time- and error-related features

Kappa: 0.86

Data Relabeling Model was coded as a Java program Had the program relabel all the 2,386 clips Generated three sets of confused-not confused sequences Correlated the percentage of the sequences of each student to their midterm exam scores We counted the number of occurrences of each state or sequence per student within each set. The total number of sequences per student varied. We there- fore normalized the data by dividing the number of occurrences of each state or sequence per student by the total number of occurrences for that student.

Not Confused-Not Confused Not-Confused-Confused Confused-Not Confused Results   Not Confused-Not Confused Not-Confused-Confused Confused-Not Confused Confused-Confused Relationship with midterm .064 .139  .144 -.229 (0.539) (0.180) (0.163) (0.026) R = above P = below

Results NNN NNC NCN NCC CNN CNC CCN CCC -.015 .014 .062 -.046 .233   NNN NNC NCN NCC CNN CNC CCN CCC Relationship with Midterm -.015 .014 .062 -.046 .233 .163 .052 -.337 (.901) (.909) (.610) (.704) (.05) (.174) (.665) (.004)

Conclusion Prolonged confusion has a negative impact on student’s performance Resolved confusion has a positive impact on student’s performance A certain amount of confusion is needed for learning

On-going Work Support the incorporation of tools for automatic detection of confusion in computer science learning environments Redoing the sampling and clipping method

Thank you Questions? Confusion = thrashing = repetitively getting errors Area for future work = go deeper to confusion literature