Abstract: The application of factor analysis to the Force Concept Inventory (FCI) has proven to be problematic. Some studies have suggested that factor.

Slides:



Advertisements
Similar presentations
Multiple Regression.
Advertisements

The Robert Gordon University School of Engineering Dr. Mohamed Amish
Covariance and Correlation
Confirmatory factor analysis GHQ 12. From Shevlin/Adamson 2005:
Some terminology When the relation between variables are expressed in this manner, we call the relevant equation(s) mathematical models The intercept and.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Quantitative Skills 4: The Chi-Square Test
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 25, Slide 1 Chapter 25 Comparing Counts.
Factor Analysis There are two main types of factor analysis:
When Measurement Models and Factor Models Conflict: Maximizing Internal Consistency James M. Graham, Ph.D. Western Washington University ABSTRACT: The.
Chi-square Test of Independence
Further Inference in the Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
Analysis of Variance & Multivariate Analysis of Variance
Statistics for Managers Using Microsoft® Excel 5th Edition
BA 427 – Assurance and Attestation Services
Relationships Among Variables
Example of Simple and Multiple Regression
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Factor Analysis PowerPoint Prepared by Alfred.
Chapter 13: Inference in Regression
Writing the Research Paper BY: DR. AWATIF ALAM Associate Professor.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 26 Comparing Counts.
Copyright © 2010 Pearson Education, Inc. Warm Up- Good Morning! If all the values of a data set are the same, all of the following must equal zero except.
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
Correlation.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Section 10-1 Review and Preview.
The Impact of the History of Physics on Student Attitude and Conceptual Understanding of Physics Often the pre and posttest show negative shift in introductory.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Section 9-2 Inferences About Two Proportions.
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Section 10-5 Multiple Regression.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Education 795 Class Notes Data Analyst Pitfalls Difference Scores Effects Sizes Note set 12.
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
ANALYSIS OF VARIANCE (ANOVA) BCT 2053 CHAPTER 5. CONTENT 5.1 Introduction to ANOVA 5.2 One-Way ANOVA 5.3 Two-Way ANOVA.
Analysis of Variance 1 Dr. Mohammed Alahmed Ph.D. in BioStatistics (011)
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Scientific Methods and Terminology. Scientific methods are The most reliable means to ensure that experiments produce reliable information in response.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 12 Testing for Relationships Tests of linear relationships –Correlation 2 continuous.
Lecture 12 Factor Analysis.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Copyright © 2010 Pearson Education, Inc. Warm Up- Good Morning! If all the values of a data set are the same, all of the following must equal zero except.
Correlation & Regression Analysis
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
© Copyright McGraw-Hill 2004
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 11 Analyzing the Association Between Categorical Variables Section 11.2 Testing Categorical.
11 Chapter 5 The Research Process – Hypothesis Development – (Stage 4 in Research Process) © 2009 John Wiley & Sons Ltd.
Ken Heller School of Physics and Astronomy University of Minnesota Assessing quantitative skills Supported in part by Department of Education (FIPSE),
AP Statistics Chapter 21 Notes
Lecture 11. The chi-square test for goodness of fit.
Chapter 13- Inference For Tables: Chi-square Procedures Section Test for goodness of fit Section Inference for Two-Way tables Presented By:
1 HETEROSCEDASTICITY: WEIGHTED AND LOGARITHMIC REGRESSIONS This sequence presents two methods for dealing with the problem of heteroscedasticity. We will.
Comparing Counts Chapter 26. Goodness-of-Fit A test of whether the distribution of counts in one categorical variable matches the distribution predicted.
BPS - 5th Ed. Chapter 221 Two Categorical Variables: The Chi-Square Test.
Principal Component Analysis
Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 8 th Edition Chapter 9 Hypothesis Testing: Single.
FACTOR ANALYSIS.  The basic objective of Factor Analysis is data reduction or structure detection.  The purpose of data reduction is to remove redundant.
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
GS/PPAL Section N Research Methods and Information Systems
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
Evaluation of measuring tools: validity
CJT 765: Structural Equation Modeling
Chi Square (2) Dr. Richard Jackson
Principal Component Analysis
Product moment correlation
COMPARING VARIABLES OF ORDINAL OR DICHOTOMOUS SCALES: SPEARMAN RANK- ORDER, POINT-BISERIAL, AND BISERIAL CORRELATIONS.
Presentation transcript:

Abstract: The application of factor analysis to the Force Concept Inventory (FCI) has proven to be problematic. Some studies have suggested that factor analysis of test results serves as a helpful tool in assessing the recognition of Newtonian concepts by students. Other work has produced at best ambiguous results. We report on our analysis of over 500 pre- and 400 post-tests. The factor structure is more pronounced in the post-test with a more readily identifiable association between factors and physical concepts. Aspects of Factor Analysis Applied to the Force Concept Inventory M. R. Semak and R. D. Dietz Department of Physics and Astronomy University of Northern Colorado, Greeley, CO Introduction The exchange of views between Huffman, Heller, Hestenes, and Halloun 1,2,3 in the pages of The Physics Teacher back in 1995 brought the concept of factor analysis to the attention of the physics education research community. Recently, Scott, Schumayer, and Gray 4 published a detailed account of how they applied factor analysis to investigate how over two thousand students answered the FCI questions after they had completed the mechanics section of an introductory physics course. They determined that five factors could be identified. This agrees in general with our unpublished results. Here we concentrate on comparing student performance on the FCI before and after instruction in mechanics and the consequent evolution of the factor pattern. Technique Factor analysis is a statistical technique used to search for response patterns in large data sets. It does this by investigating patterns of correlations among observed variables in order to identify and study constructs, or factors, that are not directly observable but hypothesized to exist. Factor analysis does not identify what the unobserved constructs are, only the pattern of responses. The researchers must try to identify them from the questions which load on each factor. Our FCI data set is the record of correct/incorrect responses of 518 pre-tests and 413 post-tests collected over a four-year period from students in introductory physics. This is the set of variables we analyze. Newtonian mechanics was the subject of study over most of the semester for all students. With dichotomous variables, the Pearson correlations used when handling continuous data with factor analysis are seen as inappropriate. A matrix of tetrachoric correlations is considered proper. Using Mplus 5, a software package capable of calculating tetrachoric correlations, a factor model was sought for both our pre- and post-test FCI data using the weighted least squares with mean and variance adjustment (WLSMV) estimator. The fit statistics and the interpretability of the factors were the major guide in choosing the optimal number of factors. Given question 1’s (this question concerns the equivalence principle) comparatively small loading values for both the pre- and post-test cases, we concluded that the factor models would be better served by leaving it out of the analysis. Given these considerations, we were led to choose five factors for the pre-test model and seven for that of the post-test. Factor Model Fit One of the fit statistics we considered is the “chi-square test for model fit” that compares the fit of a model with a certain number of factors to a hypothesized one- factor, baseline model. The baseline model will have a large chi-square value indicating a poor fit to a model with strong intercorrelations among the variables. The null hypothesis is that there isn’t any difference between the baseline model and the observed data. So, if the null hypothesis can be rejected for a model with a certain number of factors, the model can explain the correlations among the variables. For our FCI pre-test data, a five factor model has a chi-square value of (with 140 degrees of freedom and a p-value of 0.38) as compared to the baseline model’s value of (with 78 degrees of freedom and negligible p-value). For our FCI post-test data, a seven factor model has a chi-square value of (with 121 degrees of freedom and a p-value of 0.36) as compared to the baseline model’s value of (with 65 degrees of freedom and negligible p-value). In both cases, the chi-square values are considered insignificant and the null hypothesis is rejected. 6, 7 So, we take five factors for the pre-test model and seven for that of the post-test. Analysis We found that factor analysis of the pre-test data indicates no clear-cut factors that can easily be identified with introductory physics concepts other than the Third Law. However, analysis of the post-test results does show that the students’ responses have changed considerably and produces seven factors that can be associated with specific physics concepts. The table above shows the absolute values of the loadings for the seven factor model of the post-test situation. Here we find a substantial evolution from the pre-test case as factors can now be associated with specific physics concepts. That the number of factors has increased from the pre- to the post-test and that these factors have a clearer physical interpretation suggests that students have refined their ability to consider physical problems in greater detail. This is also borne out by the fact that the post-test patterns show associations among questions which have more than simply a broad physical feature in common. It’s interesting to note that we did not confirm the findings of Scott et al. 4 in which question 16 is found to load onto both a First and Third Law factor. References 1. P. Heller and D. Huffman, “Interpreting the Force Concept Inventory” Phys. Teach. 33, (1995). 2. D. Huffman and P. Heller, “What does the force concept inventory actually measure?” Phys. Teach. 33, (1995). 3. D. Hestenes and I. Halloun, “Interpreting the Force Concept Inventory” Phys. Teach. 33, (1995). 4. T. F. Scott, D. Schumayer, and A. R. Gray, “Exploratory factor analysis of a Force Concept Inventory data set,” Phys. Rev. ST Phys. Educ. Res 8, (2012). 5. L.K. Muthén and B.O. Muthén, Mplus User’s Guide 5th Ed. Los Angeles: Muthén & Muthén ( ). 6. D. Hooper et al., “Structural Equation Modelling: Guidelines for Determining Model Fit,” The Electronic Journal of Business Research Methods, 6, Issue 1, (2008) and references therein. 7. S. Hutchinson (private communication, 2014). The factor analysis was done using the statistics software package, Mplus 5, with the WLSMV estimator on tetrachoric correlations and a Quartermin rotation. Post-test Seven Factor Model First Law Applied Forces Third Law FACTORS Identify Forces Sudden Force Balanced Forces Kinemat- ics QUESTION Acknowledgement The authors offer great thanks to Susan Hutchinson, Professor of Statistics, Applied Statistics and Research at the University of Northern Colorado for many thoughtful discussions concerning this project. Conclusion The evolution of the FCI results to a situation in which students make more refined associations may be indicative of a mind on the path to grasping certain invaluable physical subtleties. We offer a way to quantify this. Also, given certain loading/correlation values or lack of interpretability, this analysis can help justify examining certain FCI questions as to their effectiveness as a measure of student understanding. Characterizing the Evolution of the Factor Pattern Each eigenvalue of the correlation matrix is the variance accounted for by a corresponding factor. With these variables being expressed in standard form, the total variance for each variable is one and, thus, that of the entire variable set is equal to the number of variables. The latter sum is invariant regardless of the success of the factor model. The individual eigenvalues can, of course, change depending on the amount of variance for which each factor accounts. In the extreme case of the factor model’s failure, all eigenvalues would equal one. Quantifying this nonuniformity in the set of eigenvalues can be seen as a measure of order found in the model by way of the formation of a viable factor pattern. We use the mathematical form of Shannon entropy to quantify the nonuniformity in the eigenvalue set. The entropy is given by where N is the number of variables,, and is the set of eigenvalues. Also, and, the maximum value being taken when for all i, the case of highest disorder. Using this idea, and. So, we find a decrease in entropy in going from the pre- to the post-test model. This offers a quantitative measure of the evolution to a situation in which more variance is accounted for by the factor pattern..