Presentation is loading. Please wait.

Presentation is loading. Please wait.

Can Institutions Really Be Compared Using Standardised Tests? Presented at the 2008 Meeting of the European Association for Institutional Research Copenhagen,

Similar presentations


Presentation on theme: "Can Institutions Really Be Compared Using Standardised Tests? Presented at the 2008 Meeting of the European Association for Institutional Research Copenhagen,"— Presentation transcript:

1 Can Institutions Really Be Compared Using Standardised Tests? Presented at the 2008 Meeting of the European Association for Institutional Research Copenhagen, Denmark August 2008 By Trudy W. Banta Professor of Higher Education and Senior Advisor to the Chancellor for Academic Planning and Evaluation Indiana University-Purdue University Indianapolis 355 N. Lansing St., AO 140 Indianapolis, Indiana 46202-2896 tbanta@ iupui.edu http://www.planning.iupui.edu

2 Standardized tests of generic skills cannot provide valid comparisons of institutional quality. © TWBANTA-IUPUI

3  Standardized tests do have a place in assessing learning.  Higher education must be accountable © TWBANTA-IUPUI

4 My History  Educational psychology  Program evaluation & measurement  Performance funding in Tennessee  1990 USDOE effort to build a national test  1992 Initiated evidence-based culture at IUPUI © TWBANTA-IUPUI

5 Group Assessment Has Failed to Demonstrate Institutional Accountability Focus on improvement at unit level Rare aggregation of data centrally Too few faculty involved HE scholars focused on K-12 assessment

6 © TWBANTA-IUPUI Now We Have the Press to Assess with a Test

7 © TWBANTA-IUPUI 2006 Commission on the Future of Higher Education  We need a simple way to compare institutions  The results of student learning assessment, including value added measurements (showing skill improvement over time) should be... reported in the aggregate publicly.

8 OECD’s AHELO for 10 HEIs from 3-4 countries 1. Generic skills (CLA) 2. Disciplines (Engineering and Economics) 3. Value added 4. Contextual information indicators © TWBANTA-IUPUI

9 Two-Pronged Strategy in Washington 1. Pressure accreditors 2. Voluntary System of Accountability - NASULGC - AASCU

10 © TWBANTA-IUPUI Voluntary System of Accountability Report Scores in critical thinking, written communication, analytic reasoning using CAAP MAPP CLA

11 © TWBANTA-IUPUI Collegiate Assessment of Academic Proficiency (6 independent modules) Reading Writing Skills Writing Essay Mathematics Science Critical Thinking

12 © TWBANTA-IUPUI Measure of Academic Proficiency & Progress Humanities Social Sciences Natural Sciences College – Level Reading College – Level Writing Critical Thinking Mathematics Total Score

13 © TWBANTA-IUPUI Collegiate Learning Assessment Performance and Analytic Writing Tasks measuring Critical Thinking Analytic Reasoning Written Communication Problem Solving

14 © TWBANTA-IUPUI Standardized tests CAN initiate conversation

15 Advantages of standardized tests of generic skills  promise of increased reliability & validity  norms for comparison © TWBANTA-IUPUI

16 Limitations of standardized tests of generic skills  cannot cover all a student knows  narrow coverage, need to supplement  difficult to motivate students to take them! © TWBANTA-IUPUI

17 TN = Most Prescriptive (5.45% of Budget for Instruction) 1. Accredit all accreditable programs (25) 2. Test all seniors in general education (25) 3. Test seniors in 20% of majors (20) 4. Give an alumni survey (15) 5. Demonstrate use of data to improve (15) ___ 100

18 © TWBANTA-IUPUI At the University of Tennessee CAAP Academic Profile (now MAPP) COMP (like CLA and withdrawn by 1990) College BASE

19 © TWBANTA-IUPUI In TN We Learned 1) No test measured 30% of gen ed skills 2) Tests of generic skills measure primarily prior learning 3) Reliability of value added =.1 4) Test scores give few clues to guide improvement actions

20 © TWBANTA-IUPUI VSA Instrument Reviews Revealed Available tests of generic skills – 1) Lack test – retest reliability 2) Lack content validation (except with ACT/SAT) 3) Have no evidence that score gain is due to college experience

21 © TWBANTA-IUPUI An Inconvenient Truth.9 = the correlation between SAT and CLA scores of institutions thus 81% of the variance in institutions’ scores is due to prior learning

22 © TWBANTA-IUPUI How Much of the Variance in Senior Scores is Due to College Impact? Student motivation to attend that institution (mission differences) Student mix based on age, gender socioeconomic status race/ethnicity transfer status college major

23 © TWBANTA-IUPUI How Much of the Variance in Senior Scores is Due to College Impact? (continued) Student motivation to do well Sampling error Measurement error Test anxiety  College effects ______ 19 %

24 Threats to Conclusions Based on Test Scores 1. Measurement error 2. Sampling error 3. Different tests yield different results 4. Different ways of presenting results 5. Test bias 6. Pressure to raise scores - Daniel Koretz “Measuring Up” Harvard U. Press - 2008 © TWBANTA-IUPUI

25 Student Motivation Samples of students are being tested Extrinsic motivators (cash, prizes) are used We have learned: Only a requirement and intrinsic motivation will bring seniors in to do their best

26 Concerns About Value Added Student attrition Proportion of transfer students Different methods of calculating Unreliability Confounding effects of maturation © TWBANTA-IUPUI

27 Recent University of Texas Experience  30 – 40% of seniors at flagships earn highest CLA score (ceiling effect)  flagship campuses have lowest value added scores © TWBANTA-IUPUI

28 Other Recent Critiques  Cross-sectional design incorporates too many confounding variables ~ OR ~  Freshman scores generally are NOT randomly drawn from same population as senior scores © TWBANTA-IUPUI

29 Word from Measurement Experts Given the complexity of educational settings, we may never be satisfied that value added models can be used to appropriately partition the causal effects of teacher, school, and student on measured changes in standardized test scores. - Henry Braun & Howard Wainer Handbook of Statistics, Vol. 26: Psychometrics Elsevier 2007 © TWBANTA-IUPUI

30 Better Ways to Demonstrate Accountability 1. Performance Indicators  Access, social mobility  Diversity  Workforce development  Economic development  Engaging student experience © TWBANTA-IUPUI

31 Better Ways to Demonstrate Accountability 2. Measures of Learning  Standardized tests in major fields  Internship performance  Senior projects  Electronic portfolios  External examiners © TWBANTA-IUPUI

32 AAC&U Essential Learning Outcomes  Knowledge of human cultures and physical and natural world  Intellectual and practical skills (writing, thinking, team work)  Personal and social responsibility  Integrative learning © TWBANTA-IUPUI

33 Expensive Alternatives ?  Agreement on outcomes  Agreement on standards of achievement  Peer review © TWBANTA-IUPUI

34 Accompanying Benefits Teach faculty how to develop better classroom assessments Involve faculty in using results to improve learning More collaboration across disciplines and institutions Closer ties with community © TWBANTA-IUPUI


Download ppt "Can Institutions Really Be Compared Using Standardised Tests? Presented at the 2008 Meeting of the European Association for Institutional Research Copenhagen,"

Similar presentations


Ads by Google