C R E S S T / U C L A Issues and problems in classification of students with limited English proficiency Jamal Abedi UCLA Graduate School of Education.

Slides:



Advertisements
Similar presentations
National Accessible Reading Assessment Projects Examining Background Variables of Students with Disabilities that Affect Reading Jamal Abedi, CRESST/University.
Advertisements

English only VS. L1 support Looking at English language acquisition of native Spanish speaking children.
Assessment, Accountability and NCLB Education 388 Lecture March 15, 2007 Kenji Hakuta, Professor.
Testing What You Teach: Eliminating the “Will this be on the final
STAR Assessments: Using data to drive your instruction 2012.
Chapter Fifteen Understanding and Using Standardized Tests.
ELL-Language-based Accommodations for Content Area Assessments The University of Central Florida Cocoa Campus Jamal Abedi University of California, Davis.
Characterizing the Complexities of Bilingualism: Issues of homogeneity and heterogeneity in language and literacy among Spanish-speaking Latino/a school-
Effect Size and Meta-Analysis
LPAC overview - Fall 2004 The Language Assessment Process in Texas  LPAC, Language Assessment Proficiency Committee  ARD is the Admissions, Review and.
Language Proficiency Assessment Commitee (LPAC)
Language Proficiency Assessment Commitee (LPAC)
1 Academic Performance of English Language Learners on Grades 3-8 ELA Tests (2007 to 2009) David Abrams Assistant Commissioner Office of Standards, Assessment.
September, 2010 Accomack County Public Schools. DEFINITION OF AN LEP STUDENT  An LEP student is one: Who was not born in the U.S. or whose native language.
Chapter 4 Validity.
Using Growth Models for Accountability Pete Goldschmidt, Ph.D. Assistant Professor California State University Northridge Senior Researcher National Center.
Are Accommodations Used for ELL Students Valid? Jamal Abedi University of California, Davis National Center for Research on Evaluation, Standards and Student.
Using Hierarchical Growth Models to Monitor School Performance: The effects of the model, metric and time on the validity of inferences THE 34TH ANNUAL.
Identification, Assessment and Re-classification of English Learners Initial Identification  Complete within 30 school days of enrollment Administer Home.
Meeting NCLB Act: Students with Disabilities Who Are Caught in the Gap Martha Thurlow Ross Moen Jane Minnema National Center on Educational Outcomes
1/16 CRESST/UCLA Alternative Assessment for English Language Learners Christy Kim Boscardin Barbara Jones Shannon Madsen Claire Nishimura Jae-Eun Park.
Jamal Abedi University of California, Davis/CRESST Presented at: The Race to the Top Assessment Program Public & Expert Input Meeting December 2, 2009.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
MEASUREMENT AND EVALUATION
Chapter 14 Understanding and Using Standardized Tests Viewing recommendations for Windows: Use the Arial TrueType font and set your screen area to at least.
Article Summary – EDU 215 Dr. Megan J. Scranton 1.
Language Proficiency Assessment Commitee (LPAC)
Creating Assessments with English Language Learners in Mind In this module we will examine: Who are English Language Learners (ELL) and how are they identified?
What ACCESS, the New Virginia Test for LEP Students, Means for School Districts LEP Caucus Presentation July 2008.
The University of Central Florida Cocoa Campus
Mini Project I-- Evaluation of a Standardized Test By Marcia Luebbe.
Chapter 6 ~~~~~ Oral And English Language Learner/Bilingual Assessment.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
National Accessible Reading Assessment Projects Research on Making Large-Scale Reading Assessments More Accessible for Students with Disabilities June.
Texas Comprehensive SEDL Austin, Texas March 16–17, 2009 Making Consistent Decisions About Accommodations for English Language Learners – Research.
CRESST ONR/NETC Meetings, July 2003, v1 1 ONR Advanced Distributed Learning Language Factors in the Assessment of English Language Learners Jamal.
C R E S S T / U C L A Impact of Linguistic Factors in Content-Based Assessment for ELL Students Jamal Abedi UCLA Graduate School of Education & Information.
Forum - 1 Assessments for Learning: A Briefing on Performance-Based Assessments Eva L. Baker Director National Center for Research on Evaluation, Standards,
Identification and Assessment of Talented Learners What is identification of the gifted? A process by which educators attempt to become aware of students.
1/27 CRESST/UCLA Research findings on the impact of language factors on the assessment and instruction of English language Learners Jamal Abedi University.
CRESST ONR/NETC Meetings, July 2003, v1 ONR Advanced Distributed Learning Impact of Language Factors on the Reliability and Validity of Assessment.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Do we have enough evidence on the validity of accommodations to justify the reporting of accommodated assessments? Jamal Abedi University of California,
Further Research Baker, E., Goldschmidt, P., Martinez, F., & Swigert, S. (February, 2002). In search of school quality and accountability: Moving beyond.
Assessment. Workshop Outline Testing and assessment Why assess? Types of tests Types of assessment Some assessment task types Backwash Qualities of a.
State Practices for Ensuring Meaningful ELL Participation in State Content Assessments Charlene Rivera and Lynn Shafer Willner GW-CEEE National Conference.
1 Children Left Behind in AYP and Non-AYP Schools: Using Student Progress and the Distribution of Student Gains to Validate AYP Kilchan Choi Michael Seltzer.
Jamal Abedi CRESST/University of California,Los Angeles Paper presented at 34 th Annual Conference on Large-Scale Assessment Boston, June 20-23, 2004.
Educational Challenges of English Language Learners.
C R E S S T / U C L A Psychometric Issues in the Assessment of English Language Learners Presented at the: CRESST 2002 Annual Conference Research Goes.
Jamal Abedi, UCLA/CRESST Major psychometric issues Research design issues How to address these issues Universal Design for Assessment: Theoretical Foundation.
Critical Issues Related to ELL Accommodations Designed for Content Area Assessments The University of Central Florida Cocoa Campus Jamal Abedi University.
Chapter 7: The Distribution of Sample Means
C R E S S T / U C L A UCLA Graduate School of Education & Information Studies Center for the Study of Evaluation National Center for Research on Evaluation,
Career Counseling: A Holistic Approach
THE CALIFORNIA ENGLISH LANGUAGE DEVELOPMENT TEST (CELDT) Poway Unified School District.
NCLB Assessment and Accountability Provisions: Issues for English-language Learners Diane August Center for Applied Linguistics.
BY MADELINE GELMETTI INCLUDING STUDENTS WITH DISABILITIES AND ENGLISH LEARNERS IN MEASURES OF EDUCATOR EFFECTIVENESS.
ELL-Focused Accommodations for Content Area Assessments: An Introduction The University of Central Florida Cocoa Campus Jamal Abedi University of California,
Barbara Foorman, Yaacov Petscher, & Chris Schatschneider, Florida Center for Reading Research, Florida State University.
Terra Nova By Tammy Stegman, Robyn Ourada, Sandy Perry, & Kim Cotton.
Exploring Data Use & School Performance in an Urban School District Kyo Yamashiro, Joan L. Herman, & Kilchan Choi UCLA Graduate School of Education & Information.
Test Validity.
Assistant Commissioner Office of Standards, Assessment and Reporting
Understanding and Using Standardized Tests
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Relationship between Standardized and Classroom-based Assessment
What Do All These Numbers Mean? Interpreting Gifted Test Scores
Chapter 3: How Standardized Test….
Presentation transcript:

C R E S S T / U C L A Issues and problems in classification of students with limited English proficiency Jamal Abedi UCLA Graduate School of Education & Information Studies Center for the Study of Evaluation National Center for Research on Evaluation, Standards, and Student Testing Paper presented at the 2002 Annual Meeting of the American Educational Research Association New Orleans, LA

C R E S S T / U C L A Classification of ELL Students There is a growing concern over the assessment and classification of language minority students. However, before developing a valid and reliable assessment system, a well-defined, objective definition of the term “LEP” or “ELL” must be obtained. Unfortunately, the criteria for identifying LEP students are not used uniformly across the nation. In several language background studies conducted at UCLA /CRESST (Abedi and Lord, 2001; Abedi, Lord, and Plummer, 1995; Abedi, Lord, and Hofstetter, 1997, Abedi, Hofstetter, Baker, and Lord, 1998), one of the major problems encountered was the lack of a commonly acceptable definition for limited English proficiency. 1

C R E S S T / U C L A 2 Classification of Students with Limited English Proficiency There are many different criteria by which a student can be classified as LEP. Among the most important of these criteria are being speaker of a language other than English and scoring low on the English proficiency tests. The first criterion, i.e., being a non-native English speaker, is defined in Los Angeles area schools based on the information from the Home Language Survey. The Home Language Survey For many schools in Los Angles area, the Home Language Survey is the only source of information used to determine the need for a student to be tested for English Proficiency. Recent dialog over the type of bilingual instruction causes reporting inaccurately for the purpose of assuring that their children be treated no differently from their Anglo classmates. Other concerns for the student whose parents may have citizenry issues have led to a more relaxed treatment of the home surveys than what the district would prefer. Questions have been raised about the accuracy of a survey completed by a parent who is illiterate or who has no familiarity with written English.

C R E S S T / U C L A Assessment of Students’ Language Proficiency in English Language proficiency and achievement tests in English are commonly used for identification and assessment of LEP students. According to Hopstock, Bucaro, Fleischman, Zehler, and Eu (1993) eighty- three percent of school districts use English language proficiency tests alone or with other techniques to decide if a student is LEP. The English proficiency tests used frequently for such purposes are the Bilingual Syntax Measure (BSM), the Idea Proficiency Test (IPT), the Language Assessment Battery (LAB), the Language Assessment Scales (LAS), the Maculaitis Assessment Program (MAC), and the Peabody Picture Vocabulary Test (PPVT). 3

C R E S S T / U C L A Achievement Tests Achievement tests in English are used by approximately 52 percent of school districts to help identify LEP students, assign them to school services,and reclassify them from LEP status. Commonly used achievement tests are: the California Achievement Test (CAT), Iowa Test of Basic Skills (ITBS), Metropolitan Achievement Test (MAT), Stanford Achievement Test (SAT), and the Comprehensive Test of Basic Skills (CTBS). Zehler, et al. (1994) did a comprehensive review of these language proficiency tests and found major differences in all the content areas in which the tests were compared. An even more serious criticism of these language proficiency and achievement tests is the problem of the validity and reliability of these tests for LEP populations and the exclusion of LEP students from the norming group for these tests. For example Abedi & Leon (1999) found that language factors may be an additional source of measurement error in the assessment of LEP which may reduce the reliability of the tests considerably. Abedi, Leon & Mirocha (2001) found that language factors in content-based assessment may seriously undermine the validity of the tests and may be considered a source of construct-irrelevant variance (Messick, 1994, p.14). 4

C R E S S T / U C L A 5 Data Sources Site 1. Site 1 is a large urban school district. ITBS test data were obtained. There were 36,065 students in the grade 3 (7,270 bilingual), In grade 6 there were 28,313 students (3,341 or 11.8% bilingual) and in grade 8, there were 25,406 students (2,306 or 9.1% were bilingual). Site 2. There were 414,169 students in the grade 2 population (125,109 or 30.2% were LEP), in grade 7 there were 349,581 students (73,993 or 21.2% LEP). In grade 9 there were 309,930 students (57,991 or18.7% LEP). Stanford 9 test data were obtained for all students in Grades 2 to 11 for the academic year. Site 3. There were 12,919 students in the grade 10 population (431 or 3.3% LEP). In grade 11 there were 9,803 students in the population (339 or 3.5% LEP). Site 4. There were 13,810 students in the grade 3 (1,065 or 7.7% LEP). In grade 6 there were 12,998 students in the population (813 or 6.3% LEP), in grade 8 there were 12,400 students (807 or 6.5% LEP).

C R E S S T / U C L A Findings Relationship between language proficiency test scores and LEP classification. Since LEP classification is based on students’ level of language proficiency and because LAS is a measure of language proficiency, one would expect to find a perfect correlation between LAS scores and LEP levels (LEP versus non-LEP). The results of analyses indicated a weak relationship between language proficiency test scores and language classification codes (LEP categories). Table 1. Correlation between LAS rating and LEP classification for Site 4 Correlation G2 G3 G4 G5 G6 G7 G8 G9 G10 G11 G12 Pearson r Sig (2-tailed) N

C R E S S T / U C L A 7

Table 4. Correlation coefficients between LEP classification code and ITBS subscales for Site 1 Grade Reading Math Concept Math Problem Math & Estimation Solving Computation Grade 3 Pearson r Sig (2-tailed) N 36,006 35,981 35,948 36,000 Grade 6 Pearson r Sig (2-tailed) N 28,272 28,273 28,250 28,261 Grade 8 Pearson r Sig (2-tailed) N 25,362 25,336 25,333 25,342 8

C R E S S T / U C L A Table 5. Correlation coefficients between LEP classification code and Stanford 9 subscales for Site 2 Grade Reading Language Science Math Spelling Social Sci Grade 3 Pearson r Sig (2-tailed) N 376, ,669 77, , ,699 62,317 Grade 5 Pearson r Sig (2-tailed) N 358, ,523 81, , ,689 73,975 Grade 7 Pearson r Sig (2-tailed) N 336, , , , ,745 86,894 Grade 9 Pearson r Sig (2-tailed) N 293, , , ,558 86, ,022 Grade 11 Pearson r Sig (2-tailed) N 225, , , ,217 58, ,891 9

C R E S S T / U C L A Table 6. Correlation coefficients between LEP classification code and Stanford 9 subscales for Site 3 Grade Reading Science Math Grade 10 Pearson r Sig (2-tailed) N 11,158 10,231 10,301 Grade 11 Pearson r Sig (2-tailed) N 8,740 7,900 8,040 10

C R E S S T / U C L A 11 Table 7. Correlation coefficients between LEP classification code and Stanford 9 subscales for Site 4 Grade Reading Math Math Computation Application Grade 3 Pearson r Sig (2-tailed) N 14,050 14,282 14,208 Grade 6 Pearson r Sig (2-tailed) N 13,354 13,364 13,299 Grade 8 Pearson r Sig (2-tailed) N 12,484 12,579 12,337 Grade 10 Pearson r NA Sig (2-tailed) N 9,499 9,778

C R E S S T / U C L A 12 Findings and Conclusions  For an effective instruction and a valid and reliable assessment for English language learners, a well-defined, objective definition of the term “ELL” and “LEP” is needed.  The results of studies nationwide suggest, however, that such definition is not provided.  The results of our analyses on large-scale data did not show a strong relationship between LEP classification and students’ level of English proficiency.  Analyses on the distribution of some English language proficiency tests showed a negatively-skewed distribution suggesting that the English proficiency test items did not have enough discrimination power.  The results of analyses also indicated that for grades 3 through 5, low scoring students tend to remain classified as LEP in most of these districts. As grade level increases, however, the variation in agreement among districts also increases. There appears to be an increasing tendency to reclassify low scoring students as grade level increases.  The results of our analyses also suggested that there was not any single criterion that highly correlates with LEP classification code. This may be due to psychometric characteristics of the measures or due to issues on the validity of LEP classification or, most likely, a combination of both.  Thus, the use of multiple criteria is recommended in assessments particularly in high- stakes assessments (e.g., in LEP classification). However, technical issues in using multiple criteria must be considered.