Operational Check of the 2010 FCAT 3 rd Grade Reading Equating Results

Similar documents
Redesign of MCAS Tests Based on a Consideration of Information Functions 1,2. (Revised Version) Ronald K. Hambleton and Wendy Lam

Longitudinal Effects of Item Parameter Drift. James A. Wollack Hyun Jung Sung Taehoon Kang

Investigating Common-Item Screening Procedures in Developing a Vertical Scale

Computer Software for IRT Graphical Residual Analyses (Version 2.1) Tie Liang, Kyung T. Han, Ronald K. Hambleton 1

A standardization approach to adjusting pretest item statistics. Shun-Wen Chang National Taiwan Normal University

UK Clinical Aptitude Test (UKCAT) Consortium UKCAT Examination. Executive Summary Testing Interval: 1 July October 2016

Effects of Selected Multi-Stage Test Design Alternatives on Credentialing Examination Outcomes 1,2. April L. Zenisky and Ronald K.

Reliability & Validity

Assessing first- and second-order equity for the common-item nonequivalent groups design using multidimensional IRT

Assessing first- and second-order equity for the common-item nonequivalent groups design using multidimensional IRT

Practitioner s Approach to Identify Item Drift in CAT. Huijuan Meng, Susan Steinkamp, Pearson Paul Jones, Joy Matthews-Lopez, NABP

proficiency that the entire response pattern provides, assuming that the model summarizes the data accurately (p. 169).

The Effects of Model Misfit in Computerized Classification Test. Hong Jiao Florida State University

Potential Impact of Item Parameter Drift Due to Practice and Curriculum Change on Item Calibration in Computerized Adaptive Testing

Linking Current and Future Score Scales for the AICPA Uniform CPA Exam i

Scoring Subscales using Multidimensional Item Response Theory Models. Christine E. DeMars. James Madison University

Glossary of Terms Ability Accommodation Adjusted validity/reliability coefficient Alternate forms Analysis of work Assessment Band Battery

ASSUMPTIONS OF IRT A BRIEF DESCRIPTION OF ITEM RESPONSE THEORY

Computer Adaptive Testing and Multidimensional Computer Adaptive Testing

Testing Issues and Concerns: An Introductory Presentation to the State Directors of Career and Technical Education

MCS Experience in Selection & Recruitment (Saudi Arabia)

MODIFIED ROBUST Z METHOD FOR EQUATING AND DETECTING ITEM PARAMETER DRIFT. Keywords: Robust Z Method, Item Parameter Drift, IRT True Score Equating

Determining the accuracy of item parameter standard error of estimates in BILOG-MG 3

IRT-Based Assessments of Rater Effects in Multiple Source Feedback Instruments. Michael A. Barr. Nambury S. Raju. Illinois Institute Of Technology

Field Testing and Equating Designs for State Educational Assessments. Rob Kirkpatrick. Walter D. Way. Pearson

Maintenance of Vertical Scales Under Conditions of Item Parameter Drift and Rasch Model-data Misfit

VQA Proficiency Testing Scoring Document for Quantitative HIV-1 RNA

Technical Report: Does It Matter Which IRT Software You Use? Yes.

What are the Steps in the Development of an Exam Program? 1

Running head: GROUP COMPARABILITY

Standard Setting and Score Reporting with IRT. American Board of Internal Medicine Item Response Theory Course

Glossary of Standardized Testing Terms

Test-Free Person Measurement with the Rasch Simple Logistic Model

STANDARD FOR THE VALIDATION AND PERFORMANCE REVIEW OF FRICTION RIDGE IMPRESSION DEVELOPMENT AND EXAMINATION TECHNIQUES

Innovative Item Types Require Innovative Analysis

Using a Performance Test Development & Validation Framework

An Automatic Online Calibration Design in Adaptive Testing 1. Guido Makransky 2. Master Management International A/S and University of Twente

Understanding the Dimensionality and Reliability of the Cognitive Scales of the UK Clinical Aptitude test (UKCAT): Summary Version of the Report

Designing item pools to optimize the functioning of a computerized adaptive test

An Introduction to Psychometrics. Sharon E. Osborn Popp, Ph.D. AADB Mid-Year Meeting April 23, 2017

Linking Errors Between Two Populations and Tests: A Case Study in International Surveys in Education

Ensuring Comparative Validity

CLEAR Exam Review. Volume XXVII, Number 2 Winter A Journal

STUDY GUIDE. California Commission on Teacher Credentialing CX-SG

Validity and Reliability Issues in the Large-Scale Assessment of English Language Proficiency

Quality Control Assessment in Genotyping Console

PRINCIPLES AND APPLICATIONS OF SPECIAL EDUCATION ASSESSMENT

Dealing with Variability within Item Clones in Computerized Adaptive Testing

Screening Score Report

Application of Multilevel IRT to Multiple-Form Linking When Common Items Are Drifted. Chanho Park 1 Taehoon Kang 2 James A.

Conjoint analysis based on Thurstone judgement comparison model in the optimization of banking products

An Evaluation of Kernel Equating: Parallel Equating With Classical Methods in the SAT Subject Tests Program

More than 70 percent of the students are likely to respond correctly. Easy

Automated Test Assembly for COMLEX USA: A SAS Operations Research (SAS/OR) Approach

A Gradual Maximum Information Ratio Approach to Item Selection in Computerized Adaptive Testing. Kyung T. Han Graduate Management Admission Council

Martin Senkbeil and Jan Marten Ihme. for Grade 9

Test Development. and. Psychometric Services

Differential Item Functioning

The Effects of Employee Ownership in the Context of the Large Multinational : an Attitudinal Cross-Cultural Approach. PhD Thesis Summary

STAAR-Like Quality Starts with Reliability

Overview. Considerations for Legal Defensibility in Standard Setting. Portland, Oregon September 16, 2016

Independent Verification of the Psychometric Validity for the Florida Standards Assessment. Executive Summary. August 31, 2015

Audit - The process of conducting an evaluation of an entity's compliance with published standards. This is also referred to as a program audit.

EIOPA Common Application Package For Internal Models: Explanatory Note

ITEM RESPONSE THEORY FOR WEIGHTED SUMMED SCORES. Brian Dale Stucky

World Journal of Engineering Research and Technology WJERT

Test Analysis Report

Texas Educator Certification Program Technical Manual

Six Major Challenges for Educational and Psychological Testing Practices Ronald K. Hambleton University of Massachusetts at Amherst

VALIDATING CRITERION-REFERENCED READING TESTS a. Daniel S. Sheehan and Mary Marcus Dallas Independent School District

An Approach to Implementing Adaptive Testing Using Item Response Theory Both Offline and Online

Report on LLQP Curriculum Survey Results. March 2014

The SPSS Sample Problem To demonstrate these concepts, we will work the sample problem for logistic regression in SPSS Professional Statistics 7.5, pa

Estimating Standard Errors of Irtparameters of Mathematics Achievement Test Using Three Parameter Model

7 Statistical characteristics of the test

Higher National Unit specification: general information. Graded Unit 3

NATIONAL CREDENTIALING BOARD OF HOME HEALTH AND HOSPICE TECHNICAL REPORT: JANUARY 2013 ADMINISTRATION. Leon J. Gross, Ph.D. Psychometric Consultant

WORK ENVIRONMENT SCALE 1. Rudolf Moos Work Environment Scale. Adopt-a-Measure Critique. Teresa Lefko Sprague. Buffalo State College

CASE STUDY OF APPLYING DIFFERENT ENERGY USE MODELING METHODS TO AN EXISTING BUILDING

Estimating Reliabilities of

INTERNAL CONSISTENCY OF THE TEST OF WORKPLACE ESSENTIAL SKILLS (TOWES) ITEMS BASED ON THE LINKING STUDY DATA SET. Report Submitted To: SkillPlan

Psychometric Issues in Through Course Assessment

A Systematic Approach to Performance Evaluation

Smarter Balanced Scoring Specifications

University of Groningen. Education in laparoscopic surgery Kramp, Kelvin Harvey

Evaluating the Technical Adequacy and Usability of Early Reading Measures

The Examination for Professional Practice in Psychology: The Enhanced EPPP Frequently Asked Questions

Elementary Education Subtest II (103)

Elementary Education Subtest II (103)

Linking DIBELS Next with the Lexile Framework

Technical Report: June 2018 CKE 1. Human Resources Professionals Association

DATA-INFORMED DECISION MAKING (DIDM)

RUNNING HEAD: MODELING LOCAL DEPENDENCE USING BIFACTOR MODELS 1

National Council for Strength & Fitness

Seton Hall University BACCALAUREATE SOCIAL WORK PROGRAM ASSESSMENT OF STUDENT LEARNING OUTCOMES. LAST COMPLETED ON July 2, 2015

Texas Educator Certification Program Technical Manual

Model construction of earning money by taking photos

New England Fishery Management Council DRAFT MEMORANDUM

Top-Down Proteomics Enables Comparative Analysis of Brain. Proteoforms Between Mouse Strains

Energy Trust of Oregon New Homes Billing Analysis: Comparison of Modeled vs. Actual Energy Usage

Transcription:

Operational Check of the 2010 FCAT 3 rd Grade Reading Equating Results Prepared for the Florida Department of Education by: Andrew C. Dwyer, M.S. (Doctoral Student) Tzu-Yun Chin, M.S. (Doctoral Student) Kurt F. Geisinger, Ph.D. May 13, 2010 Questions concerning this report can be addressed to: Kurt F. Geisinger, Ph.D. Buros Center for Testing 21 Teachers College Hall University of Nebraska Lincoln Lincoln, NE, 68588-0353 kgeisinger2@unl.edu

1 Operational Check of the 2010 FCAT 3 rd Grade Reading Equating Results Purpose of the report In order to assist in ensuring the accuracy of the scores from the 2010 FCAT examinations, the Buros Center for Testing was asked by the Florida Department of Education (FDOE) to review the procedures used in their 2010 FCAT item calibration, scaling, and equating procedures and to perform an independent check of their calibration and equating results using data and anchor item parameter estimates provided by the FDOE. This study is similar to one conducted by Buros in 2009, and this report follows a highly similar format to the 2009 report. Although we have been asked to examine the equating results of three 2010 tests (3 rd grade reading, 10th grade reading, 11 th grade science), in this draft report we only address the results from the operational check of the 3 rd grade reading test. The results pertaining to the other tests will follow in the coming weeks. This draft report (1) provides an outline of the procedures we used to operationally calibrate the items and link their parameter estimates to the FCAT reporting scale, (2) summarizes our results, and (3) compares them to results obtained by the FDOE. Operational item calibration and scaling We carried out an operational check of the item calibration, scaling and equating of the 3 rd grade reading FCAT test. This test was chosen by the FDOE to be reviewed in part because of the high stakes nature of this assessment. More specifically, the Grade 3 Reading assessment has a state-level retention cut-point that a student must pass in order

2 to advance to the next grade. This section provides a short description of the procedures we followed to arrive at our equating results. Step 1: Calibration data files The FDOE provided us with a calibration data set, but minor data cleaning was still required for each grade. For our analysis, we included only examinees who received an anchor form (i.e., forms 37, 38, 39, 40). Non-standard curriculum students and students who would not receive reported scores were excluded from the data file as well. We understand that this practice is consistent with procedures followed by the FDOE. The sample sizes of the calibration samples for each grade are reported in Table 1; we believe that these sizes are adequate for equating purposes. Table 1: Calibration sample size information Total N Form 17 Form 18 Form 19 Form 20 Grade 3 12,654 3,175 3,161 3,134 3,184 Step 2: Classical item statistics Classical item statistics were computed to identify any items that might need to be removed from the item calibration step. For our analysis, we computed item p-values and corrected item-total correlations. Items with extreme p-values or low item-total correlations were flagged for further psychometric review using the same flagging criteria described in the calibration and equating specifications document provided by the FDOE. A few individual items were flagged, but after further review of each item, including item-fit statistics, no item was judged to be problematic enough to be removed from item calibration.

3 Step 3: Item Calibration All items (i.e., forms) were calibrated concurrently using PARSCALE-4 (Muraki & Bock, 2003) using the 3PL model. All items achieved convergence. Table 2: Description of item calibration set Grade Description 3 45 core items, 27 anchor items (77 total) Step 4: Item ICC s and Item Fit Statistics Item fit statistics as well as graphs of the ICC s (expected vs. observed) were examined to determine if any item should be removed from the item calibration. PARSCALE s G 2 statistic was computed and employed to assess item fit. The ICC s and residual plots were created by the software program ResidPlots-2 (Liang, T., Han, K. T, & Hambleton, R. K., 2008). Although a few items were identified as having a significant lack of fit, PARSCALE s G 2 statistic is known to be overly sensitive with large sample sizes 1, and the ICC plots did not indicate major problems with any of the items, therefore no items were removed from the item calibration or from the subsequent item scaling based on these measures. Step 5: Year-to-year anchor item performance A number of methods were used to determine if the anchor items were behaving differently in the current administration than they had in previous administrations. All of the methods used in the 2009 check were again used in 2010, including delta plots of item difficulties, correlations and scatter plots of new and old anchor item parameter 1 Each core item, for example, was administered to all students in the calibration sample, while each anchor item was administered to roughly one fourth of the students in the calibration sample. As a result, a higher proportion of core items were artificially flagged as being misfitting.

4 estimates, and correlations and scatter plots between shifts in item difficulty and shifts in item position (within the test). The root mean squared deviations (RMSD) of the a- and b-parameters (see Gu, Lall, Monfils, & Jiang, 2010 for a detailed description) were also added. Shifts in item position were not problematic for the anchor items in any of the anchor forms, and although a few individual items flagged as performing differently across administrations based on some of the measures, none of the individual anchor items were deemed problematic enough to warrant removal from the anchor set. The correlation between the old and new a-parameters for form 40 was lower (r a,form 40 = 0.506) than for other forms (r a,form 37 = 0.850, r a,form 38 = 0.879, r a,form 39 = 0.836). This low correlation seems to be an artifact of the range restriction of the new and old a- parameters for form 40, nonetheless, the equating results were investigated both with and without the form 40 anchor items in the anchor set. When the form 40 anchors were removed, the overall anchor set was shortened (and perhaps the content representation suffered slightly), so we also investigated the results of adding in one of the sets of backup anchor items (i.e., core items that could be used as anchors). After examining items statistics for each of the backup anchor item set, we felt that the first 10 core items (from the ZUM03 passage code) formed the best replacement anchor set, but we acknowledge that this decision was based solely on item performance statistics and that the other backup anchor set may be preferred for content representation purposes. At last, we also reproduced the equating results with the final anchor set selected by FDOE. The advantage of performing this analysis done after our independent

5 analyses was that it demonstrates the variance due purely to software differences. Table 3 summarizes the possible anchor sets we chose to examine in the item parameter scaling stage. Table 3: Description of anchor item set for 3 rd grade reading Anchor item set Description Grade 3 #1 All planned anchor items used (n=27) #2 Form 40 anchor items removed (n=20) #3 Form 40 removed, Core items 1-10 added (n=30) #4* All planned anchor items, Core items 1-11 and 13-21 added * FDOE s final anchor set The Stocking and Lord (1983) equating procedure was implemented for each anchor item set described above, and the impact on final scores was examined as part of the justification for removing items from the anchor set. The equating results, including the transformation coefficients (M1 and M2) for all three anchor sets described in Table 3 are very similar. Those results are presented in more detail in Table 4. Step 6: Item parameter scaling The item parameters were placed on the FCAT reporting scale by the program IRTEQ (Han, K. T., 2007). This program uses Stocking & Lord (1983) methodology to adjust the item parameter estimates obtained in Step 3 so that they are on the FCAT reporting scale. A separate item parameter scaling run was performed for each anchor set that was considered (described in Step 5). Step 7: Examination of results For each item parameter scaling run, the linking coefficients (M1 and M2) from the Stocking & Lord (1983) procedure were compared to the final coefficients computed by the FDOE in their operational scaling and equating. Software differences and slight

6 differences in anchor sets were expected to produce slightly different linking coefficients, so differences between the scale scores that would be obtained using the FDOE linking coefficients and scale scores that would be obtained using our linking coefficients were calculated at important points along the FCAT reporting scale (specifically, the cutoff points that separate the FCAT reporting categories). The FCAT cut scores were first transformed back to the theta scale using the M1 and M2 coefficients produced by FDOE. The cut points on the theta scale were then transformed to the FCAT scale using the M1 and M2 obtained with different anchor sets from our investigation. This procedure was followed to provide a means for determining the magnitude/importance of the differences between our linking coefficients and the FDOE linking coefficients. The results are reported in Table 4. Before discussing the information in Table 4, it is important to note that just as in 2009, we expected slight differences between our results and the values provided by the FDOE due to differences in calibration and equating software. FDOE and its subcontractors used MULTILOG for calibration while we used PARSCALE in this project. It has been demonstrated in the psychometric literature that different IRT software packages usually produce different, though highly similar, parameter estimates (e.g., Childs and Chen, 1999). With a simulation study, DeMars (2002) further concluded that MULTILOG and PARSCALE could both recover item parameters well and the selection between the two software packages for a specific project can be made based on considerations other than estimation accuracy. In addition to calibration software packages, calibration options that may assist parameter estimation (e.g., prior distributions, starting values, and RIDGE estimator) and convergence criteria within a

7 given IRT calibration software package can also introduce differences in parameter estimates. Besides different calibration software, we also used a different software package (i.e., IRTEQ) for implementing the Stocking & Lord procedure in order to obtain the linking coefficients. The difference in the equating programs can also result slightly different linking coefficients. One possible advantage of using different software packages can be the additional evaluation of the stability of the parameter estimates and linking coefficients. Moreover, duplication of analyses with the same software basically only has the potential to identify errors of analysis or judgment. Using different software permits consideration of the two above-named types of errors but also considers differences across standard software packages. Because evaluating the content representation of the anchor set was outside the scope of this project, we had no reason to remove (or add) items from the anchor set based on item content. We believe, based on observation and conversations we have had with the FDOE, that that task has been handled appropriately by the FDOE. Taking software differences into consideration, our item calibration and equating results should be considered consistent with the results obtained by FDOE and its subcontractors. The last four columns of Table 4 show the magnitude of the differences between scale scores calculated by FDOE and scale scores calculated by Buros at several important points along the FCAT reporting scale. Generally speaking, the Buroscalculated scale scores are always within a few points of the FDOE-calculated scores, especially at points located towards the center of the scale. Differences this small are likely the result of using different software packages (McKinley, personal

8 communication, 2009), thus the consistency of the results in Table 4 lend support to the accuracy of the 2010 FCAT scores. Overall impressions of the calibration, scaling, and equating procedures Despite knowing that software differences would likely lead to slightly different results, carrying out the operational calibration and equating is an important step in to the evaluation of the calibration and equating procedures used by the FDOE. The FCAT 2010 Calibration and Equating Specifications (2010) document provided to us by the FDOE, which served as a guide when we performed that work, was also reviewed. Overall, our impressions were that the entire process is well-organized, the statistical analyses used to identify problematic items are thorough, the organizations involved in the operational work are all nationally recognized testing firms composed of high quality staff, and the responsibilities of those organizations are clearly defined and set up in a way to ensure the accuracy of the scores on the 2010 FCAT. Moreover, they have built effective checks and balances into the equating process, and while equating is a highly quantitative procedure, there is also considerable judgment involved.

9 Table 4: Scaling and equating results for each calibration/anchor set combination Grade 3 Description of Calibration and Scaling items Calibration items All core items used and 27 planned anchors used in calibration Transformation Coefficients Differences between Scale Scores in the area of achievement level cut points Anchor items M1 M2 Cut 1 2 Cut 2 3 Cut 3 4 Cut 4 5 FDOE values 48.565 321.371 259 284 332 394 All "planned" anchor items included in anchor set (n=27) 49.73 323.53 259.7 285.3 334.4 397.9 Form 40 anchor items removed from anchor set (n=20) Form 40 removed and Core 50.23 323.93 259.4 285.3 334.9 399.0 Items 1 10 added (n=30) 50.32 324.31 259.7 285.6 335.3 399.6 FDOE s final set : all "planned" anchor items and Core Items 1 11 and 13 21 added (n=47) 49.62 321.76 258.0 283.6 332.6 396.0

10 References FCAT 2010 Calibration and Equating Specifications (January, 2010). Final Version. Childs, R. A., & Chen, W.-H. (1999). Software note: Obtaining comparable item parameter estimates in MULTILOG and PARSCALE for two polytomous IRT models. Applied Psychological Measurement, 23, 371-379. DeMars, C. E. (2002). Recovery of Graded Response and Partial Credit Parameters in MULTILOG and PARSCALE. Paper presented at the annual meeting of the National Council for Measurement in Education, Chicago, IL. (ERIC Document Reproduction Service No. ED476138) Gu, L., Lall, V.F., Monfils, L., Jiang, Y. (2010). Evaluating anchor items for outliers in IRT common item equating: A review of the commonly used methods and flagging criteria. Paper presented at the annual meeting of the National council on Measurement in Education (NCME). April 29-May 3, 2010, Denver, CO. Han, K. T. (2007). IRTEQ [Computer software]. Amherst, MA: Center for Educational Assessment, University of Massachusetts at Amherst. Liang, T., Han, K. T, & Hambleton, R. K. (2008). ResidPlots-2 [Computer software]. Amherst, MA: Center for Educational Assessment, University of Massachusetts at Amherst. McKinley, R. (2010). Personal communication, May 20, 2010. Muraki, E., & Bock, R. D. (2003). PARSCALE-4: IRT item analysis and test scoring for ratingscale data [Computer software]. Chicago: Scientific Software International. Stocking, M. L., & Lord, F. M. (1983). Developing a common metric in item response theory. Applied Psychological Measurement, 7, 201-210.