|
TABLE OF CONTENTS |
TABLE OF CONTENTS |
LIST OF TABLES |
LIST OF FIGURES |
ACKNOWLEDGEMENTS
|
1. |
Introduction
|
|
1.1 |
|
Background |
|
1.2 |
|
Rationale for the 2010 Mod-MSA: Reading |
|
1.3 |
|
Eligibility Criteria for the Mod-MSA: Reading
|
|
1.4 |
|
Test Administration of the 2010 Mod-MSA: Reading |
|
|
|
Test Materials |
|
|
|
Test Administration Schedule |
|
|
|
Student Participation |
|
|
|
Accommodations for Assessment |
|
|
|
Test Format |
|
|
|
Security of Test Materials |
|
1.5 |
|
Quality Control Procedures |
|
1.6 |
|
Item Bank Construction
|
2. |
Test Design, Development, and Scoring of the 2010 Mod-MSA: Reading
|
|
2.1 |
|
Test Design and Structure of the 2010 Mod-MSA: Reading |
|
2.2 |
|
Development and Review of the 2010 Mod-MSA: Reading |
|
|
|
National Psychometric Council |
|
|
|
Content Review Committee |
|
|
|
Bias Review Committee |
|
|
|
Vision Review Committee |
|
2.3 |
|
Operational Form Structure of the 2010 Mod-MSA: Reading |
|
2.4 |
|
Item Analyses Undertaken Prior to the Creation of the Operational Forms |
|
|
|
Classical Item Analyses |
|
|
|
Differential Item Functioning Analyses |
|
|
|
Item Response Theory (IRT) Analyses |
|
2.5 |
|
Items Flagged for Inspection Prior to the Creation of the Operational Forms |
|
2.6 |
|
Items Selected for the 2010 Operational Tests |
|
2.7 |
|
Scoring Procedures of the 2010 Mod-MSA: Reading
|
3. |
Central Tendency Measures and Classical Item Analyses of the 2010 Mod-MSA: Reading
|
|
3.1 |
|
Measures of Central Tendency |
|
|
|
Standard Error of the Mean |
|
3.2 |
|
Item-Level Descriptive Statistics
|
4. |
Scale Creation, Equating and Raw Scores to Scale Scores Conversion via Item Response Theory Procedures
|
|
|
|
The Rasch Model |
|
4.1 |
|
Calibration and Scaling Procedures for Grades 3 to 5 |
|
4.2 |
|
Specifics for Creating the Base Scale for the Mod-MSA: Reading Grades 3-5 |
|
4.3 |
|
Calibration Equating the 2010 Mod-MSA: Reading Grades 6-8 |
|
4.4 |
|
Specifics of Linking and Equating the 2010 Mod-MSA Grades 6-8: Reading |
|
4.5 |
|
Reporting Scale Scores for the 2010 Mod-MSA: Reading |
|
4.6 |
|
Score Interpretation |
|
4.7 |
|
Final Performance Level Cut Points for the Mod-MSA: Reading
|
5. |
Comparative Classical and IRT Stats Across Years for the Common Items Used in Grades 6-8
|
6. |
IRT Item Level Statistics for the 2010 Mod-MSA: Reading
|
|
6.1 |
|
Rationale for the Use of the Rasch Model |
|
6.2 |
|
Fit Statistics for the Rasch Model |
|
|
|
Mean-Square |
|
6.3 |
|
Rasch Item Level Statistics
|
7. |
Test Reliability
|
|
7.1 |
|
Precision and Reliability (Classical Methods) |
|
|
|
Standard Error of Measurement (SEM) of the Test |
|
|
|
Cronbach’s Alpha (KR20) |
|
7.2 |
|
IRT Method in Measuring Precision of the Test |
|
7.3 |
|
Decision Accuracy and Consistency at the Cut Scores
|
8. |
Test Validity
|
|
8.1 |
|
Test Validity for the 2010 Mod-MSA: Reading |
|
|
|
Evidence from Internal Structure of the Tests |
|
8.2 |
|
Unidimensionality Analysis for the 2010 Mod-MSA: Reading
|
9. |
Summary of Operational Test Results of the 2010 Mod-MSA: Reading
|
|
9.1 |
|
Classical Descriptive Test Statistics |
|
9.2 |
|
Scale Score Descriptive Statistics |
|
9.3 |
|
Frequency of Students at Each Proficiency Level by Grade |
|
9.4 |
|
Test Characteristic Curves, Test Information Functions, and Conditional Standard Errors of Measurement
|
10. |
References
|
APPENDICES
|
A.
| Operational Items Selected After Data Review |
B.
| Item Distractor Analyses |
C.
| Frequency Distribution Histograms of Scale Scores |
D.
| Standard Setting Report |
E.
| Alignment of the 2010 Mod-MSA: Reading Items to the State Standards |
F.
| Identifying Students for Participation in the Mod-MSA: Reading Program |
G.
| Comparability Study of Paper and Pencil, and Online Administration of the Mod-MSA
|