Professional Documents
Culture Documents
Pruebas de aprovechamiento
Main Logo_IV_4c
Examiner’s Manual
Barbara J. Wendling ◆ Nancy Mather ◆ Fredrick A. Schrank
Reference Citations
■ To cite the entire Batería IV battery, use:
Woodcock, R. W., Alvarado, C. G., Schrank, F. A., McGrew, K. S., Mather, N., &
Muñoz-Sandoval, A. F. (2019). Batería IV Woodcock-Muñoz. Itasca, IL: Riverside
Assessments, LLC.
■ To cite the Batería IV Pruebas de aprovechamiento, use:
Woodcock, R. W., Alvarado, C. G., Schrank, F. A., Mather, N., McGrew, K. S., & Muñoz-
Sandoval, A. F. (2019). Batería IV Woodcock-Muñoz: Pruebas de aprovechamiento. Itasca,
IL: Riverside Assessments, LLC.
■ To cite this manual, use:
Wendling, B. J., Mather, N., & Schrank, F. A. (2019). Examiner’s Manual. Batería IV
Woodcock-Muñoz: Pruebas de aprovechamiento. Itasca, IL: Riverside Assessments, LLC.
Copyright © 2019 by Riverside Assessments, LLC. All rights reserved. No part of this work
may be reproduced or transmitted in any form or by any means, electronic or mechanical,
including photocopying and recording, or by any information storage or retrieval system
without the prior written permission of Riverside Assessments, LLC unless such copying
is expressly permitted by federal copyright law. Requests for permission to make copies of
any part of the work should be addressed to Riverside Insights, Attention: Permissions, One
Pierce Place, Suite 900W, Itasca, Illinois 60143.
Published in Itasca, IL
Batería III Woodcock-Muñoz, WJ-III, WJ-R, Woodcock-Johnson, Woodcock Language
Proficiency Battery, and Woodcock-Muñoz Language Survey are registered trademarks of
Riverside Assessments, LLC.
The Batería IV logo, Bilingual Verbal Ability Tests, Dean-Woodcock, Riverside Insights,
the Riverside Insights logo, Scales of Independent Behavior–Revised, WJ IV, WJ IV
Interpretation and Instructional Interventions Program, Woodcock Interpretation and
Instructional Interventions Program, and Woodcock Language Proficiency Battery are
trademarks of Riverside Assessments, LLC.
The MindHub is a registered trademark of the Institute for Applied Psychometrics (IAP)
and Interactive Metronome.
All other trademarks are the property of their respective owners.
The Batería IV tests are not to be used in any program operating under statutes or
regulations that require disclosure of specific item content and/or correct responses to the
public, including examinees or their parents. Any unauthorized distribution of the specific
item content and/or correct responses is prohibited by copyright law.
For technical information, please visit http://www.wj-iv.com or call Riverside Insights
Customer Service at 800.323.9540.
About the Authors of the
Batería IV
Richard W. Woodcock
Richard W. Woodcock, EdD, has an extensive background in education and psychology. He
has held a variety of positions, including elementary school teacher, school psychologist,
director of special education, university professor, and editor of a test publishing company.
He earned his doctorate degree from the University of Oregon with a dual major in statistics
and psycho-education, and he completed a post-doctoral fellowship in neuropsychology at
Tufts University School of Medicine. He has published more than 135 professional books and
articles.
Since 1957, Dr. Woodcock has held positions and appointments at Western Oregon
University, the University of Northern Colorado, George Peabody College for Teachers, the
University of Arizona, the University of Southern California, the University of Virginia, and
Vanderbilt University. He is currently serving as a research professor in the Department
of Psychology and Philosophy at Texas Woman’s University. Dr. Woodcock is a Fellow of
the American Psychological Association and the American Academy of School Psychology
and a Diplomate of the American Board of Professional Psychology. In 2013 the Texas
Statewide Evaluation Project Conference honored Dr. Woodcock with its first Lifetime
Achievement Award.
Among Dr. Woodcock’s publications are the Colorado Braille Battery, the Peabody Rebus
Reading Program, the Goldman-Fristoe-Woodcock Auditory Skills Test Battery, the Woodcock
Reading Mastery Tests®–Revised, the Woodcock-Johnson® Psycho-Educational Battery–Revised,
the Batería Woodcock-Muñoz: Pruebas de habilidad cognitiva–Revisada, the Batería Woodcock-
Muñoz: Pruebas de aprovechamiento–Revisada, the Woodcock Language Proficiency Battery®–
Revised, the Scales of Independent Behavior–Revised™, the Woodcock-Muñoz Language Survey®,
the Woodcock Diagnostic Reading Battery, the Woodcock-Johnson III Tests of Cognitive Abilities,
the Woodcock-Johnson III Tests of Achievement, the Dean-Woodcock™ Neuropsychological
Assessment System, the Bilingual Verbal Ability Tests™, the Batería III Woodcock-Muñoz®:
Pruebas de habilidades cognitivas, the Batería III Woodcock-Muñoz: Pruebas de aprovechamiento,
and the Woodcock-Muñoz Language Survey–Revised.
iii
Criselda G. Alvarado
Criselda Guajardo Alvarado received her BS degree with teaching certificates in elementary
education, psychology, bilingual education, and mathematics and her MEd as an educational
diagnostician from the University of Texas Rio Grande Valley. She received her EdD in
Curriculum and Instruction with a specialization in bilingual education at the University of
Houston.
Dr. Alvarado devoted her professional life to education and assessment. She began her
career as a bilingual education teacher in an elementary school in Donna, Texas, before
becoming a middle school mathematics and reading teacher in Robstown, Texas. Dr. Alvarado
then moved to Houston, Texas, and worked as a bilingual elementary school teacher. She
also worked as an educational diagnostician and special education supervisor in several
elementary, middle, and high schools as well as in a juvenile justice charter school. She was
an adjunct professor at the University of Houston from 1994 to 1999 and in 2016, and she
was an assistant professor at the University of Houston–Clear Lake from 2012 to 2015. In
1994, she founded Education and Evaluation Consultants, where she worked for various
school districts and state departments as a consultant on dyslexia, special education, bilingual
education, and educational assessment. For several years, Dr. Alvarado taught in and directed
an all-volunteer after-school tutorial program for children and an English as a second
language (ESL) program for adults in her local community.
From 1991 to 2015, Dr. Alvarado was the director of bilingual studies at Measurement
Learning Consultants, where she worked on several English and Spanish standardization
projects, including the Woodcock-Muñoz Language Survey, Batería Woodcock-Muñoz: Pruebas
de habilidades cognitivas–Revisada, Batería Woodcock-Muñoz: Pruebas de aprovechamiento–
Revisada, Woodcock-Johnson III Tests of Cognitive Abilities, Woodcock-Johnson III Tests of
Achievement, Batería III Woodcock-Muñoz: Pruebas de habilidades cognitivas, and Batería III
Woodcock-Muñoz: Pruebas de aprovechamiento. She is a coauthor of the Woodcock-Muñoz
Language Survey–Revised, the Bilingual Verbal Ability Tests, and the Woodcock-Muñoz
Language Survey III.
Fredrick A. Schrank
Fredrick A. (Fred) Schrank is the senior author of the Woodcock-Johnson IV (WJ IV™) family
of tests. Fred is a licensed psychologist in the state of Washington and is a board-certified
specialist in school psychology by the American Board of Professional Psychology (ABPP).
He began his professional career as a counselor and diagnostician in the Dodgeville, North
Fond du Lac, and De Forest, Wisconsin, public school districts while simultaneously earning
an EdS degree in Educational Psychology from the University of Tennessee–Knoxville. He
subsequently earned a PhD from the University of Wisconsin–Madison. Dr. Schrank then
taught in graduate-level professional preparation programs at Truman State University
(Missouri) and the University of Puget Sound (Washington) prior to a three-decade career
devoted almost exclusively to the development and publication of the Woodcock-Johnson
family of tests. As part of a commitment to professional psychology, he has served as an oral
examiner for the American Board of School Psychology (ABSP) and was elected president of
the American Academy of School Psychology (AASP).
Dr. Schrank is the lead author on several WJ IV test batteries, intervention programs,
books, and book chapters, including the Woodcock-Johnson IV Tests of Cognitive Abilities, the
Woodcock-Johnson IV Tests of Oral Language, the Woodcock-Johnson IV Tests of Achievement, the
Woodcock-Johnson IV Tests of Early Cognitive and Academic Development, the WJ IV Interpretation
iv
and Instructional Interventions Program, and Essentials of WJ IV Cognitive Abilities Assessment. In
2018, Dr. Schrank was named a J. William Fulbright Specialist by the U.S. Department of State,
Bureau of Educational and Cultural Affairs. He is hosted by the Department of Educational and
Counselling Psychology of McGill University in Montreal, Quebec.
Nancy Mather
Nancy Mather is a professor at the University of Arizona in the Department of Disability
and Psychoeducational Studies. She holds an MA in Behavior Disorders and a PhD from
the University of Arizona in Special Education and Learning Disabilities. She completed
a postdoctoral fellowship under the mentorship of Dr. Samuel Kirk at the University of
Arizona.
Dr. Mather assisted Dr. Richard Woodcock with several aspects of test development for
the Woodcock-Johnson Psycho-Educational Battery–Revised (WJ-R®), including coauthoring
the Examiner’s Manuals for the WJ-R Tests of Cognitive Ability and the WJ-R Tests of
Achievement. She has been a coauthor of both the Woodcock-Johnson III (WJ III®) and the
WJ IV and has coauthored two books on the interpretation and application of the WJ IV—
Essentials of WJ IV Tests of Achievement and Woodcock-Johnson IV: Reports, Recommendations,
and Strategies.
She has served as a learning disabilities teacher, a diagnostician, a university professor,
and an educational consultant. Dr. Mather conducts research in the areas of reading and
writing development and conducts workshops on assessment and instruction both nationally
and internationally. She has published numerous articles and has coauthored several
books linking assessment and intervention, including Learning Disabilities and Challenging
Behaviors: A Guide to Intervention and Classroom Management (3rd ed.), Evidence-Based
Interventions for Students with Learning and Behavioral Challenges, Essentials of Assessment
Report Writing (2nd ed.), Essentials of Evidence-Based Academic Interventions, Writing
Assessment and Instruction for Students with Learning Disabilities (2nd ed.), and Essentials of
Dyslexia: Assessment and Intervention.
Kevin S. McGrew
Kevin S. McGrew is director of the Institute for Applied Psychometrics (IAP), LLC, a private
research and consulting organization he established in 1998. He also is a visiting lecturer
in Educational Psychology (School Psychology Program) at the University of Minnesota
and director of research for Interactive Metronome, a neurotechnology and rehabilitation
company. He holds a PhD in Educational Psychology (Special Education) from the University
of Minnesota and an MS in School Psychology and a BA in Psychology from Minnesota State
University–Moorhead.
Dr. McGrew was a practicing school psychologist for 12 years in Iowa and Minnesota.
From 1989 to 2000, he was a professor in the Department of Applied Psychology at St.
Cloud State University, St. Cloud, Minnesota. He has served as a measurement consultant
to a number of psychological test publishers, national research studies, and organizations.
Since 2009 Dr. McGrew has served as a consulting expert, via declarations or testimony, to
the courts regarding the measurement of intelligence and psychometric issues relevant to
intellectual assessment in death penalty cases (capital punishment cases involving individuals
with intellectual disabilities). Since 2014 he has served as an intelligence theory and testing
consultant at the Dharma Bermakna Foundation and the Universitas Gadjah Mada for the
Indonesia AJT Cognitive Assessment Development Project.
v
He has authored numerous publications and made state, national, and international
presentations in his primary areas of research interest in human intelligence, intellectual
assessment, human competence, applied psychometrics, and the Cattell-Horn-Carroll (CHC)
theory of cognitive abilities. He is a frequent speaker at state, national, and international
conferences. He is an active distributor of theoretical and research information via three
professional blogs and The MindHub® web portal.
Ana F. Muñoz-Sandoval
Dr. Ana F. Muñoz-Sandoval was associate director of Measurement Learning Consultants
for two decades. During that time, she directed and authored the Spanish adaptation of
several assessment instruments. Most recently, she authored the Batería III Woodcock-
Muñoz: Pruebas de habilidades cognitivas, the Batería III Woodcock-Muñoz: Pruebas de
aprovechamiento, and the Bilingual Verbal Ability Tests™. Dr. Muñoz-Sandoval also coauthored
the Woodcock-Muñoz Language Survey, the Woodcock Language Proficiency Battery–Revised
(Spanish Form), the Batería Woodcock-Muñoz: Pruebas de habilidad cognitiva–Revisada, and
the Batería Woodcock-Muñoz: Pruebas de aprovechamiento–Revisada. In 1997, Dr. Muñoz-
Sandoval participated in the development and standardization of the Woodcock-Johnson III.
By invitation, she has presented assessment workshops in Argentina, Brazil, Costa Rica, and
Mexico.
Dr. Muñoz-Sandoval holds teaching credentials from Mendoza, Argentina, where she
lived before coming to the United States in 1970. She has also lived in Italy, Nepal, Pakistan,
and South Africa. She studied German at the Tribhuvan University in Kathmandu, Nepal.
She devoted a decade to teaching Spanish language at the university level. She received a
BA degree in Anthropology from the State University College at Buffalo. That was followed
by an MS in Student Personnel Administration. She earned her EdD from the University of
Southern California with specialization in international and multicultural education.
Contributing Author
Barbara J. Wendling coauthored the Batería IV Examiner’s Manuals with Nancy Mather and
Fredrick Schrank. Barbara is an educational consultant with expertise in assessment, test
interpretation, and academic interventions. She holds an MA in Learning Disabilities, and she
has over 17 years of experience as an educator and diagnostician in Illinois public schools
and 20 years of experience in educational and assessment publishing.
Barbara has coauthored several books on assessment and intervention, including Essentials
of Evidence-Based Academic Interventions, Writing Assessment and Instruction for Students with
Learning Disabilities, and Essentials of Dyslexia: Assessment and Intervention. In addition,
she has coauthored the following books on the use and interpretation of the Woodcock-
Johnson: Essentials of the WJ III Tests of Achievement Assessment, Essentials of the WJ III Tests
of Cognitive Abilities Assessment (2nd ed.), and Essentials of the WJ IV Tests of Achievement.
She is also coauthor of the WJ III and WJ IV versions of the Woodcock Interpretation and
Instructional Interventions Program™.
vi
Acknowledgments
The Batería IV Woodcock-Muñoz was developed from the contributions of many individuals
of all ages and from all walks of life, each motivated by a desire or call to make a valuable
contribution to Spanish-language cognitive ability, oral language, and achievement
assessment. First and foremost, there would be no Batería IV without the volunteerism of
the standardization examinees who contributed their time and effort and the examiners
who administered the tests, gathered the data, and built the initial foundation for test
interpretation.
In retrospect, a few key people have made such significant contributions that even special
mention seems inadequate as an expression of their impact. Mary Ruef provided initial
project guidance, inspiration, and encouragement that was derived from her many years of
experience developing prior editions of the Batería and Woodcock-Muñoz Language Survey.
Christina Kowalczyk commanded a remarkable effort in coordinating the preparation of the
standardization materials and scoring the test results.
Jennifer Alvarado and Vanessa Cisneros were exceptionally competent and supportive
project staff at the Texas office of Education & Evaluation Consultants. In particular, Ms.
Alvarado rose to the task of completing many of the details of manuscript and data delivery
to Riverside Insights™ following the untimely death of her mother-in-law and mentor,
Criselda Alvarado. The consistency and accuracy of the Spanish test materials is in large part
thanks to the painstaking and meticulous translation and editing work of Ms. Cisneros and
Martha Rivera.
The critical task of managing all of the nuances of the Batería IV blueprint and converting
standardization data into derived scores was accomplished through the highly skilled efforts
of Erica LaForte. She, with the assistance of Michael Custer, was primarily responsible for
leading the complex and involved data analysis procedures and score interpretation plan.
Cindy Currie provided valuable development support by managing the flow of materials
through the production and publication process and reviewing test materials in careful detail.
Brenda Gilliam and Virginia Gonzalez served as content consultants during the later
stages of the project. They assisted with the item selection for several tests, reviewed
manuscript and page proofs, and consulted on verbiage for administration instructions.
Their real-world knowledge of bilingual education and their experience administering prior
versions of the Batería Woodcock-Muñoz to English learners proved to be an invaluable asset
to the project.
Last but not least, the Batería IV never would have become a reality without the supportive
product management and development leaders at Riverside Insights, including Katy Genseke,
Jamie Whitaker, and Mark Ledbetter.
FAS
BJW
vii
Dedication
This edition of the Batería Woodcock-Munoz is dedicated to the life and memory of our
colleague and friend Criselda G. Alvarado, a consummate educator and assessment
professional, who is primarily responsible for the development of the Spanish-language
adaptation and translation of the test materials you hold in your hands today. To those of
us who knew her well, she will always be remembered for the missionary zeal that she
commanded—which we are assured will continue its expression in the form of increased
understanding and better educational planning for hundreds of thousands of Spanish-
speaking bilingual individuals.
Dr. Alvarado’s professional legacy is embodied in the Batería IV; in the minds of countless
numbers of assessment professionals she trained; as well as through the healthy, engaged lives
of the students and colleagues she encouraged, mentored, and supported. To the authors of
the WJ IV, and to the thousands of psychologists and diagnosticians whom she so positively
influenced, the Batería IV will always be viewed as a testament to a life well lived in pursuit
of making a difference in equity of educational experience for English learners.
FAS
KSM
NM
viii
Table of Contents
About the Authors of the Batería IV iii
Acknowledgments vii
Dedication viii
Chapter 1: Overview 1
Comparison to the Batería III Pruebas de aprovechamiento 5
Organization of the Batería IV Pruebas de aprovechamiento 6
Components of the Batería IV Pruebas de aprovechamiento 8
Test Book (Libro de pruebas) 9
Examiner’s Manual (Manual del examinador) 9
Technical Manual 9
Online Scoring and Reporting Program 9
Test Record (Protocolo de pruebas) 9
Response Booklet (Folleto de respuestas) 9
Relationship of the Batería IV to the CHC Theory of Cognitive Abilities 10
Uses of the Batería IV Pruebas de aprovechamiento 11
Use With the Batería IV COG 11
Use With the WJ IV OL 11
Diagnosis 11
Determination of Variations and Comparisons 11
Educational Programming 12
Planning Individual Programs 12
Guidance 12
Assessing Growth 13
Program Evaluation 13
Research 13
Psychometric Training 13
Examiner Qualifications 14
Confidentiality of Test Materials and Content 14
ix
Prueba 7: Análisis de palabras (Test 7: Word Attack) 19
Prueba 8: Lectura oral (Test 8: Oral Reading) 20
Prueba 9: Fluidez en lectura de frases (Test 9: Sentence Reading Fluency) 20
Prueba 10: Fluidez en datos matemáticos (Test 10: Math Facts Fluency) 20
Prueba 11: Fluidez en escritura de frases (Test 11: Sentence Writing Fluency) 20
Prueba 12: Rememoración de lectura (Test 12: Reading Recall) 20
Prueba 13: Números matrices (Test 13: Number Matrices) 20
Batería IV APROV Clusters 21
Reading Clusters 21
Math Clusters 22
Written Language Clusters 22
Cross-Domain Clusters 23
x
Accommodations 40
Recommended Accommodations 41
Young Children 41
English Learners 42
Individuals With Learning and/or Reading Difficulties 43
Individuals With Attentional and Behavioral Difficulties 43
Individuals With Hearing Impairments 45
Individuals With Visual Impairments 48
Individuals With Physical Impairments 52
Interpretive Cautions 52
Use of Derived Scores 52
xi
Interpreting the Written Language Tests 86
Interpreting Variations and Comparisons 90
Intra-Ability Variations 91
Ability/Achievement Comparisons 95
Comparative Language Index (CLI) 97
Discrepancy Scores 97
Implications Derived From Test Results 98
References 99
List of Tables
Table 1-1 Comparison of the WJ IV and Batería IV Tests 2
Table 1-2 Comparison of the WJ IV and Batería IV Clusters 3
Table 1-3 Organization of the Batería IV APROV Tests 7
Table 1-4 Organization of the Batería IV APROV Clusters 8
Table 2-1 Batería IV APROV Selective Testing Table 18
xii
Table 3-1 Batería IV APROV Tests Useful for Individuals With Hearing Impairments 47
Table 3-2 Batería IV APROV Tests Useful for Individuals With Visual Impairments 51
Table 5-1 Hierarchy of Batería IV APROV Test Information 68
Table 5-2 APROV Clusters That Yield a CALP Level 73
Table 5-3 CALP Levels and Corresponding Implications 74
Table 5-4 Classification of Standard Score and Percentile Rank Ranges 76
Table 5-5 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 1: Identificación de letras y palabras 79
Table 5-6 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 4: Comprensión de textos 81
Table 5-7 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 9: Fluidez en lectura de frases 82
Table 5-8 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 2: Problemas aplicados 84
Table 5-9 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 5: Cálculo 85
Table 5-10 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 10: Fluidez en datos matemáticos 86
Table 5-11 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 3: Ortografía 88
Table 5-12 Percentage by Age of Occurrence of Qualitative Observations for
WJ IV ACH Test 6: Writing Samples 89
Table 5-13 Percentage by Age of Occurrence of Qualitative Observations for
Prueba 11: Fluidez en escritura de frases 90
Table 5-14 Batería IV Intra-Ability Variation and Ability/Achievement Comparison
Procedures 91
Table 5-15 Batería IV Intra-Achievement Variations 93
Table 5-16 Batería IV Academic Skills/Academic Fluency/Academic
Applications Variations 94
Table 5-17 Batería IV Oral Language/Achievement Comparisons 97
Table E-1 Translated and Adapted Tests of the Batería IV 136
Table E-2 Distribution of the Batería IV Calibration Sample by Age Group 138
Table E-3 Distribution of Sampling Variables in the Batería IV Calibration Study 139
Table E-4 Percentage of Batería IV Calibration Items Flagged for
Potential Gender DIF 141
Table E-5 Reliability Coefficients for Batería IV Nonspeeded Tests by Age Group 143
Table E-6 Test-Retest Reliability Coefficients From the WJ IV/Batería IV
Speeded Test-Retest Study 145
Table E-7 Reliability Coefficients for Batería IV Clusters by Age Group 145
List of Figures
Figure 1-1 Components of the Batería IV APROV. 8
Figure 3-1 Recommended arrangement for administering the test. 27
Figure 3-2 Suggested Starting Points table for Prueba 2: Problemas aplicados. 30
xiii
Figure 3-3 Example of Item 1 used as the basal on Prueba 1: Identificación de letras
y palabras. 32
Figure 3-4 Determination of basal and ceiling with two apparent basals and two
apparent ceilings. 34
Figure 3-5 The “Test Session Observations Checklist” from the Test Record. 36
Figure 3-6 “Qualitative Observation” checklist for Prueba 1: Identificación de letras
y palabras. 37
Figure 4-1 Reading error types in Prueba 8: Lectura oral. 60
Figure 4-2 Example of Test Record and “Qualitative Observation Tally” for
Prueba 8: Lectura oral. 61
Figure 5-1 Comparison of the traditional and extended percentile rank scales with the
standard score scale (M = 100, SD = 15). 75
Figure 5-2 Various skills measured by the Batería IV APROV reading tests. 77
Figure 5-3 Various skills measured by the Batería IV APROV math tests. 83
Figure 5-4 Various skills measured by the Batería IV APROV writing tests. 87
Figure 5-5 Three types of intra-ability variation models in the Batería IV. 92
Figure 5-6 Four types of ability/achievement comparison models in the Batería IV. 95
xiv
Chapter 1
Overview
The Batería IV Woodcock-Muñoz (Batería IV; Woodcock, Alvarado, Schrank, McGrew, Mather,
& Muñoz-Sandoval, 2019a) is a comprehensive, Spanish-language psychoeducational
assessment system that includes two test batteries: the Batería IV Woodcock-Muñoz: Pruebas
de habilidades cognitivas (Batería IV COG; Woodcock, Alvarado, Schrank, McGrew, Mather,
& Muñoz-Sandoval, 2019b) and the Batería IV Woodcock-Muñoz: Pruebas de aprovechamiento
(Batería IV APROV; Woodcock, Alvarado, Schrank, Mather, McGrew, & Muñoz-Sandoval,
2019). Tests included in the Batería IV are either adaptations or translations of tests from the
Woodcock-Johnson IV (WJ IV; Schrank, McGrew, & Mather, 2014). The Batería IV batteries
can be used in conjunction with the Woodcock-Johnson IV Tests of Oral Language (WJ IV OL;
Schrank, Mather, & McGrew, 2014) to form a broad-based assessment of cognitive abilities,
achievement, and comparative oral language abilities. The WJ IV OL includes oral language
tests in English as well as Spanish to provide information on relative language proficiency and
language dominance.
Some of the Batería IV tests can be used with Spanish-speaking individuals as young as 24
months, but the majority of the Batería IV tests are best suited for use with individuals from 5
to 95 years of age. Spanish-language calibration data, based on a sample of 601 native Spanish
speakers, are equated to the large, nationally representative WJ IV norming sample of 7,416
individuals ranging from 2 to 90+ years of age. The Spanish standardization data were used to
calibrate the new test items and to equate the items to the scales underlying the WJ IV tests.
The equating procedure produces a psychometrically sound interpretive model that allows an
examiner to describe an individual’s performance on the Spanish tests in terms of comparable
ability in English. See Appendix E for technical information about the development, norming,
and calibration of the Batería IV. Additional information about the WJ IV norming sample is
provided in the Woodcock-Johnson IV Technical Manual (McGrew, LaForte, & Schrank, 2014),
available via download from the online scoring and reporting program.
The Batería IV is based on the Cattell-Horn-Carroll (CHC) theory of cognitive abilities,
sometimes referred to as CHC theory version 2 (McGrew et al., 2014; Schneider & McGrew,
2012, 2018). As part of the translation-adaptation process, some of the most educationally
and diagnostically useful WJ IV tests were selected for inclusion in the Batería IV.
Consequently, there are some differences in the number of tests and clusters that are available
in the Batería IV. Table 1-1 lists the tests in the WJ IV and the Batería IV. Table 1-2 lists the
clusters in the WJ IV and the Batería IV.
Overview 1
Table 1-1. WJ IV BATERÍA IV
Comparison of the WJ IV
Woodcock-Johnson IV Tests of Cognitive Batería IV Woodcock-Muñoz: Pruebas de habilidades
and Batería IV Tests Abilities (two Test Books) cognitivas (one Test Book)
Test 1: Oral Vocabulary Prueba 1: Vocabulario oral
Test 2: Number Series Prueba 2: Series numéricas
Test 3: Verbal Attention Prueba 3: Atención verbal
Test 4: Letter-Pattern Matching Prueba 4: Pareo de letras idénticas
Test 5: Phonological Processing Prueba 5: Procesamiento fonético
Test 6: Story Recall Prueba 6: Rememoración de cuentos
Test 7: Visualization Prueba 7: Visualización
Test 8: General Information Prueba 8: Información general
Test 9: Concept Formation Prueba 9: Formación de conceptos
Test 10: Numbers Reversed Prueba 10: Inversión de números
Test 11: Number-Pattern Matching Prueba 11: Pareo de números idénticos
Test 12: Nonword Repetition Prueba 12: Repetición de palabras sin sentido
Test 13: Visual-Auditory Learning (Can be administered from Batería III COG Standard Test Book
and scored using Batería IV norms)
Test 14: Picture Recognition (Can be administered from Batería III COG Extended Test Book
and scored using Batería IV norms)
Test 15: Analysis-Synthesis
Test 16: Object-Number Sequencing
Test 17: Pair Cancellation Prueba 13: Cancelación de pares
Test 18: Memory for Words
(Included from WJ IV OL Test 4: Rapid Picture Naming) Prueba 14: Rapidez en la identificación de dibujos
2 Overview
Table 1-1. (cont.) WJ IV BATERÍA IV
Comparison of the WJ IV
Woodcock-Johnson IV Tests of Achievement Batería IV Woodcock-Muñoz: Pruebas de
and Batería IV Tests
(two Test Books) aprovechamiento (one Test Book)
Test 16: Spelling of Sounds
Test 17: Reading Vocabulary
Test 18: Science
Test 19: Social Studies
Test 20: Humanities
Overview 3
Table 1-2. (cont.) WJ IV BATERÍA IV
Comparison of the WJ IV
Woodcock-Johnson IV Tests of Cognitive Batería IV Woodcock-Muñoz: Pruebas de habilidades
and Batería IV Clusters Abilities cognitivas
Cognitive Efficiency Eficiencia cognitiva
Reading Aptitude Aptitud de lectura
Math Aptitude Aptitud matemática
Writing Aptitude Aptitud de escritura
Multiple goals guided the WJ IV revision blueprint that underlies the Batería IV. First,
this comprehensive assessment system was designed to be on the cutting edge of practice.
It facilitates exploring individual strengths and weaknesses across cognitive, linguistic,
and academic abilities; complements response to intervention (RTI) models; and reframes
variations and ability/achievement comparisons. Second, the blueprint pushed the tests
beyond CHC theory as it was conceived in the Woodcock-Johnson III (WJ III ; Woodcock, ®
4 Overview
McGrew, & Mather, 2001, 2007) and the Batería III Woodcock-Muñoz (Muñoz-Sandoval,
Woodcock, McGrew, & Mather, 2005, 2007a). Whereas the third editions of these tests
focused primarily on broad CHC abilities, the fourth editions focus on the most important
broad and narrow CHC abilities for describing cognitive performance and understanding
the nature of learning problems (McGrew, 2012; McGrew & Wendling, 2010; Schneider &
McGrew, 2012, 2018). Some of the tests and clusters emphasize narrow CHC abilities, and
others were designed to reflect the importance of cognitive complexity through the influence
of two or more narrow abilities on task requirements. Additional goals addressed ease and
flexibility of use. New features allow novice examiners to use the tests with confidence while
providing experienced examiners with a rich array of interpretive options to customize and
enhance their evaluations. The structure of the WJ IV and Batería IV systems also facilitates
examiner use by creating comprehensive cognitive, achievement, and oral language batteries
that can be used in conjunction with one another or as standalone batteries.
The WJ IV and Batería IV interpretation plan includes a full array of derived scores for
reporting results. The accompanying online scoring and reporting program quickly calculates
and reports all derived scores.
This manual describes the Batería IV APROV, which can be used independently or in
combination with the WJ IV OL and/or the Batería IV COG batteries. While the manual is
written in English, all test and cluster names are presented in Spanish. Initially, these names
are in Spanish with English in parentheses. Later the names are presented only in Spanish.
Test components and types of scores are presented one time in English with Spanish in
parentheses, and all future references are in English. Examiners can refer to Tables 1-1
and 1-2 in this chapter and the glossary in Appendix D of this manual for assistance with
translation.
Overview 5
■ Several Batería III APROV tests have been relocated. The Spanish forms of Vocabulario
sobre dibujos (Picture Vocabulary), Comprensión de indicaciones (Understanding
Directions), and Comprensión oral (Oral Comprehension) are now in the WJ IV Tests
of Oral Language. These three tests may be used in conjunction with the Batería
IV tests and entered into the online scoring and reporting program for inclusion in
interpretation and reporting. Rememoración de cuentos (Story Recall) is now in the
Batería IV COG.
■ Three Batería III APROV tests have been replaced: Quantitative Concepts has been
replaced by Números matrices (Number Matrices), a stronger measure of mathematical
reasoning; Writing Samples has been replaced by Expresión de lenguaje escrito (Written
Language Expression), a test that is easier to score; and Reading Vocabulary has been
replaced by Rememoración de lectura (Reading Recall), a more comprehensive measure
of reading comprehension.
■ The addition of the new test Lectura oral (Oral Reading), when combined with Fluidez
en lectura de frases (Sentence Reading Fluency), allows a Fluidez en la lectura (Reading
Fluency) cluster score to be obtained.
■ There are no audio-recorded tests in the Batería IV APROV.
■ Scoring is available through the web-based online scoring and reporting program that is
included with the purchase of each Test Record.
■ The following Batería III APROV tests have been eliminated from the Batería IV
APROV: Editing, Spelling of Sounds, Academic Knowledge, Sound Awareness, Story
Recall–Delayed, and Punctuation and Capitalization.
■ Three test names were changed to more accurately reflect the task: Writing Fluency is
now Prueba 11: Fluidez en escritura de frases (Test 11: Sentence Writing Fluency); Math
Fluency is now Prueba 10: Fluidez en datos matemáticos (Test 10: Math Facts Fluency);
and Reading Fluency is now Prueba 9: Fluidez en lectura de frases (Test 9: Sentence
Reading Fluency).
■ The procedures for evaluating ability/achievement comparisons and intra-ability
variations have been simplified and offer increased flexibility for the examiner.
∘ Three types of intra-ability variations are available: intra-cognitive, intra-
achievement, and academic skills/academic fluency/academic applications.
∘ Four types of ability/achievement comparisons are available: general intellectual
ability (GIA), Gf-Gc composite, scholastic aptitude, and oral language ability
(requires administration of the Spanish forms of three tests from WJ IV OL).
■ The Batería III predicted achievement/achievement discrepancy procedure has been
replaced with the scholastic aptitude/achievement comparison procedure. There are four
specific aptitude clusters: two for reading, one for math, and one for writing. Each of
these four aptitude clusters contains four cognitive tests that best predict performance
in the specific achievement area.
6 Overview
An examiner seldom needs to administer all of the tests or complete all of the interpretive
options for a single person. The importance of selective testing becomes apparent as the
examiner gains familiarity with the Batería IV APROV. An analogy to craftsmanship is
appropriate: The Batería IV APROV provides an extensive tool chest that can be used
selectively by a variety of skilled assessment professionals. Different assessments require
different combinations of tools.
Table 1-3 lists the tests included in the Batería IV APROV. Icons following several tests
indicate tests that are administered using the Response Booklet ( ) and tests that are
timed ( ). The table groups the tests by content area rather than by order of appearance in
the Test Book.
Table 1-4 illustrates the 17 clusters, or groupings of tests, that are available from the
Batería IV APROV. These clusters are the primary source of interpretive information to help
identify performance levels, determine educational progress, and identify an individual’s
strengths and weaknesses.
Overview 7
Table 1-4. CURRICULAR AREA CLUSTERS
Organization of the
Reading Lectura
Batería IV APROV Clusters
Lectura amplia
Destrezas básicas en lectura
Comprensión de lectura
Fluidez en la lectura
Mathematics Matemáticas
Matemáticas amplias
Destrezas en cálculos matemáticos
Resolución de problemas matemáticos
Writing Lenguaje escrito
Lenguaje escrito amplio
Expresión escrita
Cross-Domain Clusters Destrezas académicas
Fluidez académica
Aplicaciones académicas
Aprovechamiento breve
Aprovechamiento amplio
Figure 1-1.
Components of the
Batería IV APROV.
8 Overview
Test Book (Libro de pruebas)
The Test Book is in an easel format positioned so the stimulus pictures or words face
the examinee and the directions face the examiner. Specific administration directions are
provided page by page for all tests.
Technical Manual
The WJ IV Technical Manual may be downloaded as a PDF from the online scoring and
reporting program and provides a summary of the development, standardization, and technical
characteristics of the WJ IV; most of this technical information also applies to the Batería IV.
Overview 9
addition, a worksheet is provided in the Response Booklet for Prueba 2: Problemas aplicados
(Test 2: Applied Problems) and Prueba 13: Números matrices (Test 13: Number Matrices).
10 Overview
Uses of the Batería IV Pruebas de aprovechamiento
The procedures followed in developing and standardizing the Batería IV APROV have
produced an instrument that can be used with confidence in a variety of settings. The
wide age range and breadth of coverage allow the Batería IV APROV tests to be used for
educational, clinical, or research purposes from the preschool to the geriatric level. Because
the Batería IV APROV is co-normed with both the Batería IV COG and the Spanish forms of
three tests from the WJ IV OL, accurate predictions and comparisons can be made among the
batteries.
Diagnosis
An examiner can use the Batería IV APROV to determine and describe a profile of an
individual’s academic strengths and weaknesses. Additionally, test results help determine how
certain factors affect related aspects of development. For example, a weakness in phoneme/
grapheme knowledge may interfere with overall development in reading and spelling.
Similarly, a weakness in spelling may help explain an individual’s difficulties on school
assignments requiring writing.
An examiner also can use the Batería IV APROV for a more in-depth evaluation after an
individual has failed a screening procedure (e.g., a kindergarten screening) or to substantiate
the results of other tests or prior evaluations.
Overview 11
The Batería IV intra-ability variations are useful for understanding an individual’s strengths
and weaknesses, diagnosing and documenting the existence of specific abilities and disabilities,
and acquiring the most relevant information for educational and vocational planning. Analysis
of this in-depth assessment data, which goes well beyond the historical and traditional
singular focus on ability/achievement discrepancy data, can be linked more directly to
recommendations for service delivery and the design of an appropriate educational program.
Although many unresolved issues characterize the appropriate determination and
application of discrepancy information in the field of learning disabilities, an ability/
achievement discrepancy may be used as part of the selection criteria for learning disability
(LD) programs. Even though a discrepancy may be statistically significant, this type of
comparison is rarely appropriate as the sole criterion for determining the existence or
nonexistence of a learning disability or for determining eligibility for special services.
Analyses of other abilities and an understanding of the relationships and interactions
among various abilities and skills are needed to determine whether a person does or does
not have a learning disability. Given the problems inherent in employing and interpreting
ability/achievement discrepancies, multiple sources of information, including background
information (e.g., educational history, classroom performance), as well as clinical experience,
are needed to make an accurate diagnosis.
Educational Programming
When combined with behavioral observations, work samples, and other pertinent
information, Batería IV APROV results will help the skilled clinician make decisions
regarding educational programming. The test results demonstrate a student’s most appropriate
instructional level and the types of services that may be needed. The Batería IV APROV also
can assist in vocational planning, particularly when successful job performance depends on
specific types of skills, such as reading, writing, or mathematics performance.
Guidance
The Batería IV APROV can provide guidance in educational and clinical settings. The
results of the evaluation can help teachers, counselors, social workers, and other personnel
understand the nature of an individual’s academic strengths and weaknesses and determine
the necessary levels of assistance. The Batería IV APROV also can provide valuable
information to help parents understand their child’s particular academic problems or needs.
12 Overview
Assessing Growth
The Batería IV APROV can provide a record of functioning and growth throughout an
individual’s lifetime. The Batería IV APROV also can be used to assess changes in a person’s
performance following a specific time interval, such as after a year of receiving special
educational services.
Program Evaluation
The Batería IV APROV can provide information about program effectiveness at all levels
of education, from preschool through adult. For example, Batería IV APROV tests can be
administered to evaluate the effects of specific school programs or the relative performance
levels (in a certain skill) of students in a class or school.
The continuous-year feature of the Batería IV school-age norms meets the reporting
requirements for educational programs. This feature is especially useful because it provides
norms based on data gathered continuously throughout the school year as opposed to norms
based on data gathered at, perhaps, two points in the school year and then presented as fall
and spring norms.
Research
The possibilities for using the Batería IV APROV in research are unlimited. The wide age
range and breadth of coverage are important advantages underlying its use for research at all
age levels, from preschool through geriatric. Computer scoring allows easy storage of clinical
data. Because the Batería IV APROV tests are individually administered, the researcher has
more control over the quality of the data obtained.
The Batería IV APROV provides predictor or criterion measures that can be used in
studies investigating a variety of experimental effects. Additionally, the wide age range
allows longitudinal or cohort research data to be gathered using the same set of tests and
test content. In educational research, the Batería IV APROV provides a comprehensive set
of related measures for evaluating the comparative efficacy of several programs or services
or for evaluating the effectiveness of curricular interventions. The Batería IV APROV also is
useful for describing the characteristics of examinees included in a sample or experimental
condition and for pairing students in certain experimental designs.
The range of interpretive information available for each test and cluster includes error
analysis, description of developmental status (age and grade equivalents), description of
quality of performance (relative proficiency indexes [RPIs] and instructional zones), and
comparison with grade or age mates to determine group standing (percentile ranks and
standard scores). The W score and standard score scales (discussed in Chapter 5) are both
equal-interval scales that can be used in statistical analyses based on the assumption of equal-
interval metrics. As described in the WJ IV Technical Manual, the W score is the preferred
metric for most statistical analyses.
Psychometric Training
This manual contains the basic principles of individual clinical assessment and specific
administration, scoring, and interpretive information for the Batería IV APROV, which makes
the Batería IV APROV an ideal instrument for introducing individualized assessment in
college and university courses. The Batería IV APROV provides new examiners with a broad
foundation in the administration, scoring, and interpretation of individualized assessments.
Experience in clinical assessment with the Batería IV APROV provides a solid foundation for
learning to administer and interpret other test instruments.
Overview 13
Examiner Qualifications
The examiner qualifications for the Batería IV APROV have been informed by the joint
Standards for Educational and Psychological Testing (American Educational Research
Association [AERA], American Psychological Association [APA], & National Council on
Measurement in Education [NCME], 2014).
Any person administering the Batería IV APROV needs thorough knowledge of the exact
administration and scoring procedures and an understanding of the importance of adhering
to standardized procedures. To become proficient in administering the Batería IV APROV,
examiners need to study the administration and scoring procedures carefully and follow the
procedures precisely. This Examiner’s Manual provides guidelines for examiner training and
includes specific instructions for administering and scoring each test. In addition, examiners
who administer the Batería IV APROV must be fluent and literate in Spanish.
Competent interpretation of the Batería IV APROV requires a higher degree of knowledge
and experience than is required for administering and scoring the tests. Graduate-level
training in educational assessment and a background in diagnostic decision making are
recommended for individuals who will interpret Batería IV APROV results. Only trained
and knowledgeable professionals who are sensitive to the conditions that may compromise,
or even invalidate, standardized test results should make interpretations and decisions.
The level of formal education recommended to interpret the Batería IV APROV is typically
documented by successful completion of an applicable graduate-level program of study that
includes, at a minimum, a practicum-type course covering administration and interpretation
of standardized tests of academic achievement. In addition, many qualified examiners possess
state, provincial, or professional certification, registration, or licensure in a field or profession
that includes as part of its formal training and code of ethics the responsibility for rendering
educational assessment and interpretation services.
Because professional titles, roles, and responsibilities vary among states (or provinces),
or even from one school district to another, it is impossible to equate competency to
professional titles. Consequently, the Standards for Educational and Psychological Testing
(AERA, APA, & NCME, 2014) suggest that it is the responsibility of each school district to
be informed by this statement of examiner qualifications and subsequently determine who,
under its aegis, is qualified to administer and interpret the Batería IV APROV.
14 Overview
An examiner should not inform examinees of the correct answers to any of the questions
during or after testing. When discussing test results, examiners may describe the nature of
the items included in a test, but they should not review specific test content. Examiners
should use examples similar to the test items without revealing actual items.
Questions often arise about the federal requirement that families be given access to
certain educational records. To comply with this requirement, a school or school district
may be required to permit “access” to test protocols; however, “access” does not include the
right to make copies of the materials provided. The Family Educational Rights and Privacy
Act (FERPA) provides that parents are to be given the right to “inspect and review” the
educational records of their children (U.S. Department of Education, Family Educational
Rights and Privacy Act. (1974). 20 U.S.C. § 1232g; 34 CFR §99.10). The right to inspect
and review is defined as including the right to a response from the participating agency “to
reasonable requests for explanations and interpretations of the records” (34 CFR §99.10(c))
and, if circumstances prevent inspection or review, the agency may either (a) provide a copy
or (b) make other arrangements that allow for inspection and review (34 CFR §99.10(d)).
So long as the test protocols are made available to the parent, or the parent’s representative,
for review, all requirements of the law are met without violating the publisher’s rights or the
obligations of the educational institution to keep the test materials confidential. There is,
therefore, no obligation to provide copies or to permit the parent, or the legal representative
of the parent, to make copies.
Similar concerns arise when a party seeks to introduce testing materials in a trial or
other legal proceeding. In such cases, it is important that the court take steps to protect the
confidentiality of the test and to prevent further copying or dissemination of any of the test
materials. Such steps include: (a) issuing a protective order prohibiting parties from copying
the materials, (b) requiring the return of the materials to the qualified professional upon
the conclusion of the proceedings, and (c) ensuring that the materials and all references to
the content of the materials will not become part of the public record of the proceedings. To
ensure that these protections are obtained, Riverside Insights should be contacted whenever it
appears likely that testing materials will be introduced as evidence in a legal proceeding.
Examiners or school districts with questions about copyright ownership or confidentiality
obligations should contact Riverside Insights at the toll-free telephone number listed on the
copyright page of this manual.
Overview 15
Chapter 2
Descriptions of the
Batería IV APROV Tests
and Clusters
The Batería IV Woodcock-Muñoz: Pruebas de aprovechamiento (Batería IV APROV) contains 13
tests measuring three curricular areas—reading, mathematics, and written language. Specific
combinations, or groupings, of these 13 tests form clusters for interpretive purposes. (For
administration and scoring procedures, see Chapters 3 and 4 of this manual.)
The tests combine to form 17 cluster scores, including an Aprovechamiento breve (Brief
Achievement) score and an Aprovechamiento amplio (Broad Achievement) score. Although
tests are the basic administration components of the Batería IV APROV, clusters of tests
provide the primary basis for test interpretation. Cluster interpretation minimizes the danger
of generalizing from the score for a single narrow ability to a broad, multifaceted ability or
skill. Cluster interpretation results in higher validity because more than one component
of a broad ability constitutes the score that serves as the basis for interpretation. In some
situations, however, the narrow abilities and skills that are measured by the individual
tests should be considered. This is particularly important when significant differences
exist between or among the tests in a cluster. In these cases, more information is obtained
by analyzing performance on each test, which may indicate the need for further testing.
Occasions exist when it is more meaningful to describe a narrow ability than it is to report
performance on a broad ability. To increase the validity of narrow ability interpretation,
the Batería IV provides clusters for a number of important narrow abilities. These narrow
abilities often have more relevance for informing instruction and intervention (McGrew &
Wendling, 2010).
ticos
s ma cos
temá
Leng n de prob matemáti
lectu ra
lectu
as
o
mplio
ectur ra
mpli
breve
vech académic
lema
icas
plias
ulos
s en
nto a
acion mica
Expre escrito a
Fluid s académ
iento
ta
Reso s en cálc
Fluid nsión de
ásica
uaje rito
as am
escri
p lia
amie
l
adé
esc
es
Mate ticas
a
m
l
b
ra am
a
mátic
ez en
ez ac
sión
ezas
vech
lució
uaje
eza
eza
pre
má
ra
Lectu
Lectu
Destr
Destr
Destr
Aplic
Mate
Leng
Com
Ap r o
Apro
APROV 1 Identificación de letras y palabras
APROV 2 Problemas aplicados
APROV 3 Ortografía
Batería IV APROV Battery
Reading Clusters
Five reading clusters are available.
Lectura (Reading)
The Lectura cluster is a measure of reading achievement (a reading-writing [Grw] ability),
including reading decoding and the ability to comprehend connected text while reading.
This cluster is a combination of Prueba 1: Identificación de letras y palabras and Prueba 4:
Comprensión de textos. It has a median reliability of .94 in the 5 to 19 age range and .96 in the
adult age range.
Matemáticas (Mathematics)
The Matemáticas cluster provides a measure of math achievement (quantitative knowledge
[Gq] ability), including problem solving and computational skills. This cluster includes
Prueba 2: Problemas aplicados and Prueba 5: Cálculo. It has a median reliability of .95 in the 5
to 19 age range and .96 in the adult age range.
Cross-Domain Clusters
Five cross-domain clusters are available, including two general academic proficiency cluster
scores, Aprovechamiento breve (Brief Achievement) and Aprovechamiento amplio (Broad
Achievement). Various combinations of tests are used to form three additional cluster
scores: Destrezas académicas (Academic Skills), Fluidez académica (Academic Fluency),
and Aplicaciones académicas (Academic Applications). These three clusters (skills, fluency,
and applications) contain tests of reading, math, and written language and can be used to
determine whether the person exhibits significant strengths and/or weaknesses among these
three types of tasks across academic areas.
General Administration
and Scoring Procedures
To become proficient in administering and scoring the Batería IV Pruebas de aprovechamiento
(Batería IV APROV), examiners should carefully study the general administration and scoring
procedures in this chapter and the specific procedures for each test in Chapter 4 and in
the Test Book. Additionally, two appendices of this manual provide reproducible checklists
to help examiners build competency administering and scoring the tests. Appendix B,
the “Batería IV Pruebas de aprovechamiento Examiner Training Checklist,” is a test-by-test
form that may be used as a self-study or observation tool. Appendix C is the “Batería IV
General Test Observations Checklist,” which may be used by an experienced examiner when
observing a new examiner.
Practice Administration
After thoroughly studying this Examiner’s Manual, the Test Book, the Test Record, and
the Response Booklet, both experienced and novice examiners should administer several
practice tests. When administering practice tests, try to replicate an actual testing situation,
pretending that the practice session is an actual administration. Do not discuss the test or the
answers to specific items. After completing each practice administration, record any questions
that arose during the practice session. Before administering another practice test, answer
the questions by reviewing the Examiner’s Manual or consulting an experienced examiner.
While administering practice tests, strive for these two goals: exact administration and brisk
administration.
Exact Administration
The goal of standardized testing is to see how well a person can respond when given
instructions identical to those presented to individuals in the norming sample. When learning
to administer the Batería IV APROV tests, study the contents of the Test Book, paying
particular attention to the information on the introductory page of each test, the specific
instructions on the test pages, and the boxes with special instructions.
The first page after the tab in each test provides general information and instructions
specific to that test. Review this information frequently. This page usually includes
administration information, scoring information, suggested starting points, basal and ceiling
requirements, and information about materials required to administer the test.
The directions for administering each item are located on the examiner’s side of the
pages in the Test Book. The directions include the script to be read to the examinee (printed
in bold blue type) and, if applicable, specific pointing instructions. Always use the exact
Brisk Administration
After the initial practice sessions, strive for a brisk testing pace. Inefficient testing procedures
bore the examinee, invite distraction, and increase testing time. It is not appropriate to stop
testing and visit with the examinee during the testing session. When the person has finished
responding to an item, immediately begin the next item.
In most instances, an examinee does not need a break before beginning the next test.
Each test begins with easy questions presented in a different format, thus providing a built-
in change of pace from one test to the next. Using a brisk testing pace enhances rapport and
helps an examinee maintain attention.
Continue to practice administering the tests until the two goals of exact and brisk
administration have been met.
The room should have a table (or other flat working space of adequate size) and two
chairs, one being an appropriate size for the examinee. A suitable seating arrangement allows
the examiner to view both sides of the Test Book easel, point to all parts of the examinee’s
page and the Response Booklet, and record responses on the Test Record out of the
examinee’s view. The examinee should be able to view only the examinee’s test pages. When
the Test Book easel is set up for administration, it becomes a screen allowing the examiner to
record responses on the Test Record out of the examinee’s view.
The best seating arrangement is one in which the examiner and the examinee sit
diagonally across from each other at the corner of a table. This arrangement is illustrated in
Figure 3-1 for a right-handed examiner. The arrangement (seating and setup of materials)
should be reversed for a left-handed examiner.
Another possible seating arrangement is for the examiner and the examinee to sit directly
across the table from each other. With this arrangement, the table must be narrow and low
enough so that the examiner can see over the upright Test Book easel and accurately point to
the examinee’s page when necessary.
Establishing Rapport
In most instances, the examiner will have little difficulty establishing a good relationship
with the examinee. Do not begin testing unless the person seems relatively at ease. If he
or she does not feel well or will not respond appropriately, do not attempt testing. Often
examiners begin the testing session with a short period of conversation while completing
the “Identifying Information” portion of the Test Record. A brief explanation of the test is
provided in the “Introduction” section in the front of the Test Book.
To help put the individual at ease, smile frequently throughout the testing session and call
the person by name. Between tests, let the examinee know that he or she is doing a good job,
using such comments as “Bien” (Fine) and “Muy bien” (Good). Encourage a response even
when items are difficult. It is fine to say, “Te atreves a hacer este?” (Would you like to take
a guess on that one?), but the comments should not reveal whether answers are correct or
incorrect. Do not say, “Muy bien” (Good) only after correct responses or pause longer after
incorrect responses before proceeding to the next item.
Identifying Information
For the most part, the “Identifying Information” section on the first page of the Test Record
is self-explanatory. For younger examinees, verify the date of birth using school records
or with a parent. Prior to testing, check to see if the person should be wearing glasses or a
hearing aid.
Test Selection
It is important to select tests that are appropriate for the individual being evaluated. Consider
the individual’s age, developmental level, and achievement levels as part of this test selection
process. For example, it would be inappropriate to give a test that requires reading ability to a
young child with limited reading experience. Whereas some tests, such as Prueba 1:
Identificación de letras y palabras and Prueba 4: Comprensión de textos, have a number of
prereading items, other tests do not. For example, on Prueba 9: Fluidez en lectura de frases,
Order of Administration
In most cases, administer the first six tests in the order that they appear in the Test Book.
These are the core tests (Tests 1 through 6), and they have been organized to alternate
between different tasks and achievement areas (e.g., reading and math) to facilitate optimal
attention and interest. However, the tests may be administered in any order. For example,
testing may begin with Prueba 5: Cálculo, rather than with Prueba 1: Identificación de letras y
palabras. Furthermore, testing may be discontinued between the administration of any two
tests. The decision to administer any of the remaining tests should be based upon the referral
question(s) and the examinee’s age and interests. These additional tests may be administered
in any order with one or two exceptions.
If an examinee struggles with a certain type of task, as a general rule, do not administer
two such tests in a row (e.g., timed tests, reading tests, or tests involving sustained writing,
such as Prueba 6: Expresión de lenguaje escrito or Prueba 11: Fluidez en escritura de frases).
Additionally, if you are planning to administer Prueba 9: Fluidez en lectura de frases, Prueba
10: Fluidez en datos matemáticos, and Prueba 11: Fluidez en escritura de frases, these timed
tests should be interspersed in the administration sequence rather than administered
consecutively.
Time Requirements
Always schedule adequate time for testing. Generally, experienced examiners will
require approximately 40 minutes to administer the core set of tests (Tests 1 through 6).
Administration of Prueba 6: Expresión de lenguaje escrito requires about 15 to 20 minutes,
whereas the other tests require about 5 to 10 minutes each. When administering each test,
allow a reasonable amount of time for a person to respond and then suggest moving on to the
next item. Also allow more time for a specific item if the person requests it or if more time is
allowed under the specific test directions.
Very young individuals or those who have unique characteristics that may impact
test administration may require additional testing time. These individuals may produce
a scattering of correct responses requiring administration of a greater number of items.
Some people may respond more slowly, change their answers more frequently, or require
more prompting and querying. In addition, an examiner may inadvertently begin at an
inappropriate starting point, which extends the testing time.
5 1 K 61 municipalidad 3
4
2-1
2-9
<K.0
<K.0
41
42
7-9
7-10
6 1 i 62 irregularidad 5
6
3-4
3-9
<K.0
<K.0
43
44
8-0
8-1
63 almohadón
7 1 s 64 terapéutico 7 4-0 <K.0 45 8-3
8 0 p 65 isósceles 8 4-2 <K.0 46 8-5
10 0 R 67 eufemismo
10
11
4-7
4-10
<K.0
<K.0
48
49
8-8
8-10
Timed Tests
Prueba 9: Fluidez en lectura de frases, Prueba 10: Fluidez en datos matemáticos, and Prueba
11: Fluidez en escritura de frases are timed tests. Prueba 9: Fluidez en lectura de frases and
Prueba 10: Fluidez en datos matemáticos have a 3-minute time limit, and Prueba 11: Fluidez en
escritura de frases has a 5-minute time limit. Although Tests 9 through 11 are in a numeric
sequence, it is recommended that these three timed tests not be administered consecutively.
The time limits are noted in both the Test Book and the Test Record. Administer these
tests using a stopwatch. If not using a stopwatch, write the exact starting and finishing times
in minutes and seconds in the space provided on the Test Record. For example, 17:23 would
indicate that the test started at 17 minutes and 23 seconds after the hour. The test then would
end exactly 3 minutes later at 20 minutes and 23 seconds (20:23) after the hour. A watch or
clock with a second hand is also useful for administering tests with the instruction to proceed
to the next item if an examinee has not responded to an item within a specified period of time.
Examiner Queries
For certain responses, the Query keys in the Test Book provide prompts designed to elicit
another answer from the examinee. For example, Item 36 on Prueba 5: Cálculo requires the
examinee to reduce the fraction to obtain credit. The query on this item is a reminder to
ask the examinee to simplify his or her answer. Use professional judgment when querying
responses that are not listed in the Query key. For example, if an individual provides a
response that seems to be partially correct, it is permissible to query with a comment such as
”Dame [Deme] otra respuesta” (Tell me another answer).
When using this checklist, it is necessary to possess knowledge of the behaviors that can
be considered both typical and atypical for the age or grade level of the individual who is
being assessed. A wide range of behaviors may be considered typical within any age or grade
level. The checklist is designed so that a “typical” rating in each category is easily identified.
For example, typical examinees are cooperative during the examination, seem at ease and
comfortable, are attentive to the tasks, respond promptly but carefully, and generally persist
with difficult tasks. These behaviors are indicated as “Typical for age/grade” on the checklist.
For other categories, particularly those that reveal marked differences from age to age,
examiners will need to apply a finer knowledge of age- or grade-appropriate behaviors. For
example, “typical” levels of activity or conversational proficiency would be quite different
for a 5-year-old than for a 9-year-old child. For some age or grade levels, ratings such as
“Appeared fidgety or restless at times” could be included within the range of behaviors that
is “Typical for age/grade” rather than a separate category. In such instances, it would be more
accurate to check “Typical for age/grade” than “Appeared fidgety or restless at times” because
the former conveys the concept of age- or grade-appropriate behavior.
Use the “Test Session Observations Checklist,” located on the Test Record, immediately after
test administration. Each of the items describes a category of observations. For each item, place
a check mark in the box corresponding to the quality that best describes the behavior of the
individual who was assessed. Only one category should be checked for each item. If any item
does not apply to the individual, or if the categories do not convey an adequate description of
the examinee’s test session behaviors, leave the item blank. Also note any other behaviors of
clinical interest. This type of qualitative information may affect interpretation of test results.
Be sure to respond to the question “Do you have any reason to believe this testing session
may not represent a fair sample of the examinee’s abilities?” located on the Test Record cover.
If Yes is checked in response to this question, complete the sentence “These results may not
be a fair estimate because… _______.” Examples of reasons for questioning the validity of the
test results may include suspected or known problems with an examinee’s hearing or vision,
emotional problems of a nature that interfere with the person’s ability to concentrate, and
certain background factors.
Item Scoring
With the exception of three tests (Prueba 6: Expresión de lenguaje escrito, Prueba 8: Lectura
oral, and Prueba 12: Rememoración de lectura), score each item administered by placing a 1
or a 0 in the appropriate space on the Test Record: 1 = correct response, 0 = incorrect or no
response. (Detailed scoring procedures for Prueba 6: Expresión de lenguaje escrito, Prueba 8:
Lectura oral, and Prueba 12: Rememoración de lectura are included in Chapter 4.) For items
not administered, leave the corresponding spaces on the Test Record blank. After a test has
been administered and completely scored, the only blank spaces should be items below the
basal and above the ceiling levels or items not included in the assigned block of items.
The correct and incorrect keys accompanying many of the items in the Test Book are
guides that demonstrate how certain responses are scored. Not all possible correct and
incorrect answers are listed. Judgment is required when scoring some responses. In the keys,
the first response listed is the answer given most frequently during the standardization.
Accommodations
The Batería IV is ideally suited to increase the participation of students with disabilities in
assessment and accountability systems. This section identifies several administration features
of the Batería IV that allow individuals with disabilities to participate more fully in the
evaluation process.
Setting
The individual administration format of the Batería IV APROV provides the opportunity
for standardized assessment on a one-to-one basis. Use of a separate location for testing
minimizes the distractions inherent in a classroom group-testing environment. If needed,
use noise buffers such as earplugs or headphones to mask external sounds. Also, incorporate
special lighting, special acoustics, or adaptive or special furniture if needed.
Timing
Use of basal and ceiling rules focuses the assessment on the examinee’s level of ability and
minimizes testing time. In addition, frequent breaks can be taken between tests, if needed.
With the exception of the timed tests, individuals can have extended time to complete tasks,
if required.
Presentation
All instructions are presented orally to the examinee, and the language of the instructions is
at a sufficiently simple level of linguistic complexity to minimize language comprehension
barriers. The instructions may be repeated or signed, if necessary. Special sample items on
many of the tests help clarify the person’s understanding. Use of large print, fewer items per
page, and increased space between items allows examinees to focus better on individual items
without being overwhelmed by simultaneous presentation of multiple items as would occur
during a group-administered assessment. Visual magnification devices and templates that
reduce glare also may be incorporated into the assessment without affecting validity.
Scheduling
Administration of the Batería IV APROV tests can be scheduled at a specific time of day
to accommodate individual examinee needs. The tests may be presented in any order to
maximize interest and performance. When an individual cannot sustain peak performance for
long periods of time, the test may be administered over several days.
Young Children
Assessing young children in their preschool and early school years requires an examiner who
is trained and knowledgeable in this type of assessment. Examiners must select tests that
are appropriate for the age and functional level of the examinee. Some tests may not have
an adequate floor for young or low-functioning individuals, and other tests are designed
for use with school-age children or older individuals. For example, few individuals below
age 6 would be expected to perform adequately on tests such as Prueba 9: Fluidez en lectura
de frases or Prueba 13: Números matrices. On the other hand, examinees as young as age 2
generally can perform beginning tasks on Prueba 1: Identificación de letras y palabras and
Prueba 2: Problemas aplicados.
Preparation for Testing
Some young children may be uncomfortable with unfamiliar adults and may have difficulty
separating from their caregiver or teacher. It may be necessary to spend additional time with
such a child with a familiar adult nearby prior to accompanying the child to the testing
situation. Let the young child know that the caregiver is nearby and will be around when
testing is completed. In extreme circumstances, it may be necessary to have the familiar adult
stay with the child during testing. However, under these circumstances, the caregiver must
understand the standardized conditions under which the testing must occur. Every effort
should be made to minimize the caregiver’s involvement in the test situation. If a parent must
be present during the testing session, carefully explain the testing process, including the
establishment of test basals and ceilings (i.e., that some items may be very easy for the child
and that other items may be difficult), before the actual testing begins. Also, explain to the
parent that it is important he or she not assist the child in any way during the testing session.
The parent should be asked to sit to one side behind the child so that it is not too easy for the
child to interact with the parent during the test administration.
English Learners
The most important accommodation for students who are English learners (ELs) is having
an examiner who is knowledgeable about important issues relevant to second language
acquisition, the assessment process, and the interpretation of test results for students who are
ELs. To this end, the examiner must be familiar with the second language acquisition process,
native language attrition, language shift in dominance, cross linguistic transfer of learning,
and the impact of special language programming and socioeconomic factors on language
learning (August & Shanahan, 2006; Cummins & Hornberger, 2008; de Leeuw, 2008;
pragmatics using the hands, body, and facial expressions. Sign language does not follow
the grammar of the oral language. Sign languages in each country often have their
own dialect. In the United States, American Sign Language (ASL) is used. In Canada,
ASL is used mostly by Anglophones, and Quebec Sign Language is used mostly by
Francophones. In the Spanish-speaking countries, there are different sign languages for
different countries, including Argentine, Columbian, Cuban, Mexican, Peruvian, and
Spanish sign languages. Most sign languages have sign language dictionaries that can be
accessed online.
■ Sign-Supported Speech: The use of spoken language with sign used simultaneously
all or most of the time. People using this form of communication are not able to
adequately comprehend spoken language without sign accompaniment.
■ Aural/Oral Language: The use of spoken language without sign, usually aided by some
often have a more limited oral vocabulary than their hearing peers because they do not have the same access to spoken language. Rather than
demonstrating difficulty with reading speed or recall, the examinee may not know the meaning of some of the words.
8 Prueba 6: Expresión de lenguaje escrito—Explain the directions carefully and possibly change the wording if the examinee does not appear to
understand.
9 Prueba 6: Expresión de lenguaje escrito, Prueba 11: Fluidez en escritura de frases—Spelling errors made by individuals whose primary
communication mode is manual often have little phonetic relationship to the intended word. Allow time to review the responses and, if the
response word is not understandable due to a nonphonetic misspelling, ask the examinee to sign it. Even if no credit is awarded, knowing what
word the examinee intended will help with interpretation.
10 Prueba 8: Lectura oral—Because a person must know the meaning of a word to sign it, for sign communicators, this test assesses reading
vocabulary and comprehension instead of oral reading. Consequently, responses cannot be compared with the performance of hearing/speaking
peers in the norm sample. For examinees who use speech, consider that errors in pronunciation may be secondary to the hearing impairment
(articulation) rather than indications of weak decoding skills.
11 Prueba 12: Rememoración de lectura—For examinees who use sign language, this test might indicate their comprehension and recall of written
Spanish; however, they will have to fingerspell names and other words that do not have signs. The interpreter must be alerted to the importance of
the bolded words so that he or she will voice those particular words if the examinee’s signed response appropriately represents them.
12 Prueba 13: Números matrices—Because of the complexity, signed instructions may have to deviate significantly from the standardized instructions
3: Ortografía ◆3
4: Comprensión de textos □4
5: Cálculo ◆
7: Análisis de palabras ◆
8: Lectura oral □4
Interpretive Cautions
Many test modifications, such as altering administration procedures by providing additional
cues, are appropriate in specific circumstances. Modifying test procedures requires
understanding the examinee’s condition or Spanish-speaking limitations, as well as the nature
and purpose of each test. Keep in mind that, in many instances, the purpose of an evaluation
is to determine an individual’s unique pattern of strengths and weaknesses and then to use
this assessment data to suggest appropriate classroom accommodations and to recommend
possible teaching strategies and interventions. Although a modification may improve
test performance, the resulting score may not be an accurate reflection of an examinee’s
capabilities. Note any deviation from the standardized administration on the Test Record and
always include a statement of the modified testing conditions in the written report.
Administering and
Scoring the Batería IV
APROV Tests
This chapter contains detailed administration procedures for each of the tests in the
Batería IV Pruebas de aprovechamiento (Batería IV APROV). Comparing the information in
this chapter with the actual instructions in the Test Book will help examiners learn both
administration and scoring procedures. In addition, the test-by-test “Batería IV Pruebas de
aprovechamiento Examiner Training Checklist” in Appendix B of this manual can be a helpful
tool for examiners learning to administer the Batería IV APROV. It is recommended that
examiners first learn and practice administering the core tests (Tests 1 through 6) and then
the remaining tests.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of reading
achievement. Consult the Suggested Starting Points table in the Test Book to determine an
appropriate starting point for the examinee.
Basal
Test by complete pages until the 6 lowest-numbered items administered are correct, or until
the page with Item 1 has been administered.
Scoring
Score each correct response 1 and each incorrect response 0. Score words that are not read
fluently (smoothly) on the last attempt 0. Do not penalize an examinee for mispronunciations
resulting from articulation errors, dialect variations, or regional speech patterns. Record the
total number of all items answered correctly and all items below the basal in the Number
Correct box after the last Identificación de letras y palabras item on the Test Record.
Administration Procedures
Know the exact pronunciation of each item before administering the test. The correct
pronunciation is in parentheses following more difficult items. For additional help with
pronunciation, refer to a standard dictionary. Do not tell or help the examinee with any
letters or words during this test.
If the examinee’s response to a specific item is unclear, do not ask him or her to repeat the
specific item. Instead, allow the examinee to complete the entire page and then ask him or
her to repeat all of the items on that page. Score only the item in question; do not rescore the
other items.
If the examinee pronounces words letter by letter or syllable by syllable instead of reading
them fluently, tell the examinee, “Primero lee [lea] la palabra en silencio y luego dime
[dígame] la palabra completa” (First read the word silently and then say the whole word
smoothly). Give this instruction only once during administration of this test. If the examinee
gives more than one response, score the last response. Examiners may wish to record
incorrect responses for later error analysis. In addition, examiners may wish to complete
the “Qualitative Observation” checklist on the Test Record to document how the person
performed the task.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of math
achievement. Consult the Suggested Starting Points table in the Test Book to determine an
appropriate starting point for the individual.
Basal
Test by complete pages until the 5 lowest-numbered items administered are correct, or until
the page with Item 1 has been administered.
Ceiling
Test by complete pages until the 5 highest-numbered items administered are incorrect, or
until the page with Item 50 has been administered.
Administration Procedures
If the examinee requests or appears to need it, provide the worksheet (Hoja de trabajo)
in the Response Booklet and a pencil with eraser prior to being prompted to do so. In all
cases, provide the Response Booklet and a pencil as directed at Item 27. Any question may
be repeated during the test whenever the examinee requests. Because the focal construct of
this test is not the person’s reading ability, read all items to the examinee. Completing the
“Qualitative Observation” checklist on the Test Record can help characterize the examinee’s
performance on this task.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of spelling skill.
Consult the Suggested Starting Points table in the Test Book to determine an appropriate
starting point for the person.
Basal
Test until the 6 lowest-numbered items administered are correct, or until Item 1 has been
administered.
Ceiling
Test until the 6 highest-numbered items administered are incorrect, or until Item 52 has been
administered.
Scoring
Score each correct response 1 and each incorrect response 0. Do not penalize for poor
handwriting or reversed letters as long as the letter does not form a different letter. For example,
a reversed lowercase c would not be penalized, but a reversed lowercase b would be penalized
because it becomes the letter d. Accept upper- or lowercase responses as correct unless a case is
specified. Record the total number of all items answered correctly and all items below the basal
in the Number Correct box after the last Ortografía item on the Test Record.
Administration Procedures
Know the exact pronunciation of each test item before administering the test. The correct
pronunciation is in parentheses following more difficult items. Request printed responses;
however, accept cursive responses. Completing the “Qualitative Observation” checklist on the
Test Record can help describe the examinee’s automaticity on this task.
Starting Point
Begin with the Introduction for examinees functioning at the preschool to kindergarten level.
Begin with Item 7 for all examinees functioning at the grade 1 level. For all other examinees,
administer Sample Item B and then select a starting point based on an estimate of the examinee’s
present level of reading achievement. Consult the Suggested Starting Points table following
Sample Item B in the Test Book to determine an appropriate starting point for the individual.
Basal
Test by complete pages until the 6 lowest-numbered items administered are correct, or until
the page with Item 1 has been administered.
Ceiling
Test by complete pages until the 6 highest-numbered items administered are incorrect, or
until the page with Item 54 has been administered.
Scoring
Score each correct response 1 and each incorrect response 0. Unless noted, accept only one-
word responses as correct. If an examinee gives a two-word or longer response, ask for a one-
word answer. Score a response correct if it differs from the correct response(s) listed only in
verb tense or number (singular/plural), unless otherwise indicated by the scoring key. Score
a response incorrect if the person substitutes a different part of speech, such as a noun for a
verb, unless otherwise indicated by the scoring key. Do not penalize for mispronunciations
resulting from articulation errors, dialect variations, or regional speech patterns.
Record the total number of all items answered correctly and all items below the basal in
the Number Correct box after the last Comprensión de textos item on the Test Record. Do not
include points for the introduction or sample items.
Administration Procedures
Examinees should read the passages silently; however, some individuals, especially younger
children, may read aloud. If this happens, ask the person to read silently. If the individual
continues to read aloud, do not insist on silent reading. Do not tell the examinee any words
on this test.
The examinee needs to identify the specific word that goes in the blank. If he or she reads
the sentence aloud with a correct answer, say, “Responde [Responda] con una sola palabra”
(Tell me one word). If the examinee cannot provide the word, score the item incorrect.
For Items 15 and higher, if the examinee does not respond to an item in about 30 seconds,
encourage a response. If the person still does not respond, score the item 0, point to the next
item and say, “Prueba [Pruebe] con esta” (Try this one). The 30 seconds is a guideline and
not a time limit. If an examinee requires more time to complete an item, more time may be
given. For example, if a response is encouraged after 30 seconds and the examinee indicates
he or she is still reading or needs more time, it is permissible to give more time.
Mark the one description on the “Qualitative Observation” checklist on the Test Record
that best describes the person’s performance on this task.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of computational
skill. Consult the Suggested Starting Points table in the Test Book to determine an appropriate
starting point for the person.
Basal
Test until the 6 lowest-numbered items administered are correct, or until Item 1 has been
administered.
Ceiling
Test until the 6 highest-numbered items administered are incorrect, or until Item 57 has been
administered.
Scoring
Score completed items on this test before moving to another test to verify the basal and
ceiling and to complete any queries. Score each correct response 1 and each incorrect
response 0. If the examinee skips an item before the last completed item, score the item 0.
Score poorly formed or reversed numbers correct on this test. Score transposed numbers (e.g.,
12 for 21) incorrect. Record the total number of all items answered correctly and all items
below the basal in the Number Correct box after the last Cálculo item on the Test Record. Do
not include points for sample items.
Administration Procedures
If testing begins with Sample Item A and the examinee responds incorrectly to one or both
of the sample items, discontinue testing and record a score of 0 for this test. Make sure
to complete any queries listed in the Test Book, such as for the items involving reducing
fractions. Do not point to the signs or remind the examinee to pay attention to the signs
during this test. Use the “Qualitative Observation” checklist on the Test Record to help
describe the person’s rate and automaticity on this task.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of writing ability.
Administer the appropriate block of items as indicated in the Suggested Starting Points table
in the Test Book, on the page after the Expresión de lenguaje escrito tab.
Continuation Instructions
This test uses continuation instructions instead of basal and ceiling rules. Administer all
items in the selected block and then follow the continuation instructions. The continuation
instructions appear at the end of each block of items in the Test Book and on the Test Record.
sentence, score the item 0 and say, “Recuerda [Recuerde] que debes [debe] escribir una
oración completa” (Remember to write a complete sentence).
■ If the examinee changes a stimulus word in any way, score the item 0 and say,
“Recuerda [Recuerde] que no debes [debe] cambiar las palabras de ninguna forma”
(Remember, do not change the word(s) in any way).
■ If the examinee does not use all the stimulus words, score the item 0 and say, “Recuerda
[Recuerde] que debes [debe] usar todas las palabras.” (Remember to use all the words.)
■ If the examinee responds with two or more sentences, score the item 0 and say,
“Recuerda [Recuerde] que debes [debe] escribir solo una oración” (Remember to use
the words in one sentence).
■ If the examinee appears to be struggling to use the words in only the order presented,
say, “Recuerda [Recuerde] que puedes [puede] usar las palabras en cualquier orden”
(Remember, you can use the words in any order). While this reminder may be given
only once, it does not result in an automatic item score of 0.
You may want to write down these reminders on a card and keep it visually accessible
during testing. Although the reminders may not be repeated more than once, you may repeat
the instructions for the test items, if necessary.
Record the total number of points for each administered block in the appropriate boxes on
the Test Record.
Administration Procedures
If an examinee’s response to an item is illegible or difficult to read, ask him or her to write
as neatly as possible. The examiner may read any words to the examinee during this test
or repeat the instructions, if necessary. When an examinee asks if spelling is important or
how to spell a word, encourage the individual to just do the best he or she can. Do not spell
any words for the examinee. The overall quality of the individual’s written sentences can be
described by completing the “Qualitative Observation” checklist on the Test Record.
This test may be administered simultaneously to a small group of two or three individuals
if, in the examiner’s judgment, this procedure will not affect any examinee’s performance.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of reading skill.
The table in the Test Book presents suggested starting points.
Basal
Test by complete pages until the 6 lowest-numbered items administered are correct, or until
the page with Item 1 has been administered.
Ceiling
Test by complete pages until the 6 highest-numbered items administered are incorrect, or
until the page with Item 34 has been administered.
Scoring
Score each correct response 1 and each incorrect response 0. Score words that are not read
fluently (smoothly) on the last attempt 0. Do not penalize an examinee for mispronunciations
resulting from articulation errors, dialect variations, or regional speech patterns. Record the
total number of all items answered correctly and all items below the basal in the Number
Correct box after the last Análisis de palabras item on the Test Record. Do not include points
for sample items.
Administration Procedures
It is essential to know the exact pronunciation of each test item before administering the test.
The correct pronunciation is in parentheses following more difficult items. Say the phoneme
(the most common sound of the letter), not the letter name, when letters are printed within
slashes, such as /p/.
If the examinee has any special speech characteristics resulting from articulation errors or
dialect variations, become familiar with the examinee’s speech pattern before administering
this test.
If the examinee’s response to a specific item is unclear, do not ask him or her to repeat the
specific item. Instead, allow the person to complete the entire page and then ask him or her
to repeat all of the items on that page. Score only the item in question; do not rescore the
other items.
If the examinee pronounces words letter by letter or syllable by syllable instead of reading
them fluently, tell the individual, “Primero lee [lea] la palabra en silencio y luego dime
[dígame] la palabra completa” (First read the word silently and then say the whole word
smoothly). Give this instruction only once during the administration of this test. Score the
examinee’s last response. The examiner may wish to record incorrect responses for later
error analysis. In addition, the examiner may wish to complete the “Qualitative Observation”
checklist on the Test Record to document how the person performed the task.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of reading skill.
Consult the Suggested Starting Points table in the Test Book to determine an appropriate
starting point for the individual.
Continuation Instructions
This test uses continuation instructions instead of basal and ceiling rules. Follow the
continuation instructions to determine which additional sentences should be administered
and when to discontinue testing. The continuation instructions are located at the bottom of
the examiner pages in the Test Book and on the Test Record.
Scoring
When the examinee reads a sentence with no errors, score the item 2. If the examinee makes
one error on the sentence, score the item 1. When the examinee makes two or more errors,
score the item 0. Types of reading errors include mispronunciations, omissions, insertions,
substitutions, hesitations of more than 3 seconds, repetitions, transpositions, or ignoring
punctuation. If the examinee self-corrects within 3 seconds, do not count the word as an
error. Do not penalize the examinee for mispronunciations resulting from articulation errors,
dialect variations, or regional speech patterns. Record the number of points earned in the
Number of Points box after the last Lectura oral item on the Test Record.
Administration Procedures
It is essential to know the exact pronunciation of each test item. The correct pronunciation is
in parentheses following more difficult words.
Become familiar with the types of reading mistakes that count as errors on this test. Figure
4-1 lists the types of reading errors that are shown in the Test Book. Sentences are reproduced
on the Test Record to facilitate scoring. During the test, follow along on the Test Record as the
examinee reads each sentence and mark each error with a slash (/) at the point in the sentence
where the error occurs. In most cases, the slash will be placed on the printed word that was the
error (i.e., mispronunciation, omission, substitution, transposition, hesitation, or repetition).
For an inserted word, place the slash between the two printed words where the insertion
occurred. If the examinee ignores punctuation (e.g., does not pause at a comma or raise his or
her voice for a question mark), place the slash on the punctuation mark that was ignored. The
examiner can also record and total each type of error in the “Qualitative Observation Tally” on
the Test Record. Figure 4-2 illustrates a portion of a completed Test Record and tally.
Mala pronunciación
La calificación se en basa la administración de un grupo específico de ítems.
Ignora puntuación
Tally” for Prueba 8: Lectura
Transposición
Sustitución
Repetición
Vacilación
Inserción
oral.
Omisión
Calificación 2, 1, 0
1
1 Una historia sobre ballenas. Un 1
2 2 Pueden nadar.
1 muchos 1
5 Comen peces pequeños.
Ítems 1-5 8A: Número de Número de errores
6 puntos (0-10) Ítems 1-5 2 1 1
5 o menos puntos: Suspenda la prueba
6 o más puntos: Administre Ítems 6-10, si no han sido administrados
10 Un amigo puede preguntar: “¿Alguna vez has visto a una ballena saltar
alto fuera del agua?”
Ítems 6-10 8B: Número de Número de errores
puntos (0-10) Ítems 6-10
5 o menos puntos: Administre Ítems 1-5, si no han sido administrados
6 o más puntos: Administre Ítems 11-15, si no han sido administrados
11 Las ballenas usan sonidos para comunicarse; estos sonidos pueden ser
escuchados por millas bajo el agua.
When prompted, give the examinee the Response Booklet and a pencil with an eraser. This
test requires a stopwatch or a watch or clock with a second hand.
Starting Point
All examinees complete the sample items and practice exercise and then begin with Item 1.
Time Limit
Discontinue testing after exactly 3 minutes and collect the examinee’s pencil and Response
Booklet. Record the exact finishing time in minutes and seconds on the Test Record. It is
Scoring
Score each correct response 1 and each incorrect response 0. Ignore skipped items. Use the
Fluidez en lectura de frases Scoring Guide overlay to score this test. Record both the total
number of items answered correctly and the total number of items answered incorrectly
within the 3-minute time limit in the Fluidez en lectura de frases Number Correct and
Number Incorrect boxes on the Test Record. To obtain the estimated age and grade
equivalents on the Test Record, subtract the Number Incorrect from the Number Correct.
Enter both the Number Correct and the Number Incorrect into the online scoring and
reporting program. Do not include points for sample items or practice exercises.
Administration Procedures
If the examinee has 2 or fewer correct on Practice Exercises C through F, discontinue testing
and record a score of 0 in the Fluidez en lectura de frases Number Correct box on the Test
Record.
The sentences are intended to be read silently. Remind the examinee to read silently if he
or she begins reading aloud. If the person appears to be answering items without reading the
sentences, remind him or her to read each sentence. If the individual stops at the bottom of
a page, remind him or her to continue to the top of the next column or to the next page. If
the examinee starts to erase a response, provide a reminder to cross out the answer he or she
does not want.
This test may be administered simultaneously to a small group of two or three individuals
if, in the examiner’s judgment, this procedure will not affect any person’s performance.
However, do not administer this test to individuals who cannot read.
Starting Point
All examinees begin with Item 1.
Time Limit
Discontinue testing after exactly 3 minutes and collect the examinee’s pencil and Response
Booklet. Record the exact finishing time in minutes and seconds on the Test Record. It is
important to record the exact finishing time because examinees who do well and finish in less
than 3 minutes will receive a higher score than individuals who continue to work for the full
3 minutes.
If the examinee has 3 or fewer correct after 1 minute, discontinue testing, and record a
time of 1 minute and the Number Correct (0 to 3) on the Test Record.
Scoring
Score each correct response 1 and each incorrect response 0. Use the Fluidez en datos
Administration Procedures
Do not point to the signs or remind the examinee to pay attention to the signs during testing.
Watch to make sure the examinee is going from left to right, row by row, down the page.
Some examinees may choose to work left to right on the first row, right to left on the second
row, and so on, which is acceptable. However, if the examinee starts skipping around, remind
him or her to proceed across the page, one row at a time. If the examinee stops at the bottom
of the page, remind him or her to continue to the top of the next page. If the examinee starts
to erase a response, remind the examinee to cross out the answer he or she does not want.
This test may be administered simultaneously to a small group of two or three individuals
if, in the examiner’s judgment, this procedure will not affect any person’s performance.
Starting Point
All examinees complete the sample items and then begin with Item 1.
Time Limit
Discontinue testing after exactly 5 minutes and collect the examinee’s pencil and Response
Booklet. Record the exact finishing time in minutes and seconds on the Test Record. It is
important to record the exact finishing time because examinees who do well and finish in less
than 5 minutes will receive a higher score than individuals who continue to work for the full
5 minutes.
If an examinee has 3 or fewer correct responses within the first 2 minutes, discontinue
testing. Record a time of 2 minutes and the Number Correct (0 to 3) on the Test Record.
Scoring
Score each correct response 1 and each incorrect response 0. Score any skipped items
incorrect. Do not penalize an examinee for errors in punctuation, capitalization, or spelling or
for poor handwriting unless the response is illegible. Score illegible items incorrect.
Sometimes it may not be immediately apparent whether to score an item correct or
incorrect. A few general guidelines will assist in scoring the Fluidez en escritura de frases test.
To receive credit for an item, the examinee must use all three stimulus words in a complete
sentence. As noted in the Test Book instructions, the examinee may not change the stimulus
word in any way. If, for example, the examinee alters the tense of a verb or changes a noun
from singular to plural, score the item incorrect. A minor change in a word may make it
easier for the examinee to write a sentence, thus altering the difficulty level of the item.
However, if a stimulus word is miscopied or misspelled, the item can still receive credit as
long as the miscopying did not result in a change in tense, part of speech, or number.
To receive credit, the response must be a reasonable sentence. Some examinees may
Administration Procedures
If the examinee receives a 0 on Sample Items B through D after the error correction
procedure, discontinue testing and record a score of 0 in the Fluidez en escritura de frases
Number Correct box on the Test Record. If the examinee stops at the bottom of a page,
remind him or her to continue to the top of the next page.
In this test, the examiner may read any of the stimulus words to the examinee if requested
by the examinee. This test may be administered simultaneously to a small group of two or
three individuals if, in the examiner’s judgment, this procedure will not affect any person’s
performance.
Starting Point
Select a starting point based on an estimate of the examinee’s present level of reading ability.
Consult the Suggested Starting Points table in the Test Book to determine an appropriate
starting point for the examinee.
Continuation Instructions
This test uses continuation instructions instead of basal and ceiling rules. Follow the
continuation instructions in the Test Book to determine which additional stories should be
administered and when to discontinue testing. Because the continuation instructions on the
Test Record are abbreviated, consult the complete continuation instructions in the Test Book.
Scoring
On the Test Record, the elements to be scored are separated by slash marks (/). Place a check
mark above each element that the examinee recalls correctly during the retelling. Score each
correctly recalled element 1 and each incorrectly recalled element 0. Score elements not
recalled at all (correctly or incorrectly) 0. Scoring is based on a key word (shown in bold
Administration Procedures
Direct the examinee to read the story once silently. If necessary, remind the examinee of
this rule. Turn the page after the examinee has finished reading the story once. Prompt the
examinee as directed to retell the story. Do not tell the examinee any words on this test. It is
important to be familiar with the stories and required elements before administering this test.
This will facilitate scoring elements, particularly if the examinee retells them out of sequence.
Starting Point
Select the appropriate sample item based on an estimate of the person’s present achievement
level. Begin with Sample Item A for examinees functioning at the kindergarten to grade 8
level. For all other examinees, administer Sample Item B and then select a starting point
based on an estimate of the examinee’s present level of ability. Consult the Suggested Starting
Points table following Sample Item B in the Test Book to determine an appropriate starting
point for the individual.
Basal
Test by complete pages until the 6 lowest-numbered items administered are correct, or until
the page with Item 1 has been administered.
Ceiling
Test by complete pages until the 6 highest-numbered items administered are incorrect, or
until the page with Item 30 has been administered.
Scoring
Score each correct response 1 and each incorrect response 0. To be correct, an answer must
solve the problem both horizontally and vertically. Record the total number of all items
answered correctly and all items below the basal in the Number Correct box after the last
Números matrices item on the Test Record. Do not include points for sample items.
The four levels of test information are cumulative; that is, each successive level builds on
information from the previous level. Information from all four levels is necessary to describe
a person’s performance completely. Level 1 provides qualitative data that are often used to
support a clinical hypothesis. Levels 2, 3, and 4 include a variety of score options from which
to select.
Level 1 information is obtained through behavioral observations during testing and
through analysis of erroneous responses to individual items. Observation of an examinee’s
Types of Scores
This section discusses the variety of scores available for test interpretation. Included among
these scores are grade equivalents (GE), age equivalents (AE), relative proficiency indexes
(RPI), cognitive-academic language proficiency (CALP) levels, percentile ranks (PR), and
standard scores (SS). Most of these scores will be familiar to examiners who have used the
Batería III Woodcock-Muñoz or the Woodcock-Johnson IV. Several optional standard score
scales, including the normal curve equivalents (NCE) scale, also are discussed.
Table 5-4. Standard Score Range Percentile Rank Range Batería IV Classification
Classification of Standard
131 and above 98 to 99.9 Very Superior
Score and Percentile Rank
Ranges 121 to 130 92 to 97 Superior
111 to 120 76 to 91 High Average
90 to 110 25 to 75 Average
80 to 89 9 to 24 Low Average
70 to 79 3 to 8 Low
69 and below 0.1 to 2 Very Low
The online scoring and reporting program provides the option to report an additional
standard score from a selection of four other types of standard scores: z scores, T scores,
stanines, and normal curve equivalents (NCEs). The basic standard score is the z score with
a mean of 0 and a standard deviation of 1. The T score has a mean of 50 and a standard
deviation of 10. Although T scores have been frequently used in education and industry,
they have been replaced by the deviation-IQ scale (M = 100, SD = 15) for most clinical
applications. Another standard score scale is the traditional stanine scale. Stanines have a
mean of 5 and a standard deviation of 2 and are most useful in applications in which a single-
digit gross scale of measurement is desired. The normal curve equivalent scale (Tallmadge
& Wood, 1976) has a mean of 50 and a standard deviation of 21.06 and has been used most
often for evaluating student performance in certain federally funded programs.
Interpreting Tests
This section contains details on interpretation of the tests in each of the curricular areas.
Chapter 2 contains functional definitions of the abilities measured by each test. In evaluating
the practical significance of differences among test performance, consider any extenuating
circumstances that may explain these differences, as well as any unusual behaviors or
responses obtained on those tests. This information may have useful diagnostic implications.
LESS COMPLEX
LESS COMPLEX
In terms of complexity, the skills measured in the four Batería IV math tests range from
the lower-level ability of recognizing math symbols and vocabulary to the higher-level ability
of mathematical reasoning and problem solving. Based on CHC theory, the math tests are
primarily measures of quantitative knowledge (Gq), although some math tests measure other
aspects of processing, particularly fluid reasoning (Gf ) or processing speed (Gs).
Prueba 10: Fluidez en datos matemáticos (Test 10: Math Facts Fluency)
This test is a measure of math achievement and number facility requiring the examinee
to solve simple addition, subtraction, and multiplication problems rapidly (Gq, Gs). Low
performance on Fluidez en datos matemáticos may be a function of limited knowledge of basic
math facts or lack of automaticity. The “Qualitative Observation” checklist for this test helps
document how the examinee approached the task. Table 5-10 provides information about the
percentage of age mates who were assigned each rating in the norming sample. For example,
while 17% of the 8-year-olds whose performance was rated solved the problems slowly, only
6% of the 11-year-olds had that same rating. Using this information can help determine how
typical or atypical the examinee’s performance is compared to age mates.
Rate/Automaticity Prueba 11: Fluidez en A picture and three Writing short sentences
escritura de frases words to form into quickly—requires correct
a sentence syntax and automaticity
Isolated Letters Prueba 3: Ortografía Orally presented letters Writing letter names
(Sublexical Level)
LESS COMPLEX
Prueba 11: Fluidez en escritura de frases (Test 11: Sentence Writing Fluency)
This test measures the examinee’s ability to write rapidly with ease (automaticity; Grw, Gs).
The task requires the production of legible, simple sentences with acceptable syntax. Minimal
analytic attention or problem solving is necessary.
Performance on Prueba 11: Fluidez en escritura de frases may be related to several factors,
including muscular or motor control, response style, ability to sustain concentration, and
reading or spelling skills. When an examinee’s attention is focused on the mechanics of
writing rather than on the formulation or expression of ideas, writing is not automatic. Poor
muscular control may contribute to a concentration on the mechanics of the writing task and
contribute to low scores. In addition, an examinee’s response to timed tasks can influence the
quality of automaticity. A range of different response styles to this task has been observed.
Some examinees complete all tasks at a slow, consistent pace, regardless of imposed time
constraints. Other examinees work very rapidly but tend to make a lot of careless errors. In
an interpretation of the examinee’s response style, an examiner may want to define whether
the examinee worked (a) slowly but inaccurately, (b) slowly and accurately, (c) rapidly but
Intra-Ability Variations
Intra-ability variation models are bidirectional comparisons (as represented by the two-
headed arrows in Figure 5-5) that allow comparison of performance among skills and
abilities. There are three types of intra-ability variations in the Batería IV: intra-achievement
(determined with the Batería IV APROV), academic skills/academic fluency/academic
applications (determined with the Batería IV APROV), and intra-cognitive (determined with
the Batería IV COG). The two variation procedures discussed in detail here are the ones
available when using the Batería IV APROV. While a summary is presented here, consult the
Batería IV COG Examiner’s Manual for further information about the intra-cognitive variation
procedure.
Intra-Achievement Variations
This variation procedure allows comparison of one area of academic achievement to
the examinee’s expected or predicted performance as determined by his or her average
performance on other achievement areas. An intra-achievement variation is present within
individuals who have specific achievement strengths or weaknesses, such as superior
math skills relative to their expected achievement based on their average performance
in other areas of achievement. Individuals with a significant intra-achievement variation
exhibit specific strengths or weaknesses in one or more areas of achievement. This type of
information is an invaluable aid in instructional planning and can be used, for example,
to support the hypothesis of a specific difficulty as compared to generally low academic
performance across achievement domains. For example, a student may perform poorly in
mathematics but may have average abilities on tasks involving reading.
As indicated in Table 5-15, intra-achievement variations can be calculated if Batería IV
APROV Tests 1 through 6 are administered. Each test is compared to the examinee’s predicted
or expected test score based on his or her average performance on the other five tests. For
example, when considering Prueba 1: Identificación de letras y palabras, the individual’s
average performance on the remaining five tests (Tests 2 through 6) is used as the predictor
to determine his or her expected score on Prueba 1: Identificación de letras y palabras. This
expected score is then compared to the person’s obtained Prueba 1: Identificación de letras
y palabras score. If the individual’s expected score is higher than his or her actual score, a
relative weakness is identified. If the expected score is lower than the actual score, a relative
strength is identified.
Intra-Achievement
Variations
Cognitive Abilities Achievement
As an option, other tests can be included in the variation procedure. When including any
additional tests, the corresponding cluster or clusters that are created also are included in
the variation procedure. For example, if Prueba 7: Análisis de palabras is administered, the
Destrezas básicas en lectura cluster is available when combined with Prueba 1: Identificación
de letras y palabras. Therefore, both Prueba 7: Análisis de palabras and the Destrezas básicas
en lectura cluster are compared to the same expected score based on the same predictor as
Prueba 1: Identificación de letras y palabras in the variation procedure. No matter how many
tests are administered, the predictor score is always based on five tests from Batería IV
APROV Tests 1 through 6. An intra-achievement variation is present within individuals
who have specific academic strengths or weaknesses, such as superior Destrezas básicas en
lectura (Grw) relative to their expected performance based on their average performance on
the remaining five tests. If any of the optional additional tests are included in the variation
procedure, the variation is labeled Intra-Achievement (Extended).
Prueba 3: Ortografía
Prueba 6: Expresión de lenguaje escrito Uses same predictor as Expresión de lenguaje escrito
Prueba 11: Fluidez en escritura de frases
Expresión escrita
Intra-Cognitive Variations
This variation is present within individuals who have specific cognitive strengths or
weaknesses, such as high fluid reasoning (Gf ) or poor short-term working memory (Gwm).
Equal interest exists in either a strength or a weakness in one ability relative to an individual’s
average performance in other cognitive abilities. This profile of variations can document areas
of relative strength and weakness, provide insights for program planning, and contribute to
a deeper understanding of the types of tasks that will be especially easy or difficult for an
individual compared to his or her other abilities.
Based on Batería IV COG Tests 1 through 7, this variation procedure allows comparison
of one area of cognitive ability to the examinee’s expected or predicted score based on his
or her average performance on six of the first seven cognitive tests, each measuring some
aspect of a different CHC cognitive ability (Gc, Gf, Gwm, Gs, Ga, Glr, Gv). For example,
when considering Prueba 1: Vocabulario oral, the individual’s average performance on the
remaining six tests (Prueba 2: Series numéricas, Prueba 3: Atención verbal, Prueba 4: Pareo
de letras idénticas, Prueba 5: Procesamiento fonético, Prueba 6: Rememoración de cuentos, and
Prueba 7: Visualización) is used as the predictor to determine the person’s expected Prueba 1:
Vocabulario oral score. This expected score is then compared to the person’s obtained Prueba
1: Vocabulario oral score. An intra-cognitive variation is present within individuals who have
specific cognitive strengths or weaknesses, such as superior comprehension-knowledge (Gc)
relative to their expected performance based on their average performance in other areas of
cognitive ability. If administered, the three Spanish language tests in the WJ IV OL can be
entered into the intra-cognitive variation. See the Batería IV COG Examiner’s Manual for
more information.
Gf-Gc Comparisons
Matemáticas
Matemáticas amplias
Destrezas en cálculos matemáticos
Resolución de problemas matemáticos
Lenguaje escrito
Lenguaje escrito amplio
Expresión escrita
Destrezas académicas
Fluidez académica
Aplicaciones académicas
Discrepancy Scores
The online scoring and reporting program includes two scores for use in interpreting
the presence and severity of any variation, comparison, or discrepancy. These are called
the discrepancy percentile rank (discrepancy PR) and the discrepancy standard deviation
(discrepancy SD). These scores are based on actual difference scores computed for each
individual in the norming sample. (See the WJ IV Technical Manual for more information.)
August, D., & Shanahan, T. (2006). Developing literacy in second-language learners: Report of
the national literacy panel on language minority children and youth. Mahwah, NJ: Lawrence
Erlbaum.
Carroll, J. B. (1993). Human cognitive abilities: A survey of factor-analytic studies. New York,
NY: Cambridge University Press.
Carroll, J. B., & Maxwell, S. E. (1979). Individual differences in cognitive abilities. Annual
Review of Psychology, 30, 603–640.
Cattell, R. B. (1950). Personality: A systematic theoretical and factoral study. New York, NY:
McGraw-Hill.
Corn, A. L., & Lusk, K. E. (2010). Perspectives on low vision. In A. L. Corn & J. N. Erin
(Eds.), Foundations of low vision: Clinical and functional perspectives (2nd ed., pp. 3–34).
New York, NY: AFB Press.
Cummins, J. (1984). Bilingualism and special education: Issues in assessment and pedagogy.
Austin, TX: Pro-Ed.
Cummins, J., & Hornberger, N. H. (2008). Bilingual education. New York, NY: Springer.
de Leeuw, E. (2008). When your native language sounds foreign: A phonetic investigation into
first language attrition (Unpublished doctoral dissertation). Queen Margaret University,
Edinburgh, Scotland. Retrieved from https://eresearch.qmu.ac.uk/handle/20.500.12289/7436.
Flege, J. E., Schirru, C., & MacKay, I. R. (2003). Interaction between the native and second
language phonetic subsystems. Speech Communication, 40, 467–491.
Grosjean, F. (2001). The bilingual’s language modes. In J. Nicol (Ed.), One mind, two
languages: Bilingual language processing (2nd ed., pp. 1–22). Oxford, England: Blackwell.
Herschell, A. D., Greco, L. A., Filcheck, H. A., & McNeil, C. B. (2002). Who is testing
whom: Ten suggestions for managing disruptive behavior of young children during testing.
Intervention in School and Clinic, 37, 140–148.
References 99
Horn, J. L. (1965). Fluid and crystallized intelligence. (Unpublished doctoral dissertation).
University of Illinois, Urbana-Champaign, IL.
Horn, J. L. (1989). Models for intelligence. In R. Linn (Ed.), Intelligence: Measurement, theory
and public policy (pp. 29–73). Urbana, IL: University of Illinois Press.
Horn, J. L., & Cattell, R. B. (1966). Refinement and test of the theory of fluid and crystallized
general intelligences. Journal of Educational Psychology, 57, 253–270.
Horn, J. L., & Stankov, L. (1982). Auditory and visual factors of intelligence. Intelligence, 6,
165–185.
Linacre, J. M. (2002). What do infit and outfit, mean-square and standardized mean? Rasch
Measurement Transactions, 16(2), 878.
Mather, N., & Wendling, B. J. (2014). Examiner’s Manual. Woodcock-Johnson IV Tests of Oral
Language. Rolling Meadows, IL: Riverside Publishing.
McArdle, J. J., & Woodcock, R. W. (1998). Human cognitive abilities in theory and practice.
Mahwah, NJ: Lawrence Erlbaum.
McGrew, K. S., LaForte, E. M., & Schrank, F. A. (2014). Technical Manual. Woodcock-Johnson
IV. Rolling Meadows, IL: Riverside Publishing.
100 References
McGrew, K. S., & Wendling, B. J. (2010). Cattell-Horn-Carroll cognitive-achievement
relations: What we have learned from the past 20 years of research. Psychology in the
Schools, 47, 651–675.
Muñoz-Sandoval, A. F., Woodcock, R. W., McGrew, K. S., & Mather, N. (2005, 2007a).
Batería III Woodcock-Muñoz. Rolling Meadows, IL: Riverside Publishing.
Muñoz-Sandoval, A. F., Woodcock, R. W., McGrew, K. S., & Mather, N. (2005, 2007b).
Batería III Woodcock-Muñoz Normative Update: Pruebas de aprovechamiento. Rolling
Meadows, IL: Riverside Publishing.
Muñoz-Sandoval, A. F., Woodcock, R. W., McGrew, K. S., & Mather, N. (2005, 2007c).
Batería III Woodcock-Muñoz Normative Update: Pruebas de habilidades cognitivas. Rolling
Meadows, IL: Riverside Publishing.
Prifitera, A., Saklofske, D. H., & Weiss, L. G. (Eds.). (2008). WISC-IV clinical assessment and
intervention. (2nd ed., pp. 217–235). San Diego, CA: Academic Press.
Rasch, G. (1960). Probabilistic models for some intelligence and attainment tests. Copenhagen,
Denmark: Danish Institute for Educational Research.
Sackett, P. R., & Yang, H. (2000). Correction for range restriction: An expanded typology.
Journal of Applied Psychology, 85, 112–118.
Sattler, J. M., & Hoge, R. D. (2005). Assessment of children: Behavioral, social, and clinical
foundations (5th ed.). San Diego, CA: Author.
Schrank, F. A., Mather, N., & McGrew, K. S. (2014). Woodcock-Johnson IV Tests of Oral
Language. Rolling Meadows, IL: Riverside Publishing.
Schrank, F. A., McGrew, K. S., & Mather, N. (2014). Woodcock-Johnson IV. Rolling Meadows,
IL: Riverside Publishing.
Smith, J. K. (1999). The effects of practice on the reading speed, accuracy, duration, and visual
fatigue of students with low vision when accessing standard size print with optical devices.
(Unpublished doctoral dissertation). University of Arizona, Tucson, AZ.
Stevens, S. S. (1951). Handbook of experimental psychology. New York, NY: John Wiley.
Tallmadge, G. K., & Wood, C. T. (1976, October). User’s guide, ESEA Title I evaluation and
reporting system. Mountain View, CA: RMC Research.
Thomas, W. P., & Collier, V. P. (2002). A national study of school effectiveness for language
minority students’ long-term academic achievement. Santa Cruz, CA: Center for Research on
Education, Diversity, and Excellence, University of California-Santa Cruz. Retrieved from:
http://www.thomasandcollier.com/assets/2002_thomas-and-collier_2002-final-report.pdf.
References 101
U.S. Department of Education. Family Educational Rights and Privacy Act. (1974). 20 U.S.C.
§ 1232g; 34 CFR Part 99.10(c) and (d).
Wendling, B. J., Mather, N., & Schrank, F. A. (2019). Examiner’s Manual. Batería IV Pruebas
de habilidades cognitivas. Itasca, IL: Riverside Assessments, LLC.
Wolfe, E. W. (2004). Equating and item banking with the Rasch model. In E. V. Smith, Jr. &
R. M. Smith (Eds.), Introduction to Rasch measurement. Maple Grove, MN: JAM Press.
Wolfe, F., MacIntosh, R., Kreiner, S., Lange, R., Graves, R., & Linacre, J. M. (2006). Multiple
significance tests. Rasch Measurement Transactions, 19, 1044.
Woodcock, R. W. (1973). Woodcock Reading Mastery Tests. Circle Pines, MN: American
Guidance Service.
Woodcock, R. W. (1982, March). Interpretation of the Rasch ability and difficulty scales
for educational purposes. Paper presented at the meeting of the National Council on
Measurement in Education, New York, NY.
Woodcock, R. W. (1987, 1998). Woodcock Reading Mastery Tests–Revised. Circle Pines, MN:
American Guidance Service.
Woodcock, R. W. (1988, August). Factor structure of the tests of cognitive ability from the 1977
and 1989 Woodcock-Johnson. Paper presented at the Australian Council on Educational
Research Seminar on Intelligence, Melbourne, Australia.
Woodcock, R. W. (1999). What can Rasch-based scores convey about a person’s test
performance? In S. E. Embretson & S. L. Hershberger (Eds.), The new rules of
measurement: What every psychologist and educator should know (pp. 105–128). Mahwah,
NJ: Lawrence Erlbaum.
Woodcock, R. W., Alvarado, C. G., Schrank, F. A., Mather, N., McGrew, K. S., & Muñoz-
Sandoval, A. F. (2019). Batería IV Woodcock-Muñoz: Pruebas de aprovechamiento. Itasca, IL:
Riverside Assessments, LLC.
Woodcock, R. W., Alvarado, C. G., Schrank, F. A., McGrew, K. S., Mather, N., & Muñoz-
Sandoval, A. F. (2019a). Batería IV Woodcock-Muñoz. Itasca, IL: Riverside Assessments, LLC.
102 References
Woodcock, R. W., Alvarado, C. G., Schrank, F. A., McGrew, K. S., Mather, N., & Muñoz-
Sandoval, A. F. (2019b). Batería IV Woodcock-Muñoz: Pruebas de habilidades cognitivas.
Itasca, IL: Riverside Assessments, LLC.
Woodcock, R. W., & Dahl, M. N. (1971). A common scale for the measurement of person ability
and test item difficulty (AGS Paper No. 10). Circle Pines, MN: American Guidance Service.
Woodcock, R. W., McGrew, K. S., & Mather, N. (2001, 2007). Woodcock-Johnson III. Rolling
Meadows, IL: Riverside Publishing.
Wright, B. D., & Stone, M. H. (1979). Best test design. Chicago, IL: MESA Press.
Zieky, M. (1993). Practical questions in the use of DIF statistics in test development. In P.
W. Holland & H. Wainer (Eds.), Differential item functioning (pp. 337–347). Hillsdale, NJ:
Lawrence Erlbaum.
References 103
Appendix A
Bessemer Phoenix
Alaska Pima
Birmingham
Anchor Point Pirtleville
Center Point
Homer Portal
Clay
Nikolaevsk Sahuarita
Crestline
Cullman San Simon
Arizona Scottsdale
Fairfield
Avondale Sierra Vista
Florence
Benson Surprise
Forestdale
Bisbee Tempe
Gardendale
Bouse Tolleson
Hamilton
Bowie Tucson
Helena
Buckeye Vail
Homewood
Chandler Willcox
Hoover
Douglas
Huntsville
El Mirage Arkansas
Moody
Gilbert Arkadelphia
Mountain Brook
Glendale Blytheville
Pelham
Goodyear Lowell
Riverchase
Hereford Pine Bluff
Roebuck Plaza
Laveen Refield
Scottsboro
Mesa Springdale
Selma
Oro Valley
Tarrant
Appendix A 105
California Forest Ranch Monrovia
106 Appendix A
Santa Cruz Castle Rock Winsted
Santa Maria Centennial Woodbury
Santa Monica Colorado Springs
Santa Rosa Denver Delaware
Santee Englewood Bear
Scotts Valley Evans
Shasta Greeley District of Columbia
Shasta Lake Greenwood Village Washington
Sherman Oaks Highlands Ranch
Florida
South San Francisco Larkspur
Spring Valley Littleton Apopka
Stockton Loveland Boca Raton
Studio City Parker Boynton Beach
Sun City Thornton Bradenton
Sylmar Westminster Brandon
Tarzana Brooksville
Temecula Connecticut Clearwater
Temple City Chester Clearwater Beach
Thousand Oaks Clinton Clermont
Tiburon Durham Coconut Creek
Turlock East Haven Cooper City
Tustin Essex Coral Gables
Upland Groton Coral Springs
Valencia Ivoryton Crystal River
Van Nuys Litchfield Davie
Venice Middlefield Deerfield Beach
Ventura New Britain Dunedin
Vista New Haven Fort Lauderdale
West Hollywood New London Fort Myers
Willows Oakdale Fruitland Park
Winchester Oakville Gainesville
Windsor Southington Glen Saint Mary
Woodland Stamford Green Cove Springs
Woodland Hills Stratford Greenacres
Yorba Linda Torrington Hallandale
Yreka Waterford Hallandale Beach
Watertown Hernando
Colorado West Granby Hialeah
Aurora West Haven Holiday
Boulder Westport Hollywood
Castle Pines Windsor Hudson
Appendix A 107
Jacksonville Ponte Vedra Beach Columbus
Jensen Beach Port Orange Comer
Kenneth City Port Richey Conyers
Lady Lake Port Saint Lucie Cordele
Lakewood Ranch Redington Shores Cumming
Land O Lakes Riverview Cuthbert
Largo Safety Harbor Dacula
Lauderdale Lakes Saint Augustine Decatur
Lauderhill Saint Petersburg Doraville
Leesburg Sarasota Douglasville
Lutz Seffner Duluth
Margate Southwest Ranches Dunwoody
Miami Stuart Ellenwood
Miami Beach Sunrise Fayetteville
Middleburg Tallahassee Flintstone
Miramar Tamarac Flowery Branch
Myakka City Tampa Fort Gaines
New Port Richey Tarpon Springs Fort Oglethorpe
Newport Temple Terrace Gainesville
Nokomis The Villages Grayson
North Lauderdale Trinity Helena
North Miami Valrico Kennesaw
North Miami Beach Wellington Lagrange
Oakland Park Wesley Chapel Lawrenceville
Ocala West Palm Beach Lilburn
Ocklawaha Weston Lincolnton
Ocoee Lithonia
Odessa Georgia Loganville
Oldsmar Alpharetta Lula
Orange Park Athens Marietta
Palm Beach Atlanta McDonough
Palm Beach Gardens Blakely Meansville
Palm City Bogart Milton
Palm Coast Bonaire Monroe
Palm Harbor Brooks Morganton
Palmetto Buford Morris
Parkland Calhoun Norcross
Pembroke Pines Canon Oxford
Plant City Carlton Ringgold
Plantation Chamblee Riverdale
Pompano Beach College Park Rock Springs
108 Appendix A
Rossville Nampa Evergreen Park
Rydal Rexburg Fairview Heights
Sandy Springs Rigby Frankfort
Smyrna Saint Anthony Geneva
Snellville Star Gilberts
Social Circle Teton Glendale Heights
Stockbridge Twin Falls Glenview
Stone Mountain Hampshire
Summerville Illinois Hanover Park
Suwanee Addison Harwood Heights
Sylvester Alsip Hazel Crest
Trenton Arlington Heights Highland Park
Tucker Aurora Highwood
Union City Bartlett Hinsdale
Warner Robins Batavia Hoffman Estates
Watkinsville Bensenville Hometown
Winder Berwyn Joliet
Woodstock Blue Island Justice
Bolingbrook LaGrange Park
Hawaii Buffalo Grove Lake in the Hills
Hauula Carpentersville Lansing
Honokaa Cary Libertyville
Honolulu Chicago Lincolnwood
Kaaawa Chicago Heights Lisle
Kailua Chicago Ridge Lockport
Kaneohe Cicero Lynwood
Kapolei Country Club Hills Lyons
Laupahoehoe Crest Hill Mahomet
Pearl City Crestwood Matteson
Wahiawa Crystal Lake McHenry
Waimanalo DeKalb Midlothian
Des Plaines Mokena
Idaho Dixon Montgomery
Ammon Dolton Mundelein
Boise Downers Grove Naperville
Bonners Ferry East Hazel Crest New Lenox
Caldwell East Saint Louis Oak Forest
Eagle Elgin Oak Lawn
Idaho Falls Elk Grove Village Orland Hills
Kuna Elmwood Park Orland Park
Meridian Evanston Palatine
Appendix A 109
Palos Heights Kansas Thomaston
Palos Hills Bonner Springs Westbrook
Park Ridge Chanute
Pingree Grove Maryland
De Soto
Plainfield Baltimore
Edwardsville
Prospect Heights Bel Air
Frontenac
Richton Park Bethesda
Gardner
River Grove Brandywine
Girard
Riverdale Catonsville
Gypsum
Riverside Clinton
Kansas City
Robbins Columbia
Lawrence
Rockton Cumberland
Leavenworth
Rolling Meadows Darlington
Leawood
Sauk Village Dayton
Lenexa
Schaumburg Edgewood
Linn Valley
Shirland Elkridge
Merriam
Skokie Elkton
Olathe
South Beloit Ellicott
Overland Park
South Holland Ellicott City
Pittsburg
Stickney Essex
Prairie Village
Streamwood Forest Hill
Salina
Tinley Park Fork
Scammon
Vernon Hills Gaithersburg
Shawnee
Waukegan Glenwood
Weir
Westmont Hampstead
Wheeling Kentucky Havre de Grace
Winnetka London La Plata
Woodridge Laurel
Woodstock Maine Pasadena
Worth Bucksport Perry Hall
Cape Elizabeth Pomfret
Indiana Rockville
Kennebunk
East Chicago Limington Salisbury
Fishers Naples Severn
Hammond Orland Silver Springs
Monticello Orrs Island Stevensville
Westfield Portland Sykesville
Whiting Rockland West Friendship
South Portland Woodbine
Iowa Woodstock
Spruce Head
Fort Dodge Tenants Harbor
110 Appendix A
Massachusetts Mendon Bloomfield Hills
Malden Kalkaska
Michigan Kentwood
Manchester
Ann Arbor Kewadin
Manomet
Auburn Hills Kingsford
Marion
Battle Creek Kingsley
Marlborough
Berkley Lake Orion
Marstons Mills
Birmingham Lambertville
Medway
Appendix A 111
Lathrup Village Troy Maple Lake
Lawrence Twin Lake Mapleton
Lincoln Park Walled Lake Minneapolis
Livonia Warren Minnetonka
Luna Pier Waterford Mounds View
Macomb Wayland New Brighton
Madison Heights Wayne New London
Mancelona West Bloomfield North Mankato
Manton Westland Norwood Young America
Marysville White Lake Oakdale
Melvindale Williamsburg Owatonna
Mesick Wixom Plymouth
Milford Ypsilanti Rochester
New Baltimore Rockford
Northville Minnesota Roseville
Norway Andover Saint Clair
Oak Park Big Lake Saint Cloud
Oakland Blaine Saint Francis
Ottawa Lake Brooklyn Center Saint Paul
Petoskey Buffalo Shakopee
Plainwell Cambridge Shoreview
Plymouth Centerville Spring Lake Park
Pontiac Champlin Stacy
Redford Chaska Stillwater
River Rouge Chisago City Vadnais Heights
Rochester Circle Pines Wayzata
Rochester Hills Coon Rapids Wells
Romulus Delano White Bear Lake
Roseville Duluth Woodbury
Royal Oak Forest Lake Wyoming
Saint Clair Shores Fridley
Sault Sainte Marie Golden Valley Mississippi
Shelby Ham Lake Bay Springs
South Boardman Hopkins Brandon
Southfield Hugo Decatur
Spring Lake Hutchinson Forest
Springfield Lexington Hickory
Sterling Heights Lindstrum Lawrence
Taylor Lino Lakes Little Rock
Temperance Mahtomedi Meridian
Traverse City Maple Grove Newton
112 Appendix A
Pearl Park Hills New Hampshire
Richland Peculiar Allenstown
Union Pleasant Hope Ashland
Raytown Atkinson
Missouri Republic Bedford
Arnold Richmond Chichester
Ballwin Riverview Claremont
Barnhart Rogersville Concord
Battlefield Saint Ann Derry
Belton Saint Charles Dover
Bonne Terre Saint Clair Epsom
Buffalo Saint John Goffstown
Chesterfield Saint Louis Hampstead
Clarksville Saint Peters Henniker
Fair Grove Springfield Hooksett
Ferguson Strafford Hudson
Florissant Troy Laconia
Foley Villa Ridge Londonderry
Garden City Walnut Shade Manchester
Hazelwood Warsaw Merrimack
High Ridge Webster Groves Milford
Hollister Wentzville Nashua
Imperial Willard New Boston
Independence Wright City Pembroke
Joplin
Plaistow
Kansas City Montana
Portsmouth
Labadie Billings
Rye
Lake Saint Louis Livingston
Swanzey
Lebanon Missoula
Lee’s Summit New Jersey
Lowry City Nebraska
Allendale
Manchester Firth
Bloomfield
Maplewood Lincoln
Carlstadt
Marshfield Omaha
Cliffside Park
Nixa Roca
Closter
O’Fallon Seward
Dumont
Olivette Valparaiso
East Orange
Oronogo East Rutherford
Osceola
Nevada
Elmwood Park
Ozark Reno
Englewood
Pacific Fair Lawn
Appendix A 113
Fairview Wenonah Dexter
Fort Lee West New York Dix Hills
Franklin Westville East Jewett
Garfield Westwood East Northport
Hackensack Woodland Park East Rochester
Haledon Wyckoff Eastchester
Hamilton Elmsford
Jersey City New Mexico Fairport
Landing Espanola Farmington
Lincroft Las Cruces Freedom
Linden Los Alamos Gansevoort
Lodi Santa Fe Garden City
Mahwah Garden City South
Midland Park
New York Geneseo
Newark Acra Geneva
North Arlington Albany Glen Park
North Bergen Ardsley Glenmont
Northvale Arkville Glenville
Oradell Ashland Harrison
Palisades Park Astoria Hemlock
Paramus Auburn Hempstead
Paterson Ballston Lake Hensonville
Pequannock Bedford Hills Homer
Ridgefield Park Beaver Falls Honeoye Falls
Ridgewood Bethpage Huntington Station
River Vale Bloomfield Hurley
Riverdale Boiceville Irvington
Rochelle Park Brewster Jamaica
Rutherford Brownville Jewett
Saddle Brook Canandaigua Kenmore
Secaucus Castleton Lake Peekskill
Sewell Castleton on Hudson Lakeville
South Orange Churchville Lancaster
Springfield Clifton Park Lexington
Succasunna Clifton Springs Lima
Teaneck Clyde Liverpool
Titusville Cohoes Lowville
Union Beach Cortland Lyons
Wallington Croghan Macedon
Wanaque Croton on Hudson Malverne
Wayne Dansville Manchester
114 Appendix A
Maplecrest Saugerties Brevard
Marathon Savannah Cary
Marion Scarsdale Catawba
Maspeth Schenectady Chapel Hill
Massapequa Scotia Charlotte
Merrick Seneca Falls Cornelius
Middlesex Shandaken Cullowhee
Miller Place Shortsville Davidson
Mineola Sleepy Hollow Denver
Monroe Smithtown Durham
Mount Vernon Sodus Franklin
Naples Somers Garner
New Hyde Park Sparkill Gastonia
New Rochelle Spencerport Grover
New York Stillwater Hendersonville
Newark Tappan Huntersville
Niagara Falls Truxton Indian Trail
Niskayuna Uniondale Kitty Hawk
North Rose Verplanck Louisburg
North Syracuse Victor Matthews
Nunda Walworth Mint Hill
Nyack Waterloo Mount Holly
Ontario Watertown Otto
Orchard Park Watkins Glen Raleigh
Ossining Webster Roanoke Rapids
Palmyra West Hempstead Salisbury
Pelham West Henrietta Shelby
Penfield Westbury Southern Shores
Phelps White Plains Stanley
Pittsford Whitestone Statesville
Plainfield Williamson Sylva
Port Gibson Windham Vale
Prattsville Wolcott Wake Forest
Rennselaer Woodhaven Waxhaw
Rexford Woodstock Wilmington
Ridgewood Yonkers Youngsville
Rochester Yorktown Heights
Rockville Centre Ohio
Romulus North Carolina Akron
Rotterdam Apex Baltimore
Rye Brook Asheville Bay Village
Appendix A 115
Bedford Beaverton Irwin
Brookfield Bend Johnstown
Canal Winchester Cannon Beach Kingston
Canfield Clackamas Langhorne
Chagrin Falls Cornelius Lititz
Cleveland Eugene Malvern
Columbus Grants Pass Media
Elyria Gresham Milton
Fredericksburg Happy Valley Montgomery
Fremont Lake Oswego Montoursville
Guysville Manzanita Mount Pleasant
Highland Heights Nehalem Nesquehoning
Hudson Oregon City Newton
Ironton Portland North Wales
Jefferson Sandy Oakmont
Kent Seaside Oreland
Kitts Hill Tillamook Palmerton
Lakewood Tolovana Park Perkasie
Lithopolis Troutdale Philadelphia
Logan Warrenton Pittsburgh
Mentor Pittston
Miamisburg Pennsylvania Pottstown
Painesville Allison Park Quakertown
Pedro Bechtelsville Richeyville
Solon Bensalem Scranton
Strongsville Blairsville Shavertown
Tiffin Bradford Shelocta
Toledo Brockway Souderton
Twinsburg Brookville Trumbauersville
Westerville Claysburg Upper Darby
Wickliffe Conshohocken Wallingford
Wooster Dallas Warminster
Everson Warrington
Oklahoma Fort Washington West Chester
Drumright Forty Fort West Pittston
Sallisaw Freeport Wexford
Stillwater Furlong Wilkes-Barre
Gettysburg Williamsport
Oregon Homer City Williamstown
Aloha Hummelstown Wynnewood
Astoria Indiana Yardley
116 Appendix A
Rhode Island Tennessee Soddy Daisy
Appendix A 117
Euless McKinney The Colony
Farmers Branch Mesquite Tyler
Farmersville Mission Watauga
Flint Missouri City Weatherford
Fort Worth Murphy Wichita Falls
Fredericksburg New Braunfels Winona
Frisco New Ulm Wolfe City
Garland North Richland Hills Wylie
Georgetown Odessa
Gonzales Olney Utah
Grand Prairie Pasadena Clarkston
Grapevine Pearland Clearfield
Haltom City Pflugerville Ephraim
Helotes Pharr Layton
Hereford Pinehurst Lehi
Highlands Plano Logan
Holliday Porter Midvale
Houston Princeton Murray
Huffman Richardson Orem
Humble Richland Hills Pleasant Grove
Iowa Park Richmond Provo
Irving Rio Grande City Salt Lake City
Katy Roma Sandy
Keller Rosharon Santaquin
Kilgore Round Rock Saratoga Springs
Killeen Rowlett Spanish Fork
Kingwood Sachse Springville
La Feria San Antonio Taylorsville
Lake Dallas San Juan West Jordan
Lancaster San Saba West Valley City
LaPorte Schertz
Leander Seguin
Vermont
Lewisville Selma Essex Junction
Liberty Hill Spring Highgate
Lindale Spring Branch Lyndonville
Live Oak Stafford Passumpsic
Llano Stephenville Rochester
Longview Sterling City South Burlington
Louisville Sugar Land Swanton
Lucas Sunnyvale White River Junction
McAllen Taylor
118 Appendix A
Virginia Washington Puyallup
Fairfax Duvall
West Virginia
Falls Church Eatonville
Anmoore
Ferrum Edmonds
Bridgeport
Forest Everett
Buckhannon
Fredericksburg Federal Way
Charleston
Glade Hill Ferndale
Clarksburg
Hampton Freeland
Fairmont
Hardy Gig Harbor
Fort Ashby
Lynchburg Kenmore
Franklin
New Castle Kennewick
Grafton
Newport Kent
Harpers Ferry
Newport News Lacey
Hedgesville
Norfolk Lake Tapps
Huntington
North Tazewell Lakewood
Keyser
Oak Hill Langley
Lost Creek
Reston Liberty Lake
Mineral Wells
Roanoke Lynnwood
Monongah
Rocky Mount Milton
Morgantown
Salem Moses Lake
Nutter Fort
Springfield Mountlake Terrace
Ona
Sterling Newcastle
Parkersburg
Troutville Oak Harbor
Ridgeley
Warrenton Okanogan
Shinnston
Winchester Olympia
Springfield
Woodbridge Packwood
West Milford
Port Angeles
Port Hadlock
Port Townsend
Appendix A 119
Wisconsin
Bonduel
Burlington
Cecil
Cedarburg
Cudahy
Delavan
Eagle River
Elkhorn
Fontana
Franklin
Germantown
Glendale
Greenfield
Hales Corners
Hartford
Lake Geneva
Lakewood
Manawa
Marathon
Milwaukee
Muskego
New Berlin
Oak Creek
Oconomowoc
Racine
Saint Francis
Salem
South Milwaukee
Superior
Townsend
Union Grove
Wabeno
Waterford
Wauwatosa
West Allis
Winneconne
120 Appendix A
Appendix B
Batería IV Pruebas
de aprovechamiento
Examiner Training
Checklist
Name of Examiner:__________________________________ Date:___________________________________________
Y N N/O 3. Asks examinee to reread all items on page if response is unclear and then scores only item
in question.
Y N N/O 4. Does not tell examinee any letters or words during test.
Y N N/O 5. Gives reminder to pronounce words smoothly only once during test.
Y N N/O 7. Encourages examinee to try next word after 5 seconds unless examinee is still actively
engaged in trying to pronounce word.
Appendix B 121
Y N N/O 3. Reads all items to examinee.
Y N N/O 4. Provides Response Booklet and pencil at any time if examinee requests it or appears to need
it (e.g., uses finger to write on table or in air).
Y N N/O 8. Scores item incorrect if numeric response is wrong or if examinee provides incorrect label
(required or not).
Prueba 3: Ortografía
Y N N/O 1. Uses Response Booklet and pencil.
Y N N/O 4. Does not penalize for poor handwriting or reversed letters as long as letter does not form
different letter (e.g., reversed b becomes d and would be an error).
Y N N/O 4. Does not insist on silent reading if examinee persists in reading aloud.
Y N N/O 6. Accepts only one-word responses as correct, unless otherwise indicated by scoring key.
Y N N/O 7. Asks examinee to provide one word that goes in blank when he or she reads item aloud and
provides answer in context.
Y N N/O 8. Scores responses correct if they differ in verb tense or number, unless otherwise indicated.
Y N N/O 9. Scores responses incorrect if examinee substitutes different part of speech, unless otherwise
indicated.
122 Appendix B
Y N N/O 10. Tests by complete pages.
Y N N/O 11. Counts all items below basal as correct.
Prueba 5: Cálculo
Y N N/O 1. Uses Response Booklet and pencil.
Y N N/O 3. Discontinues testing and records score of 0 if examinee responds incorrectly to both
sample items.
Y N N/O 3. Administers prescribed block of items and then follows the Continuation Instructions.
Y N N/O 5. Knows the five reminders to use during administration and provides them each just once
during the test at the first occurrence of that error type.
Y N N/O 6. Does not penalize for spelling, punctuation, capitalization, or usage errors unless otherwise
indicated.
Y N N/O 7. Asks examinee to write as neatly as possible if responses are illegible or difficult to read.
Y N N/O 8. Does not penalize for handwriting errors unless words are illegible.
Y N N/O 9. Does not penalize for spelling errors unless the misspelling interferes with understanding
the examinee’s response or forms another real word.
Y N N/O 13. Does not ask examinee to read his or her response to score item.
Y N N/O 14. Enters score for each block administered and enters an X for any block not administered
into online scoring and reporting program.
Appendix B 123
Prueba 7: Análisis de palabras
Y N N/O 1. Uses suggested starting points.
Y N N/O 3. Says most common sound (phoneme) for letters printed within slashes (e.g., /p/), not
letter name.
Y N N/O 4. Reminds examinee to say words smoothly only once during test if examinee pronounces
nonword phoneme by phoneme or syllable by syllable.
Y N N/O 5. Asks examinee to reread all items on page if response is unclear and then scores only item
in question.
Y N N/O 6. Does not tell examinee any letters or words during test.
Y N N/O 6. Marks slash (/) at each point on Test Record where error occurs.
Y N N/O 7. After hesitation of 3 seconds, marks word as incorrect and tells examinee to go on to
next word.
Y N N/O 8. Knows that self-corrections within 3 seconds are not counted as errors.
Y N N/O 9. Scores each sentence as 2 (no errors), 1 (one error), or 0 (two or more errors).
Y N N/O 3. Begins with sample items and practice exercise for all examinees.
Y N N/O 4. Discontinues testing if examinee has 2 or fewer items correct on Practice Exercises C–F and
records score of 0 on Test Record.
124 Appendix B
Y N N/O 6. Records exact starting and stopping times if stopwatch is unavailable.
Y N N/O 7. Records exact finishing time in minutes and seconds on Test Record.
Y N N/O 8. Reminds examinee to read each sentence if he or she appears to be answering items
without reading.
Y N N/O 14. Subtracts Number Incorrect from Number Correct when obtaining estimated AE/GE from
Test Record.
Y N N/O 4. Discontinues testing if examinee has 3 or fewer items correct after 1 minute and records
time of 1 minute and Number Correct (0 to 3) on Test Record.
Y N N/O 7. Records exact finishing time in minutes and seconds on Test Record.
Y N N/O 8. Does not draw attention to mathematical signs or remind examinee to pay attention to signs
during test.
Y N N/O 9. Does not penalize for poorly formed or reversed numbers.
Y N N/O 10. Reminds examinee to proceed across page from left to right, row by row, if he or she starts
skipping around.
Y N N/O 11. Reminds examinee to continue if he or she stops at bottom of first page.
Appendix B 125
Y N N/O 4. Discontinues testing if examinee has score of 0 on Sample Items B–D after error correction
and records score of 0 on Test Record.
Y N N/O 5. Discontinues testing if examinee has 3 or fewer correct after 2 minutes and records time of
2 minutes and Number Correct (0 to 3) on Test Record.
Y N N/O 8. Records exact finishing time in minutes and seconds on Test Record.
Y N N/O 11. Scores as correct all responses that are complete, reasonable sentences using all target words.
Y N N/O 12. Knows target words may not be changed in any way (e.g., verb tense or nouns changed
from singular to plural).
Y N N/O 13. Does not penalize for spelling, punctuation, or capitalization errors.
Y N N/O 14. Does not penalize for poor handwriting or spelling unless response is illegible.
Y N N/O 15. Scores skipped items as incorrect.
Y N N/O 16. Scores responses that omit critical words as incorrect.
Y N N/O 17. Scores responses that omit less meaningful words (e.g., la or este) as correct if all other
criteria are met.
Y N N/O 18. Accepts symbols (e.g., & for y) if all other criteria are met.
Y N N/O 2. Follows Continuation Instructions to determine when to continue testing or when to stop.
Y N N/O 6. Scores element as correct if examinee uses key word (in bold) or close synonym during
retelling.
Y N N/O 7. Does not penalize for mispronunciations resulting from articulation errors, dialect
variations, or regional speech patterns.
Y N N/O 8. Scores response correct if it differs from correct response listed only in possessive case, verb
tense, or number (singular/plural), unless otherwise indicated in scoring key.
Y N N/O 9. Knows that any number that is a key word (in bold) must be recalled exactly.
Y N N/O 10. Scores derivations of names as correct (e.g., Annie for Ann).
126 Appendix B
Prueba 13: Números matrices
Y N N/O 1. Gives examinee worksheet in Response Booklet and pencil when directed.
Y N N/O 5. Allows 30 seconds for Items 1–11 and 1 minute for Items 12–30 before moving to next item.
Appendix B 127
Appendix C
Y N N/O 2. Develops seating arrangement in which examiner can see both sides of Test Book but
examinee can see only examinee pages.
Administration
Y N N/O 3. Keeps Test Record behind Test Book and out of examinee’s view.
Y N N/O 5. Points with left hand while recording responses with right hand (reversed for left-handed
examiner).
Y N N/O 12. Moves to next item after allowing examinee appropriate, but not excessive, amount of time
to respond.
Y N N/O 13. Is familiar with contents of all examiner page boxes containing supplementary instructions.
Y N N/O 15. When testing backward to obtain basal, starts with first item on preceding page and presents
all items on page if stimuli are visible to examinee.
Appendix C 129
Y N N/O 16. Administers all items on page when stimuli are visible to examinee rather than stopping in
middle of page when ceiling is reached.
Scoring
Y N N/O 22. Does not penalize examinee for mispronunciations resulting from articulation, speech, or
dialectical differences.
Y N N/O 23. Uses item-scoring procedures specified in manual (e.g., 1 = correct response, 0 = incorrect
response, and blanks for items not administered).
Y N N/O 27. Uses optional “Qualitative Observation” checklists for Tests 1–11, as appropriate.
Y N N/O 28. Enters all identifying information and scores correctly into online scoring and
reporting program.
Comments:
130 Appendix C
Appendix D
Glossary of Batería IV
Terms in English and
Spanish
Tests in the Cognitive Battery
Test 1: Oral Vocabulary Prueba 1: Vocabulario oral
Test 1A: Oral Vocabulary–Synonyms Prueba 1A: Vocabulario oral – Sinónimos
Test 1B: Oral Vocabulary–Antonyms Prueba 1B: Vocabulario oral – Antónimos
Test 2: Number Series Prueba 2: Series numéricas
Test 3: Verbal Attention Prueba 3: Atención verbal
Test 4: Letter-Pattern Matching Prueba 4: Pareo de letras idénticas
Test 5: Phonological Processing Prueba 5: Procesamiento fonético
Test 5A: Phonological Processing–Word Prueba 5A: Procesamiento fonético – Acceso
Access de palabras
Test 5B: Phonological Processing–Word Prueba 5B: Procesamiento fonético – Fluidez
Fluency de palabras
Test 5C: Phonological Processing– Prueba 5C: Procesamiento fonético –
Substitution Sustitución
Test 6: Story Recall Prueba 6: Rememoración de cuentos
Test 7: Visualization Prueba 7: Visualización
Test 7A: Visualization–Spatial Relations Prueba 7A: Visualización – Relaciones
espaciales
Test 7B: Visualization–Block Rotation Prueba 7B: Visualización – Rotación de
bloques
Test 8: General Information Prueba 8: Información general
Test 8A: General Information–Where Prueba 8A: Información general – Dónde
Test 8B: General Information–What Prueba 8B: Información general – Qué
Test 9: Concept Formation Prueba 9: Formación de conceptos
Test 10: Numbers Reversed Prueba 10: Inversión de números
Test 11: Number-Pattern Matching Prueba 11: Pareo de números idénticos
Test 12: Nonword Repetition Prueba 12: Repetición de palabras sin sentido
Appendix D 131
Test 13: Pair Cancellation Prueba 13: Cancelación de pares
Test 14: Rapid Picture Naming Prueba 14: Rapidez en la identificación de
dibujos
Clusters
General Intellectual Ability (GIA) Habilidad intelectual general (GIA)
Brief Intellectual Ability (BIA) Habilidad intelectual breve (BIA)
Gf-Gc Composite Gf-Gc combinado
Comprehension-Knowledge (Gc) Comprensión-conocimiento (Gc)
Fluid Reasoning (Gf) Razonamiento fluido (Gf)
Short-Term Working Memory (Gwm) Memoria de trabajo a corto plazo (Gwm)
Cognitive Processing Speed (Gs) Velocidad de procesamiento cognitivo (Gs)
Auditory Processing (Ga) Procesamiento auditivo (Ga)
Long-Term Storage and Retrieval (Glr) Almacenamiento y recuperación a largo plazo
(Glr)
Visual Processing (Gv) Procesamiento visual (Gv)
Quantitative Reasoning Razonamiento cuantitativo
Auditory Memory Span Alcance de la memoria auditiva
Number Facility Destreza numérica
Perceptual Speed Rapidez perceptual
Vocabulary Vocabulario
Cognitive Efficiency Eficiencia cognitiva
Reading Lectura
Broad Reading Lectura amplia
Basic Reading Skills Destrezas básicas en lectura
132 Appendix D
Reading Comprehension Comprensión de lectura
Reading Fluency Fluidez en la lectura
Mathematics Matemáticas
Broad Mathematics Matemáticas amplias
Math Calculation Skills Destrezas en cálculos matemáticos
Math Problem Solving Resolución de problemas matemáticos
Written Language Lenguaje escrito
Broad Written Language Lenguaje escrito amplio
Written Expression Expresión escrita
Academic Skills Destrezas académicas
Academic Fluency Fluidez académica
Academic Applications Aplicaciones académicas
Brief Achievement Aprovechamiento breve
Broad Achievement Aprovechamiento amplio
Scholastic Aptitude Aptitud académica
Reading Aptitude Aptitud de lectura
Math Aptitude Aptitud matemática
Writing Aptitude Aptitud de escritura
Broad Oral Language Amplio lenguaje oral
Oral Language Lenguaje oral
Listening Comprehension Comprensión auditiva
Appendix D 133
Incorrect Incorrecto
Query Si responde
Error or No Response Error o falta de respuesta
Test Record Protocolo de pruebas
Identifying Information Datos personales
Test Session Observations Checklist Información del contacto y el uso del idioma
score entry anotación de puntuaciones
Qualitative Observation Observación cualitativa
Scoring Table Tabla de puntuaciones
Response Booklet Folleto de respuestas
Scoring Guide Plantilla de calificación
audio equipment equipo de audio
audio recording grabación de audio
headphones audífonos
speakers altavoces
Scores
raw score Puntaje bruto
W score Puntuación W
W Difference score Diferencia W
age equivalent (AE) Equivalente de edad (AE)
grade equivalent (GE) Equivalente de grado (GE)
developmental zone Zona de desarrollo
instructional zone Zona de instrucción
relative proficiency index (RPI) Índice de proficiencia relativa (RPI)
Comparative Language Index (CLI) Índice comparativo de lenguaje (CLI)
CALP Levels Niveles CALP
normal curve equivalent (NCE) Equivalente de la curva normal (NCE)
percentile rank (PR) Rango percentil (PR)
standard score (SS) Puntuación estándar (SS)
standard error of measurement (SEM) Error estándar de midicion (SEM)
discrepancy score Puntuación de la discrepancia
discrepancy percentile rank Rango percentil de las discrepancias
discrepancy SD score Puntuación de la discrepancia del SD
134 Appendix D
Appendix E
Batería IV Technical
Supplement
The Batería IV Woodcock-Muñoz (Batería IV) is the parallel Spanish version of the Woodcock-
Johnson IV (WJ IV), and both batteries rely upon the same norming sample for the derivation
of norm-referenced scores. Support for the use and interpretation of the Batería IV scores
draws from the large body of validity evidence gathered for the prior versions of the battery
and from evidence presented in this manual and in the Woodcock-Johnson IV Technical Manual
(McGrew, LaForte, & Schrank, 2014). This evidence includes documentation of the goals and
objectives of the Batería IV revision and the procedures used for test development, norming,
and equating. Users should consult the WJ IV Technical Manual for detailed information
about the norming study and the battery’s technical characteristics. This appendix is a
supplement to that manual and includes information specific to the development, calibration,
and equating of the Batería IV forms of the tests.
Translation/Adaptation
All of the Batería IV tests are either translations or adaptations of the parallel tests in the
WJ IV. Tests that are direct translations contain the same items as the WJ IV forms of the
tests; for these tests, only the item directions were translated into Spanish. Batería IV COG
Prueba 2: Series numéricas is an example of a translated test. In this test, the stimulus material
is exactly the same on the WJ IV and the Batería IV; the directions are precisely parallel
but are in different languages. In contrast, some tests could not be translated directly and
needed to be adapted for use with Spanish-speaking individuals. A test is considered an
adaptation when the measured construct is the same in English and Spanish, but the items
were changed or adapted to be appropriate for Spanish-speaking examinees. For example, in
Batería IV APROV Prueba 3: Ortografía, most Batería IV items are different from the WJ IV
items, but the test measures the same broad and narrow abilities using the same procedure.
Table E-1 contains a list of the Batería IV tests and indicates whether each test was translated
or adapted. In general, most of the visual processing, fluid reasoning, processing speed, and
quantitative ability tests were translated, whereas the comprehension-knowledge, auditory,
long-term storage and retrieval, reading, and writing tests required adaptation.
Appendix E 135
Table E-1. Test Name Translated Adapted
Translated and Adapted Pruebas de habilidades cognitivas
Tests of the Batería IV
Prueba 1: Vocabulario oral ■
Prueba 2: Series numéricas ■
Prueba 3: Atención verbal ■
Prueba 4: Pareo de letras idénticas ■
Prueba 5: Procesamiento fonético ■
Prueba 6: Rememoración de cuentos ■
Prueba 7: Visualización ■
Prueba 8: Información general ■
Prueba 9: Formación de conceptos ■
Prueba 10: Inversión de números ■
Prueba 11: Pareo de números idénticos ■
Prueba 12: Repetición de palabras sin sentido ■
Prueba 13: Cancelación de pares ■
Prueba 14: Rapidez en la identificación de dibujos ■1
Pruebas de aprovechamiento
Prueba 1: Identificación de letras y palabras ■
Prueba 2: Problemas aplicados ■
Prueba 3: Ortografía ■
Prueba 4: Comprensión de textos ■
Prueba 5: Cálculo ■
Prueba 6: Expresión de lenguaje escrito ■2
Prueba 7: Análisis de palabras ■
Prueba 8: Lectura oral ■3
Prueba 9: Fluidez en lectura de frases ■
Prueba 10: Fluidez en datos matemáticos ■
Prueba 11: Fluidez en escritura de frases ■
Prueba 12: Rememoración de lectura ■
Prueba 13: Números matrices ■
1 T his test is a direct translation of the WJ IV test with the exception of Item 104, which was changed from a football to a soccer ball in the Batería IV
form.
2 The WJ IV Writing Samples test was replaced with Written Language Expression (Expresión de lenguaje escrito) in Batería IV. The two tests contain
very similar item types; however, Written Language Expression scoring is simpler and does not require the examiner to use a separate scoring
guide.
3 Two items in the English form of this test did not translate well into Spanish; therefore, these items are slightly different in Batería IV.
136 Appendix E
The Batería IV test translation and adaptation work was performed by, or under the direction
and supervision of, Dr. Criselda Alvarado. Some of the tests included in the Batería IV were
translated or adapted during the development of the earlier editions of the Batería; other tests
were brand new in the WJ IV and were translated into Spanish for the first time during the
Batería IV development. An example of one such test is APROV Prueba 8: Lectura oral.
For some adapted tests, Dr. Alvarado and her project team wrote new items to augment the
existing Spanish item pools so that the Batería IV tests would contain new content and would
be relevant for a wide range of Spanish-speaking examinees representing different linguistic
and cultural backgrounds. For instance, for COG Prueba 6: Rememoración de cuentos,
Dr. Alvarado and her team wrote 10 new stories containing a total of 113 new test items.
Calibration Study
Several tests required calibration, either because they were new tests in the Batería IV or
because they contained new items. A calibration study was conducted that included six
Batería IV tests: COG Prueba 5A: Procesamiento fonético – Acceso de palabras, COG Prueba 5C:
Procesamiento fonético – Sustitución, COG Prueba 6: Rememoración de cuentos, COG Prueba
12: Repetición de palabras sin sentido, APROV Prueba 8: Lectura oral, and APROV Prueba 12:
Rememoración de lectura. The primary goals of the study were to determine the difficulty
levels of the new Spanish items and to equate those items to the scales underlying the English
forms of the tests.1
1 Severaladditional adapted tests were not included in the calibration study because adequate item data from prior Spanish calibration studies existed to support
construction of Batería IV test forms. These tests included Prueba 1: Vocabulario oral and Prueba 8: Información general in the COG battery and Prueba 1: Identificación
de letras y palabras, Prueba 2: Problemas aplicados, Prueba 3: Ortografía, Prueba 4: Comprensión de textos, Prueba 6: Expresión de lenguaje escrito, and Prueba 7: Análisis
de palabras in the APROV battery. The extant data from these earlier studies were used to equate the Spanish items for these tests to the scale underlying the English
WJ IV tests, following the procedures described under “Calibration and Equating of Items” below. In addition, the extant item data for Batería III COG Prueba 2:
Aprendizaje visual-auditivo and Prueba 13: Reconocimiento de dibujos were used to equate these two Batería III tests to the scales underlying the WJ IV forms of these
tests so that the tests (and clusters that utilize the tests) can be scored with WJ IV/Batería IV norms.
Appendix E 137
Calibration Study Data Collection
The Batería IV calibration study was conducted between December 2017 and April 2018.
In this study, the six Batería IV tests were administered to a sample of 601 native Spanish-
speaking examinees between the ages of 2 and 81 years.
All calibration study data were collected by trained examiners. At the outset of the study,
Riverside Insights project staff recruited examiners via e-mail targeted to customer databases
and member databases from the National Association of School Psychologists and the
National Latino/Latina Psychological Association. All professional norming study examiners
completed a 2-hour online training course consisting of test-by-test video modules with
embedded practice exercises and a summative quiz. Examiners were required to achieve a
minimum passing score on the quiz to be approved for participation. After completion of
online training, examiners completed one practice test administration. Practice administration
protocols were reviewed by project staff to ensure that examiners were proficient in the
administration and scoring of all norming tests. Riverside Insights project staff provided
feedback to examiners on any issues or concerns that were noted on their practice cases.
After the practice case was approved, examiners could begin recruiting and testing
calibration study participants. Riverside Insights project staff reviewed all submitted protocols
for completion and accuracy of administration procedures (e.g., adherence to basal and
ceiling rules and continuation rules). Riverside Insights project staff continually monitored
the sample acquisition to ensure adherence to the demographic variable distributions of the
sampling plan. All calibration data, including demographic information, item scores, and item
responses, were manually key-entered and verified prior to analysis.
Table E-2 presents the distribution of the calibration sample by age group.
The calibration study examinees were selected from all regions of the United States. The
sample was chosen to ensure a broad representation of sex, parent or examinee education
level, and country of Hispanic origin/nativity. Table E-3 contains the distribution of these
sampling variables in the calibration study.
138 Appendix E
Percentage of
Table E-3. Number in Calibration Study
Distribution of Sampling Sampling Variable Calibration Study Sample
Variables in the Batería IV Sex
Calibration Study
Male 246 40.9
Female 355 59.1
Parent1 or Examinee Education
High School or Less 301 50.1
> High School 300 49.9
Geographic Location
Arizona 3 0.5
California 22 3.7
Connecticut 1 0.2
Florida 5 0.8
Illinois 99 16.5
New Jersey 5 0.8
New York 7 1.2
Tennessee 67 11.2
Texas 375 62.4
Virginia 17 2.8
Hispanic Origin
Cuban 9 1.5
Dominican 5 0.8
Guatemalan 12 2.0
Mexican 421 70.1
Puerto Rican 33 5.5
Salvadoran 17 2.8
Other/Mixed 104 17.3
1 Parent education is reported for examinees who are less than 18 years old.
2 Wolfe
(2004) terms this type of equating the “equating constants” method, while Linacre (2012) refers to it as the “Fahrenheit-Celsius” method. This method differs
from the Rasch common-item-anchor equating design employed in the WJ IV norming (and described in the WJ IV Technical Manual) in that the item difficulty
parameters for each data set are estimated separately, and the difficulty measures from one set of items are then transformed onto the other scale outside of the
estimation process.
Appendix E 139
1. Identify stable common linking items. For each test, the separate Spanish and English
item difficulties for the common items were cross-plotted. Extreme outliers, identified
using a linear regression procedure, revealed some items with very different relative
W-difficulty estimates in Spanish and English. These outlier items were removed from
the common item linking set.
2. Apply the scale transformation equation. For each test, item W-difficulty means
(Ms) and standard deviations (SDs) were computed for the subsets of common
items from the Spanish and English item pools. Spanish item W-difficulty values
were then adjusted to the scale of the English item pools using the following unit
transformation equation:
SD
De′ = ____
e
(D – Ms) + Me , (E.1)
SDs s
where De′ is the item difficulty of any Spanish item transformed onto the English
item difficulty scale, SDe is the standard deviation of the English common-item
difficulties, SDs is the standard deviation of the Spanish common-item difficulties,
Ds is the difficulty of the Spanish item to be transformed, Ms is the mean of the
Spanish common-item difficulties, and Me is the mean of the English common-item
difficulties. Application of this transformation equation placed the Spanish items onto
the scale of the WJ IV English item pools.
140 Appendix E
between two means (Linacre, 2012). Items were flagged if the DIF contrast between males
and females was greater than or equal to 5.82 W points.3 Items were also flagged if significant
(p <.05) Rasch-Welch t-test4 or Mantel-Haenszel DIF5 statistics were reported. Items were
flagged regardless of the direction of the apparent bias. The percentages of flagged items with
both DIF contrast greater than or equal to 5.82 points and significance at the p <.05 level are
reported for the Batería IV calibration study tests in Table E-4.
Reliability
Reliability refers to the precision of a test score. High reliability indicates that an individual’s
measure on a test would be unlikely to change if he or she were retested under similar
conditions. Reliability is a necessary, but not sufficient, condition for validity. Although high
reliability does not necessarily imply that a test score is valid for a specific purpose, reliability
is an important element of the overall validity argument for a test. The reliability coefficient
can be thought of as an index of the precision with which relative standing or position in a
group is measured.
3 A DIF contrast with a W-point difference greater than or equal to 5.82 W points (i.e., 0.64 × 9.1024 W points, which is the value of 1 Rasch logit) corresponds to the
commonly used Educational Testing Service (ETS) “C” classification for moderate to severe DIF (Linacre, 2016; Zieky, 1993).
4 In a test of 20 items, one would expect one item to exhibit significant DIF by chance (p <.05, the Type I error rate). Several authors (e.g., Linacre, 2012; Wolfe et
al., 2006) suggest the use of the Bonferroni correction to adjust for Type I error when performing multiple statistical tests. Because the purpose of this DIF analysis
was exploratory—items exhibiting significant DIF contrast were not rejected outright but rather were flagged for further review—no correction was applied in
these analyses. The numbers of pairwise t tests in the analysis of DIF for each test suggests that some unbiased items were likely flagged; however, this potential
overidentification was deemed acceptable for the purposes of this DIF study.
5 The Mantel-Haenszel procedure is a statistical approach that utilizes a contingency table to test the significance of score differences between a referent and a focal
Appendix E 141
Test Reliabilities
For the six tests that were included in the Batería IV calibration study, reliability coefficients
were calculated using item-level data from the calibration study. For all other tests in the
Batería IV battery, reliability coefficients were calculated using item-level data from the
norming study. For all nontimed, or nonspeeded, tests, internal consistency reliabilities were
calculated using the split-half procedure. Raw scores were computed based on the odd- and
even-numbered items, and correlations were computed between these sets of scores.
The split-half procedure is inappropriate6 for tests containing multiple-point items
(e.g., APROV Prueba 8: Lectura oral). The reliabilities for these tests were calculated using
information provided by the Rasch model.
As described in Chapter 2 of the WJ IV Technical manual, all Batería IV speeded tests (e.g.,
COG Prueba 4: Pareo de letras idénticas, COG Prueba 5B: Procesamiento fonético – Fluidez
de palabras) were calibrated using a rate-based metric. Although this rate-based metric is
useful for calibrating items and rank-ordering examinees, it yields inflated standard errors
for ability measures due to the limited number of possible scores for each time interval. For
this reason, the procedures for calculating Rasch-based reliability coefficients that were used
for the tests with multiple-point items were not appropriate for the speeded tests. Instead, a
test-retest study was conducted during the WJ IV norming for all speeded tests. Examinees in
three separate age groups were administered the norming form of each speeded test, followed
by a second administration of the same form of the test 1 day later. The retest interval in
this study was intentionally short to minimize changes in test scores due to changes in the
examinee’s state or latent trait. Correlations between the first and second administrations
were computed, and a correction was applied for restriction of range in the study samples
(Sackett & Yang, 2000).
For the tests with subtests (COG Prueba 1: Vocabulario oral, COG Prueba 5: Procesamiento
fonético, COG Prueba 7: Visualización, and COG Prueba 8: Información general), test
reliabilities were computed using Mosier’s (1943) formula for reliability of composite scores.
Details of the procedures for computing reliabilities are included in Chapter 4 of the WJ IV
Technical Manual.
All reliability coefficients were corrected for published test length using the Spearman-
Brown correction formula. Table E-5 presents the median reliability coefficients by age group
for the nonspeeded tests included in the Batería IV. Table E-6 presents the results of the
speeded test-retest study for several age groups.
Cluster Reliabilities
Cluster reliabilities were also computed using Mosier’s (1943) formula for composite
reliability. Table E-7 presents the median cluster reliabilities for all Batería IV clusters by age
group.
6 Internalconsistency reliability methods, such as the split-half procedure, assume that the average correlation between items within a test is the same as the average
correlation between items from the hypothetical alternative forms created by splitting the test into two smaller tests (e.g., odd and even items). This assumption
is violated when tests contain items that produce a different range of scores for each item (as in the Batería IV tests with multiple-point item scoring). In this case,
splitting the test in half may produce tests that are no longer equivalent; the items on one half of the test may have a higher maximum possible total score than the
items on the other half.
142 Appendix E
Table E-5. Age
Reliability Coefficients for Test Name 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Batería IV Nonspeeded Tests
Pruebas de habilidades
by Age Group
cognitivas
Prueba 1: Vocabulario 0.89 0.90 0.90 0.89 0.89 0.86 0.86 0.84 0.84 0.89 0.89 0.88 0.88
oral
Prueba 2: Series 0.91 0.91 0.92 0.92 0.90 0.90 0.87 0.87 0.84 0.84 0.91 0.91
numéricas
Prueba 3: Atención 0.88 0.88 0.90 0.90 0.89 0.89 0.82 0.82 0.86 0.86 0.83 0.83 0.85 0.85
verbal
Prueba 5: 0.82 0.82 0.85 0.85 0.85 0.85 0.81 0.81 0.80 0.80 0.83 0.83 0.86 0.86
Procesamiento fonético
Prueba 6: 0.96 0.96 0.96 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95
Rememoración de
cuentos
Prueba 7: Visualización 0.90 0.90 0.90 0.87 0.87 0.83 0.83 0.79 0.79 0.81 0.81 0.84 0.84 0.82 0.82
Prueba 8: Información 0.90 0.90 0.90 0.85 0.85 0.83 0.83 0.84 0.84 0.80 0.80 0.85 0.85 0.87 0.87
general
Prueba 9: Formación de 0.84 0.84 0.84 0.94 0.94 0.94 0.94 0.94 0.94 0.93 0.93 0.93 0.93 0.91 0.91
conceptos
Prueba 10: Inversión de 0.80 0.80 0.83 0.83 0.84 0.84 0.82 0.82 0.84 0.84 0.89 0.89 0.87 0.87
números
Prueba 12: Repetición 0.91 0.91 0.92 0.92 0.93 0.93 0.94 0.94 0.94 0.94 0.91 0.91 0.87 0.87
de palabras sin sentido
Age
Test Name 17 18 19 20–29 30–39 40–49 50–59 60–69 70–79 80+ Median
Pruebas de habilidades
cognitivas
Prueba 1: Vocabulario 0.88 0.88 0.88 0.89 0.90 0.92 0.92 0.92 0.93 0.93 0.89
oral
Prueba 2: Series 0.92 0.92 0.92 0.86 0.88 0.89 0.90 0.90 0.93 0.93 0.91
numéricas
Prueba 3: Atención 0.91 0.91 0.91 0.82 0.86 0.87 0.83 0.83 0.82 0.82 0.86
verbal
Prueba 5: 0.86 0.86 0.86 0.87 0.88 0.91 0.91 0.91 0.90 0.90 0.85
Procesamiento fonético
Prueba 6: 0.95 0.95 0.95 0.96 0.96 0.96 0.96 0.96 0.96 0.96 0.95
Rememoración de
cuentos
Prueba 7: Visualización 0.85 0.85 0.85 0.83 0.85 0.87 0.87 0.87 0.87 0.87 0.85
Prueba 8: Información 0.88 0.88 0.88 0.88 0.92 0.91 0.91 0.91 0.95 0.95 0.88
general
Prueba 9: Formación de 0.92 0.92 0.92 0.91 0.94 0.93 0.95 0.95 0.96 0.96 0.93
conceptos
Prueba 10: Inversión de 0.91 0.91 0.91 0.91 0.93 0.94 0.90 0.90 0.91 0.91 0.88
números
Prueba 12: Repetición
de palabras sin sentido 0.87 0.87 0.87 0.85 0.85 0.90 0.90 0.90 0.90 0.90 0.91
Appendix E 143
Table E-5. (cont.) Age
Reliability Coefficients for Test Name 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Batería IV Nonspeeded Tests
Pruebas de
by Age Group
aprovechamiento
Prueba 1: Identificación 0.97 0.97 0.97 0.98 0.98 0.96 0.96 0.94 0.94 0.92 0.92 0.88 0.88 0.90 0.90
de letras y palabras
Prueba 2: Problemas 0.92 0.92 0.92 0.91 0.91 0.92 0.92 0.89 0.89 0.87 0.87 0.89 0.89 0.90 0.90
aplicados
Prueba 3: Ortografía 0.92 0.92 0.92 0.93 0.93 0.92 0.92 0.92 0.92 0.90 0.90 0.88 0.88 0.89 0.89
Prueba 4: Comprensión 0.88 0.88 0.88 0.98 0.98 0.94 0.94 0.89 0.89 0.81 0.81 0.84 0.84 0.87 0.87
de textos
Prueba 5: Cálculo 0.93 0.93 0.94 0.94 0.91 0.91 0.89 0.89 0.89 0.89 0.93 0.93
Prueba 6: Expresión de 0.99 0.99 0.99 0.99 0.99 0.94 0.94 0.89 0.89 0.79 0.79 0.79 0.79
lenguaje escrito
Prueba 7: Análisis de 0.96 0.96 0.94 0.94 0.93 0.93 0.91 0.91 0.89 0.89 0.87 0.87
palabras
Prueba 8: Lectura oral 0.92 0.92 0.92 0.92 0.91 0.91 0.91 0.91 0.90 0.90 0.89 0.89
Prueba 12: 0.96 0.96 0.97 0.97 0.97 0.97 0.97 0.97 0.95 0.95 0.94 0.94
Rememoración de
lectura
Prueba 13: Números 0.78 0.78 0.91 0.91 0.94 0.94 0.91 0.91 0.94 0.94 0.89 0.89
matrices
Age
Test Name 17 18 19 20–29 30–39 40–49 50–59 60–69 70–79 80+ Median
Pruebas de
aprovechamiento
Prueba 1: Identificación 0.91 0.91 0.91 0.91 0.91 0.93 0.95 0.95 0.94 0.94 0.94
de letras y palabras
Prueba 2: Problemas 0.92 0.92 0.92 0.92 0.92 0.90 0.91 0.91 0.94 0.94 0.91
aplicados
Prueba 3: Ortografía 0.89 0.89 0.89 0.90 0.90 0.90 0.94 0.94 0.93 0.93 0.92
Prueba 4: Comprensión 0.89 0.89 0.89 0.91 0.91 0.91 0.92 0.92 0.93 0.93 0.89
de textos
Prueba 5: Cálculo 0.94 0.94 0.94 0.93 0.93 0.95 0.93 0.93 0.94 0.94 0.93
Prueba 6: Expresión de 0.79 0.79 0.79 0.79 0.79 0.79 0.79 0.79 0.79 0.79 0.79
lenguaje escrito
Prueba 7: Análisis de 0.88 0.88 0.88 0.89 0.87 0.88 0.94 0.94 0.93 0.93 0.91
palabras
Prueba 8: Lectura oral 0.89 0.89 0.89 0.89 0.89 0.90 0.90 0.90 0.90 0.90 0.90
Prueba 12: 0.94 0.94 0.94 0.96 0.96 0.97 0.97 0.97 0.97 0.97 0.97
Rememoración de
lectura
Prueba 13: Números 0.92 0.92 0.92 0.90 0.92 0.91 0.95 0.95 0.93 0.93 0.92
matrices
144 Appendix E
Table E-6. Age
Test-Retest Reliability Test Name 7–11 14–17 26–79
Coefficients From the
Pruebas de habilidades cognitivas
WJ IV/Batería IV Speeded
Test-Retest Study Prueba 4: Pareo de letras idénticas 0.91 0.88 0.91
Prueba 11: Pareo de números idénticos 0.85 0.84 0.88
Prueba 13: Cancelación de pares 0.89 0.89 0.95
Prueba 14: Rapidez en la identificación de dibujos 0.90 0.79 0.90
Pruebas de aprovechamiento
Prueba 9: Fluidez en lectura de frases 0.95 0.93 0.93
Prueba 10: Fluidez en datos matemáticos 0.95 0.97 0.95
Prueba 11: Fluidez en escritura de frases 0.83 0.76 0.88
Appendix E 145
Table E-7. (cont.) Age
Reliability Coefficients for
Cluster Name 17 18 19 20–29 30–39 40–49 50–59 60–69 70–79 80+ Median
Batería IV Clusters by Age
Group Pruebas de habilidades
cognitivas
Habilidad intelectual 0.97 0.97 0.97 0.96 0.97 0.97 0.97 0.97 0.97 0.97 0.97
general
Habilidad intelectual 0.95 0.95 0.95 0.92 0.94 0.95 0.94 0.94 0.95 0.95 0.94
breve
Gf-Gc combinado 0.95 0.95 0.95 0.94 0.96 0.96 0.96 0.96 0.97 0.97 0.95
Comprensión- 0.93 0.93 0.93 0.93 0.95 0.95 0.95 0.95 0.97 0.97 0.93
conocimiento (Gc)
Comprensión- 0.95 0.95 0.95 0.96 0.96 0.96 0.96 0.96 0.97 0.97 0.95
conocimiento – Extendida
Razonamiento fluido 0.95 0.95 0.95 0.92 0.94 0.94 0.95 0.95 0.96 0.96 0.94
(Gf)
Memoria de trabajo a 0.94 0.94 0.94 0.92 0.94 0.95 0.92 0.92 0.92 0.92 0.91
corto plazo (Gwm)
Velocidad de 0.94 0.94 0.94 0.93 0.93 0.94 0.94 0.94 0.94 0.94 0.94
procesamiento cognitivo
(Gs)
Procesamiento auditivo 0.90 0.90 0.90 0.90 0.90 0.93 0.93 0.93 0.93 0.93 0.92
(Ga)
Destreza numérica 0.92 0.92 0.92 0.92 0.93 0.93 0.91 0.91 0.91 0.91 0.90
Rapidez perceptual 0.93 0.93 0.93 0.93 0.93 0.93 0.93 0.93 0.92 0.92 0.93
Vocabulario 0.93 0.93 0.93 0.94 0.94 0.94 0.94 0.94 0.94 0.94 0.93
Eficiencia cognitiva 0.94 0.94 0.94 0.94 0.95 0.95 0.94 0.94 0.94 0.94 0.93
Eficiencia cognitiva – 0.96 0.96 0.96 0.95 0.96 0.96 0.95 0.95 0.95 0.95 0.95
Extendida
Aptitud de lectura 1 0.94 0.94 0.94 0.95 0.95 0.96 0.95 0.95 0.95 0.95 0.93
Aptitud de lectura 2 0.94 0.94 0.94 0.94 0.94 0.95 0.94 0.94 0.94 0.94 0.93
Aptitud matemática 0.94 0.94 0.94 0.94 0.94 0.95 0.95 0.95 0.95 0.95 0.94
Aptitud de escritura 0.92 0.92 0.92 0.94 0.94 0.95 0.94 0.94 0.93 0.93 0.92
146 Appendix E
Table E-7. (cont.) Age
Reliability Coefficients for Cluster Name 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Batería IV Clusters by Age
Pruebas de
Group
aprovechamiento
Lectura 0.96 0.96 0.96 0.99 0.99 0.97 0.97 0.95 0.95 0.93 0.93 0.92 0.92 0.94 0.94
Lectura amplia 0.99 0.99 0.97 0.97 0.97 0.97 0.96 0.96 0.96 0.96 0.96 0.96
Destrezas básicas en 0.98 0.98 0.97 0.97 0.96 0.96 0.95 0.95 0.93 0.93 0.94 0.94
lectura
Comprensión de lectura 0.99 0.99 0.97 0.97 0.94 0.94 0.91 0.91 0.92 0.92 0.92 0.92
Fluidez en la lectura 0.96 0.96 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95
Matemáticas 0.95 0.95 0.96 0.96 0.94 0.94 0.93 0.93 0.94 0.94 0.95 0.95
Matemáticas amplias 0.97 0.97 0.97 0.97 0.97 0.97 0.96 0.96 0.96 0.96 0.97 0.97
Destrezas en cálculos 0.96 0.96 0.97 0.97 0.96 0.96 0.96 0.96 0.95 0.95 0.97 0.97
matemáticos
Resolución de 0.91 0.91 0.94 0.94 0.95 0.95 0.93 0.93 0.95 0.95 0.94 0.94
problemas matemáticos
Lenguaje escrito 0.99 0.99 0.97 0.97 0.95 0.95 0.93 0.93 0.90 0.90 0.91 0.91
Lenguaje escrito amplio 0.99 0.99 0.96 0.96 0.95 0.95 0.94 0.94 0.92 0.92 0.93 0.93
Expresión escrita 0.99 0.99 0.95 0.95 0.92 0.92 0.91 0.91 0.87 0.87 0.87 0.87
Destrezas académicas 0.99 0.99 0.97 0.97 0.95 0.95 0.93 0.93 0.92 0.92 0.94 0.94
Fluidez académica 0.97 0.97 0.97 0.97 0.97 0.97 0.97 0.97 0.97 0.97 0.97 0.97
Aplicaciones académicas 0.99 0.99 0.98 0.98 0.95 0.95 0.93 0.93 0.93 0.93 0.94 0.94
Aprovechamiento breve 0.98 0.98 0.98 0.98 0.98 0.97 0.97 0.96 0.96 0.96 0.96 0.95 0.95 0.96 0.96
Aprovechamiento amplio 0.99 0.99 0.99 0.99 0.99 0.99 0.98 0.98 0.98 0.98 0.98 0.98
Appendix E 147
Table E-7. (cont.) Age
Reliability Coefficients for
Cluster Name 17 18 19 20–29 30–39 40–49 50–59 60–69 70–79 80+ Median
Batería IV Clusters by Age
Group Pruebas de
aprovechamiento
Lectura 0.94 0.94 0.94 0.95 0.95 0.95 0.96 0.96 0.96 0.96 0.95
Lectura amplia 0.96 0.96 0.96 0.96 0.96 0.97 0.97 0.97 0.97 0.97 0.96
Destrezas básicas en 0.94 0.94 0.94 0.94 0.94 0.95 0.97 0.97 0.96 0.96 0.95
lectura
Comprensión de lectura 0.94 0.94 0.94 0.94 0.94 0.94 0.95 0.95 0.96 0.96 0.94
Fluidez en la lectura 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95
Matemáticas 0.96 0.96 0.96 0.96 0.96 0.96 0.96 0.96 0.97 0.97 0.96
Matemáticas amplias 0.97 0.97 0.97 0.97 0.97 0.97 0.97 0.97 0.98 0.98 0.97
Destrezas en cálculos 0.97 0.97 0.97 0.96 0.96 0.97 0.97 0.97 0.97 0.97 0.97
matemáticos
Resolución de 0.95 0.95 0.95 0.95 0.95 0.94 0.96 0.96 0.96 0.96 0.95
problemas matemáticos
Lenguaje escrito 0.90 0.90 0.90 0.91 0.91 0.91 0.93 0.93 0.92 0.92 0.92
Lenguaje escrito amplio 0.93 0.93 0.93 0.93 0.93 0.93 0.95 0.95 0.94 0.94 0.94
Expresión escrita 0.87 0.87 0.87 0.87 0.87 0.87 0.88 0.88 0.88 0.88 0.88
Destrezas académicas 0.94 0.94 0.94 0.95 0.95 0.95 0.96 0.96 0.96 0.96 0.95
Fluidez académica 0.96 0.96 0.96 0.96 0.97 0.97 0.97 0.97 0.97 0.97 0.97
Aplicaciones académicas 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.95 0.96 0.96 0.95
Aprovechamiento breve 0.96 0.96 0.96 0.96 0.96 0.96 0.97 0.97 0.97 0.97 0.96
Aprovechamiento amplio 0.98 0.98 0.98 0.98 0.98 0.99 0.99 0.99 0.99 0.99 0.98
148 Appendix E
800.323.9540
wj-iv.com