You are on page 1of 745
Essentials of Psychological Testing Essentials of Psychological Testing Fifth Edition Lee J. Cronbach Stanford University 1817 HARPER & ROW, PUBLISHERS, New York Grand Rapids, Philadelphia, St. Louis, San Francisco London, Singapore, Sydney, Tokyo Sponsoring Editor: Laura Pearson Project Editor: Thomas R. Farrell Design and Cover Coordinator: Heather A. Ziegler Cover Design: Delgado Design, Inc. Production: Willie Lane ESSENTIALS OF PSYCHOLOGICAL TESTING, Fifth Edition Copyright © 1990 by Harper & Row, Publishers, Inc. SAT and Advanced Placement test questions reprinted from Taking the SAT, 1978, and Advanced Placement Course Description: History, 1988 (respectively), College Entrance Examination Board. Reprinted by permission of Educational Testing Service, the copyright owner. Permission to reprint the SAT and Advanced Placement questions and descriptive material does not constitute review or endorsement by Educational Testing Service or the College Board of this publication as a whole or of any other testing information it may contain. All rights reserved. Printed in the United States of America. No part of this book may be used or reproduced in any manner whatsoever without written permission, except in the case of brief quotations embodied in critical articles and reviews. For information address Harper & Row, Publishers, Inc., 10 East 53d Street, New York, NY 10022. Library of Congress Cataloging-in-Publication Data Cronbach, Lee J. (Lee Joseph), (date)- Essentials of psychological testing / Lee J. Cronbach. — 5th ed. Pp. cm. ISBN 0-06-041418-9 1. Psychological tests. _ I. Title. BF173.C76 1990 150'.28'7—dc20 89-27772 CIP 9 9192 987654321 Contents List of Figures xiii List of Tables xvii Preface xxi Notes to Students xxv PART ONE. BASIC CONCEPTS 1 1. Who Uses Tests? And for What Purposes? 3 Testing and Social Policy 5 Reformers’ Hopes for Tests 5 Critics’ Discontents 8 How the Testing Enterprise Has Evolved 10 Purchasing Tests 16 Sources of Information 16 Who May Obtain Tests? 18 Decisions for Which Tests Are Used 24 Classification 25 Promoting Self-Understanding 27 Program Evaluation 28 Scientific Inquiry 29 vi CONTENTS Varieties of Tests and Test Interpretations 31 What Is a Test? 32 Psychometric and Impressionistic Styles of Testing 34 Classification of Procedures 37 Tests of Maximum Performance 38 Measures of Typical Response 41 Testing in the Computer Age 46 Item Banks 46 Automated Administration 47 Inferences, Short and Long 51 Administering Tests 55 Two Specimen Tests 57 Block Design 57 Mechanical Comprehension 59 Procedure for Test Administration 63 Testing a Group 64 Policies Regarding Guessing 66 Establishing Rapport in Testing an Individual 69 Testing as a Social Relationship 72 Characteristics of the Tester 74 How Testers Distort Results 76 Motivation of the Test Taker 79 ‘Advance Preparation and Coaching 82 Pros and Cons of Coaching 84 Dynamic Assessment 87 Rights of the Test Taker 89 Providing for Physical Handicaps 95 Scores and Score Conversions 97 Obtaining Raw Scores 99 Scoring Procedures 99 Limitations of Raw-Score Scales 102 Domain-Referenced Interpretations and the Alternatives 104 Three Kinds of interpretation 104 Reading Power: Illustration of a Content Scale 106 Setting Standards 107 Conversions of Raw Scores 109 Percentiles 109 Standard Scores 114 The Normal Distribution 119 Comparison of Score Conversions 121 Profiles 123 CONTENTS vii Norms 125 Characteristics Desired in Norms 125 Differentiated Norms 131 Norms by Calibration 134 Unidimensional Scaling 136 Stretching-theScore Scale 137 Applications 140 _-~ Cautions 142 How to Judge Tests: Validation 144 The Need for Critical Scrutiny 145 Sources of Information and Criticism 147 The Testing Standards 148 Methods of Validation Inquiry 150 Introduction to Criterion-oriented Inquiries 152 Introduction to Content-related Inquiries 157 Construct Validation: Explaining Scores and Their Relations 159 Validation Against a Criterion: A Closer Look 160 The Criterion 160 Correlation and Regression Coefficients 161 Validity in Classification 167 Checking Out Test Content 170 Selection of Content 170 Format 174 Statistical Properties of Items 176 More on Construct Validation 178 “Inquiries Contributing to Explanation 180 Strong and Weak Construct Validation 183 Validation as Persuasive Argument 185 How to Judge Tests: Reliability and Other Qualities 190 Errors of Measurement 191 True-Score Theory 192 Generalizability Theory 195 Estimation Methods 197 Simple Correlational Studies 197 Estimation from a Two-Way Score Table 199 Internal-Consistency Formulas 202 Interpreting Findings 206 Considering the Range of Scores 206 Number of Observations: Effect on Accuracy 207 viii CONTENTS Accuracy for Particular Individuals 210 Effect of Measurement Error on Correlations 213 Further Considerations in Choosing Tests 214 Evaluating a Test 217 PART TWO. TESTS OF ABILITY 223 Ze General Ability: Appraisal Methods 225 The Emergence of Mental Testing 227 {Measurement Before Binet ‘227 |Binet Defines Intelligence 1229 Contrasting British Ideas 231 Developments in the United States 232 Appreciating the Diversity of Abilities 236 Score Scales of Mental Tests 240 Illustrative Mental Tests 243 The Wechsler and Its Cousins 243 The Kaufman Batteries 253 The Cognitive Abilities Test and the Stanford-Binet 258 Tests for Admission to Advanced Education 264 Testing in the Early Years 269 The Meanings of General Ability 274 Reliability of Scores and Score Differences 276 Test and Subtest Scores 276 Differences and Gains 278 Divergence of Constructs 280 Findings on Part Scores of Mental Tests 280 Changes with Age 282 Change in Level of Performance 283 Consistency of Individual Differences 284 Ability and Later Accomplishment 288 Three Lines of Inquiry 293 How Unitary Is General Ability? 294 Consistency Across Tasks 294 The Fluid-Crystallized Spectrum 296 How Central Is Speed? 301 Ability to Learn 303 Performance Through the Clinician’s Eyes 305 Information Processing: The View from the Laboratory 307 Two Illustrative Studies 308 Piaget’s Account of Intellectual Development 311 Modeling Bead-Chain Performance 312 Sternberg’s Componential Analysis 316 Marks of Excellence 318 CONTENTS ix Influences on Intellectual Development 320 How Nature and Nurture Combine 321 The Continuity of Development 321 What Kinship Correlations Do Not Mean 324 Physiological Bases for Intellectual Differences 325 Hemispheric Specialization 326 Evidence from PET Scans and Electric Potentials 327 Cultural Influences 328 How Culture Channels Motivation 328 Adapting Testing to the Culture 329 Cultural Shaping of Profiles 331 Cultural Loading of Ability Tests 334 Comparisons of U.S. Ethnic Groups 334 Should Cultural Experience Count? 336 Detecting Questionable Items 337 Intervention to Modify Abilities 340 Testing and Educational Differentiation 347 Decisions About Learners with Intellectual Handicaps 348 Fitting Instruction to the Child 352 Problems in Evaluating Assignment Schemes 356 Achievement Tests in Educational Management 358 The Role of Objectives 359 Standard Tests 362 An Advanced Placement Test 365 Constructed-Response vs. Choice-Response Tests 368 10. Multiple Abilities and Their Role in Counseling 370 Sorting Tests by Factor Analysis 372 How Factor Analysis Is Carried Out 372 Pictures of the Makeup of CogAT 375 Confirmatory Analysis 378 Multilevel Structures 379 Variables and Tests in the Thurstone Tradition 383 Guilford’s “Structure of Intellect” 386 Divergent Thinking 387 Multiscore Aptitude Batteries 390 The Differential Aptitude Tests 390 The Armed Services Vocational Aptitude Battery 393 The General Aptitude Test Battery 394 Limitations on Interpretation of Profiles 399 Possibilities in Microanalysis 404 x CONTENTS 11. Personnel Selection 406 Developing a Selection Procedure 407 Job Analysis 409 Choice of Tests for Tryout 411 Experimental Trial 412 The Criterion 413 Translating Predictions into Decisions 417 Two Examples 421 Selecting Power-Plant Operators 421 ‘An Aptitude Test for Programmers 423 Interpreting Predictive Validities 428 What Level of Validity Is Acceptable? 428 Adjusting Validity Coefficients 432 Validity Generalization 433 Using Multiple Predictors 437 Linear Composites 437 Patterns and Configurations 440 Cross-Validation 443 Classification Decisions 445 Faimess in Selection 446 Some History and Law 446 Psychometric Issues 450 Main Points Reviewed 455 PART THREE. MEASURES OF TYPICAL RESPONSE 459 12. Interest Inventories 461 Functions of Interest Measurement 462 Dimensionalizing Interests 464 Choices in Constructing Inventories 467 Simplicity vs. Complexity 468 Item Form 469 Organizing Items into Scales 470 Score Reporting 473 Predictive Significance 476 Stability 476 Agreement of Inventories with Claimed Interests 478 Relations with Criteria 478 Interest Measures in Counseling 480 Technical Issues 483 Two Types of Criterion Key 483 Changes from Decade to Decade 484 Erasing Sex Differences 485 13. 14 15. CONTENTS — xi General Problems in Studying Personality 487 The Range of Inquiries 488 Alternatives to Self-Report 490 Should Methods Agree? 493 Conceptualizations of Individuality 497 Traits as Response Tendencies in Situations 499 Systems of Traits 503 cal Inferences 506 Interpreting Without Theory: Criterion Keying 510 Stability of Personality 512 Self-Description: Report of Typical Behavior? 513 Limitations of Inventory Responses 515 Coping with Biases and Distortions 519 Ethical Issues 522 Respecting the Dignity of Those Tested 523 How Permissible Is Deception? 524 Limits on Scientific Inquiry 525 Personality Measurement Through Self-Report 528 Specimen Instruments 530 Evaluative Scales 530 Multiscore Descriptions 537 Validity for Evaluating Disorders 549 Screening 549 Classification 552 Forecasting Response to Psychological Intervention 554 Predictions in the General Population 555 Occupational Predictions 555 Educational Predictions 560 Structure of the Individual 564 Human Judges as Test Interpreters 569 The Computer as Test Interpreter 573 Reports at Three Levels 573 Hazards and Safeguards 578 Judgments and Systematic Observations 582 Ratings by Supervisors and Professional Observers 583 Sources of Error 583 Improvement of Ratings 586 Information from Acquaintances and Associates 595 Mothers’ Reports on Children 595 Peer Ratings 601 xii 16. CONTENTS Participants’ Reports on Settings 603 Direct Observation 606 Choice of the Situations 606 The Schedule for Sampling 607 Observer Error and Its Reduction 610 Inferences from Performance 612 Performance Tests 613 Procedures 613 Problems of Design and Validity 620 Two Stylistic Tests 623 The Bender Gestalt 623 The Rorschach 628 Thematic Tests 631 The TAT and Its Meaning 631 Correlates of Responses 633 Integrative Assessment 636 History of Assessment Programs 636 Specialized Tests for Organizational Behavior 639 Validity, Past and Present 641 Toward an Ecological View 646 Appendix A. Selected Publishers and Test Distributors 649 Appendix B. Classified Index of Tests and Inventories 651 References 657 Glossary 701 Name Index 709 Subject Index 719 List of Figures ll 12 13 14 15 16 21 2.2 2.3 2.4 3.1 3.2 3.3 3.4 3.5 3.6 4.1 Selecting appropriate tests: Responsibility of test users 15 Illustrative catalog entries 15 Embedded figures 18 Information requested from individuals seeking to purchase tests 22 Expansion of opportunity through classification 27 Changes in goals of college freshmen over twenty years 29 Two of the Porteus mazes 39 Complex coordination test 40 A self-report of mood and behavior at a sampled moment 44 Micropat system for delivering perceptual and motor tests 50 Block Design materials 58 Mechanical Comprehension items 60 Student anxiety on examination days 80 Maze performance with and without ego involvement 81 Teachers’ impressions from conventional testing and dynamic testing 88 Recommendations for protection of test takers 90 Part of the Ayres scale for scoring handwriting 99 xiii xiv 4.2 4.3 44 4.5 4.6 47 48 49 4.10 5.1 5.2 5.3 54 5.5 5.6 5.7 5.8 5.9 5.10 6.1 6.2 71 7.2 7.3 7A 75 76 dat 78 79 7.10 711 712 LIST OF FIGURES Scoring rules for a Social Reasoning item 100 Determining percentile equivalents 110 Distributions of raw scores and percentile equivalents 112 The normal distribution 120 DAT profile of Robert Finchley 122 Report on four variables from a self-report inventory 124 Score reports used to plan reading instruction 125 Stages in scaling test items 138 Progress of two students through five items of an adaptive test 141 Expectancy charts for aircraft-armorer trainees 153 Scatter diagram for Test 1 and Criterion C 162 Scatter diagrams yielding large and small coefficients 165 Predictive correlations in groups with wide and narrow range 166 Possible findings in an experiment on pupil classification 171 Checklist for reviewing multiple-choice items in editing an achievement test 173 Four items from the Mathematics section of the SAT 176 Characteristic curves for the items in Figure 5.7 177 Considerations supporting validity argument for a selection program 187 Item from a test for fire fighters 188 Test and retest scores on pitch discrimination 211 Distributions for three tests given to the same group 212 Matrix items at three levels of difficulty 232 Alpha scores of soldiers, by rank 234 Alpha scores of soldiers, by level of education 235 Wechsler performance tasks 247 Profiles based on general and differentiated norms 250 Grouping of subtests within KABC scores 254 Item types used in three Simultaneous Processing subtests 256 Selected tasks from CogAT 259 CogaT score report for Patrick Branch 260 Easy and hard items for a Copying subtest 262 Validity with which college marks are predicted 267 Rank order of SAT means of states, with an indicator of self- selection by test takers 268 8.1 8.2 8.3 8.4 8.5 8.6 8.7 8.8 8.9 8.10 8.11 9.1 9.2 9.3 9.4 9.5 9.6 10.1 10.2 10.3 10.4 10.5 10.6 IL 11.2 11.3 114 11.5 11.6 1.7 11.8 11.9 12.1 LIST OF FIGURES xv How overlap of tests reduces the reliability of their difference 279 Trends over time in general ability 283 Prediction of mental-test standings from earlier tests 285 Test standings of three children from age 2to18 286 Educational and occupational histories of 97 boys 290 Spectrum of general abilities 297 Choice format for block design task 308 Two styles of response to a paper-folding item 309 Successive stages in command of bead-chain tasks 313 Actions making up bead-chain performance at the preoperational level 314 Actions making up bead-chain performance at the operational level 315 Learning curves for pups of four breeds 323 Block Design responses accepted by non-Western test takers 330 Findings in an experiment on compensatory education 341 Steps in evaluating a pupil referred for possible assignment to special education 352 Items matched to specific objectives of language instruction 364 Multiple-choice questions on European history 366 Two plots of CogAT factor loadings 376 A two-dimensional mapping of diverse tests 381 Hierarchy of relations among abilities 382 Guilford’s categories for test tasks 386 Items representing DAT subtests 391 Two GATB profiles for a student engineer 399 Predictive validity with three sets of criteria 414 Engineering grades tabulated against a predictor 418 Success in engineering as a function of test score 419 Diagramming task for prospective programmers 426 Benefit from selection 430 Possible nonlinear relations of outcome to test score 441 Expected outcome as a function of two predictors 442 Prediction within sexes from two aptitude measures 450 Change in adequacy of work force with increase in minority hiring 452 Themes in the RIASEC system 464 12.2 12.3 12.4 12.5 12.6 13.1 13.2 13.3 13.4 13.5 14.1 14.2 14.3 14.4 14.5 146 14.7 14.8 14.9 14.10 15.1 15.2 15.3 15.4 16.1 16.2 16.3 16.4 16.5 16.6 LIST OF FIGURES A map of the world of work 466 Direct and norm-referenced profiles 475 Part of the basic Strong profile 477 Part of the Strong occupational profile 477 Comparison of two criterion-keying techniques 484 Data reflecting four personality types found in first-grade girls 496 Sources of variation in student behavior 502 An interpersonal circle 504 Relation of patient characteristics to assaultive behavior: Clinicians’ estimates and actual validity 509 “Self-confidence” scores of students playing two roles 518 Area self-esteem scores of a 16-year-old girl on a percentage scale 532 The same scores on a percentile scale 533 PSI profile of a rebellious adolescent girl 536 JPI profile of a 22-year-old male student 538 The “Clinical syndrome” profile of Case 168252 546 The “Clinical personality pattern” of Case 168252 546 Prediction of military criteria from personality and ability 557 Joint influence of teacher style and student personality 562 Meaning systems of the two Eves 566 Relation of candidates’ traits to hiring recommendations of two interviews 571 Descriptive graphic rating scales 588 An employee-rating scale anchored in behavior 589 A pupil-rating scale anchored in behavior 590 How students and teachers perceive classrooms 605 Standardized arrangement for observing mother and child 615 Stylistic differences in maze performance 616 Bender-Gestalt patterns to be copied 624 David’s Bender reproductions 625 Thematic pictures to measure reactions of parachutists 635 Stages in assessment and in criterion development 642 List of Tables Ll 3.1 41 4.2 43 44 5.1 5.2 5.3 5.4 5.5 5.6 5.7 5.8 5.9 6.1 6.2 Part of a scoring plan for the form in Figure 1.4 23 Tricks of the SAT-taking trade 83 Percentile norms for DAT Mechanical Reasoning 114 Determining standard scores 116 A table of standard-score norms 117 Standard-score scales 118 Experience table for drafting courses 152 Elements in a study of test-criterion relations 154 MCT correlations with criteria 156 Elements in ensuring or investigating adequacy in a content sample 158 Data on ten salespersons 161 Computing the product-moment correlation 163 Rated innovativeness correlated with ability and personality measures 168 Correlations of tests similar in form or in content 175 Kinds of evidence offered to MCT interpreters 180 Analysis for five observations of typing speed 200 Data for illustrative reliability analyses 202 xviii 6.3 6.4 6.5 6.6 6.7 6.8 6.9 6.10 71 72 7.3 7A 8.1 8.2 8.3 9.1 9.2 10.1 10.2 10.3 10.4 10.5 10.6 10.7 10.8 10.9 LIST OF TABLES Calculating a from variance components 203 Calculating a from item variances or means 204 Calculating a split-half coefficient 204 What “error” means under several procedures 205 Effect on error of making additional observations 207 Effect on validity of making additional observations 214 A form for evaluating tests 218 Illustrative summary evaluation 220 Subtests of the Wechsler scales for various ages 244 Wechsler verbal tasks at three levels 246 Mercer's three assessment models 252 Practical consequences of selective admissions 267 Measurement error of WISC-R 276 Independence in adulthood of persons who scored low on mental tests in childhood 289 Correlations among WISC-R subscores 295 Responses of pupils classified by home background 338 Dialogue to orient a laggard youth to an easy matrix item 345 Intercorrelations of four measures for adult workers 373 Factor loadings of four tests 374 Factor analysis of six CogAT subtests 375 Exploratory and confirmatory factor loadings compared 379 Notes on the Kit of Factor-Referenced Cognitive Tests 385 Stability and consistency of convergent and divergent abilities 389 Correlations within the DAT 392 DAT correlations with course marks about 6 months later 392 Job families and their aptitude requirements 397 10.10 Grade-11 scores predicted from Grade-5 tests 402 10.11 Stability of GATB scores 403 IL 11.2 113 114 115 12.1 Validity coefficients for the CPAB 428 Mean validity coefficients in two categories of clerical job 435 Validity data and combining weights in aircrew classification 439 Two strategies for referring employment service applicants to an employer 454 Test bias against minorities: Some questions and answers 456 How Engineer(m) scoring weights were established 472 12.2 13.1 13.2 13.3 14.1 14.2 14.3 14.4 14.5 14.6 14.7 15.1 15.2 15.3 15.4 15.5 16.1 16.2 LIST OF TABLES xix Raw scores and normed scores of college women as predictors of occupational choice 474 Correlations of ratings in a multitrait-multimethod study 495 The “Big Five” personality dimensions 505 Description of a Cattell factor 507 Convergent correlations for JPI scores 537 Phrases used in describing persons high on MMPI scales 4 and 9 542 Reputations and self-descriptions associated with MMPI peaks of college women 544 Success of a cut score in distinguishing mental patients 550 Accuracy with which a recommended cut score identifies patients with severe anxiety 553 Correlations of judged potential of women buyers with an ability measure and 13 self-report scores 556 Criterion statements associated with high DLQ scores 575 Intercorrelations of ratings on power plant operators 585 Characteristics associated with creativity in women mathematicians 594 Hierarchical mapping of problem behavior in children 596 Average maladjustment scores of children, from mothers’ reports 600 Number of children with each combination of IS and IQ scores 600 Physiological response of parachutists to thematic pictures 635 Correlation of criterion ratings with assessments of marketing managers 644 Preface This book is about current and emerging practices, some of them radi- cally new. But it is also a book about a mature discipline, one whose basic principles have stood firm under the pressures of new social de- mands and the emergence of new technology. After 40 years in print this book is, I believe, the oldest psychology textbook that is still being kept up to date. To be sure, a revision of Guil- ford’s Fundamental Statistics for Psychology and Education (1942; sixth edition, with Fruchter, 1977) could recapture the crown any day it comes along. Good company, as is the Anastasi-Foley Differential Psychology (1937; last revised 1958), still found in Books in Print. My aim in teaching, now as it was 40 years ago, has been to commu- nicate general concepts and principles of interpretation and criticism that can apply to most psychological assessment—including the tests and suggested meanings that will appear long after the course is over. Be- yond that, I want to prepare students to evaluate the broad views about human affairs that are continually being derived, logically and illogi- cally, from test data—views about race differences, for example, or a sus- pected decline in educational standards. When I started to teach “Mental Measurements” regularly in 1946, I was distressed by the textbooks available. Earlier there had been re- flective texts on mental testing, notably one by Frank N. Freeman (one of my teachers at Chicago), but these were left behind by the emergence of the Wechsler-Bellevue, wartime experiences in personnel testing and xxii PREFACE in evaluation of neuropsychiatric casualties, and the appearance of many clinical tests (including MMPI) and of multiple-aptitude batteries. Most texts of those days were primarily descriptions of one test after another— number of minutes allowed per section, number and sources of cases in the norm group, encyclopedic facts on reliability and (sometimes) exter- nal correlates. Having stripped such forgettable detail from my text (save as needed to make examples concrete), I titled the book Essentials. I would have more pointedly expressed my aim with the title How to Think About Psychological Tests and Test Findings. That is still my con- cern. Over time, the tone of Essentials has changed somewhat, pursuing in increasing depth such central issues as “test bias” and the limitations of profile interpretation. But, mindful of the student for whom the book is a first introduction, I subordinate subtleties. I expect readers to get the big picture if they choose merely to scan some page that goes deeper into a subtopic than they would wish. Testing was once taught as a technology of assessing individuals and programs. As long ago as 1960 attention began to shift toward policy and ethics. Each successive edition of this book has given more attention to conflicts among the rights, needs, and purposes of tester and tested. The 1984 edition stressed social justice as the proper aim of test use, and this edition extends that discussion. Surely few contemporary policymakers notice books like this, but the policymakers of the future are sitting in our classes, and the courses should prepare them to see many angles in each slogan-driven policy dispute that arises in the future. The big ideas of the 1984 edition remain here, with little change. The 1984 edition took advantage of an intensive study of policy issues by the National Research Council’s Committee on Testing, and of the deliberations of the committee revising the Testing standards. (I was a member of the first committee and a consultant to the second.) The views those reports advanced have now been extended by a study of test user qualifications, by a code of fair testing practice, and by guidelines for computer-based test interpretations. The old issues continue to take new twists, of course. This edition looks at the following controversial matters, among others: the Golden Rule, the Lake Wobegon effect, the Department of Education “wall charts,” and a race-conscious selection rule proposed by the U.S. Em- ployment Service (USES). New principles are not needed to think about the current controversies; rather, these are occasions to apply thinking that matured in the previous decade. Test descriptions in this book are vehicles for large ideas, not them- selves the primary content. It is ironic, therefore, that need to update descriptions is what motivated me to revise the 1984 edition for a 1990 copyright (a 6-year cycle, compared with an average interval of 12 years previously). I did not want to leave instructors with the burden of de- scribing the prominent, but much-changed Stanford-Binet of 1986 (or of PREFACE xxiii conveying specifics of the 1985 Testing standards). After I started revis- ing, much else happened, hence this edition covers much that is new in the testing field. The Test of Mechanical Comprehension now has a computer-adap- tive version (which makes it even more suitable as a basic example for Part One). The Strong Interest Blank has gone through a metamorphosis. The revisers made a fresh start with Strong’s methodology, and for inter- pretative purposes meshed the many new keys with Campbell’s and Holland’s conceptualizations of interests. The USES, pressed to extend minority opportunity and to recognize Jack Hunter’s recommendations on validity generalization, has proposed radically altered rules for trans- lating its aptitude profiles into decisions. My 1984 edition described the Kaufmans’ then-new ability test, but evaluation was superficial because the technical manual became available only while the book was in pro- duction. Now I take into account the Kaufmans’ research and that re- ported by others. The computer is changing most aspects of testing, bringing benefits and costs. Automation of test administration and interpretation is gaining ground, although it tends to reduce interaction between assessor and examinee—with consequent impoverishment of information on both sides. Supplying computer-generated “self-interpreting” reports to cli- ents seems likely to invite misunderstanding, and to encourage overreli- ance on the test as authority. Persons not qualified as test interpreters are writing software for start-up companies in the field; the risks are ob- vious. On the positive side, if the computer is kept in an adjunct role it reduces testing costs, and its reports stimulate thought in the prof sional interpreter. Adaptive testing is paying off, and we are beginning to see new test tasks, resembling video arcade displays, which have ob- vious potential. The computer, then, comes in for extensive discussion in every section of this book. “Dynamic assessment” is developing along many lines. One of sev- eral approaches that link continuing assessment closely with instruction, it challenges psychometric traditions in order to provide more help for individuals in difficulty. Ideas about personality appraisal are changing notably. Twenty years ago, psychologists measuring personality traits were thrown onto the defensive by Mischel’s famous book and other attacks on the trait concept. That controversy has dropped to a murmur. Mischel acknowl- edges that there is appreciable stability and consistency in a person’s style of behavior. And trait theorists now recognize more explicitly that traits do not shape actions without regard to the situation. Clinicians who for a time refused to consider any evidence except for overt “behaviors” now seek information on the client’s feelings and self-concept, as well as on acts, Today’s balanced outlook is producing better theory and better experimentation. Simultaneously, personnel testers are returning enthusiastically to xxiv PREFACE development (and validation) of personality inventories for employees. ‘As always, the published literature is mostly a mishmash of small and noncumulative studies, with isolated positive and negative results. Much more substantial studies, with large samples and careful collection of criteria, are appearing in current technical reports, especially from the Army’s mammoth Project A. The optimism of the latest reports may not persist; but selfreports on personality are coming back into personnel decisions. Movement on other fronts continues. I have added to the previous edition’s treatment of cognitive psychology, validity generalization, cross-validation, item response theory and tailored testing, learning dis- ability, and several other topics. In revising, I of course took the opportunity to clarify here and there, to tighten the organization, and to remove whatever errors my friends and I have caught. I paid attention also to book reviewers. They gener- ally treated the content respectfully, but on one matter of style several reviewers were so forcefully negative that I have bowed to them. In the 1984 edition I eliminated as many formulas as possible. Students often shy off from algebra, and I believe that the concepts they need can be explained without formulas. In deference to the reviews, quite a few formulas are offered this time, though not enough to make this a book on mathematical psychometrics, which was never its mission. Still, I try to write so that even students who are slow readers of algebra will under- stand what the formula accomplishes. ‘As usual, many informed persons have helped me, and I thank them. In particular, I thank the following reviewers: Winifred Arthur, Texas A&M University; Donald Bubenzer, Kansas State University; John Cor- nell, Baylor University; Joseph Gorczynski, Trenton State University; John Poggio, University of Kansas; Peter J. Rowe, College of Charleston; Samuel Seaman, Baylor University; and Allen Shue, California State University, Hayward. I also offer special thanks to those who went be- yond my requests and suggested further material—in particular, Susan Barnes, David Bartram, Wayne Camara, Marvin Dunnette, Kevin More- land, Jerry Murphy, Robert Most, Gary Robertson, Richard Snow, and Alexandra Wigdor. LEE J. CRONBACH Notes to Students A few features designed to assist students (and other users) require com- ment. Superquestions. The five questions at the start of each chapter have several purposes. They give an initial impression of the content to come, pointing out some main concerns of the chapter. They suggest connec- tions to be made during reading across the subdivisions of the chapter. Some readers will want to jot down, as they proceed, points to consider in answering this or that question. And, obviously, a student would be wise to return to the questions after reaching the end of the chapter. The superquestions are appropriate for essay examinations; a reader who can write sensible, meaty paragraphs on one of them must have a grasp of that main theme. I emphasize, however, that the superquestions do not exhaust the significant ideas of any chapter. Orientation. It is a good idea to examine a chapter’s Table of Con- tents before starting to read its text. The titles of the subdivisions seen all together indicate how ideas progress; and the number of pages ina section usually hints at its complexity and importance. That is not always a good basis for distinguishing big ideas from small ones, however. Sometimes a longish section merely presents an example not deserving close study, and sometimes a brief section is vi- tal. Therefore, paragraphs headed Orientation appear in most chapters. These paragraphs suggest what is most worth studying in the chapter, xxv xxvi NOTES TO STUDENTS and what is presented chiefly by way of illustration or as material for future reference. Obviously your interests, or the emphasis in your train- ing program, may cause you to depart from my suggestions. Discussion questions. A second kind of question is scattered through each chapter. Typically, these are invitations to think realisti- cally about problems of test developers and users of tests. Usually, sev- eral principles and values can be brought to bear on a question. Toa question about testing practice or decisions based on tests, the best an- swer will usually be “Yes and no” (or the like), followed by remarks about circumstances that affect the decision, Not many questions have a single “right” answer that applies everywhere. That is the case with questions that arise in practice. Abbreviations. Test titles are long, and some tests are referred to often. Therefore, initials are much used in talking about tests. Abbrevia- tions are listed in Appendix B; but to provide even quicker retrieval a footnote at the start of each chapter lists abbreviations appearing in it. Certain exceptions help to keep the lists short. I do not list an abbrevia- tion introduced in a subsection of the chapter and used only in the two or three next pages, nor do I list common statistical abbreviations (s.d. for standard deviation and r for correlation). Glossary and index. Following the appendices is a glossary that re- views briefly the meaning of most of the technical terms used in the text. These short and simple statements are less formal than a dictionary definition, but they should be easy to hold in mind. The text discussion of a term is usually more complete, and to locate the discussion you should turn to the Subject Index. Often, the discus sion is divided, appearing at two or more places. This reflects the policy by which this book is organized. Most topics are treated in more than one place. That permits a comparatively simple introduction, followed at a later point by greater detail or by a presentation that ties the idea to a new context. Such division of material should help in learning, be- cause it provides for review of ideas presented early. PART One BASIC CONCEPTS Chapter 1 Who Uses Tests? And for What Purposes? 1A. 1B. 1C. 1D. IE. “Testing of abilities has always been intended as an impartial way to perform a political function.”” Explain. Test production is largely centralized nowadays. What are the good and bad consequences of centralization? What information about a test is found in a catalog? Where can more complete information by obtained? An inadequately qualified tester may misuse a test. Whose responsibility is it to make that less likely? Illustrate each of these uses of tests: classification; promoting self- understanding; program evaluation; scientific inquiry. 4 PART ONE BASIC CONCEPTS ‘Assessment of personal characteristics plays a large role in modern life, shaping individuals’ schooling, careers, and treatment by institutions and caregivers. Tests make many kinds of assessment more dependable. They can survey a student body to identify the talented and those who should have special help. They can check out the effectiveness of social programs and generate hints regarding better policies. In guidance and clinical work, tests evaluate each person’s assets and needs. Almost never should a test score by itself determine what will be done by or for the person. In examining an adult, for example, the clini- cal psychologist is likely to consider—alongside the test scores—the past and present family history, the educational and employment history, the medical record, feelings expressed in an interview, and opinions elicited from the person’s acquaintances, as well as the resources for help that the community can offer. Tests have been much criticized because misconceptions and misap- plications have led to unfortunate decisions. Professional leaders have bad ideas as well as good ones. Some of today’s “best” ideas will surely be rejected by our successors. It is this corrective process that sets test- ing apart from other sources of information about persons and institu- tions. Tests are almost unique in being reproducible and explicit. Be- cause the validity of interpretations and decision rules can be checked out, information from tests is more likely to be steadily improved than is appraisal of other kinds. In this book, the concept of “test” encompasses Socratic questioning ofa student, use of an apparatus in which a pigeon obtains a grain of rice by pecking at the brighter of two light spots, and a round-robin chess tournament. The book will concentrate, however, on tasks designed to be presented in the same way to many persons in many places. These “standardized” tests are important in themselves, and they are espe- cially suitable for textbook treatment because research on them accumu- lates. Most of the concepts used to analyze these tests apply also to un- standardized tests. One main topic is ability tests. The book also takes up systematic observation of behavior, questions about beliefs and feelings, and probes that search out motives of which the examinee may be unaware. Test interpretation will receive more attention here than test con- struction, and particular tests will be described primarily as illustrations. The book will discuss the kinds of tests used in schools, treatment cen- ters, industry, and military settings. Use of tests in research will also be considered. How does a certain new drug affect mood and alertness? How does the makeup of a group affect its teamwork? At what age is a child able to “see things through the eyes of another person”? In a developing country, how do new atti- tudes about work styles and social institutions spread? Each such in- quiry requires its own form of test. CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 5 TESTING AND SOCIAL POLICY I shall discuss public policy and public debates, as well as test technique and scientific findings. Month after month, disputes about tests and test results make headlines. You may recall some of the controversies: Does compensatory education work? Do services that coach for the Law School Admission Test give an unfair advantage to applicants who can afford time and money for the coaching? Can a medical school properly put an applicant of minority origin ahead of a nonminority applicant with a higher test score? Although specific disputes fade into history, the ba- sic conflicts recur. Recent debates have echoed others dating back to the beginnings of modern testing (Cronbach, 1975; Haney, 1984). The same misunderstandings about tests and the same value conflicts surface again and again. There are two sides to most of the issues. For example, many critics oppose using tests to judge whether a child is “retarded.” The truth is that some children have profited from being placed in a class for the “educable mentally retarded” (EMR)! and some have not. At times un- sound testing or bad interpretations of tests have assigned children to the retarded category who did not belong there. On the other hand, the majority of children whose failure in schoolwork leads teachers to sus- pect retardation are saved from the EMR class by the positive evidence of the psychological examination (Lambert, in Glaser & Bond, 1981). Public controversy deals in stereotypes, never in subtleties, Al- though it is easy for Luddites to smash up a device, to improve a system requires calm study. So, following the advice of a professional study committee, the same California legislature that voted to outlaw group mental testing by local districts instituted a new, carefully safeguarded, statewide test of mental ability for first-graders, these data being needed as a baseline in evaluations of reading instruction that the legislature desired. On the advice of the same committee, a statewide mental test in Grades 6 and 10 that had no proper function was canceled. Sound policy is not for tests or against tests; what matters is how tests are used, Unfortunately, the public is most often exposed to angry clamor from extremists who either see no good in any test or see no evil. Reformers’ Hopes for Tests “Testing has been politicized in recent years,” some professionals com- plain. They can point to biased treatment of testing issues, even in re- spected newspapers (Herrnstein, 1982). Although the bias of journalists is objectionable, the passionate interest of citizens is not. ‘Testing of abil- ities has always been intended as an impartial way to perforiiva political ‘On abbreviations, see “Notes to Students” (p. xxv). 6 PART ONE BASIC CONCEPTS function—that of determining who gets what. Some positions in society inevitably carry more responsibility and influence than others; ambi- tious individuals seek those posts, and each subcommunity with a spe- cial interest seeks to increase its share. Power and privilege, responsibility and reward are allocated by for- mal and informal processes. Broadly speaking, it is through connections or through competence that individuals move into rewarding roles. In- heritance of a family firm, am uncle’s influence at city hall; the acquain- tances a poor youth makes at a rich kids’ university—these are forms of advancement-through-connections. Coming up with a slogan that cap- tures customers, handling a football well enough to attract professional scouts, listening carefully enough to citizens that one’s next campaign speech addresses their concerns—that is moving ahead through compe- tence. A society can give great scope to connections or little. Shifting that balance is the aim of many a political movement. Those who want to match responsibility to competence often advo- cate testing as a superior way t( distribute op} Proposing re- forms for education in-VirginiasThomias Jefferson said that the state should start everyone off with a few years of education close to home. Then a traveling examiner should choose, each year, “the boy of best genius”; if his parents were poor, he would be sent forward to a regional grammar school at state expense.” After his 2 years there, another test would select who should continue. Ultimately half of these would be subsidized to study at college, the less excellent being left to find posts (very likely as schoolmasters). Since children of wealthier parents could attend these same schools at their own expense, Jefferson was not eras- ing privilege. But his intent was to bring forward a “natural aristocracy” of virtue and talent, to displace a “tinsel aristocracy” founded on wealth and birth (quoted from a letter to John Adams, October 28, 1813). In this spirit, civil service systems introduced after 1830 were intended to place public business in the hands of qualified workers rather than of appli- cants with powerful patrons. (The history of this kind of testing can be traced back to at least 1000 B.c., in China.) Perhaps the high point among meritocratic proposals was one from John Stuart Mill. Although Mill wanted to extend voting rights from property owners to all ranks of British society, he placed a premium on competence. The first necessity, as he saw it, was universal education. For the sake of what we would now call pluralism, Mill wanted numer- °The gender here is Jefferson's. In the remainder of this book, I adopt a convention that will annoy some readers until they come to appreciate its virtues. Such gender-equalizing devices as “he or she” and speaking only in plurals are awkward, and in psychology they can blur meanings. Therefore, I arbitrarily shall use female nouns and pronouns when referring to a tester or investigator (other than an identified male). Except where special emphasis on both sexes is required or where I identify a female by name, I assign males to all other roles—test taker, a person who engages the tester’s services, and so forth. At times I change pronouns within a quotation to conform to the style. CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 7. ous private local schools rather than a state system. State funds would help poor parents pay for tuition, however. Achievement tests should be given annually to make certain that every child was being properly schooled; if the child failed the test, the parent (!) would be fined (Mill, 1859/1977, vol. 20, pp. 302ff.). In elections, every adult male was to have one vote; but, to qualify, he would have to pass a test of reading, writing, and arithmetic. Mill also favored giving extra votes to those whose opin- ions deserved greater weight, so he proposed giving the banker more votes than the tradesman. For equity, he proposed to offer a harder, op- tional test by which the tradesman might prove himself the mental equal of the banker and gain a banker’s number of votes (Mill, 1859/1977, vol. 19, pp. 323ff.). Tests have helped to extend educational opportunity. The presti- gious eastern U.S. colleges had little need for entrance examinations in the days when they drew their applicants mostly from nearby academies. Knowing the curricula and standards of those schools, the colleges could judge the credentials supplied by the academies. When in the 1920s they sought applicants from the entire nation, the colleges found tran- scripts from distant high schools hard to interpret. Tests in school sub- jects, they felt, would be a poor basis for comparirig students when course content differed from school to school. To permit fairer compari- sons, they established the Scholastic Aptitude Test. Ability pretests came into the lower schools for a similar reason. Prior to 1900 the schools were Procrustean. Students kept up with their class or repeated the grade; those with poor records left school early. Once the schools accepted the challenge of providing 12 years of educa- tion to as many youngsters as possible and of keeping children with classmates of their own age, it became necessary to propose a level and pace of instruction to fit each student. Ability grouping, assignment of some children to “special” education, and allocation of high school stu- dents into precollege, commercial, and technical courses were all part of this attempt. Tests were a favored basis for such decisions because they seemed blind to sex, to color, and to family origins. John Gardner's comment (1961, p. 48) recaptures some of the thinking of those times: An acquaintance of mine who recently visited a provincial school in France reported, “The teacher seemed to find it impossible to separate his judg- ment of a pupil’s intelligence from his judgment of the pupil’s cleanliness, good manners, neatness of dress and precision of speech. Needless to say, his students from the upper and upper middle social classes excelled in these qualities.” Before the rise of objective tests American teachers were susceptible—at least in some degree—to the same social distortion of judg- ment. Against this background, modern methods of mental measurement hit the educational system like a fresh breeze. The tests couldn’t see whether the youngster was in rags or in tweeds, and they couldn’t hear the accents of the slum. The tests revealed intellectual gifts at every level of the popu- lation. 2 PART ONE BASIC CONCEPTS Competence is the key to productivity. Tests for workers pay their way by bringing greater skills and greater talent into a firm, and by help- ing to fit the assignment to the employee. Tests in the military reduce costs by shortening training time, reducing failures, and making sure that those who complete training can do the job. Employers, the armed services, schools, and mental hospitals all em- ploy tests to keep institutions running smoothly. But do not be misled by the fact that the largest sales of tests are for such essentially conservative, stabilizing purposes. Tests also serve—indeed, are sometimes of great value to—the reformer. The pioneering tests of Galton, a hundred years ago, advanced the idea that one’s place should depend on competence and not on family position. The same hope animates today’s pioneers who devise educational activities for children who do not prosper under the usual instruction (Scarr, in Glaser & Bond, 1981). Tests call adminis- trators’ attention to schools where children from impoverished homes do less well than similar children elsewhere. When a school’s approach works, tests can provide convincing evidence of the achievement and can encourage other schools to adopt similar practices. A private practice is not “an institution,” but tests pay off there also—by contributing to the client’s insight, by directing the professional’s effort, and by assessing progress. Over and over, evidence from tests has challenged familiar practices and ideas. The very first use of behavioral measures to study character (see p. 618) found that traditional “character-building” programs did not immunize youngsters against temptation to cheat and steal. By this stan- dard, both the Boy Scout organization and the Sunday school failed. A little later, the origins of hostility, distrust, and bigotry were studied. It turned out that many of the bigots were prosperous citizens, respected and chosen for responsibility in social organizations. The tests disclosed inner fears and feelings of inadequacy in many persons whom acquain- tances thought of as secure and trustworthy adults (Adorno et al., 1950). Critics’ Discontents Responsibility and opportunity are not distributed as critics of society would wish. Evaluating the criticisms fits properly into later chapters, but here I note three main complaints. Matching opportunity to competence, some say, is exaggerated into winner-take-all competition. Everyone agrees that persons who are as- signed responsibilities should be competent to carry them. Above the level of adequacy, however, should small differences in apparent com- petence determine who gets the opportunity? Second, it is said that evidence of competence or adjustment col- lected at one age is allowed to determine the person’s remote future. How heavily should one’s past performance determine future prospects? ‘Third, there is discontent that youngsters born into favorable circum- CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 9 stances tend to do better in the educational system and to reach higher- level occupations. They also score higher on ability tests. One humanist (Barzun, 1959, p. 142) explained the connection this way: There is no mystery about it: the child who is familiar with books, ideas, conversation—the ways and means of the intellectual life—before he begins school, indeed, before he begins consciously to think, has a marked advan. tage. He is at home in the House of Intellect just as the stableboy is at home among horses, or the child of actors on the stage. If less educated parents or those from a certain subculture are not equip- ping the child with the competence the school builds on, then matching opportunity to competence may be perpetuating a social hierarchy, And yet: Should evidence of competence be discounted if the performer was given advantages by his family? A society must sort. Specialized roles have to be filled. So individu- als should follow somewhat different paths in order to accumulate train- ing and experience relevant to this or that role. But for which roles should sorting replace open access? At what ages should selection and self-selection into lines of concentration (and levels of intellectual de- mand) take place? What evidence about individuals should enter these routing decisions? And, once the evidence is in hand, what policies should govern the sorting? No policy applying to all opportunities can be found. Even with re- gard to a single specialty, there is bound to be disagreement. Among those shaping policy for admission to pediatrics, for example, some will emphasize scientific competence; others will give priority to skill in de- veloping positive emotional relationships. Each generation has to think afresh about basic principles—about the meaning of “equal opportunity,” for example. Members of the new generation may differ from their elders because fresh knowledge has emerged or priorities have shifted. To understand and evaluate old prac- tices and new proposals, we will have to give some attention to legal decisions about testing practice, to the laws that impose tests on schools and students in the name of “accountability,” and even to moth-eaten arguments about the inheritance of abilities (Glaser & Bond, 1981; Heller et al., 1982). Beyond the three complaints listed a few paragraphs back, there is another type of discontent. Predicting a person’s fate is seen to bea less worthy accomplishment than helping the person to a better fate. Tests sometimes do help in career guidance, for example, or in guiding day- to-day microadjustment of speech therapy. But the use of tests to shape educational or clinical services for a particular person on a particular day is much less advanced than their use for prediction and after-the-fact evaluation (Linn, 1989a, p. 9). A reader might reasonably ask about my biases. I am critical of many practices in testing; on the other hand, I could not write this book if I 10 PART ONE BASIC CONCEPTS opposed testing. Individuals ought to develop understanding of them- selves, and institutions have to reach decisions; tests, I think, can help. So I shall be pointing out how suitable tests can be helpful and warning against questionable practices. One further preliminary remark: The words “bias” and “fairness” score points in debate, but make poor terms for scholarly analysis. “Bias” can refer to a deliberate effort to harm an individual or group, to a com- pletely honest misinterpretation, or to a sound interpretation that the speaker dislikes. The meaning of “fairness” changes from year to year in legal circles, and in reviewing testing practice even the Supreme Court occasionally splits down the middle. I shall avoid the words “fairness” and “bias” where I can, but I take up the pertinent research and value questions in connection with specific uses and interpretations. 1. What do these terms mean to you? a. equality among citizens b. equality of opportunity c. elitism d. merit system 2. Some courts have screened prospective jurors by testing their verbal com- prehension and reasoning. What can be said for and against such a pro- cedure? » . The licensing examination for drivers is applied mechanically without tak- ing facts about the person's background and past opportunities into ac- count. Is it proper to be so “objective” when the applicant's quality of life and earning power are at stake? . It is conceivable that by the year 2040 women, blacks, and other identifi ble groups in the United States will have the same range of education and occu- pational responsibility as any other group and will have wage rates to ‘match. Assume that the sorting processes (not necessarily based on “tests”) are appropriate to the various responsibilities. What objections to that soci- ety would be likely to come from an egalitarian? B HOW THE TESTING ENTERPRISE HAS EVOLVED Testing had informal, even casual beginnings. A psychologist or physi- cian wanted to observe some type of motor, intellectual, or emotional response and set up a situation that gave a good opportunity for observa- tion. As she mentioned her findings to others, they copied the technique in their own clinics and laboratories. Soon there was a small market for equipment (tachistoscopes for studying flash perception, formboards for testing perception and reasoning, etc.). Before and after 1900 a few books were written, each describing one investigator's procedures. The 1920s saw the publication of tests in editions of hundreds of thousands of copies. This bespoke the enthusiasm for testing in schools CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 11 that followed the publicity given to mass testing of recruits in World War I. To psychologists of those days, the great advantage of published tests was the norms they provided. Until norms were available, the clinician’s basis for identifying especially strong and weak performance was lim- ited to her own experience. The school administrator likewise found it advantageous to give a test that other school systems were using; then the local averages could be compared with experience elsewhere (D. Resnick, in Wigdor & Garner, vol. 2, 1982). Independent Test Authorship Test publishing was publishing and nothing more in those days. After a psychologist had prepared a test, copies were printed for general sale, perhaps through a firm selling apparatus to psy- chology laboratories. As the demand for tests grew, some textbook pub- lishers began to handle tests, and some firms specializing in school tests or industrial tests were established. Until about 1945, the typical test was developed by an author or team of authors who completed the test and offered it to the publisher. The publisher assisted in the final stages of research and in editing the test manual, but the author had the main responsibility for designing the test, preparing items, and suggesting how to interpret results. Testing was largely decentralized. Every institution—college, busi- ness, clinic, or whatever—planned and administered its own program. Each school system was free to adopt tests or not and to choose the ones used. Counseling agencies purchased different tests, and sometimes each psychologist within the agency chose the tests she preferred. This decentralization encouraged publication of a great variety of tests. With published tests available, the industrial psychologist no longer thought it necessary to make up new tests for her own factory. Even a great na- tional agency such as the Veterans Administration relies on published tests in its clinical and counseling services. Centralized Test Development Centralization began slowly. Specialists at the state university might develop an “every-pupil” test to be given in the lower grades throughout the state, so that local officials could com- pare the local score range with that of other schools teaching much the same curriculum. In the 1920s a national association of colleges commis- sioned L. L. Thurstone to test linguistic and quantitative aptitudes of high school seniors. The aim was guidance rather than selection; a sec- ondary use was to tell college officials and faculties how student intake varied from college to college. Likewise, the College Board introduced its Scholastic Aptitude Test. To assist employers, the U.S. Employment Service developed entry tests for many occupations. A wave of expansion occurred after 1945—in clinical psychology, in vocational guidance, in selection for schools and jobs, and ultimately in policy-oriented surveys. Centralization proceeded apace. The Educa- 12. PARTONE BASIC CONCEPTS tional Testing Service, formed in 1948, became a facility for developing tests for college admission, for testing of student accomplishment at all levels, and for policy research—for example, research on the effect of “Sesame Street” on children’s development. Commercial publishers of tests merged, and some became subsidiaries of giant corporations (Hol- men & Docter, 1972). Testing programs were more often centralized; nowadays we have some industry-wide personnel testing, for example, and statewide tests for high school graduation.» Development of technically refined tests requires concentrated ef- fort and large resources, especially when a test must be prepared for just one season of use. The Law School Admission Test, for example, must be given several times per year to accommodate applicants, and there must be a new form for each testing. The effort is enormous: item writ- ing; tryout of items and review of the resulting data; assembly, printing, delivery, and administration of the final test; scoring; reporting of scores to those tested and to the schools of their choice; plus follow-up research to profit from the year’s experience. For popular types of tests, the market is large and competitive. The publisher of an established test fights to keep it prominent and to domi- nate the auxiliary market for scoring services; the sponsors of a new test ora competing service must make a major effort to break in. The increas- ingly numerous software packages that permit a user to administer, score, or interpret a test on her own microcomputer have raised competi- tion to a higher power. “Test publishers” evolved into test developers. A developer's staff of professionals identifies a kind of instrument that can serve a client agency or be marketed more broadly, designs and constructs the instru- ment (perhaps with help from teachers, clinicians, or other experts), ¢: ries out research before and after releasing the test, and organize: ing-and-reporting service. Other staff members advertise the test and arrange seminars to train interpreters. Development of software for psy- chologists’ microcomputers has emerged as a separate craft. Centralized testing brings benefits. Qualified staffs produce steady technical improvement. A central organization can enlist qualified sub- ject-matter consultants. A large data-processing facility provides prompt ‘and extensive score reports and permits elaborate, instructive analyses of data. Information accumulating in the central file of a developer pro- vides an increasingly rich base for interpretation and for making next year’s test comparable to earlier editions. Centralization is an enemy of diversity. Some tests become promi- nent while worthy competitors go out of print. Large firms are under- standably reluctant to invest in a test that will have a small market, and a small firm cannot afford to develop an extensive research base. The surest way to find a large market is to bring out a test that resembles the popular ones. An industry with a large investment cannot move far ahead CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 13 of its customers. Only if the next generation of test purchasers appreci- ates the possibilities in uncommon types of tests will someone be enter- prising enough to keep them available. Along with centralization of test production has come an increasing influence of central authorities on test use. Some tests are imposed on schools to obtain information for state or national planners and perhaps to press schools to focus on curriculum areas or types of students that concern the central authority. The National Collegiate Athletic Associ tion, in the late 1980s, has imposed on member colleges a uniform eligi bility rule that factors in scores from the SAT (and similar tests). The tests were not intended for such use and are only tangentially relevant. Charged with enforcing civil rights laws, the Equal Employment Oppor- tunity Commission has filed complaints against many employers, charg- ing bias in selection. As one consequence, fewer employers test nowa- days than did in the 1970s. There is lively debate about whether the tests imposed on schools are elbowing worthwhile topics out of the in- struction, and whether the reliance on less-controlled information in place of tests for hiring is holding back productivity. Computerized Testing The nature of tests and testing is changing rapidly as new possibilities of the computer are developed. It is now possible for the computer to give complicated tests and to score responses as they are made. Moments after the test ends, a report can be printed out for the test taker or for the tester’s file. The computer may supplement the scores with a verbal interpretation that is similar to what a psychologist would write after viewing the same record. The computer's precision makes it a major resource for laboratory studies of memory, perception, and brain functioning, and these measurements are beginning to spill over into clinical examinations. For checking on effects of glaucoma, for example, there is now an automated psychophysical test that maps pe- ripheral vision. With auxiliary equipment, the computer can present lifelike tasks of many kinds. Physicians’ skill in diagnosis, for example, can be tested with simulated patients (Melnick & Clyman, 1988). After reading a brief account of symptoms, the physician being tested requests specific infor- mation such as blood pressure or an X ray. That prerecorded information is displayed, and the physician recommends a next step in treatment. Further information on the patient’s progress and further proposals for action alternate. The events of several days are simulated in 30 minutes or so. At present, this procedure is used as a self-test during medical training, but there are plans to use it in licensing. The Role of Professional Associations There is rather little governmental regulation of test publication and use. A few pieces of federal legislation have been significant—notably those having to do with education of 14 PART ONE BASIC CONCEPTS handicapped children and with equitable employment testing—and there are scattered state laws including those that license psychologists. Court decisions, actions by administrative agencies, and governmental investigations have left some marks on testing practice and I shall come to them. (One example that falls outside the scope of this book is a Labor Department ruling in 1988 that largely prohibits the use of lie detectors by private employers.) Mostly, however, it is professional associations that have studied the adequacy of tests and testing and have distributed guidelines for sound practice. The greatest influence has come from three organization: the American Psychological Association (APA), the American Educational Research Association (AERA), and the National Council on Measure- ment in Education (NCME). Important statements have also come from divisions of APA, from organizations chiefly concerned with client inter- ests, and from others. Associations direct their activities toward educa ing professionals, administrators, and the public. (An official APA state- ment criticizing the inadequate scientific basis for lie detectors encouraged the Labor Department ban.) The Ethical Standards that APA issues from time to time (American Psychological Association, 1981) are a statement for study by members of the profession, persons in training, and employers of psychologists. “Enforcement” is minimal. Gross unethical violations may, to be sure, be grounds for canceling a person’s membership in APA, or for a state’s canceling a practitioner's license. The most influential of the many reports related to testing is Stan- dards for educational and psychological testing (often referred to as the Testing standards or the Standards). APA, AERA, and NCME collabo- rated on the original version in the early 1950s and on three revisions, the latest adopted in 1985 (American Educational Research Association et al., 1985; cited as “Standards 1985” hereafter in this book). This again is an educational document, not an enforceable regulation. Pub- lishers have taken the Testing standards into account in the materials they produce and in their advertising, and courts have cited the Stan- dards as a basis for judging test use (Bersoff, in Glaser & Bond, 1981; Novick, in Wigdor & Gamer, vol. 2, 1982). I shall draw upon the Stan- dards in connection with particular topics as we proceed and shall say more about their function early in Chapter 5. For specimen standards, see p. 90. Figure 1.1 serves both to illustrate such recommendations and to preview the topic of the next section. The Code from which the figure comes is a presentation of ideas from the Standards for the benefit of test takers and the interested public (see also p. 5.8). The message is, “This is the level of practice you should expect from persons and agen- cies that give tests.” (Single copies of the full document are available without charge from NCME, 1230 17th St., Washington, DC 20036.) CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 15, Test users should select tests that meet the purpose for which they are to be used and that are appropriate for the intended test-taking populations, Test Users Should: 1, First define the purpose for testing and the population to be tested. Then, select a test for that Purpose and that population based on a thorough review of the available information. 2. Investigate potentially useful sources of information, in addition to test scores, to corroborate the information provided by tests, w - Read the materials provided by test developers and avoid using tests for which unclear or incomplete information is provided, 4. Become familiar with how and when the test was devel- oped and tried out. on - Read independent evaluations of a test and of possible alternative measures. Look for evidence required to sup- Port the claims of test developers. a - Examine specimen sets, disclosed tests or samples of questions, directions, answer sheets, manuals, and score reports before selecting a test. 7. Ascertain whether the test content and norms group(s) or comparison group(s) are appropriate for the intended test takers. 8. Select and use only those tests for which the skills needed to administer the test and interpret scores cor- rectly are available. Figure 1.1. Selecting appropriate tests: responsibility of test users. (Source: Code of Fair Testing Practices in Education, 1988.) 16 PART ONE BASIC CONCEPTS PURCHASING TESTS Sources of Information Someone who thinks that testing could serve one of her purposes must first survey the tests available. Books such as this cannot list the many tests in each category; even a specialized book on, for example, disorders of speech is unlikely to be exhaustive. Persons choosing tests should consult the Mental Measurements Yearbooks (MMY; BI) and the Test Critiques (TCA) (see p. 147), both of which carry test reviews. Reference listings appear in Tests in Print (BI) and Tests (TCA). One further resource from TCA is Psychware, a listing of software for administering and scoring tests by microcomputer. The Educational Testing Service provides significant services: News on Tests covers recent offerings of publishers; Tests on Microfiche repro- duces specialized tests that are not commercially distributed; the Test Collection maintains a file of tests, manuals, and technical reports (and publishes bibliographies on testing topics). Increasingly, services such as those listed in this paragraph are available in central data bases that can be called up via computer. Those data bases can be kept up to the minute. Catalogs of test publishers (free on request) are another source of up-to-date detail. An exhaustive list of publishers appears in the MMY. Appendix A lists some publishers that persons who do much testing should know about. Also, this appendix identifies the publishers who are referred to in the text by shortened names or initials. To illustrate catalog content I present, in Figure 1.2, one publisher’s listing for the embedded figures task (EFT) shown in Figure 1.3). This task, also called hidden figures, has been prominent in research. Publishers generally offer a “sample set” or “specimen set,” as seen in the entry for the Group EFT. This set includes a copy of the test and the scoring key. The manual (here priced separately) is often part of a specimen set. Professionals using many tests collect specimen sets and supplementary technical manuals. Most universities maintain collec- tions that can be examined by persons with a professional need-to-know. Before deciding to adopt a test for some use, one should study its manual and any relevant research summary. Whereas the catalog de- scription is a few paragraphs long, the manual offers several pages of information on purposes, methods, and limitations. Sometimes a sepa- rate book provides the research summary. That is the case with EFT. A bibliography of 2800 references can also be purchased. The research summary is entitled Cognitive styles (Witkin & Good- enough, 1981). Witkin investigated for many years the idea, suggested by Gestalt psychology, that some people are dominated by any strong frame of reference or pattern in a stimulus field, to such an extent that they have trouble in perceiving elements that cut across the pattern. Re- search ultimately pushed aside Witkin’s suggestion that EFT measures Embedded Figures Test (b) Herman A. Witkin The Embedded Figures Test (EFT), designed originally for research in cognitive functioning and cogni- tive styles, has been used exten- sively in assessment studies relat- ing performance on the EFT to analytic ability in other tasks, social behavior, body concept, preferred defense mechanisms, and problem solving style. The EFT is available in alternate forms. Each requires subjects to find simple geometric figures in each of 12 complex colored de- signs. For individual administra- tion to older children, adolescents, and adults. Time: No limit. 10 to 45 minutes. CODE DESCRIPTION PRICE/UNIT 3009 «EFT Kit $26.50 Card Sets, Stylus. 50 Recording Sheets 3001 Manual $7.50 3004 Card Sets: Forms A&B $18.00 with Simple Figures & Practice items 3061__Stylus $2.50 3024 Recording Sheets 1 pad of 50 $7.00/pad Sofmore pads of 50____—$6.50/pad Manual for the Embedded Figures Tests (b) Herman A. Witkin, et al This 32-Manual contains directions for administration and scoring, plus normative, reliability, and validity data for three tests in the series—the EFT, CEFT, and GEFT. 131 references. 1971 3001 EFT Manual $7.50 Children’s Embedded Figures Test (b) Stephen A. Karp Norma Konstadt The Children’s Embedded Figures Test (CEFT) is an individually administered figures test for children aged 5 to 12. It is used to assess field dependence in devel- opmental Studies of psychological differentiation. The test series con- tains 25 items, preceded by 13 practice items. Time limit. Testing stops after several failures. CODE DESCRIPTION PRICE/UNIT 3309 Test Kit $26.00 simple cardboard forms, 38 color plates, clear plastic envelopes, Star Rubber Stamp, & pad of Recording Sheets ‘3324 Recording Sheets 1 pad of 50 10 of more pads of 50 $7.50/pad $7.00/pad Group Embedded Figures Test (b) Philip K. Oltman Evelyn Raskin Herman A. Witkin A 25-item embedded figures test in a 32-page booklet, the Grow Embedded Figures Test (GEFT) is designed for group administration. Subjects find one of 8 simple figures in the 18 complex designs and mark them. Time limit: No limit. 20 minutes. CODE DESCRIPTION PRICE/UNIT 3100 Sample Set Scoring Key, Test Booklet 3001 Manual $7.50 3105 Scoring Key $1.50 3113 Test Booklets (expendable) 1 pkg of 25: $18.50/pkg 10 or more pkgs of 25 $16.75/pkg Figure 1.2. Illustrative catalog entries. The code (b) indicates that the purchaser must have completed courses in testing or have equivalent training. (Source: Reproduced with minor modifications from the 1989 catalog of Consulting Psy- chologists Press.) 18 PARTONE BASIC CONCEPTS woe ch Le ZZ Figure 1.2. Embedded figures. a “style of field independence” requiring a theory of its own, separate from ability theory. On page 183 I shall say a bit about this issue. EFT is interpreted in this book as one of many measures of general mental ability, but the publisher lists it under “Personality.” Many times in this book it will be pointed out that personality and ability are not truly sepa- rable. Who May Obtain Tests? In the hands of persons with inadequate training, tests can do harm. Un- trained users may administer a test incorrectly. They may rely too much on inaccurate measurements, and they may misunderstand what the test measures and so reach unsound conclusions. Therefore, the user should confine herself to tests that she can handle properly and that are known to be relevant for her purposes. To see the implications of this remark, consider industrial personnel testing. To a manager, it may appear simple to give a printed reasoning test, score it with a punched-out key, tabulate the scores, and hire appli- cants with top scores. A personnel psychologist, however, knows that applicants of middling ability are better bets for some routine jobs than applicants who score very high but are likely to become bored and quit. She knows that a general mental test may not measure the abilities most important in a factory job and that even experts make errors when they try to guess which tests will predict success in a given job. Levels of Qualification Let us pursue this example. Industrial personnel workers in the United States are qualified at various levels: + Diploma in industrial-organizational psychology. A diploma is awarded by the American Board of Professional Psychology to an applicant who possesses (among other qualifications) the training CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? —- 19 and experience required for carrying out all phases of an industrial testing program. Doctoral degree in personnel psychology. A psychologist at this level (who may have been trained in a university department of psychology, education, or business management) should be able to plan a selection procedure and verify its soundness by research. If she has limited experience, she may need to consult a better- qualified person, especially in planning the program. Numerous consulting firms provide assistance in planning. Limited specialized training. Workers who have training in per- sonnel methods equivalent to a master’s degree can carry out spe- cialized functions within a general plan. They can administer com- plicated tests, collect data on the performance of employees, and make some decisions about individuals. A psychologist can train an intelligent assistant to perform such functions under super- vision. Intelligent workers without psychological training. A person with- out psychological training can learn to administer many group tests, take charge of the scoring of objective tests, and apply clearly stated rules for reducing the list of applicants to a list of finalists, Ordinary clerical workers. These workers should be used only for routine scoring under competent supervision and for assisting in test administration. When tests are used in a vocational counseling service, a school test- ing program, or a diagnostic service in a mental hospital, we would ob- serve similarly varied roles are to be played. There is need for some routine handling of tests and test data, for responsible supervision, and for high-level framing of policies and programs. Being a trained psychologist does not automatically qualify one to handle all types of psychological tests. Being expert in personnel selec- tion or in the analysis of poor readers does not necessarily qualify one for other test applications. Being a psychiatrist, social worker, teacher, or school administrator does not by itself imply readiness to use person- ality tests or even to make sound use of standardized achievement tests. Persons other than the tester come into the picture. If an excellent proce- dure is hard to explain to nonprofessionals, the psychologist may be wise to select a simpler test that is less informative. Some tests can be administered and interpreted by responsible per- sons who have no specialized training. Other tests serving the same gen- eral purpose can be used only by well-qualified psychologists. For ex- ample, two tests that might have some value in selecting prospects for “The board also grants diplomas in clinical psychology, in school psychology, in counsel- ing, and in other specialties. 20. PART ONE BASIC CONCEPTS training as junior executives are the Concept Mastery Test and the The- matic Apperception Test (TAT). The former is a difficult test of word knowledge. The directions for administering and scoring are simple enough for a high school graduate to follow. An employer with no psy- chological training can easily understand the results. To administer and interpret the TAT, a person must have graduate training in the psychol- ogy of personality and should have supervised experience with the TAT itself. It is used to investigate the motives of the test taker and his feel- ings about other persons. Serious errors would result if the test were interpreted by anyone save a cautious psychologist who is acquainted with the logic of its interpretation and the misinterpretations that experts on the test have warned against. Final decisions about hiring or promotion or treatment ought not, in general, be left to psychologists. Other members of the organization bring needed perspectives. After discussion in a broad group, a policy can sometimes be reduced to rules whose administration is properly left to the testing team. Test Distribution Misuse of tests is to be discouraged by every feasible means, and restricting a test to qualified users would be desirable. Re- stricting tests to psychologists will not do. That would be a disservice to persons served by, for example, speech therapists and primary teachers. Moreover, a trained psychologist is not a qualified user of all tests in all circumstances; and no credential—not even board certification—is a guarantee against misuse. Because it is impractical for a publisher or other central agency to check on each user’s practices, the ultimate safeguard has to be self- regulation. It is inevitably the purchaser's responsibility to judge which test records she can interpret soundly and where she is out of her depth. A school psychologist may need to make such judgments about the mem- bers of the school staff, and help them recognize and overcome their limitations. Self-regulation is the key idea behind the Testing standards. The Standards call more specifically for the test user to match her own qualifications to those required for proper administration and interpreta- tion of the test and to make a similar check before delegating any testing responsibility to another. Here is part of a statement about who may properly use a test:* The test user, in selecting or interpreting a test, should know the purposes of the testing and the probable consequences. The user should know the procedures-necessary to facilitate effectiveness-and to-reduce bias in test use. Although the test developer and publisher should provide information ‘on the strengths and weaknesses of the test, the ultimate responsibility for appropriate test use lies with the test user. The user should become knowl- “From Standards for educational and psychological testing, p. 41. Copyright © 1985 by the American Psychological Association. Reprinted by permission. CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 21 edgeable about the test and its appropriate uses and also communicate this information, as appropriate, to others. This point of view is elaborated in the Code quoted in Figure 1.1. Most publishers have tried to screen orders to keep tests out of the hands of unsuitable purchasers. Any industry-wide roster of approved purchasers, however, would very likely be regarded as unfair restraint of trade. Indeed, a system of screening purchasers that was in place in the late 1960s was overturned after an investigation by the Federal Trade Commission (Holmen & Docter, 1972). The investigation appar- ently was triggered by complaints of business consultants who wanted to provide a mail-order testing service to help their clients screen job applicants, and whose purchase orders were rejected. (Professionals dis- approve of attempts to offer opinions on persons the analyst has not in- teracted with.) A new system is emerging from the efforts of a “Test User Qualifica- tion Working Group,” a collaboration of test publishers and professional associations (Eyde et al., 1988). The group has conducted research to identify the most troublesome misuses of each major type of test. Promi- nent misuses include inappropriate choice of tests, faulty administration, interpretations that are insensitive to the limitations of scores, and poor communication of test findings. These concerns will reappear at many places in this book. Here it is worthwhile to cite five “principles of effective test use.” {Y Sound, professional use of educational and psychological tests means that “~~ all test users must: 1. maintain the security of testing materials before and after testing. 2. avoid labeling students based on a single test score. 3. adhere strictly to the copyright law and under no circumstances photo- copy or otherwise reproduce answer forms, test books, or manuals, 4. administer and score tests exactly as specified in the manual. 5. release results only to authorized persons and in a form in keeping with accepted principles of test interpretation. (This wording is appended to the form in Figure 1.4. These topics, all included in the Standards, are highlighted in the form because compara- tively inexperienced testers need these basic reminders.) The Working Group tried several types of questionnaires to be filled out by would-be purchasers, and then it produced suggestions that each firm will apply in its own way in deciding which orders to fill. There is no standard qualification form, both to avoid antitrust complaints and because firms distributing different kinds of tests or serving different types of purchasers probably should emphasize different questions. Still, all publishers are likely to collect much the same kind of infor- mation. One firm that produces ability tests, primarily for school psycholo- gists to use, developed the form in Figure 1.4 in the light of the Working a Figure 1.4. Information requested from individuals seekin; Test Purchaser Qualification Form 1. Your purpose for using ests (check as many as apply) [5 covcnional siagnosistomediation [Clinical dagnosisidassifcaon C} Counseling [] Therapy [1] Personnel selecton (Research C Learning disabilities screening [J Other 2. Your level of training (check as many as apply) (Ll Bachelors Degree: Year Insttuton Majr fl of study (masters Degree: Year Institution Major tld of study Oi doctorate: Year Insituson Major fla of study 3. Your professional credentials (check as many as apply) CO ticensed in: Area State Ucense # TD Member of professional organization(s): (AACD, AERA, APA, ASHA, CEC, NASP, etc.) i Formally recognized professional competence (fellow, dplomate, special cortiiats): C1] felow —C) diplomate other cartifiation Organization 4. Your educational background (courses and other study) ‘a, Courses (check each course completed and circle level at which course was completed.) U = Undergraduata, G = Graduate, O = Other (special course you have completed, workshop, in-service training, etc.) (circle) (circle) C. Basic tests and measurements u Gao C. Use of tests in counseling u Gao CZ. Descriptive statistics u Go Career assessment u Go Li Intettigence testing uGo (Ci Neuropsychological assessment u Go Ci speech, hearing, language assessment u Go Bother (ist below) Ci Educational diagnostics uGco u Go (7 Assessment course in major field: u Go uGao uco b. Other Study (check each type of program completed) C1 Practicum in test administration and interpretation Bi Internship (school psychology, counseling, ete.) typex———— '5. Your updating of professional knowledge and skills (check activities below and give an example) attending workshops, seminars, conferences: Gi Reading professional publications (journals, test manuals, etc): 6. Your special competence (List one test which you use regularly that best ilustrates your skil in test administration and interpretation.) g to purchase tests. This Test Purchaser Qualification Form is being used by the American Guidance Service (Robertson, 1988). Several introductory lines ask about the applicant's position, address, and supervisor's name. CHAPTER 1 WHO USES TESTS? AND FOR WHAT PURPOSES? 23. Table 1.1_ PART OF A SCORING PLAN FOR THE FORM IN FIGURE 1.4 Qualification Credits Element 1. Level of training a. Advanced degree (master's, doctorate) in field relevant to individual intellectual assessment 4 b. Bachelor's degree only 0 Element 2. Professional credentials a. State licensing in area relevant to individual intellectual assessment 2 (no state certification) 0 b. Professional membership in organization relevant to individual intellectual assessment 2 Other organizations 0 c. Professional competence Fellow or diplomate in relevant organization 2 Recognition of competence in other organization not relevant to individual intellectual assessment 0 Element 3. Educational background Courses a. Intelligence testing b. Neuropsychological assessment c. Educational diagnosis d. Basic tests and measurements, descriptive statistics e. Use of tests in counseling; career assessment Other study a. Internship in relevant professional field b. Practicum in test use and interpretation c. Relevant supervised experience 3 aNona AO The credits listed are intended to evaluate would-be users of individual intelligence tests (Robertson, 1988). The complete table of credits (which is at the tryout stage) also includes sections that assign up to 6 points for study of and experience with intelligence tests. Group’s findings. The complete form asks the applicant to pledge that she will adhere to the principles quoted above and to the Testing stan- dards Instead of reaching a decision after inspecting the form, the pub- lisher intends to apply a scoring key. The tentative key, partially illus- trated in Table 1.1, must be tried further before the number of credits needed for approval of a purchase order is finally settled. It will be possi- ble to develop one scoring formula and cutting score for comparatively simple tests and other rules for tests that require substantial professional judgment. 5. An employer without psychological training decides to buy personality tests and use them on applicants. What is gained by refusing to sell the tests, in 24 PARTONE BASIC CONCEPTS view of the fact that without them the employer will base judgments en- tirely on superficial impressions gained through an interview? > Psychologists do not favor distributing tests to people who wish to assess their own (or their children’s) aptitudes or personality characteristics. Why? 7. Five “principles of effective test use” are quoted above. Explain how viola- ting principles 1, 2, 4, and 5 could reduce effectiveness of testing, for the present examinee(s) or in the future. (With regard to principle 3: Copyright should be respected, but photocopying interferes with testing only if imper- fect copying alters the test or causes inaccurate scoring.) How might the credits in Table 1.1 be modified for purchasers of a group test of word recognition skills of primary-grade children? 2 © . After 100 or more purchase requests for individual intelligence tests have been collected, it will be possible for the publisher to collect data on the scoring system of Table 1.1. Consider, for example, the question of whether individuals with scores of 18-19 are good risks. Assuming that investigation cannot go beyond telephone calls to the applicant and her professional ac- quaintances, what questions would help to evaluate a proposal to accept or reject, hereafter, would-be purchasers with this score? DECISIONS FOR WHICH TESTS ARE USED ‘A personnel manager decides whom to hire; a teacher decides whether a pupil is ready to take up long division; a physician decides how a patient should be treated. For every such decision made about the person by others, he makes dozens about himself. He decides to try for an A in chemistry or to settle for a B, to go to the community college or to take a job, to seek work as a salesperson or as a taxi driver. Even 10-year-olds are making significant decisions. The child who sees himself as capable of excellence will spend long hours at the workbench or the piano bench; the child who expects little of himself turns to pastimes. Reaching decisions or conclusions or policies is a broader process than “testing.” Even when matters are so routinized that a test is allowed to dictate action (as when a person who gives too many incorrect answers on a driver's test is denied a license), the decision rule reflects much reasoning and judgment. Terms such as “assessment” and “evaluation” properly suggest a combination of information with value judgments that go far beyond testing. Every decision rests on a forecast. A test tells about performance at this moment. That information would not be worth knowing if one could not then predict something about later performance: Consider atest of visual recognition. We flash a row of letters on the screen for an instant, and the person reports what he has seen. Some people recognize four letters; others grasp seven in the same brief inter- val. This difference is intriguing, but it is unimportant until it can be related to other behavior. The applied psychologist sees that this task

You might also like