You are on page 1of 8

Hayward, Stewart, Phillips, Norris, & Lovell 1

Test Review: Brigance Diagnostic Comprehensive Inventory of Basic Skills-Revised (CIBS-R)

Name of Test: Brigance Diagnostic Comprehensive Inventory of Basic Skills-Revised


Author(s): Albert H. Brigance (CIBS-R Standardization and Validation Manual by Frances Page Glascoe)
Publisher/Year: Curriculum Associates, Inc. 1976, 1983, 1999
Forms: Forms A (pre-test) and B (post-test)
Age Range: kindergarten through Grade 6, ages 5 through 12 years
Norming Sample

A. Brigance undertook revisions to items and skill sequences in 1996. The national study was undertaken between 1997 and 1998 to
“define the range of students’ performance across grade, race, and socioeconomic variables, and to generate a range of standard
scores…” (Brigance, 1999, p. 27). Sampling procedures are described and efforts were made to ensure the participation of a range of
students reflecting U.S. demographics from the 1998 U.S. Census Bureau figures. Examiners were classroom teachers at each site
who “assessed their own students in their own classrooms. This ensured that standardization conditions reflected how the CIBS-R is
most often used-by teachers working in classroom settings. Three separate samples of students were collected at distinct points
during the school year” during the fall, winter, and spring” (Brigance, 1999, p. 28).

Total Number: 1, 121


Number and Age: The sample consisted of 1 121 children from ages 5 years (n=183), 6 years (n=102), 7 years (n=153), 8 years
(n=189), 9 years (193), 10 years (n=158), and 11 to 12 years (n=143).
Location: Norming samples were collected at 6 sites in four regions of the United States (North, Central, South, and West).
Demographics: Demographics were reported by four geographic regions, gender, race/ethnicity, residence (urban-suburban/rural),
ethnicity, family income, educational attainment of parents, and disability. The sample characteristics were compared to the U.S.
census information.
Rural/Urban: yes
SES: SES was determined by family income and by “participation of subjects or siblings in Federal School Lunch Program”
(Brigance, 1999, p. 29).
Other (Please Specify): no information

Summary Prepared By (Name and Date): Eleanor Stewart, October and November 2007
Test Description/Overview
Hayward, Stewart, Phillips, Norris, & Lovell 2

The Brigance Diagnostic Comprehensive Inventory of Basic Skills-Revised (CIBS-R) provides tools for assessing students’ skills
from kindergarten through the elementary and middle school years. It also allows teachers to identify children who may need further
diagnostic assessment or may be eligible for special programs. Students’ strengths and weaknesses as well as mastery can be
profiled.

The new edition, CIBS-R, includes norm-referenced scores for certain subtests thus permitting interpretation and comparison with
other measures of achievement using standard scores.

The test kit consists of an examiner’s “Technical and Validation” Manual (by Glascoe, 1999), teacher’s class record book, 40 page
student record, scoring sheets, CD conversion program, and a 3-inch binder compilation of tests across eight areas.

Purpose of Test: The purpose of the test is to assess children’s school-based skills in the early elementary and middle grades. CIBS-
R identifies strengths and weaknesses in an objectives-referenced format and can be used to refer eligible students for special
programs.

Areas Tested: CIBS-R includes a total of 154 assessments across eight areas which include: Readiness, Speech, Listening, Reading,
Spelling, Writing, Research, Study Skills, and Mathematics.

Subtests are grouped as follows to create composites: Basic Reading Composite, Reading Comprehension Composite, Math
Composite, Written Expression Composite, Listening Comprehension Indicator. Information Processing is addressed as Math
Information Processing, Writing Information Processing, and Reading Information Processing.

 Oral Language Vocabulary Grammar Narratives Other (Please Specify)


 Print Knowledge Environmental Print Alphabet Other (Please Specify)
 Phonological Awareness Segmenting, divides words into syllables Blending Elision Rhyming Other
identifies sound in spoken word, states word having same initial/final sounds, substitutes initial consonant sounds
 Reading Single Word Reading/Decoding Comprehension-passage
 Spelling Other (Please Specify)
 Writing Letter Formation Capitalization Punctuation Conventional Structures Word Choice Details
Hayward, Stewart, Phillips, Norris, & Lovell 3

Other (Please Specify) sentence writing,


 Listening Lexical Syntactic Other (Please Specify) sentence
Comment: The binder containing the 154 assessments is extensive. The organization of skills was difficult to follow as some skills
were grouped in unfamiliar ways. For example, Assessment of Speech contained both speech and language skills (e.g., #B-2
Responds to Picture, #B-3 Articulates Initial Sounds of Words). Later, in the section, “Word Analysis”, I found many items that I
would consider to relate to phonological awareness (e.g., #G-3 “Identifies Initial Consonants in Spoken Words”).

Who can Administer: Teachers and assistants (with supervision) may administer this test.

Administration Time: The testing time varies according to the number and selection of tests administered but is estimated to be
approximately one hour however, assessment can be conducted over several sessions/days. The technical manual states, “If
administering all standardized assessments, the Readiness battery requires approximately 1 ¼ hours and the CIBS-R first-grade
through sixth-grade battery takes 45 to 60 minutes” (Brigance, 1999, p. 9). Comment: In my experience, individual administration is
time consuming for classroom teachers, especially in early education settings where all students demonstrate delays. Nonetheless,
some school districts require that the CIBS-R be administered.

Test Administration (General and Subtests):

The examiner can select specific tests to administer in any order. A quick screener is available for grades one through six consisting
of: Comprehends Passages, Sentence Writing, and Computational Skills. The author states, “Each is a strong predictor of overall
success in school, and all tap critical school skills” (Brigance, 1999, p. 2). These screening tests can also be timed so that scores for
“Reading Information Processing, Writing Information Processing, and Math Information Processing” are available. Standardized
administration is urged only when considering using the norm-referenced scores.

General instructions are found in the manual, in Chapter 2, and in the binder preceding each test. Entry points are identified and
usually fall two grades below the student’s current grade. Basals are established with two to five successfully completed items.
Ceilings or discontinuation points are similarly identified in each test. The author states that the basal and ceilings can be used to
determine the instructional range. Specific instructions for each test are outlined in the test binder.

Comment from Buros reviewer: “Overall, the process of determining basal and ceiling levels and recording student responses seems
cumbersome and unnecessarily susceptible to error. In addition to the need to recall and transfer correct scoring marks into the
Hayward, Stewart, Phillips, Norris, & Lovell 4

record book, the examiner must also keep in mind the basal and ceiling demarcations, which vary from assessment to assessment…
No information is provided to validate the use of the basal and ceiling cutoffs” (Cizek & McLellan, 2001, p. 173).

Test Interpretation:

The CIBS-R generates a list of objectives-referenced information, however, this new revision allows for norm-referenced
information to be obtained as well. Instructions for using the standardized scores following a standardized administration are found in
the technical manual (Glascoe, 1999, pp. 14-18). Alternate form use is described. A section defining and interpreting raw scores,
percentiles, quotients, age, and grade equivalents follows. The CD can assist with scoring. A sample educational evaluation is
presented in the final pages of the chapter on administration in the technical manual.

Comment from Buros reviewer: “Two examples of this reporting are included in the manual. One example shows a norm-referenced
report with instructional ranges, percentile ranks, and quotients for several areas. The other is a model of a narrative description of
a student’s performance, developed by the test administrator, that summarizes the skill levels and adds commentary on behavioural
observations. The model narrative is nearly four pages in length, and is thorough in its explanation of results and recommendations
for remediation strategies and materials. On the one hand, it is encouraging to see this type of narrative synthesis included as truly
useful information that parents and teachers receive following administration of a standardized test. On the other hand, there are at
least two causes for serious concern related to this information. First, the behavioral and other information summarized in the report
could only come from the person who actually administered the CIBS-R. Thus, this type of reporting would be precluded in situations
where a teaching assistant conducted the administration, as such a person would ordinarily lack the training necessary to develop
the narrative. Second, and more importantly, the conclusions in the model report seem to go way beyond the data collected by the
assessment….Because this narrative information might likely be more salient to parents and teachers than the quantitative scores, it
is essential that it be based on more careful measures of those behaviors” (Cizek & McLellan, 2001, p. 173).

Standardization: Age equivalent scores Grade equivalent scores Percentiles Standard scores Stanines
Other (Please Specify) Quotients –scale with mean of 100 and SD of 15
A comment from the Buros reviewer regarding the supplemental composite scores addresses information processing in the areas of
math, writing, and reading: “All three composite scores are based on production rates and appear to have been named for user
appeal rather than actual utility. Test users are strongly encouraged to use caution when judging the validity of these three
composites” (Cizek & McLellan, 2001, p. 176).
Hayward, Stewart, Phillips, Norris, & Lovell 5

Reliability:

Internal consistency of items: This information is presented last in the series of reliability sections (Glascoe, 1999, p. 35). Because
the CIBS-R is a criterion-referenced measure, Guttman scalability coefficients (Guttman Lambda Coefficient [R]) were used.
Glascoe states, “the coefficients serve as an indicator that each assessment and its items are hierarchical, unidimensional, and
homogenous measures of academic and readiness skills” (p. 35). The resulting coefficients are lower than those expected with norm-
referenced tests (.80 and above). Nonetheless, Glascoe claims that the coefficients are still sufficient for eliminating test items that
lack discriminating power. She further states that the resulting R values are high (most are at or above .70). SEMs are also presented
in the section on internal consistency though these more properly belong in the section on reliability. SEMs appear to have a wide
range across the different tests (.3 to 8.5).

Comment from Buros reviewer: “Users should also be cautioned about the inaccurate definition of standard errors which it is
claimed, can be ‘added to and subtracted from each student’s raw score in order to provide a theoretical, error-free indicatory of
true performance’( p. 35). Additionally, the suggested interpretations of the lower SEM as the student’s independent skill level and
the upper SEM as a failure level are unsupported” (Cizek & McLellan, 2001, p. 174).

Test-retest: Forty-one (41) students, in kindergarten through grade six, were retested after a two-week interval, with the second
administration being performed by an educational diagnostician. Results are presented in Table 4.1, “Test-Retest Reliability of CIBS-
R First Grade through Sixth Grade Assessments”, and Table 4.2, “Test-Retest Reliability of CIBS-R Readiness Assessments”. All
correlations reported exceed .80 with only Listening Comprehension Indicator at .79, Writing Information Processing at .63, Reading
Information Processing F-2 timed Comprehends Passages at .78, and Prints personal data at .73.

Inter-rater: Data for calculating inter-rater reliability was taken from the same set as those used for test-retest reported above. The
authors state, “Since the 1998 study involved two different examiners (classroom teachers administered the CIBS-R at time one, and
an educational examiner administered the CIBS-R at time two), the previously reported test-retest reliability figures also reflect inter-
rater reliability” (Glascoe, 1999, p. 34). Since the coefficients were high, inter-rater reliability demonstrates that the “directions are
sufficiently clear to enable different examiners to obtain virtually identical results” (p. 34).

Other (Please Specify): Equivalence of forms: The procedure for determining equivalence of forms included analysis of covariance.
The author noted that the two samples used (n=492 children administered Form A and n=446 children for Form B) differed on key
characteristics related to school achievement, the univariate F statistic, and corresponding alpha. No significant difference was found
Hayward, Stewart, Phillips, Norris, & Lovell 6

between the two forms (at p< .05). Therefore, only a single set of norms was necessary to develop (p. 30).

Alternate forms reliability: The alternate forms section is unclear and no reliable interpretation is possible.

Validity:

Content: The author claims that “there is abundant support for the content validity of the CIBS-R and for its applicability in
educational settings” (Glascoe, 1999, p. 39). The author continues with the claim that the test author, A. Brigance, read the literature
in the areas addressed by the CIBS-R and that he consulted with educators throughout the U.S. on item selection. Comment: While
this approach has some merit, it is not sufficient by test development standards and in comparison to other tests I have reviewed.
Indeed, the Buros reviewer noted the same in his review where he stated, “…in an objectives-referenced instrument, content validity
evidence would seem to be of utmost importance. For some CIBS-R tests (called validated assessments), the test binder provides
references to textbook series and grade levels in which the skill is normally encountered; references to scholarly references are also
occasionally provided. For other tests, content validity information is not included. Overall, the level of content validity evidence
provided falls short of ‘abundant’ as claimed in the technical manual” (Cizek & McLellan, 2001, p. 174).

Criterion Prediction Validity:


Correlations are presented in table format, with brief commentary from the author in the text, for the following:
1. PIAT and WRAT in special needs groups.
2. Partial correlations (adjusted for students’ ages) between reading and written language tests: Iowa Test of Basic Skills,
Stanford Achievement Test, and California Achievement Test for grades one through six.
3. Partial correlations (adjusted for students’ ages) between math, general information, and the total battery scores of Iowa Test
of Basic Skills, Stanford Achievement Test, and California Achievement Test for grades one through six.
4. Woodcock-Johnson Psycho-Educational Battery: Tests of Achievement and the CIBS-R Readiness Assessments, restricted
ranges.
5. Teacher ratings and students’ performance on Readiness assessments demonstrating “close agreement” (Glascoe, 1999, p.
49).
6. WISC-III, Woodcock-Johnson Psycho-Educational Battery: Tests of Achievement, Weschler Individual Achievement Test,
DAB-2, Kaufman Test of Educational Achievement, or Peabody Individual Achievement Test were administered by
psychologists and educational diagnosticians responsible for testing children referred for special education services.
Hayward, Stewart, Phillips, Norris, & Lovell 7

Construct Identification Validity:


Age differentiation is demonstrated by increasing median raw scores and standard deviations for students in grades one through six
on the CIBS-R Assessments and Composites. Median raw scores for Readiness for children in kindergarten show progression from
early, midyear, and yearend. Numbers and sample characteristics are not provided nor is there information about subgroups of
interest. In terms of test and composite correlations, the author provides commentary on the information provided in Tables 5-3 to
5.6. These tables present correlations as follows: Intercorrelations Among Readiness Assessments and Composites, Assessments and
Composites for Grades One through Six, Readiness Assessments with Age and Grade, and CIBS-R Grades One through Six
Assessments with Age and Grade. Again, the specifics of the sample are not stated here. Comment: Overall, the author’s discussion
of this data is scant, leaving the reader to interpret. I found the comments by the Buros reviewer illuminating. The Buros reviewer
states, “Some construct validity evidence is also provided, although it does not always confirm the intended use of the test. A table of
CIBS-R intercorrelations is provided; these all are fairly high and positive. However, scores on the listening vocabulary
comprehension grade placement test correlated equally well with scores on reading, math, and written expression composites;
spelling grade-placement test scores correlate more strongly with the math composite than with reading comprehension. Other
correlations reveal that scores on the CIBS-R are related to age and grade level, with higher scores associated with advanced age
and grade; scores on the CIBS also correlate moderately to fairly strongly with the Weschler Intelligence Scale for Children-Third
Edition (WISC-III) full scale IQ scores. Base on the preponderance of the construct validity evidence provided, a reasonable case
could be made that the CIBS-R functions nearly as effectively as a measure of general cognitive ability as it does a measure of
discrete skill mastery” (Cizek & McLellan, 2001, p. 174).

Differential Item Functioning: not provided


Other (Please Specify): none

Summary/Conclusions/Observations:
The sheer size of the binder is daunting. There is a lot of text material to wade through in order to be ready to administer the
assessments. The examiner must use the specific information outlined in each assessment section especially when considering
standardized administration.

Comments from Buros reviewers:


“Until further developmental work remedies some of these concerns, the CIBS-R is probably best used as an adjunct information-
gathering tool. It is possible that the CIBS-R attempts to do too much by attempting to satisfy the needs of those interested in both
objectives- and norm-referenced assessment” (Cizek & McLellan, 2001, p. 175).
Hayward, Stewart, Phillips, Norris, & Lovell 8

“Continued use of the criterion-referenced application of this measure is strongly recommended, but the standardized version should
be used with caution when determining eligibility of students for special education services (Cizek & McLellan, 2001, p. 176).

Clinical/Diagnostic Usefulness:
Based on what I learned from reviewing the CIBS-R, I would be inclined not to recommend it. However, I think that it deserves our
attention because it is so widely used. I suspect that educational decisions might also be structured around the results. Perhaps in
this way, the CIBS-R may be profoundly influential. I encountered use of the test in early education where teachers were required to
complete certain subtests. Given that the children in the program were already identified using stronger diagnostic tests, I felt that
administration of the CIBS-R was wasting the teacher’s valuable time and not providing information that was particularly useful in
that context, however; district requirements prevailed.

References

Brigance, A. H. (1999). Brigance diagnostic comprehensive inventory of basic skills – revised (CIBS-R). North Billerica, MA:
Curriculum Associates.

Cizek, G. J. & McLellan, M. J. (2001). Review of the Brigance Diagnostic Comprehensive Inventory of Basic Skills - Revised. In B.
S. Plake & J. C. Impara (Eds.), The fourteenth mental measurements yearbook (pp. 172-175). Lincoln, NE: Buros Institute of
Mental Measurements.

Glascoe, F. P. (1999). Standardization and validation manual for the Brigance Comprehensive Inventory of Basic Skills-Revised.
North Billerica, MA: Curriculum Associates.

To cite this document:

Hayward, D. V., Stewart, G. E., Phillips, L. M., Norris, S. P. , & Lovell, M. A. (2008). Test review: Brigance diagnostic
comprehensive inventory of basic skills-revised (CIBS-R). Language, Phonological Awareness, and Reading Test Directory
(pp. 1-8). Edmonton, AB: Canadian Centre for Research on Literacy. Retrieved [insert date] from
http://www.uofaweb.ualberta.ca/elementaryed/ccrl.cfm.

You might also like