You are on page 1of 7

Frontier of Higher Education | Volume 1 | Issue 1 | September 2019

The Reliability and Validity of Language Proficiency Assessments for


English Language Learners
Xiaoyan Jing
Shandong Shengli Vocational College, Dongying, Shandong, 257097, China

Abstract: English language learners (ELLs) have become one of the fastest growing groups in elementary schools in the United States
of America. This paper will research the English proficiency assessments currently used for ELLs in the early-exit TBE programs in el-
ementary schools and the validity and reliability of these English language proficiency assessments.
Keywords: reliability; validity; ELLs
DOI: 10.36012/fhe.v1i1.893

1. Introduction proficiency annually in these four language domains


(listening, reading, writing, and speaking) to measure
nglish language lear ners (ELLs) have be-

E come one of the fastest growing groups in


elementary schools in the United States of
America. According to the National Center for Educa-
their progress; they are also required to report on "a
composite comprehension measure" (Kenyon et al.,
2011, p. 383). The test results will be used to establish
"Annual Measurable Achievement Outcomes (A-
tion Statistics in 2010, English language learners
MAOs), which are reported by states to the federal
(ELLs) are increasing in the public schools. It is esti-
government" (Kenyon et al., 2011, p. 384). No Child
mated that in 2011 to 2012, 4.4 million students in the
Left Behind also required that it is mandatory for Eng-
U.S., which accounted for 9.1% of the school-aged
lish language learners to participate in large-scale as-
population, were classified as ELLs while the number
sessments.
of ELLs varies in states, ranging from 0.7% (West Vir-
There are researches focused on the large-scale
ginia) to 23.2% (California) (Carroll & Bailey, 2016).
assessments with the main stream students. However,
The Hispanic ELLs account for 90% of all the ELLs in
it is worth examining these proficiency assessments
the United States (Umansky & Reardon, 2014). Al-
the ELLs take and the validity, reliability and effec-
bers, Kenyon, and Boal (2009) stated that ELLs aca-
tiveness of these assessments as they play a crucial
demically lag behind the non-ELLs, and have 野higher
role in measuring the English proficiency of ELLs and
rates of negative occupational, economic, social, and
classifying the ELLs into different levels and deciding
physical outcomes冶 (p. 76).
whether the ELLs can exit the programs or not.
According to the mandatory federal requirements Gottlieb and Nguyen (2007) listed fives purposes
of No Child Left Behind (NCLB) in 2001, speaking, to assess the language proficiency of students in differ-
writing, listening, and reading proficiencies of all Eng- ent language education programs:
lish lan guage learners in K-12 should be assessed. Contribute to the identification and placement of
Kenyon, MacGregor, Li and Cook (2011) stated that ELLs. Monitor the progress of ELLs爷 English lan-
"One of the mandates of the No Child Left Behind Act guage development to inform teaching and differenti-
is that states show adequate yearly progress in their ate instruction. Document growth of English language
English language learners爷 (ELLs) acquisition of Eng- development over time for state accountability. Con-
lish language proficiency. Traditionally language abili- tribute to the reclassification of ELLs when deemed
ty has been composed of four skills, listening, reading, English language proficient. Contribute to the evalua-
speaking, and writing (Bachman & Palmer, 2010). tion of language program services (Gottlieb & Nguyen,
States are required to assess ELLs爷 English language 2007, p. 33).

36
Examination and Assessment

2. Significance 2. What are the validity and reliability of these


English language proficiency assessments?
Bachman and Palmer (2010) stated that the conse-
quences of language assessment will influence not on- 3. Review of the Literature
ly the individuals but also the educational system and There are studies explained the different English lan-
society. Former researches have proved that the ELLs guage proficiency tests adopted to test the ELLs (Al-
have been misplaced and improperly classified in the bers et al., 2009), which compared the pro-NCLB tests
English education programs or even in the special edu- and post-NCLB tests. There are also studies defined
cation. Solano-Flores (2008) believed that the current the construct in the English language proficiency tests
tests used for ELLs are based on "deterministic views" (Bailey & Huang, 2011), and the influences exerted by
and "erroneous assumptions" (p. 189). For example, the score report after the English language proficiency
the ELLs are classified into a few language proficiency tests (Abedi, 2008). The study conducted by MacSwan
categories, and only the English proficiency of the and Pray (2005) focused on the linguistic requirements
ELLs is assessed without any assessment of their first in the assessments while Bailey (2017) proposed inte-
language proficiency. gration of linguistic requirements with academic con-
Most of the bilingual programs focus mainly on tent requirements.
English proficiency development while failing to show MacSwan and Pray (2005) examined 89 ELLs in
their first language proficiency development. Second, bilingual programs and found out that they did not
Solano-Flores (2008) also stated that the assessments learn English slower than students in only English pro-
for ELLs are rarely comprehensive enough to show grams and they believed that English proficiency
their listening, speaking, reading and writing develop- should involve phonology, morphology, semantics,
ment of first language and second languages. "As a re- syntax and pragmatics. Bailey (2007) integrated the
sult, language tests may provide fragmented and in- linguistic requirements with the academic English
consistent information about the linguistic proficiency skills to assess the ELLs爷 linguistic ability and aca-
of ELLs" (p. 189). Third, the definition of ELLs is re- demic content, thus forming the academic English pro-
lated to their demographic characteristics and tends to ficiency assessment, including, listening, reading,
misclassify the ELLs. The English language proficien- speaking and writing.
cy tests for ELLs are designed according to the mono- Hauck, Wolf and Mislevy (2016) stated that
lingual, English-speaking students instead of the ELLs NCLB changed the education of ELLs from two as-
who learn English as a second language, which also pects. First, the ELLs are required to be included in the
leads to misclassification of the ELLs. standardized assessments as the non-ELLs to meet the
To examine the English proficiency assessments accountability requirement of NCLB. Second, it is re-
can, in the short term, reduce the misplacement of quired that new English language proficiency stan-
ELLs, improve the English proficiency assessment of dards should be established by every state which is
ELLs, help ELLs exit the program and help to make supposed to be corresponding to content standards,
decision and policy. In the long run, improvement of and the standardized English language proficiency as-
the English proficiency tests for ELLs can guide bilin- sessments should test the ELLs annually.
gual teachers to interpret English proficiency test 3.1 The Current English Proficiency Assessments
scores and assess the ELLs effectively, compare the Used in the Bilingual Program
English proficiency tests currently used and explore Before the No Child Left Behind in 2002, there were
the effectiveness of these English proficiency tests, and also English language proficiency assessments in dif-
to search the more effective methods to assess the ferent states. However, they were not uniform and
ELLs爷 English proficiency. mainly focused on the social language proficiency of
The following research questions will be ad- ELLs (Albers et al., 2009).
dressed in this library-based research paper: Albers et al. (2009) compared the pre-NCLB ELP
1. What are the English proficiency assessments tests in which social language proficiency was tested,
currently used for ELLs in the early-exit TBE pro- while post-NCLB ELP tests in which academic lan-
grams in elementary schools? guage proficiency was included in the tests. Whether

37
Frontier of Higher Education | Volume 1 | Issue 1 | September 2019

ELP tests can benefit classroom instructio ns received standards-based, and criterion-referenced English lan-
less attention. The fundamental ELP standards include guage proficiency test administered in the USA annu-
communication for social and instructional purposes, ally to more than 840,000 English Language Learners
and the communication of "information, ideas and (ELLs), in K-12 classrooms" (p. 425). ACCESS for
concepts for concept success" (Albers et al., 2009, p. ELLs is a test system developed by World-Class In-
77) in content areas of language arts, mathematics, sci- structional Design and Assessment (WIDA) Consor-
ence and social studies. tium by three states, Wisconsin, Delaware, and
Findings showed that ACCESS for ELLs, though Arkansas to meet the assessment requirements in the
measuring the similar constructs as the pre-NCLB ELP NCLB Act. ACCESS for ELLs first, used a vertical
tests, has components not cover in these pre-NCLB scaling to show the ELLs爷 proficiency in listening,
ELP tests. ACCESS for ELLs, as a new-generation speaking, reading and writing in the different grades
ELP measurement, incorporated the academic lan- starting from the first grade. The vertical scaling en-
guage proficiency to utilize the data obtained, enhance ables to demonstrate even the small differences in the
the instructions in the classrooms, and provide inter- ELLs爷 performances, especially when they move
ventions to the ELLs (Albers et al., 2009). across different grade levels (Fox & Fairbairn, 2011).
According to Wolf, Kao, Griffin, Herman, Bach- According to Bunch (2011), the four states, Penn-
man, Chang, and Farnsworth (2008), English language sylvania, Utah, Nevada, and Wisconsin were funded
learners are linguistically and culturally varied, and under NCLB and developed the English proficiency
with diverse academic levels and language proficiency, tests to meet the requirement of NCLB. They all have
while 44% of the ELLs are in pre-kindergarten to reading, listening, reading and writing tests, and em-
Grade 3. According to the NCLB, "a home language ploy various types of test items. There is an oral test in
survey, an ELP assessment, and academic achievement these assessments scored on the basis of standardized
assessment(s) in content areas" are employed to identi- rubric. Classical response theory and item response
fy ELL students (Wolf, et al., 2008, p. 3). theory are utilized, and formal standardized setting is
Wolf et al., (2008) explained the consortia under included. They only differ in the specifics.
the NCLB and stated that "nearly 40 states were origi- According to Kenyon, MacGregor and Cook
nally represented in the four consortia, which are (2011), it is mandatory for ELLs to take English profi-
known as Mountain West Assessment Consortium ciency tests, and the results of English proficiency tests
(MWAC), Pennsylvania Enhanced Assessment Grant enable the districts and elementary schools to place
(PA EAG), State Collaborative on Assessment and ELLs, evaluate the effectiveness of programs and help
Student Standards (SCASS) Consortium, and the ELLs make progress in their English acquisition
World-Class Instructional Design and Assessment and exit the ELLs. Solano-Flores (2008) analyzed the
(WIDA) Consortium" (p. 6). The English language problems existed in the testing for ELLs and consid-
proficiency assessments have developed based on ered testing as "a communication process between as-
these consortia. There are some states that do not use sessment systems and ELLs" (p. 189). Bunch (2011)
these assessments cooperated with test publishers to examined the currently used four commercial tests for
assess the ELLs. For example, WIDA standards are ELLs, ACCESS for ELLs, CELLA, ELDA, and
the basis and theoretical constructs for the "Assessing MWA. He pointed out that the four tests are similar to
Comprehension and Communication in English each other because they all cover listening, speaking,
State-to-State for English Language Learners (AC- reading and writing. The test items adopted in these
CESS for ELLs)" (p. 13) with the purpose to deter- four tests are primarily multiple choice questions and
mine the language proficiency of ELLs and place them open-ended questions.
into different programs with different language sup- 3.2 Constructs of English Language Proficiency
port. Assessments
Fox and Fairbairn (2011) examined Assessing According to American Educational Research Associ-
Comprehension and Communication in English ation, American Psychological Association and Na-
State-to-State for English Language Learners (AC- tional Council on Measurement in Education (2014),
CESS for ELLs), which is "a large-scale, high-stakes, the construct measured in the tests should be included

38
Examination and Assessment

in the interpretation, which refers to "the concept or However, there are limited empirical researches about
characteristic that a test is designed to measure" (p. 11). academic English development. Future researches are
Bailey and Huang (2011) defined academic Eng- recommended to validate the current standards, and
lish construct as "the vocabulary, sentence structures, create new state standards.
and discourse associated with language used to teach Wolf and Faulkner‐Bond (2016) proposed that
academic content as well as the language used to navi- the academic and social language proficiency con-
gate the school setting more generally" (p. 347). They structs operated in the English proficiency assessment
examined the standards of the English proficiency tests should be "next-generation" (p. 17) issue. Future re-
and examined the role the construct plays in the stan- searches can examine the alignment between English
dards of the English language proficiency test do- language proficiency standards and English language
mains. It is mandated by NCLB that standards should proficiency assessments to understand the construct of
be applied in the accountability system of these tests. English language proficiency, which will also influ-
According to NCLB, one state or collaborating states ence the social and academic language proficiency in-
in one consortium can create their own English lan- cluded in the English language proficiency.
guage proficiency standards. The constructs measured in the English language
Abedi (2008) stated that the different standards proficiency assessment should be explained explicitly
used in the English language proficiency assessments to the policy makers to help them choose the proper
produced inconsistent classification of ELLs. In addi- assessment, and teachers to enable proper instructions
tion, the performance levels of listening, reading, to the ELLs. It is also suggested that ELLs from differ-
speaking and writing are inconsistent, which makes it ent grades should be included in the future study to ex-
difficult to interpret the results of students爷 perfor- amine the relationship between the English language
mance. proficiency and academic content assessment, espe-
According to Abedi (2008), score report and in- cially in the upper grades (Wolf & Faulkner‐Bond,
terpretation can be influenced by the number of con- 2016).
structs measured in the tests. The scores from reading, 3.3 Validity of English Language Proficiency Tests
listening, speaking and writing can be combined when According to Abedi (2008), to ensure the reliability
they all measure the same construct. However, it and validity of the English language proficiency as-
would be difficult to combine the scores when they sessments, the academic English is required to be in-
have different constructs to measure. The English lan- cluded in the assessments, which also help the ELLs to
guage proficiency assessments currently used integrate meet the academic content standards. According to the
studies about psychometric aspects and validation of Department of Education, the link between English
the assessments. For example, "a criterion-related va- language proficiency assessments and academic con-
lidity approach 噎 content validation through align- tent assessments is to enable the ELLs to achieve pro-
ment to ELP content standards and construct valida- ficiency not only in English but also in other subjects
tions using the confirmatory factor analytic approach" such as math and science. The English language profi-
(Abedi, 2008, p. 200). ciency assessments are on the basis of academic Eng-
Bailey and Huang (2011) emphasized the incor- lish assessment to facilitate ELLs爷 content learning.
poration of the academic content construct into the Wolf and Faulkner ‐Bond (2016) claimed that
English language proficiency tests. First, teachers for English language proficiency assessment is considered
non-ELLs can be involved in the teaching of ELLs to to be high-stake and in large scale, the primary and the
help ELLs to take academic content tests. ELL teach- most important criterion in classifying and exiting
ers and mainstream class teachers are supposed to co- ELLs. English language proficiency assessment is not
operate with each other so that linguistic knowledge only crucial to ELLs but also to teachers and the
and content knowledge can be used to improve the school districts due the funding locations distribution.
standards of the test. According to Bailey and Huang It is vital to examine the validity of the assessment, es-
(2011), twenty six states have adopted WIDA stan- pecially the constructs of the English language profi-
dards, published in 2007 and augmented by TESOL in ciency assessment. First, the academic language profi-
2006, which provide directions for the ACCESS. ciency has been included as one of the constructs so as

39
Frontier of Higher Education | Volume 1 | Issue 1 | September 2019

to decide whether the students have been academically Expanding, and Bridging" (p. 14). There were 6662
prepared to exit the program. participants in this testing, among which 61% were
According to American Educational Research Spanish ELLs. First, sixteen experts and teachers ex-
Association, American Psychological Association and amined whether the content of ACCESS aligned with
National Council on Measurement in Education the construct in WIDA. Evidence based on the ELLs爷
(2014), "validity refers to the degree to which evidence listening and reading scores showed that test items can
and theory support the interpretations of test scores for assess the English language proficiency of ELLs in
proposed uses of tests. Validity is, therefore, the most these five levels. Then, they examined the interrela-
fundamental consideration in developing tests and e- tions between scores from ACCESS and the four es-
valuating tests" (p. 11). Besides, the construct mea- tablished tests, "IDEA Proficiency Test (IPT), the Lan-
sured in the tests should be included in the interpreta- guage Assessment Scale (LAS), the Language Profi-
tion, which refers to "the concept or characteristic that ciency Test Series (LPTS), and the Revised Maculaitis
a test is designed to measure" (p. 11). The following II (MAC II)" (p. 15), and found the correlations be-
researches will examine the validity of English lan- tween them proved the construct validity of ACCESS.
guage proficiency, including "evidence based on test "Rasch Modeling" and "Infit and Outfit" (p. 15) statis-
content", "evidence based on response processes", "ev- tics were used to examine the four subsets in ACCESS
idence based on internal structure", and "evidence and proved that the internal structure was acceptable.
based on relations to other variables" (p.14-16). The validity of ACCESS has been approved.
It is vital to test the English language proficiency Wolf, Farnsworth and Herman (2008) reviewed
assessment validity and technical quality because if the 49 states on the systems of their ELP assessment and
assessments about ELLs are inaccurate or improper, the according validity. Findings showed that it is es-
the individual ELLs can be inappropriately placed or sential to align the ELP standards and assessment to
misclassified. Second, the conclusions about ELLs ensure the validity. Findings suggested that academic
ELP proficiency will elicit the evaluations about the English should be included in the ELP assessments
programs in which ELLs are studying. Subsequently, construct. "NCLB legislation stipulates that the con-
the programs, schools and event the districts or states structs of an ELP assessment be aligned with the state爷s
will be sanctioned due to these conclusions. Further- ELP and content standards to measure the progress of
more, the assessments not only measure ELLs爷 learn- appropriate English language development" (Wolf et
ing but also fundamentally promote reform to improve al., 2008, p. 98).
learning and teaching (Wolf, Farnsworth & Herman, Wolf and Faulkner‐Bond (2016) analyzed the
2008). linguistic content of the items in English language as-
Wolf, Guzman ‐ Orth and Hauck (2016) also sessments in the three states and discovered that social
claimed that it is crucial to investigate the validity of and academic constructs have been represented in the
criteria to reclassify and designate the ELLs. Multiple assessment, and in the reading, listening, speaking and
criteria were utilized to reclassify ELLs, which include writing language domains. Findings showed that the
scores from ELP assessment, content assessment, and representation of academic language in English lan-
local criteria including GPA, grades, and recommen- guage proficiency assessment can support the under-
dations from teachers. However, the criteria used in re- standing of ELLs爷 English language proficiency. Wolf
classification are not suitable or designed to measure and Faulkner‐Bond (2016) adopted the "hierarchical
language proficiency, which is detrimental to the re- linear modeling" (p. 8) which is to investigate the rela-
classification of ELLs. tionship between the scores from four domains of the
Miley and Farmer (2017) stated that "to ensure language assessment and those from the academic as-
the reliability and validity of the WIDA-Access, the sessment, and revealed that the English language profi-
assessment is annually field-tested to validate its abili- ciency assessment scores consistently predict the
ty to equitably assess English proficiency" (P. 202). ELLs爷 performance in content assessment in these
Based on the field testing of Kenyon (2005) in 2004, three states. The validity of the English language profi-
which was about ACCESS for ELLs, including "five ciency assessment has been approved. According to
proficiency levels: Entering, Beginning, Developing, Wolf and Faulkner ‐ Bond (2016), this can be ex-

40
Examination and Assessment

plained by the fact tha t the two assessments share Abilities" to see whether the ELLs爷 English proficien-
some English language construct. cy is related to their cognitive abilities. NYSESLAT is
Results also demonstrated that social language a language proficiency test developed under the guid-
scores can also be predictive in the performance in ance of NCLB for the ELLs, and ELLs爷 raw scores
content assessment (Wolf & Faulkner‐Bond, 2016). will be converted to scaled scores. There are four lev-
ELLs爷 social language skills are relevant to the els "Beginner, Intermediate, Advanced, and Proficient"
progress and proficiency assessment. Except for the (p. 787), and it is proved to be reliable and valid. Find-
constructs taken in the English language proficiency ings showed that the fact that ELLs are tested through
assessment, the various language skills should also be verbal or nonverbal items should be avoided. Instead,
taken into consideration. the ELLs爷 current developmental English proficiency
3.4 Reliability of English Language Proficiency As- should be evaluated because the ELLs爷 English profi-
sessment ciency has been developed just to cater to the tests, but
According to Reynolds, Livingston, Willson and Will- actually they cannot meet the demands of tests with
son (2009), the context of measurement reliability higher linguistic requirements and cultural back-
refers to 野consistency or stability of assessment re- grounds.
sults冶 (p. 91). There are different types of reliability, Findings also revealed that the application of ver-
such as test-retest reliability, alternative-form reliabili- bal or nonverbal tests on ELLs was superficial and the
ty internal-consistency reliability and inter-rater relia- fairness and equity cannot be ensured (Sotelo‐Dyne-
bility (Reynolds et al., 2009). ga et al., 2013). The ELLs who are tested to have pro-
According to Longabach and Peyton (2017), it is ficient English levels may not be same as the native
of great importance to assess the sub-score reliability English speaking peers.
and total score reliability when assessing the English
4. Findings
proficiency of ELLs because the ultimate goal of as-
sessment is to improve the education for ELLs, which The four commercial tests for ELLs, ACCESS for
relies on the accuracy of the measurement and appro- ELLs, CELLA, ELDA, and MWA were found to be
priate purpose to assessment ELLs. The reliability of similar to each other including listening, speaking,
English language proficiency tests should also be ex- reading, and writing (Bunch, 2011). ACCESS for
amined. ELLs used a vertical scaling to demonstrate the ELLs爷
Longabach and Peyton (2017) used Cronbach爷s proficiency in listening, speaking, reading and writing
alpha and standard error of measurement respectively in the different grades, which enables the demonstra-
to estimate the reliability and precision of the four tion of the small differences in the ELLs爷 perfor-
methods. Findings showed that CTT and UIRT were mances, especially when ELLs move across different
similar but lower than A-IRT and MIRT in reliability grade levels (Fox & Fairbairn, 2011).
and precision. MIRT was found to be the most reliable With the examination of the content, interrela-
one among the four methods to score the sub-domains tions between scores from ACCESS and four other es-
for all grade levels. A-IRT followed MIRT closely, tablished tests, and found the correlations between
ranking second in reliability. These score methods are them proved the construct validity of ACCESS. The
highly consistent on the sub-scores reported from each examination of four subsets, listening, reading, speak-
grade. The new English language proficiency assess- ing, and writing in ACCESS proved that the internal
ments are being developed in the states, especially the structure was acceptable. The validity of ACCESS has
states, such as California, Texas and New York, which been approved.
claim the large number of ELLs. The four scores methods were also compared and
Sotelo ‐ Dynega, Ortiz, Flanagan and Chaplin examined; these score methods are highly consistent
(2013) conducted an empirical research and examined on the sub-scores reported from each grade. According
the relationship between the performances of ELLs in to Longabach and Peyton (2017), MIRT was found to
the second grade in "New York State English as a Sec- be the most reliable one among the four methods to
ond Language Achievement Test (NYSESLAT)" (p. score the sub-domains for all grade levels.
787) and the "Woodcock-Johnson III Test of Cognitive

41
Frontier of Higher Education | Volume 1 | Issue 1 | September 2019

References Bilingual Research Journal, 29(3), 653–678.


Abedi, J. (2008). Measuring students' level of English Miley, S. K., & Farmer, A. (2017). English Language
proficiency: Educational significance and assessment Proficiency and Content Assessment Performance: A
requirements. Educational Assessment,13 (2-3), 193-214. Comparison of English Learners and Native English
doi: 10.1080/10627190802394404 Speakers Achievement. English Language Teaching,10(9),
Albers, C. A., Kenyon, D. M., & Boals, T. J. (2009). Measures 198. doi: 10.5539/elt.v10n9p198.
for determining English language proficiency and the Reynolds, C. R., Livingston, R. B., Willson, V. L., & Willson,
resulting implications for instructional provision and V. (2009). Measurement and assessment in education.
intervention.?Assessment for Effective Intervention,?34(2), Upper Saddle River, NJ: Pearson Education International.
74-85. Solano-Flores, G. (2008). Who is given tests in what language
Bachman, L. F., & Palmer, A. (2010). Language assessment in by whom, when, and where? The need for probabilistic
practice. Oxford: Oxford University Press. views of language in the testing of English language
Bailey, A. L., & Huang, B. H. (2011). Do current English learners. Educational Researcher, 37 (4), 189-199. doi:
language development/proficiency standards reflect the 10.3102/0013189X08319569
English needed for success in school?.Language Testing,28 Sotelo‐ Dynega, M., Ortiz, S. O., Flanagan, D. P., & Chaplin,
(3), 343-365. doi: 10.1177/0265532211404187 W. F. (2013). English language proficiency and test
Bunch, M. B. (2011). Testing English language learners under performance: An evaluation of bilingual students with the
no child left behind. Language Testing, 28 (3), 323-341. Woodcock ‐ Johnson iii tests of cog nitive abilities.
doi: 10.1177/0265532211404186 Psychology in the Schools, 50 (8), 781-797. doi:
Carroll, P. E., & Bailey, A. L. (2016). Do decision rules matter? 10.1002/pits.21706
A descriptive study of English language proficiency Umansky, I. M., & Reardon, S. F. (2014). Reclas sification
assessment classifications for English-language learners patterns among Latino English learner students in bilingual,
and native English speakers in fifth grade. Language dual immersion, and English immersion classrooms.?
Testing, 33(1), 23-52. doi: 10.1177/0265532215576380 American Educational Research Journal, 51 (5), 879-912.
Fox, J., & Fairbairn, S. (2011). Test review: ACCESS for ELLs R . doi: 10.3102/0002831214545110
Language testing,28(3), 425-431. doi: 10.1177/02655322 Wolf, M. K., & Faulkner‐Bond, M. (2016). Validating English
11404195 language proficiency assessment uses for English learners:
Gottlieb, M. H., & Nguyen, D. (2007).Assessment and Academic language proficiency and content assessment
accountability in language education programs: A guide for performance. Educational Measurement: Issues and
administrators and teachers. Philadelphia, PA: Caslon. Practice, 35(2), 6-18.
Hauck, M. C., Wolf, M. K., & Mislevy, R. (2016). Creating a Wolf, M. K., Farnsworth, T., & Herman, J. (2008). Validity
Next ‐ Generation System of K – 12 English Learner issues in assessing English language learners' language
Language Proficiency Assessments. ETS Research Report proficiency. Educational Assessment, 13(2-3), 80-107. doi:
Series, 2016(1), 1-10. doi:10.1002/ets2.12092 10.1080/10627190802394222.
Kenyon, D. M., MacGregor, D., Li, D., & Cook, H. G. (2011). Wolf, M. K., Kao, J., Griffin, N., Herman, J. L., Bachman, P. L.,
Issues in vertical scaling of a K-12 English language Chang, S. M., & Farnsworth, T. (2008). Issues in Assessing
proficiency test. Language Testing, 28 (3), 383-400. doi: English Language Learners: English Language Proficiency
10.1177/0265532211404190 Measures and Accommodation Uses. Practice Review (Part
Longabach, T., & Peyton, V. (2017). A comparison of reliability 2 of 3). CRESST Report 732. National Center for Research
and precision of subscore reporting methods for a state on Evaluation, Standards, and Student Testing (CRESST).
English language proficiency assessment. Language Wolf, M. K., Guzman‐Orth, D., & Hauck, M. C. (2016). Next
Testing, 1(21). doi: 10.1177/0265532217689949 ‐ Generation Summative English Language Proficiency
MacSwan, J., & Pray, L. (2005). Learning English bilingually: Assessments for English Learners: Priorities for Policy and
Age of onset of exposure and rate of acquisition among Research.?ETS Research Report Series,2016(1), 1-23.
English language learners in a bilingual education program.

42

You might also like