Professional Documents
Culture Documents
Department of Educational Administration, Foundations, and Psychology, University of Manitoba, Winnipeg, Manitoba, Canada R3T 2N2 Department of Psychology, University of Western Ontario, London, Ontario, Canada N6A 5C2
article
Article history:
info
abstract
The majority of studies reporting gains in college students critical thinking due to instruc tional process variables measured critical thinking with subject-specic questions rather than general or decontextualized questions. However, it is uncertain whether these gains were attributable to the use of subject-specic questions or to other distinc tive aspects of these studies (e.g., methodological). The present study provides a direct, controlled comparison between general and subject-specic test questions in the context of a laboratory-based true experiment assessing the effect of higher order review questions on gains in critical thinking. A stronger effect was found when the tests of critical thinking contained questions that were subject-specic (e.g., introductory psychology) rather than questions that focused on general topics. 2008 Elsevier Ltd. All rights reserved.
Received 1 August 2007 Received in revised form 18 March 2008 Accepted 26 March 2008 Available online 1 April 2008
Keywords:
Critical thinking Assessment Subject-specic General
1. Introduction
Corresponding author. Tel.: +1 204 474 6786; fax: +1 204 474 7564. E-mail addresses: renaudr@ms.umanitoba.ca (R.D. Renaud), murray@uwo.ca (H.G. Murray).
1
1871-1871/$ see front matter 2008 Elsevier Ltd. All rights reserved. doi:10.1016/j.tsc.2008.03.005
86
R.D. Renaud, H.G. Murray / Thinking Skills and Creativity 3 (2008) 8593
87
Whitlow, Stover, & Johnson, 1996; McMillan, 1987; Pascarella & Terenzini, 2005), relatively little is known about what goes on during the college experience that contributes to this. Gains in critical thinking have been compared with several process variables pertaining to either the institution itself as a whole or a specic instructor. In general, while there is little evidence to suggest that any particular institutional variable is consistently linked with improved critical thinking skills measured either with general or subject-specic content, one particular instructional variable that may be worth further attention is the frequency of higher order questions. Recently, in a series of related studies with varied methods, Renaud and Murray (2007) found that frequency of higher order questions was related to gains in students critical thinking skills. While the Renaud and Murray studies included both general and subject-specic measure of critical thinking skills, these two formats were not compared with one another. Looking at general content, Terenzini, Springer, Pascarella, and Nora (1995) found that both class-related and out-of-class student experiences were related to scores on the critical thinking module of Collegiate Assessment of Academic Prociency. In a more controlled setting, Gadzella, Ginther, and Bryant (1996) found signicant pretest to posttest gains in critical thinking using the WGCTA for a single group of students enrolled in a course where students were given critical thinking problems. However, Gadzella et al. did not have an equivalent control group, which makes it dif cult to know how much the critical thinking problems actually contributed to observed gains. In contrast, several studies focusing on higher order questions asked by the instructor while teaching found little relation with gains in general critical thinking (Foster, 1983; Smith, 1977, 1980). For example, Smith (1977) correlated frequency of higher order questions asked by the instructor during class with gain scores on the WGCTA, and found that while critical thinking was enhanced by student participation, encouragement, and peer interaction, use of higher order questions by the instructor had little effect. However, a serious limitation of the Smith (1977) study was a severe restriction in variance in that questioning occurred in only 2.6% of the total class time. With almost no class time spent on asking questions, it would be difcult to determine if asking higher order questions really did have an impact on critical thinking skills. According to both McMillan (1987) and Pascarella and Terenzini (2005), gains in critical thinking are more likely to be found when critical thinking items focus on content relevant to a particular course, or when the content is consistent with a students academic major. Looking at questions asked in student tests or assignments, the aspect most studied appears to be the effect of having students write their own higher order test questions. The relation between students writing higher order questions and critical thinking was found to be positive when students created their questions either in class (King, 1989, 1990, 1995), or as an assignment out of class (Keeley, Ali, & Gebing, 1998; King, 1995). For example, King (1995) provided students with generic question stems such as What is the best. . .and why? and What is the difference between. . .and. . .? and had students generate higher order questions right after class in a peer questioning session or on assigned readings before class. In the King studies, critical thinking was measured either in the form of self-evaluation (King, 1995), subject-specic pre- and posttests of lecture comprehension (King, 1989, 1990), or pre- and posttest measures of the number of higher order questions generated from reading a short passage. Using a different design, Logan (1976) gave a 10-item critical thinking test to one group of students near the end of a one-semester course in which the professor analyzed material in a critical manner during lectures and to a second group at the beginning of the same course in the following semester. He found that the mean score for the rst group was higher than that for the second group. While Logans results are encouraging as the measure of critical thinking was course-specic, and thus, more sensitive, it is difcult to know just how much the students in the rst group actually increased their critical thinking skills without knowing their incoming levels. Looking at one particular instructional variable, Wesp and Montgomery (1998) measured gains in critical thinking with pre- and posttests of 300-word articles in which students were asked to nd and explain several aws. They found that compared to the control class, the class in which students were engaged in critical thinking exercises showed higher gains in critical thinking skills. More recently, Williams, Oliver, and Stockdale (2004) found larger pretestposttest gains in critical thinking skills among students in classes that were given practice exam questions that involved critical thinking compared to those in the control classes, which received practice exam questions on similar topics, but without the critical thinking focus. Conversely, Tierney, Soter, OFlahavan, and McGinley (1989) conducted an experiment in which students wrote an initial draft of a letter to the editor on an assigned controversial topic. Students in the experimental group were then given higher order questions to answer before revising their initial drafts, while control students were not. According to several criteria followed by trained judges, the quality of revised drafts from the group answering higher order questions did not differ appreciably from those of the group answering no questions. Using a stronger methodology, Garside (1996) compared the effects of group discussions in an interpersonal communications class on critical thinking skills. Controlling for pretest scores and prior preparation (i.e., assigned reading before class), students engaged in group discussions did not obtain signicantly higher scores on a critical thinking test focusing on interpersonal communication concepts than those from class sections that followed the lecture format. Pertinent to this study, there are several limitations that may have prevented researchers from nding a more consistent link between specic process variables and critical thinking that could help determine whether critical thinking should be measured either with general or subject-specic items. One concern has to do with ensuring that the processes believed to have an impact on critical thinking are experienced at the intended level. While Smith (1977) was able to determine the low frequency of higher order questioning during class, the level of treatment exposure in other studies is difcult to conrm (e.g., Garside, 1996). Another methodological limitation is that without basic design controls such as a pretest (e.g., Logan, 1976) or a control group (e.g., Gadzella et al., 1996), it is unclear if the signicant relation between instructional processes and critical thinking are due to the intended treatment or other confounding variables. While many studies have explored
88
the impact of instructional processes on either general or subject-specic gains of critical thinking, two recent studies have directly compared the scores obtained from both types of tests. In a eld experiment that compared the pretestposttest gains of three different critical thinking tests across three sections of an educational psychology course, Williams (2003) found larger gains from the critical thinking test focusing on psychology compared to two general measures. Similar ndings were obtained in a later study (Williams et al., 2004). Although the methods of Williamss studies addressed many of the above limitations of previous studies, other threats to the internal validity of the ndings (e.g., contamination) were unavoidably present.
2.1. Participants
The participants in this study consisted of 190 undergraduate students registered in a rst -year introductory psychology course. Students who chose to participate in this study were given credit toward fullling the research participation require ment in this course. Before the experiment began, all participating students signed a consent form indicating that they were taking part in a study that compares study strategies. Participants were randomly assigned to either an experimental group or a control group with 96 and 94 students, respectively. To minimize the degree to which the data used in the analyses were contaminated with outlying values, each subjects data had to meet the following criteria for inclusion. Separate analyses were performed for each of the general and subject-specic critical thinking tests. Looking at the general test, there were three criteria for inclusion. First, because the maximum score on both the general pretest and posttest was 10 points, a subjects data were r emoved if he or she obtained a perfect score on the general pretest, thus precluding pretest-to-posttest gain. During the treatment phase of the experiment, to help ensure that a subject had put forth a reasonable effort toward reading the passage and answering the review questions that pertained to the passage, each subject had to obtain a minimum score of 3 out of 8. Although this criterion was applied to all subjects, it was intended more for those in the higher order group as an indication that they had engaged in a sufcient level of higher order thinking. Of the 26 subjects data who were taken out for failing to meet this criterion, 25 came from the higher order group. Finally, as an indication that a subject had put as much effort into answering the questions on the general posttest as on the pretest, and considering how a subjects scores could decline slightly from pretest to posttest because of factors other than critical thinking ability (e.g., the different questions in each versio n), a subjects data were removed if his or her posttest score was at least 4 points below his or her pretest score. Based on these three criteria, 157 subjects were retained for analyses on the general test with 66 and 91 subjects in the experimental and control conditions, respectively. With respect to the course-specic psychology portion of the critical thinking test, there were two further criteria for inclusion in the data analysis. First, as with the general subtest, each subject had to obtain a minimum score of 3 out of 8 on the review questions. Because the review exercise (i.e., treatment) was the same in both the general and psychology analyses, the number of subjects taken out from each group was the same. Regarding the level of student motivation in completing the posttest, it is possible that a student who is less motivated, and wanting to complete the experiment as quickly as possible, may be inclined to select answers without carefully reading the multiple-choice questions. With each of the 15 items in the psychology subtest having 4 options, the probability of someone obtaining up to 4 out 15 correct by picking answers at random is .69. Therefore, to help ensure that a posttest score was more indicative of a genuine effort rather than a random selection, and given that the students were expected to do reasonably well on the posttest as they had just completed the review exercise, the second criterion for inclusion was that a subjects postte st score had to be at least 5 out of 15. Based on these two criteria, 95 subjects were retained for analyses on the psychology test with 37 and 58 subjects in the experimental and control conditions, respectively.
2.2. Materials
All subjects were given a pretest and a parallel posttest measure of critical thinking, consisting of both a general subtest and a course-specic subtest. Each critical thinking test consists of 25 multiple-choice questions intended to measure the degree to which a student can engage in a particular aspect of critical thinking. Each item is followed by 4 or 5 options representing varying degrees of correctness or applicability from which students selected the most appropriate option. Scores on either test could range from a minimum of 0 to a maximum of 25. T o measure general critical thinking ability, the rst ten questions in the pretest and posttest were adapted from the Watson-Glaser Critical Thinking Appraisal (WGCTA). These questions focus on everyday situations that most people would likely be familiar with. To concur with the overall format of the WGCTA, the general items used in this study reected each of the ve main components of critical thinking covered by the WGCTA (i.e., inference, recognition of assumptions, deduction, interpretation, and evaluation of arguments), with two items reecting
89
each component. Because many of the studies reporting signicant gains in college students critical thinking used domain or subject-specic measures (see review by McMillan, 1987), the remaining 15 questions focused on a selected passage from a chapter on personality theory from the introductory psychology textbook that was used in the course. For example, one item reads Which conclusion is most valid concerning research on the ve key dimensions of personality? The chapter on personality theory was chosen because it was not scheduled to be covered in class for at least a month after the completion of this study. With respect to estimating the reliability of the scores obtained from both the general and course-specic subtests, the typical reliability estimates, namely, internal consistency and parallel measures would not be applicable for at least two main reasons. First, the items within each subtest represent multiple dimensions. A principal-components analysis with Promax rotation revealed the following number of largely unrelated factors for each subtest: pretest-general (5), pretestpsychology (6), posttest-general (4) and posttest-psychology (7). Within each dimension or factor, there are too few items to obtain a meaningful estimate of internal consistency. Second, the reliability estimates of dichotomously scored items are inversely related to the range of item difculties (i.e., the proportion of students who answered a particular item correctly) (Nunnally & Bernstein, 1994). In this study, the item difculties ranged from .12 to .90. Additional concerns surrounding the inappropriateness of internal consistency and parallel measures reliabilities of a complex performance such as critical thinking are outlined in Linn, Baker, and Dunbar (1991) and Shavelson, Baxter, and Pine (1992). Between the pre- and posttest measures of critical thinking, all subjects were asked to read a passage on personality theory. This passage consisted of 9 textbook pages, which most students read within 25 min. Along with the assigned reading, each subject was given a set of review questions that pertained to the passage. The independent variable in this experiment was the level of review questions subjects answered as they read the passage. Subjects assigned to the experimental condition were given four higher order critical thinking questions, each of which was in a short answer essay format and answerable within half a page. Subjects in the control condition were given eight lower order recall questions, each of which was also in a short answer open format and answerable within a quarter of a page. The reason for assigning a greater number of lower order questions than higher order questions was to ensure that students in both groups were spending about the same length of time on the reading assignment before starting the posttest measure of critical thinking. It was suspected that the lower order questions would be easier and, therefore, take less time and space to answer compared to the higher order questions.
2.3. Procedure
Subjects were tested in groups ranging in size from 3 to 10. To ensure that experimental and control groups had roughly equal numbers of subjects, and that subjects in both groups were tested at the same times, half of the subjects in each testing session were randomly assigned to the experimental condition and half to the control condition. Before the experiment began, each participant signed a consent form indicating his or her understanding of the experiment and willingness to participate. To encourage students to put forth their best effort, students were told before the experiment began that anyone who scored higher than the predicted population average on both critical thinking tests and the chapter questions would receive a one-dollar lottery ticket at the end of the experiment. Actually, every participant had received a lottery ticket. The predicted population average was a ctitious goal that enabled the experimenter to justify giving every participant a lottery ticket more easily than would an absolute goal such as 50% correct. To begin, all subjects were given the critical thinking pretest, which most subjects completed in about 20 min. After all subjects had completed the pretest, they were given the assigned reading and review questions at the same time. Before the students began this phase, they were instructed to briey look over the questions before reading the passage so they could identify more readily which parts of the passage to pay close attention to in order to answer the questions. Most subjects took approximately 40 min to read the passage and answer all of the questions. After each subject had nished answering his or her review questions based on the assigned reading, they were given the critical thinking posttest. After completing the posttest, each subject received a debrieng form outlining the purpose of this experiment in detail and a lottery ticket. 3. Results
90
pretest was primarily due to the majority of omitted subjects, who tended to score relatively low on this test, coming from the experimental group as explained earlier. Using the ANCOVA procedure with pretest scores as the covariate, there was no signicant difference between the groups posttest scores, with group membership accounting for only 1% of the total variance (2 = .010).
F(1,92) = 7.94, p < .01, and the effect of the treatment accounted for 7.9% of the total variance ( 2 = .079).
91
4. Discussion The present study compared gains in students critical thinking measured with general and subject -specic questions. In agreement with McMillans (1987) review and in recent studies by Williams (2003) and Williams et al. (2004), larger pretestposttest gains were found from tests that contained questions that relate to the course in which the tests were given (e.g., introductory psychology) than tests with questions that focused on general topics. Perhaps one reason for this nding is that although a student is likely to be aware of the topics in general measures (e.g., politics), they may not be required as often to engage in the types of critical thinking in these everyday topics as they would in studying particular subjects (e.g., psychology). In other words, it may be that as instructors teach the skills that foster critical thinking, much of the focus is within a subject-specic context. In addition, the length of time between the pretest and posttest (i.e., the treatment) was only about 40 min and the subjects were rst-year students. Therefore, the opportunity to engage in a variety of learning experiences over a longer time span such as 34 years, ought to contribute to gains in critical thinking skills that may be more detectable with a general measure. In addition to being signicant both in statistical and practical terms, this nding is noteworthy for another reason. Like intelligence, it is not unreasonable to think that an apprecia ble gain in critical thinking skills would occur after having been exposed to a variety of courses and experiences over an extended period. Therefore, given that one could expect only a small effect under the restricted conditions of this experiment, namely minimal exposure to the independent variable in such a short time frame, these results are encouraging. Arguably, such a minimal exposure would lead to a narrower range of the degree to which subjects had experienced the treatment, which would tend to attenuate its relationship with the scores on the dependent variable (Prentice & Miller, 1992). On a related limitation concerning the short duration of the intervention, it becomes more difcult to determine the degree to which the students pretest and posttest scores actually reect the level of critical thinking that was inu enced by the treatment. Alternatively, given that the entire experiment (i.e., pretest, intervention, and posttest) lasted approximately only 90 min, it is possible that these ndings may be a function of other inuences such as the stu dents orientation to what was required on the assessments. In future research, it will be helpful to reduce the effects of these confounding inuences by assessing these skills more comprehensively (i.e., varied measures) and over a longer period. An ongoing concern in controlled experiments like this is the degree to which students are putting forth a genuine effort to complete the tasks involved in the experiment. The students who took part in this study did so to help fulll their research participation requirement in the introductory psychology course. This requirement is based solely on the number of research studies in which the student participates, and has nothing to do with the quality of the participation. As an incentive to make the students in this study try their best in completing the critical thinking tests and the review questions, each student was promised a lottery ticket if his or her scores were above a hypothetical standard on all three tasks (i.e., pretest, review questions, posttest). Admittedly, in this study, the effectiveness of this incentive is questionable for two reasons. First, although both groups showed a pretestposttest improvement on the course-specic psychology subtest, it was somewhat surprising that there was not a larger improvement given that the critical thinking tests referred to a relatively short passage that students had read through for about 45 min. Second, it is possible that the lower scores obtained on review questions by students in the higher order condition (4.35 out of 8, on average) may suggest that these students were not engaged in higher order thinking as much as was expected. One way to address the concern regarding the level of student motivation to try their best in each of the tasks in the experiment might be to provide a more valuable incentive. At the post-secondary level, one of the clearest, most immediate incentives is grades. Therefore, one possibility would be to include the experimental tasks as a small part of a course with a corresponding weight of the nal grade. To deal with the ethical issues involved, such an experiment would require a within-subjects design such that each student receives exactly the same materials from which he or she will be assigned a grade. For example, a study could take place during a class period with the pretest and posttest critical thinking measures focusing only on relevant course material. The review questions could consist of both lower order questions that pertain to one half of the passage (e.g., based on one chapter), and higher order questions that pertain to the rest (e.g., based on another chapter in the text). The ndings in this study suggest that gains in students critical thinking skills are more clearly detected with items focusing on specic course content rather than on general issues assumed to be familiar to a student in any discipline. Concerning research that examines the link between institutional and instructional processes and student outcomes, two implications are worth noting. While Pascarella and Terenzini (2005) conclude in their extensive review that controlling for incoming ability and maturatio nal effects, most studies found a signicant gain in critical thinking going from freshman to senior year, the format of assessing critical thinking used in these studies includes general, subjectspecic, and self-ratings. Thus, it is possible that some of the studies considered in their review that used general measures of critical thinking may have underestimated the extent of student gains. Using subject-specic measures of critical thinking may help colleges to better assess gains in students critical thinking skills both within depart ments (e.g., chemistry, economics) and entire institutions based on measures obtained from several departments. From another perspective, Halpern (2001) suggests that because critical thinking ought to be a skill that students should be able to use indenitely after graduation, and therefore should be transferable to novel contexts outside of a particular course, gains over a longer enrollment (e.g., from rst-year to fourth-year) might be more appropriately assessed with
92
a general measure. Conversely, shorter term gains (e.g., one semester), might be more detectable with subject-specic measures. The second implication has to do with identifying valid institutional (e.g., campus tutoring programs) and instructional processes (e.g., higher order questioning in class) in terms of their relation with intended outcomes including gains in students critical thinking skills. Pascarella and Terenzini (2005) point out that while signicant gains in critical thinking have been found as a result of attending college, we are only beginning to learn what actually contributes toward this improvement. A more sensitive, course-based measure of critical thinking could help better determine the effectiveness of various educational processes. This would help to provide better empirical evidence concerning the factors that contribute toward student learning and development, which can be used to make more informed and justied choices both at the institutional level and within a particular course. Acknowledgement This study was supported in part by the Centre for Higher Education Research and Development, University of Manitoba, Winnipeg, Canada. References
Adams, M. H., Whitlow, J. F., Stover, L. M., & Johnson, K. W. (1996). Critical thinking as an educational outcome: An evaluation of current tools of measurement. Nurse Educator, 21(3), 2332. Beyer, B. (1987). Practical strategies for the teaching of thinking. Boston, MA: Allyn and Bacon. Chickering, A. (1972). Undergraduate academic experience. Journal of Educational Psychology, 63, 134143. Dressel, P. L., & Mayhew, L. B. (1954). General education: Explorations in evaluation. Washington, DC: American Council on Education. Ennis, R. H. (1985). A logical basis for measuring critical thinking skills. Educational Leadership, 43, 4448. Facione, P. A. (Ed.). (1990). Critical thinking: A statement of expert consensus for purposes of educational assessment and instruction. American Philosophical Association, ERIC ID 315 423. Foster, P. (1983). Verbal participation and outcomes in medical education: A study of third-year clinical-discussion groups. In C. Ellner & C. Barnes (Eds.), Studies in college teaching: Experimental results, theoretical interpretations and new perspectives . Lexington, MA: D.C. Heath. Furedy, C., & Furedy, J. (1985). Critical thinking: Toward research and dialogue. In J. Donald & A. Sullivan (Eds.), Using research to improve teaching (New Direction for Teaching and Learning No. 23). San Francisco: Jossey-Bass. Gadzella, B. M., Ginther, D. W., & Bryant, G. W. (1996). Teaching and learning critical thinking skills. In Paper presented at the International Congress of Psychology. Garside, C. (1996). Look whos talking: A comparison of lecture and group discussion teaching strategies in developing critical thinking skills. Communication Education, 45, 212227. Gouran, D. S., & Hirokawa, R. Y. (1983). The role of communication in decision-making groups: A functional perspective. In M. S. Mander (Ed.), Communications in transition. New York, NY: Praeger. Griggs, R. A., Jackson, S. L., Marek, P., & Christopher, A. N. (1998). Critical thinking in introductory psychology texts and supplements. Teaching of Psychology , 25, 254265. Halpern, D. F. (1996). Thought and knowledge: An introduction to critical thinking (3rd ed.). Mahwah, NJ: Erlbaum. Halpern, D. F. (1998). Teaching critical thinking for transfer across domains. American Psychologist, 53, 449455. Halpern, D. F. (2001). Assessing the effectiveness of critical thinking instruction. The Journal of General Education, 50, 270286. Keeley, S. M., Ali, R., & Gebing, T. (1998). Beyond the sponge model: Encouraging students questioning skills in abnormal ps ychology. Teaching of Psychology, 25, 270274. King, A. (1989). Effects of self-questioning training on college students comprehension of lectures. Contemporary Educational Psychology, 14, 116. King, A. (1990). Enhancing peer interaction and learning in the classroom through reciprocal questioning. American Educational Research Journal, 27, 664687. King, A. (1995). Inquiring minds really do want to know: Using questioning to teach critical thinking. Teaching of Psychology, 22, 1317. Lawson, T. J. (1999). Assessing psychological critical thinking as a learning outcome for psychology majors. Teaching of Psychology, 26, 207209. Linn, R. ,L., Baker, E. L., & Dunbar, S. B. (1991). Complex, performance-based assessment: Expectations and validation criteria. Educational Researcher, 20, 1521. Logan, G. H. (1976). Do sociologists teach students to think more critically? Teaching Sociology, 4, 2948. McMillan, J. H. (1987). Enhancing college students critical thinking: A review of studies. Research in Higher Education, 26, 329. McPeck, J. (1981). Critical thinking and education. New York, NY: Saint Martins Press. Moseley, D., Baumeld, V., Elliott, J., Gregson, M., Higgins, S., Miller, J., et al. (2005). Frameworks for thinking: A handbook for teaching and learning. Cambridge: Cambridge University Press. Nordvall, R. C., & Braxton, J. M. (1996). An alternative denition of quality of undergraduate education: Toward usable knowl edge for improvement. Journal of Higher Education, 67, 483497. Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric theory (3rd ed.). New York: McGraw-Hill. Pascarella, E. T., & Terenzini, P. T. (2005). How college affects students: A third decade of research San Francisco: Jossey-Bass. Prentice, D. A., & Miller, D. T. (1992). When small effects are impressive. Psychological Bulletin, 112, 160164. Renaud, R. D., & Murray, H. G. (2007). The validity of higher-order questions as a process indicator of educational quality. Research in Higher Education, 48(3), 319351. Shavelson, R. J., Baxter, G. P., & Pine, J. (1992). Performance assessments: Political rhetoric and measurement reality. Educational Researcher, 21, 2227. Smith, D. G. (1977). College classroom interactions and critical thinking. Journal of Educational Psychology, 69, 180190. Smith, D. G. (1980). College instruction: Four empirical views. Instruction and outcomes in an undergraduate setting. In Paper presented at the Annual Meeting of the American Educational Research Association. Terenzini, P. T., Springer, L., Pascarella, E. T., & Nora, A. (1995). Inuences affecting the development of students critic al thinking skills. Research in Higher Education, 36, 2339. Tierney, R. J., Soter, A., OFlahavan, J. F., & McGinley, W. (1989). The effects of reading and writing on thinking criticall y. Reading Research Quarterly, 24, 134173. Wagner, T. A., & Harvey, R. J. (2006). Development of a new critical thinking test using item response theory. Psychological Assessment, 18(1), 100105.
R.D. Renaud, H.G. Murray / Thinking Skills and Creativity 3 (2008) 8593
Watson, G. B., & Glaser, E. M. (1980). Watson-Glaser critical thinking appraisal. San Antonio: The Psychological Corporation. Wesp, R., & Montgomery, K. (1998). Developing critical thinking through the study of paranormal phenomena. Teaching of Psychology, 25, 275278. Williams, R. L. (2003). Critical thinking as a predictor and outcome measure in a large undergraduate educational psychology course . (ERIC Document Reproduction Service No. ED 478 075). Williams, R. L., Oliver, R., & Stockdale, S. (2004). Psychological versus generic critical thinking as predictors and outcome measures in a large undergraduate human development course. Journal of General Education, 53, 3758. Winter, D., & McClellend, D. (1978). Thematic analysis: An empirically derived measure of the effects of liberal arts education. Journal of Educational Psychology, 70, 816.
93