Professional Documents
Culture Documents
To cite this article: Paul Orsmond & Stephen Merry (2017) Tutors’ assessment practices and
students’ situated learning in higher education: chalk and cheese, Assessment & Evaluation in
Higher Education, 42:2, 289-303, DOI: 10.1080/02602938.2015.1103366
Download by: [UNAM Ciudad Universitaria] Date: 04 April 2017, At: 12:43
Assessment & Evaluation in Higher Education, 2017
Vol. 42, No. 2, 289–303, http://dx.doi.org/10.1080/02602938.2015.1103366
This article uses situated learning theory to consider current tutor assessment and
feedback practices in relation to learning practices employed by students outside
the overt curriculum. The case is made that an emphasis on constructive
alignment and explicitly articulating assessment requirements within curricula
may be misplaced. Outside of the overt curriculum students appear to be
interdependent learners, participating in communities of practice and learning
networks, where sense-making occurs through negotiation and there is identity
development. Such negotiation may translate curriculum requirements articulated
by tutors into unexpected meanings. Hence, tutors’ efforts might be better placed
on developing students’ ability to self-assess and to effectively evaluate and
negotiate information, rather than primarily on their own delivery of the
curriculum content and feedback. Tutors cannot be fully effective if they fail to
consider students’ learning outside the overt curriculum, and ways to facilitate
such learning processes are suggested together with future research directions.
Keywords: higher education; situated learning; communities of practice;
assessment
Introduction
While teaching sessions in higher education are undoubtedly important, and have
been the focus of much learning research, they are not the sole location of student
learning at university. This article seeks to show the importance of students’ learning
outside the overt higher education curriculum (Havnes 2008). In so doing, there is
an emphasis on students’ community learning practices, and there is consideration
of the consequences of these student practices in terms of tutor assessment, feedback
and curriculum design.
Specifically, this article endeavours to contribute to the literature on student
learning and assessment in higher education by seeking to use situated learning the-
ory, with specific reference to learning in communities of practice, to: (1) consider
how the use of tutor constructed artefacts such as assessment marking criteria and
feedback may be understood; (2) evaluate the impact of student learning outside the
overt curriculum; and (3) consider the role of negotiated identity development in stu-
dent learning. The analysis is largely theoretical, but illustrative student quotes from
two previously published interview studies are provided, and the implications of the
findings for curriculum design are considered.
By way of making explicit the difference between tutors’ teaching and learning
practice and students’ situated learning, the English expression as different as chalk
and cheese will be employed. This expression means ‘Two things that are very dif-
ferent from each other’ (http://www.phrases.org.uk/meanings/chalk-and-cheese.
html), and reflects the very different approaches to learning considered by tutors and
students.
research with undue emphasis on what occurs in the classroom, and with learning as
an individual event. This is perhaps most recognisable through constructive align-
ment, a process that uses constructivist theory and draws on the deep/surface meta-
phor. The principal components of constructive alignment, which are designed to
encourage deep engagement, are outlined by Biggs (2003): (1) the curriculum that
we teach, (2) the teaching methods that we use and (3) the assessment procedures
that we use. An ‘imbalance in the system will lead to poor teaching and surface
learning’ (26). An example of non-alignment would be seen in terms of ‘practices
that contradict what we preach’. Within constructed alignment, ‘students are
entrapped in [a] web of consistency, optimising the likelihood that they will engage
the appropriate learning activities, but paradoxically leaving them free to construct
their knowledge their way’ (27).
The final phase of constructive alignment, assessment, is often seen as the most
powerful student learning tool available to higher education tutors. For example,
Gibbs (1999, 41) discusses how assessment can be used strategically as the most
‘powerful lever teachers have to influence the way students … behave as learners’.
It is noteworthy that assessment within higher education, particularly when high
stakes, is almost exclusively an individual process, and following such assessment
students are given primarily individual feedback designed to enhance their learning
(Poulos and Mahony 2008).
Overall, it would appear that student learning in higher education is often under-
stood in terms of a tutor constructed curriculum, with assessment and feedback facil-
itating cognitive development of the individual learner. Learning is seen in terms of
the deep/surface metaphor within a social constructivist learning theory, although
this may masquerade, particularly in terms of meeting outcomes, as a behaviourist
activity (Torrance 2012).
developing their learning strategies and professional identities. Since the higher
education curriculum is designed such that the bulk of student learning occurs out-
side formal taught sessions, student communities of practice would seem to be a key
facet of students’ development. It is therefore surprising that facilitating students’ sit-
uated learning within communities of practice has received so little attention when
higher education curricula are designed and implemented.
(1) Marking criteria are translated rather than transferred, and as such individual
students interpret and create their own reality as to how the criteria can be
used. Marking criteria enter student communities of practice and their mean-
ing is negotiated, through mutual engagement within the community. As
Wenger (1998, 80) discusses, the power that institutions have over the prac-
tice of communities is always mediated. It is ‘the community that negotiates
its enterprise’.
(2) Students may understand marking criteria in a particular way when they are
received, but as the students change as learners, the criteria may develop
new meanings.
294 P. Orsmond and S. Merry
These two factors may explain why the question ‘How much clarification of
marking criteria is required?’ is unanswerable. Meaning can be lost in translation
because each student is engaged, with varying degrees of success, in an ongoing
process of negotiating meaning. Thus, Sadler (2013) discusses how it is not possible
to tell what should or should not be noticed in different contexts, and Knight (2002,
280) noted that ‘criteria used prevent or even impede communication between
communities’. Torrance (2012) discusses the requirements for what he calls transfor-
mative assessment. He argues that most assessment is conformative, in that it
emphasises marking criteria compliance so making students aware of the what and
how of criteria. Transformative assessment requires additionally for students to be
made aware of the why of criteria, ‘so that they can critique, de-construct and reject
them, and understand that other criteria might be more appropriately invoked. If you
[students] wish to pursue a different learning agenda and can justify this divergence,
so be it.’ (338). Understanding the why of criteria is part of the negotiation process.
In transformative assessment, students are encouraged to make judgements; an attri-
bute which has been an aim of higher education for over 60 years (Rogers 2003).
Hussey and Smith (2002) have previously expressed the opinion that learning out-
comes have been misappropriated as measurement of higher education teaching. A
similar argument might be made for marking criteria in terms of misappropriating
learning in higher education.
A characteristic of transformative assessment is that it does require tutors to
make more holistic judgements concerning the quality of students’ work, and such
judgements can be perceived as less robust and, possibly, more norm-referenced
than judgements based on specific individual criteria (Sadler 2009). The need for
robust measurement of student learning in higher education is acknowledged, but, as
Sadler discusses, marking criteria designed to assess complex works are often, in
reality, interdependent rather than distinct and may themselves be interpreted in dif-
ferent ways by different tutors. Sadler (2009) concludes that properly done holistic
appraisals can generate a truer representation of students’ performance than preset
criteria, so the empirical study of Bloxham, Boyd, and Orr (2011) is reassuring.
They found that a ‘high proportion of the tutors did not make use of written criteria
in their marking and, where they were used, it was largely a post hoc process in
refining, checking or justifying a holistic decision’. Clearly, more work is required
in this area, but further discussion is beyond the scope of this paper.
That knowledge can only be learnt through a student’s own experiences within a
situated learning environment, and meaning gained through negotiated practices,
would be a pathway towards self-monitoring; an important goal of higher education
learning (Boud and Associates 2010). This is an important consideration because
there has been much debate over marking criteria and their meaning, as if meaning
can be directly transferred to students or sufficient clarification can ever be given.
Students do undertake tasks which they know are self-generated and hence situated
within their own learning environment. This is illustrated by a quote from a high-
achieving student (Orsmond and Merry 2013, 744), who seems to have made a
judgement concerning their work without reference to tutor provided criteria.
If you take a piece of coursework that I did in the first year and compare it to a piece,
like an essay, that [I’ve] written this year [Year 3], you’d see a total change in [my]
style of writing and that is not something that’s been instilled in me by anyone else,
that is something that I’ve acquired.
Assessment & Evaluation in Higher Education 295
this context, it is interesting to note that Nicol, Thomson, and Breslin (2014),
writing from a different perspective, also recognised the need to strengthen ‘inner
feedback processes’.
data, with individual learning considered via the deep/surface metaphor, it is clear
that situated learning occurred. Students in the experimental group recognised that
to succeed they needed to work and talk together, and the language used is very
much of a community nature: ‘We still had problems’, ‘We still had moments of
doubt’, ‘We still felt sometimes we were walking up the down escalator, but I think
We all knew that We had something that worked and that was a start’ (226). This
social learning process carried on until the end of the year: ‘We continued in trios
and We spent time discussing each stage of the assessment’ (226). Our emphases
are in bold and capitalised.
McCune and Hounsell (2005) investigated final-year biology undergraduate stu-
dents’ grasp of the distinctive ways of thinking and practising (WTP) in the disci-
pline. From this study, it was evident that ‘WTP was most clearly evident
in situations in which students were pursuing active rather than passive learning
tasks’ (285). McCune and Hounsell interpreted their findings in terms of construc-
tive alignment and, in particular, Biggs (2003) level-three thinking about teaching
which considers: (a) what it means to understand content in the way we want it to
be understood, and (b) what kind of teaching/learning activities are required to reach
those kinds of understanding. Apart from the dilemma that we, the community of
tutors, may not have a common understanding, it has already been noted that stu-
dents are very capable of working within the paradigm of what tutors’ want, and
this may not be what they, the students, really think themselves. When considered in
terms of social learning, Boyd and Cowan and McCune and Hounsell provide evi-
dence that students utilise ways of learning outside the overt curriculum that are not
acknowledged or explored. This has important consequences.
While the argument can be made that tutors’ assessment can influence learning,
the study of Boyd and Cowan showed that the development of self-assessment pro-
cesses are crucial to changing approaches to learning. It is apparent that there are
more fundamental factors than assessment in bringing about a change in learning; a
key factor being the context of learning, the negotiation of meaning and identity
development. Self-assessment can change the learning context because it requires a
reappraisal of what it means to learn, and then to find a direction and ultimately to
change practice. Boyd was one of the student participants in the experimental group
of the study described by Boyd and Cowan, and when the study was complete she
returned to conventional teaching and assessments. She retained her commitment to
managing her own learning only when she ‘opted to work on her own criteria (her
context) rather than [that of] her supervisors (institutional context)’ (234). Our inter-
pretation is given in brackets.
Some of them [tutors] actually put it [criteria] on a cover sheet on the front and their
[standards] are circled.
Looking back over previous grades and what comments they made about it, I kind of
got in my mind what level I think my essays are generally at.
Learning implies becoming a different person. It involves, for students, the construc-
tion of professional identities that is ‘ways of being and relating in a professional
context’ (Goldie 2012, 641). McCune and Hounsell (2005, 267) well illustrate this
notion of becoming within active student learning activities. Students undertaking
placements stated: ‘You’ve gone up a level, you’re not a student anymore’ and ‘the
placement has definitely taught me to think more like a scientist’. Orsmond, Merry
and Callaghan (2013) show how, for students, a sense of identity can result from
discussions in social learning networks: ‘I felt really like a scientist … I was talking
to a family member about my dissertation and they seemed really interested … I
was going into … depth’ (899), or from communities of practice: ‘I like to talk to
different types of groups [to find out what they know] and if they’re not interested
I’ll try and change my personality … to fit’ (896). It seems that students’ identity
can develop as a result of a range of experiences.
Students’ developing identities can have broad effects on their approaches to the
curriculum and academic learning, as indicated by a student is responding to tutor
feedback and making their own judgement as to its worth in terms of how she per-
ceived herself as a creative writer:
feedback on … a writing style, I do take it, and superimpose it onto my creative side
but, obviously that doesn’t quite work … the academic and creative … If you’re aim-
ing to be a research scientist who’s going to be putting out papers and publications …
then I think you’re more likely to … take the feedback on an essay [academic writing]
… whereas for me, personally, I don’t, really have any, intentions to be writing papers.
(Orsmond and Merry, unpublished work)
Brown, Bull, and Pendlebury (1997, 7) claim that ‘assessment defines how they
[students] come to see themselves as students’. At onelevel assessment does define
students, for example, as being able to achieve or not in terms of passing
examinations. However, students define themselves in other ways in the absence of
assessment. If tutors fail to recognise this, they will not consider the link between
identity and learning in their curriculum design. When students enrol onto a degree
programme they are enrolled on a specific degree programme, for example, Animal
Biology. That is a statement of identity. Even if students do not fully understand
what an animal biologist is, they have identified that as their area of study, and
tutors should therefore design a curriculum that allows development into an animal
biologist; that is that allows students to become members of the community of
animal biological practice, and learning to talk the language of animal biology is
key to that. For Lave and Wenger (1991), learners cannot learn from talk as a
replacement for legitimate peripheral participation, they learn to talk as a key com-
ponent of legitimate peripheral participation. Students do not always understand
their progression through three years of study in terms of identity development and
the notion of becoming. While Boud (1988) is correct, in making the link between
student responsibility for their own learning and their use of prior experiences and
knowledge, it might be more helpful rather than considering experiences to think in
terms of identity development.
Assessment & Evaluation in Higher Education 299
Packer and Goicoechea (2000, 235) consider the shift from family member to
student as an ontological transformation. The new individual does not replace the
old one, but ‘assumes different modes’ of subjectivity in a different context. This is
reflected within the situated learning experience within higher education in terms of
changing community practice and social learning networks; students are recognised
as subject experts by family and friends and identity is given to them (Orsmond,
Merry, and Callaghan 2013). Thus, identity is a more powerful force in
understanding learning than experiences per se.
Acknowledgements
The authors would like to thank both professor Harry Torrance and Dr Karen Handley for
their useful discussions concerning drafts of the manuscript. Thanks are also due to the two
anonymous referees for their helpful comments on our initial submission.
Disclosure statement
The authors declare that they have no competing interests.
Assessment & Evaluation in Higher Education 301
Notes on contributors
Paul Orsmond is a senior lecturer within the Faculty of Computing, Engineering and Sciences
at Staffordshire University. His interests are in learning processes.
Stephen Merry is a visiting research fellow within the Faculty of Computing, Engineering
and Sciences at Staffordshire University. His interests include a) the use of cell culture to
investigate disease processes and b) how students use feedback to develop their learning.
References
Adams, K., S. Hean, P. Sturgis, and J. M. Clark. 2006. “Investigating the Factors Influencing
Professional Identity of First Year Health and Social Care Students.” Learning in Health
and Social Care 5: 55–68.
Baker, G. P., and P. M. S. Hacker. 1984. An Analytical Commentary on Wittgenstein’s
Philosophical Investigation. Oxford: Blackwell.
Barnacle, R., and I. Mewburn. 2010. “Learning Networks and the Journey of ‘Becoming
Doctor’.” Studies in Higher Education 35: 433–444.
Biggs, J. 2003. Teaching for Quality Learning at University. Maidenhead: Open University
Press.
Bloxham, S., P. Boyd, and S. Orr. 2011. “Mark My Words: The Role of Assessment Criteria
in UK Higher Education Grading Practices.” Studies in Higher Education 36: 655–670.
Boud, D. 1988. “Moving towards Autonomy.” In Developing Student Autonomy in Learning,
edited by D. Boud, 17–39. London: Kogan Page.
Boud, D. 1995. “How Does Self-Assessment Relate to Ideas about Learning?” In Enhancing
Learning through Self Assessment, edited by D. Boud, 24–35. London: Kogan.
Boud, D., and Associates. 2010. Assessment 2020: Seven Propositions for Assessment
Reform in Higher Education. Sydney: Australian Learning and Teaching Council.
Boyd, H., and J. Cowan. 1985. “A Case for Self-Assessment Based on Recent Studies of
Student Learning.” Assessment & Evaluation in Higher Education 10: 225–235.
Brown, G., J. Bull, and M. Pendlebury. 1997. Assessing Student Learning in Higher
Education. London: Routledge.
Cowan, J. 2010. “Developing the Ability for Making Evaluative Judgements.” Teaching in
Higher Education 15: 323–334.
Draper, S., and D. Nicol. 2013. “Achieving Transformational or Sustainable Educational
Change.” In Reconceptualising Feedback in Higher Education, edited by S. Merry, M.
Price, D. Carless and M. Taras, 190–203. London: Routledge.
Durning, S. J., and A. R. Artino. 2011. “Situativity Theory: A Perspective on How
Participants and the Environment Can Interact: AMEE Guide No 52.” Medical Teacher
33: 188–199.
Elsbach, K., P. S. Barr, and A. B. Hargadon. 2005. “Identifying Situated Cognition in
Organisations.” Organization Science 16: 422–433.
Entwistle, N., and F. Marton. 1994. “Knowledge Objects: Understandings Constituted
through Intensive Academic Study.” British Journal of Educational Psychology 64:
161–178.
Fraser, S. P., and A. M. Bosanquet. 2006. “The Curriculum? That’s Just a Unit Outline, Isn’t
It?” Studies in Higher Education 31: 269–284.
Gibbs, G. 1999. “Using Assessment Strategically to Change the Way Students Learn.” In
Assessment Matters in Higher Education, edited by S. Brown and A. Glasner, 41–53.
Buckingham: Open University Press.
Goldie, J. 2012. “The Formation of Professional Identity in Medical Students: Considerations
for Educators.” Medical Teacher 34: e641–e648.
Havnes, A. 2008. “Peer-Mediated Learning beyond the Curriculum.” Studies in Higher
Education 53: 194–204.
Hedegaard, M. 1998. “Situated Learning and Cognition: Theoretical Learning and
Cognition.” Mind, Culture and Activity 5: 114–126.
302 P. Orsmond and S. Merry
Hendry, G. D., and J. Anderson. 2013. “Helping Students Understand the Standards of Work
Expected in an Essay: Using Exemplars in Mathematics Pre-Service Education Classes.”
Assessment & Evaluation in Higher Education 38: 754–768.
Hughes, G. 2010. “Identity and Belonging in Social Learning Groups: The Importance of
Distinguishing Social, Operational and Knowledge-Related Identity Congruence.” British
Educational Research Journal 36: 47–63.
Hughes, G. 2011. “Towards a Personal Best: A Case for Introducing Ipsative Assessment in
Higher Education.” Studies in Higher Education 36: 353–367.
Hussey, T., and P. Smith. 2002. “The Trouble with Learning Outcomes.” Active Learning in
Higher Education 3: 202–233.
Ibarra, H. 1999. “Provisional Selves: Experimenting with Image and Identity in Professional
Adaptation.” Administrative Science Quarterly 44: 764–791.
Knight, P. 2002. “Summative Assessment in Higher Education: Practices in Disarray.”
Studies in Higher Education 27: 275–286.
Lave, J., and E. Wenger. 1991. Situated Learning: Legitimate Peripheral Participation.
Cambridge: Cambridge University Press.
Marton, F. 1981. “Phenomenography – Describing Conceptions of the World around Us.”
Instructional Science 10: 177–200.
Marton, F., and R. Säljö. 1976. “On Qualitative Differences in Learning: I – Outcome and
Process*.” British Journal of Educational Psychology 46: 4–11.
McCune, V., and D. Hounsell. 2005. “The Development of Students’ Ways of Thinking and
Practising in Three Final-Year Biology Courses.” Higher Education 49: 255–289.
Miller, C. M. L., and M. Parlett. 1974. Up to the Mark: A Study of the Examination Game.
London: Society for Research into Higher Education.
Nicol, D., A. Thomson, and C. Breslin. 2014. “Rethinking Feedback Practices in Higher
Education: A Peer Review Perspective.” Assessment & Evaluation in Higher Education
39: 102–122.
Orsmond, P., and S. Merry. 2013. “The Importance of Self-Assessment in Students’ Use of
Tutors’ Feedback: A Qualitative Study of High and Non-High Achieving Biology
Undergraduates.” Assessment & Evaluation in Higher Education 38: 737–753.
Orsmond, P., S. Merry, and A. Callaghan. 2013. “Communities of Practice and Ways to
Learning: Charting the Progress of Biology Undergraduates.” Studies in Higher
Education 38: 890–906.
Orsmond, P., S. Merry, and K. Handley. 2013. “Students’ Social Learning Practice as a Way
of Learning from Tutor Feedback.” In Reconceptualising Feedback in Higher Education,
edited by S. Merry, M. Price, D. Carless and M. Taras, 123–132. London: Routledge.
Orsmond, P., and R. Zvauya. 2015. “Community of Learners: Charting Learning in First Year
Graduate Entry Medical Students during Problem-Based Learning (PBL) Study.”
Advances in Health Sciences Education 20: 479–497.
Packer, M., and J. Goicoechea. 2000. “Sociocultural and Constructivist Theories of Learning:
Ontology, Not Just Epistemology.” Educational Psychologist 35: 227–241.
Poulos, A., and M. J. Mahony. 2008. “Effectiveness of Feedback: The Students’
Perspective.” Assessment & Evaluation in Higher Education 33: 143–154.
Rogers, C. 2003. Client Centred Therapy. London: Constable and Robinson.
Rommetveit, R. 2003. “On the Role of ‘a Psychology of the Second Person’ in Studies of
Meaning, Language and Mind.” Mind, Culture and Activity 10: 205–218.
Sadler, D. R. 2009. “Indeterminacy in the Use of Preset Criteria for Assessment and
Grading.” Assessment & Evaluation in Higher Education 34: 159–179.
Sadler, D. R. 2013. “Opening up Feedback: Teaching Learners to See.” In Reconceptualising
Feedback in Higher Education, edited by S. Merry, M. Price, D. Carless and M. Taras,
54–63. London: Routledge.
Torrance, H. 2012. “Formative Assessment at the Crossroads: Conformative, Deformative
and Transformative Assessment.” Oxford Review of Education 38: 323–342.
Trigwell, K., and M. Prosser. 1996. “Congruence between Intention and Strategy in
University Science Teachers’ Approaches to Teaching.” Higher Education 32: 77–87.
Trigwell, K., M. Prosser, and F. Waterhouse. 1999. “Relations between Teachers’ Approaches
to Teaching and Students’ Approaches to Learning.” Higher Education 37: 57–70.
Assessment & Evaluation in Higher Education 303