You are on page 1of 12

Medical Teacher

ISSN: 0142-159X (Print) 1466-187X (Online) Journal homepage: http://www.tandfonline.com/loi/imte20

Competency milestones for medical students:


Design, implementation, and analysis at one
medical school

Kimberly D. Lomis, Regina G. Russell, Mario A. Davidson, Amy E. Fleming,


Cathleen C. Pettepher, William B. Cutrer, Geoffrey M. Fleming & Bonnie M.
Miller

To cite this article: Kimberly D. Lomis, Regina G. Russell, Mario A. Davidson, Amy E. Fleming,
Cathleen C. Pettepher, William B. Cutrer, Geoffrey M. Fleming & Bonnie M. Miller (2017)
Competency milestones for medical students: Design, implementation, and analysis at one medical
school, Medical Teacher, 39:5, 494-504, DOI: 10.1080/0142159X.2017.1299924

To link to this article: http://dx.doi.org/10.1080/0142159X.2017.1299924

View supplementary material

Published online: 10 Mar 2017.

Submit your article to this journal

Article views: 501

View related articles

View Crossmark data

Full Terms & Conditions of access and use can be found at


http://www.tandfonline.com/action/journalInformation?journalCode=imte20

Download by: [New York University] Date: 08 June 2017, At: 06:40
MEDICAL TEACHER, 2017
VOL. 39, NO. 5, 494–504
http://dx.doi.org/10.1080/0142159X.2017.1299924

Competency milestones for medical students: Design, implementation, and


analysis at one medical school
Kimberly D. Lomisa, Regina G. Russella, Mario A. Davidsonb, Amy E. Flemingc, Cathleen C. Pettephera,
William B. Cutrerd, Geoffrey M. Flemingd and Bonnie M. Millere
a
Office of Undergraduate Medical Education, Vanderbilt University School of Medicine, Nashville, TN, USA; bDepartment of Biostatistics,
Vanderbilt University School of Medicine, Nashville, TN, USA; cOffice of Medical Student Affairs, Vanderbilt University School of Medicine,
Nashville, TN, USA; dDivision of Pediatric Critical Care, Vanderbilt University Medical Center, Nashville, TN, USA; eOffice of Health
Sciences Education, Vanderbilt University School of Medicine, Nashville, TN, USA

ABSTRACT
Competency-based assessment seeks to align measures of performance directly with desired learning outcomes based upon
the needs of patients and the healthcare system. Recognizing that assessment methods profoundly influence student motiv-
ation and effort, it is critical to measure all desired aspects of performance throughout an individual’s medical training. The
Accreditation Council for Graduate Medical Education (ACGME) defined domains of competency for residency; the subse-
quent Milestones Project seeks to describe each learner’s progress toward competence within each domain. Because the
various clinical disciplines defined unique competencies and milestones within each domain, it is difficult for undergraduate
medical education to adopt existing GME milestones language. This paper outlines the process undertaken by one medical
school to design, implement and improve competency milestones for medical students. A team of assessment experts devel-
oped milestones for a set of focus competencies; these have now been monitored in medical students over two years.
A unique digital dashboard enables individual, aggregate and longitudinal views of student progress by domain. Validation
and continuous quality improvement cycles are based upon expert review, user feedback, and analysis of variation between
students and between assessors. Experience to date indicates that milestone-based assessment has significant potential to
guide the development of medical students.

Introduction
Competency-based medical education seeks to directly link Practice points
the outcomes of a learning program with the needs of  An assessment program should be designed to
patients (Frank et al. 2010a). Competencies are trainable promote all desired learning outcomes.
attributes of an individual that must be developed in order  Milestones serve as a standardized articulation of
to successfully perform professional duties, and these are expectations, providing clarity for both assessors
often grouped within domains encompassing similar skills. and learners, and are useful to describe learner
This manuscript draws upon the common taxonomy of development across competency domains.
competencies recommended by Englander et al. (2013).  A portfolio approach involving professional judg-
Milestones describe the typical developmental pathway for ment in the review of performance evidence miti-
gates against the multiple confounders of a
a given competency (Nasca et al. 2012). Milestones are
complex system.
expressed using behavioral anchors that describe escalating
 To advance students’ future learning efforts,
levels of performance from novice (“entry”) to aspirational
meaningful assessment evidence from multiple
(Dreyfus 2004) (see Figure 1). These milestones enable the
settings should be integrated and targeted learn-
learner to recognize current status and to understand what
ing opportunities should be provided.
behaviors are necessary to achieve higher levels of per-
 Extensive collaboration and training, as well as
formance (Schumacher et al. 2013; Swing et al. 2013). continual critical appraisal of outcomes, is neces-
Milestones also provide guidance to assessors, generating a sary to successfully implement a competency-
shared mental model across disciplines regarding the based assessment system.
knowledge, skills and behaviors expected of learners.
Competency-based assessment does not seek to ensure
only minimal competence, but rather promotes each
learner’s trajectory toward excellence across multiple development throughout one’s medical career. The incorp-
domains of performance (Carraccio & Englander 2013; oration of more active and team-based learning activities
Frank et al. 2010b). It is advantageous to introduce stu- in the pre-clerkship curriculum enables demonstration
dents to the competency domains and milestones early in of domains of competency that were previously under-
training and to emphasize the need for continual emphasized at early stages of training. As a result, the

CONTACT Kimberly D. Lomis kim.lomis@vanderbilt.edu Office of Undergraduate Medical Education, Vanderbilt University School of Medicine, 201
Light Hall, Nashville TN 37232, USA
Supplemental data for this article can be accessed here.
ß 2017 Informa UK Limited, trading as Taylor & Francis Group
MEDICAL TEACHER 495

Figure 1. Example of milestone rating form and display.

competency domains are now relevant throughout all of Design


undergraduate medical education and align with expecta-
Beginning in 2012, a committee of faculty members with
tions in graduate medical education and beyond (Anderson
experience teaching learners of all levels began drafting
2012).
milestones specific to medical students. To focus our initial
To foster learner development, the assessment process
implementation, the team defined core behaviors that
must inform future learning. The framework of milestones
should be continually re-assessed throughout a student’s
is best utilized to describe competencies that develop over
training. This selection was based upon faculty judgment of
time and must be demonstrated repeatedly throughout
relative importance; priorities represented in multiple exist-
one’s career in increasingly complex settings (Krupat &
ing assessment forms; consideration of areas in which stu-
Pelletier 2016). Traditional normative assessments are lim-
dents historically struggle; and areas amenable to
ited in this regard; performance ratings relative to one’s
measurement starting in the first year. A modified Delphi
peers provide little guidance about specific behaviors that
technique (Clayton 1997) was employed to create consen-
are desired or areas for improvement. Milestones are criter-
sus around priority competencies.
ion-referenced, offering a common set of performance
Work groups for each domain of competency were pro-
benchmarks applicable across multiple courses and set-
vided background materials, including internal grading
tings. This enhances the identification of trends, providing
forms from multiple courses and graduate medical educa-
stronger evidence to prompt the student to take action.
tion (GME) milestones published at that time (medicine,
Our competency milestones were designed to continually
pediatrics, and emergency medicine). Entry-level GME
populate each learner’s digital portfolio (Chertoff et al.
expectations informed undergraduate medical education
2016). This performance evidence supports a cyclical
(UME) graduation targets (Swing 2007).
informed self-assessment process in which learners, with
The following guidelines were provided to milestone
coaches, identify relative strengths and weaknesses and
writing teams:
create appropriate learning plans.

 Create behavioral anchors that describe readily observ-


able actions (Regehr et al. 2012)
Methods
 Avoid normative language: “better than average” or “as
Design and implementation involved several steps. An expected for level”
existing assessment committee was tasked with drafting  Avoid vague language: “good” or “fairly accurate”
and improving milestones language. Initial implementation  Create five (5) anchors for all competencies
focused on two cohorts of assessors who were trained  The first level should reflect expected behaviors at entry
extensively, were given protected time for student assess- to medical school; fifth is aspirational (high-performing
ment, and shared experiences at monthly faculty develop- intern at 3 months)
ment meetings. An assessment sub-committee was charged  Lower-level anchors should not simply describe the
with the continual review of outcomes from the milestones absence of behaviors associated with higher levels, but
system using both quantitative and qualitative approaches. rather should describe typical early behaviors as part of
496 K. D. LOMIS ET AL.

a normal developmental progression (Holmboe & A customized assessment form was generated for each
Batalden 2015). course, including metadata regarding contextual informa-
tion such as setting, assessor, and level of difficulty.
Anticipating the potential need for promotions commit- Assessments were captured in a digital learning portfolio
tees to render definitive decisions regarding unacceptable designed to facilitate the collation and visualization of data
performance trends in some students, a separate process (Spickard et al. 2016).
later defined undesirable behaviors as distinct from devel- A developmental framework of assessment is intended
opmentally expected, lower-level behaviors. to direct future learning; therefore, the process must have
Several iterative rounds of editing were completed. meaningful consequences for the student. The milestones
Work groups critiqued each other’s language and chal- serve as one source of performance evidence supporting a
lenged assumptions about typical developmental steps. cyclical, faculty-facilitated self-assessment process. At pre-
Stakeholder input from course directors, potential assessors, scribed intervals, the student reviews evidence with a port-
and students was solicited at multiple points. folio coach to identify relative strengths and weaknesses
Early pilots were executed to enhance construct align- and develop appropriate personalized learning plans.
ment (Crossley et al. 2011). Based upon feedback, the Coaches support students in setting and monitoring goals
assessment committee made changes to the wording of to improve across competency domains.
some anchors. An option to indicate an intermediate level The student promotions process was redesigned to
of performance between anchors was added, indicating include the review of competency milestone evidence in
that the student intermittently displayed the higher level. addition to course grades. All domains are tracked from the
The initial layout of the milestone form for each compe- beginning of a student’s medical education. By using the
tency mimicked the charts used in GME, which include mul- same metrics across courses, trends in performance can be
tiple behaviors within each level of performance (Swing identified more readily. Given the context-specificity of
et al. 2013). Assessors found that some of the grouped ele- many skills (Linn et al. 1991), we did not envision that each
ments develop independently; therefore, each competency learner would make linear progress in each competency;
was separated into more discrete units. rather we expected fluctuating performance given various
Having completed these refinements, the assessment settings and complexities, with general improvement over
committee resolved to avoid editing the milestones lan- time (see Figure 2). Multiple milestone data points can pro-
guage for two years after implementation. Accumulated vide a signal about areas of excellence and areas for
outcomes data would provide insight into the performance improvement, but no single rating event is considered high
of the milestones for each competency, which would guide stakes for a student. There was deliberate avoidance of cal-
subsequent revisions. culating average scores. Indeed, numeric reference to mile-
stone levels within the digital platform was discouraged,
with the intent to focus on the words in each behavioral
Implementation
anchor. It was deemed essential that the student and coach
The competency milestones were implemented with the interpret the relative merits and import of each data point.
onset of a new curriculum in 2013. Implementation was ini- Trends identified after such critical review did trigger action
tially limited to two faculty assessor groups with existing by the promotion committee; most commonly this involved
responsibility for direct observation of students. These fac- a targeted learning plan directed at the competency
ulty cohorts had protected time for student interactions domain of concern.
and attended standing monthly faculty development meet-
ings. One cohort was based in the pre-clerkship phase as
Analysis
facilitators in case-based learning sessions (SGFs) and spent
6 hours per week in a classroom setting with rotating The milestones are intended to measure medical student
groups of 8 first year students. The other assessor cohort, performance in core competency domains and to provide
master clinical teachers (MCTs), was based in the clerkship developmentally-based feedback. It is critical that the mile-
phase to provide direct observation in the workplace stones represent the domains of interest and that the
regarding skills in history, physical exam, and clinical rea- anchors accurately describe the developmental trajectory of
soning. Each student spent at least 12 hours interacting these learners. The assessment tool should be “fit for
with various MCTs over the clerkship year. purpose,” valid for the intended application (van der
Students were introduced to the new assessment pro- Vleuten & Schuwirth 2005). Since individual performance
cess in a pre-matriculation online course, and received add- does not occur in a vacuum, there is inherent complexity in
itional training at orientation, in town halls, and during assessing competency development. Ongoing variations in
practice sessions. First year students were trained to use the environment, context, and level of difficulty will result
the milestones for peer evaluation, which served to solidify in fluctuations of performance ratings for an individual; this
the student awareness of competency expectations. creates challenges for validation of the milestones.
Each course or program was asked to select only those Quality improvement and validation of the milestones
competencies most applicable to the specific setting. No has been an iterative process involving evaluation of con-
course was expected to utilize all of the established mile- tent, the impact of rater variation, and qualitative evidence
stones. Course directors were not, however, permitted to of feasibility and user satisfaction (Hicks et al. 2010a, 2010b;
change the milestone anchor language; when a given com- Englander et al. 2012). Messick’s (1989) approach to validity
petency was being evaluated in different settings, the criteria for educational testing provided guidance to these
standardized performance descriptors were maintained. efforts. He conceptualized validity as a holistic summary of
MEDICAL TEACHER 497

Figure 2. One student’s performance evidence for multiple competencies in one domain displayed in the digital portfolio.

the evidence for using a particular test as well as the inter- understanding of the best approaches to ensure high-qual-
pretations of test scores relative to possible consequences. ity measurements of the intended constructs and appropri-
Messick reminds us that validity claims must engage the ate, meaningful interpretation of performance data.
entire assessment process, not only the item structure. Incorporating insights from traditional psychometric and
In the “post-psychometric era” (Hodges 2013), medical post-psychometric views, our data collection and analysis
education assessment experts are much more aware of the follows a holistic approach. We have incorporated both
“conceptual and practical challenges” (Whitehead et al. quantitative and qualitative methods to investigate the
2015) in the measurement of competencies and validation impact of milestone ratings in the complex system of med-
of performance rating systems in unstandardized, real- ical education (Baartman et al. 2007). Quantitative analysis
world contexts (Boursicot et al. 2011). Approaches that rec- of student performance on the assessments as well as
ognize the importance (and statistically confounding informal focus groups and surveys were used to assess the
nature) of the “patient-by-doctor-by-context interactions” impact and function of the milestones language.
(van der Vleuten 2014) are critical to understanding the Descriptive statistics were calculated, and graphic visual-
development of complex competencies such as communi- izations were created to consider the variability in the
cation, professionalism, and systems-based practice: aggregate distribution of milestone ratings for each compe-
tency in relation to the medians. These data were consid-
The new understanding that we are currently documenting is
that assessors are not passive, perfectly calibrated measurement
ered longitudinally, and from the perspective of individual
instruments, but active agents constructing judgments using students as well as individual assessors. During this explora-
information from situational factors and personal experiences. tory analysis of the milestones, statistical inference and
From this perspective, different assessors are not expected to analysis was not conducted. Variability was the main con-
make similar judgments, and this may actually be desirable!” sideration in investigating rater discrimination and was indi-
(pg. 235)
cated by reviewing the interquartile range. Atypical
Downing and Haladyna (2004) use a medical education patterns in the distributions of milestone levels assigned,
context to highlight the two most significant threats to such as minimal variability or extreme outliers, were
construct validity outlined by Messick: construct under-rep- explored in discussions with raters. Stakeholder input for
resentation and construct irrelevant variance. Rater vari- milestone improvement was driven by quantitative evi-
ation is often categorized as “construct irrelevant variance.” dence and remains critical for the continued development
However, the authors note statistical difficulties in differen- and effective use of the new milestone language.
tiating rater error from rater variation that is appropriate to
particular situations and contexts. Additionally, new
Results
approaches to assessing competencies through big data
and learning analytics (Pecaric et al. 2017) and qualitative Medical students enrolled from 2013 to 2015 were
methods from the social and behavioral sciences assessed using the milestone framework. First year students
(Whitehead et al. 2015) hold promise for improving our underwent approximately 36 milestone-based primary
498 K. D. LOMIS ET AL.

Figure 3. Distribution of all students in one phase of training on one competency.

Figure 4. Student performances as compared to predicted benchmark for one competency.

assessments from peers and faculty involving 3–18 compe- demonstrate a mixture in milestone levels assigned (see
tencies depending on context. In the clerkship phase (year Figure 3). This suggests that the milestones elucidate differ-
2), students received approximately 14 milestone-based ences in development among students.
assessments from faculty, with 13–20 competencies repre- The committee that drafted the milestones predicted
sented. In 2015, the assessment committee conducted an levels of performance that students at various stages of
intensive review of the milestone results at the level of training would attain. For some competencies, early stu-
individual students, individual assessors, and aggregate dents reached higher levels of performance than predicted
results within and across levels of training. Qualitative feed- (first year students were assessed in a controlled classroom
back was collected from all users (students, faculty, and setting for which they had significant time to prepare).
administrators). Aggregate data did confirm, however, that performances
below predicted benchmarks were low as compared to the
peer group (see Figure 4).
Validity evidence
Cross-sectional comparisons of students at different
Several trends in the data affirm the validity of the stages of training showed an anticipated shift to higher
milestones. milestone achievements as learners advance temporally
Aggregate data regarding the performance of learners at (see Figure 5). This is suggestive that the milestone anchors
the same level of training for each competency do describe developmental progression.
MEDICAL TEACHER 499

Figure 5. Cross-sectional view of two different classes on one competency shows generally higher ratings for the more senior class.

Figure 6. Example of a faculty member’s ratings of a single student in an observed patient encounter regarding specific aspects of the history and physical
exam.

Review of the performance ratings for individual stu- multiple competencies, very few assessors were identified
dents across competencies revealed variations; a student as consistently grading low or high across all competencies.
does not receive similar performance ratings in all compe- Box plots demonstrate that most raters applied a mix of
tencies (see Figure 6). This indicates that milestone-based milestones levels to describe the performance of a group
feedback is discriminating for discrete aspects of of students in a given competency. Additionally, each
performance. assessor used multiple milestone levels to describe each
individual student’s performance, indicating that the mile-
stone language is discriminating for discrete aspects of
Rater variation
performance.
It is possible that the aggregate milestones distributions
could be attributable to rater effect rather than true differ-
Peer assessors
ences in student performance; therefore, variation in rater
Student ratings of peer performance also show a diversity
results was investigated.
of milestone levels when describing a given student’s per-
formance. This indicates that the milestone language is dis-
Faculty assessors criminating for discrete aspects of performance.
Box plots and bar graphs were created for each compe- In the pre-clerkship small groups, both faculty and peers
tency to show the ratings used by all assessors within an rate students based upon interactions in the case-based
assessor cohort (see Figure 7). Upon reviewing this for learning sessions. The aggregate distribution of milestone
500 K. D. LOMIS ET AL.

Ratings by Master Clinical Teacher for Interpersonal and


Communication Skills: Rapport with Patients and Families
Aspirational
Milestone Level

Entry

Undesirable

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17

Master Clinical Teacher ID

Figure 7. Example of comparative assessor milestone ratings data provided to Master Clinical Teachers as part of faculty development.

Figure 8. Distribution of class milestones ratings by faculty and by peers for one competency.

levels from peer and faculty evaluations is generally similar, scale to questions regarding user understanding of mile-
indicating a shared interpretation of the milestone lan- stones and their influence on learner development trended
guage (see Figure 8). For certain competencies, some differ- in the range of neutral to agree (see Table 1). It is apparent
ences between faculty and peer patterns occur; review of from comments that some students had a positive experi-
associated qualitative comments indicates that the students ence using milestone evidence in the portfolio process,
have unique insights into these aspects of performance while others struggled to make sense of the feedback and
based upon interaction outside of class. use it for improvement. In addition to potential limitations
Each student is evaluated by two peers at each review of the milestone language and lack of familiarity with this
point. Generally, milestone levels align between the two novel criterion-based approach, this variation in meaning-
raters, or differences are clarified via comments. In one situ- fulness may have been influenced by characteristics of the
ation of a known personal conflict between two peers, student, the portfolio coach, and their interaction.
these students limited the use of low performance ratings Formal focus groups were conducted with students,
to the specific competency of concern and rated each portfolio coaches, and promotions committee members.
other more highly in other aspects of performance. Results were similarly mixed across participants: some
found the milestone feedback important to student devel-
User feedback opment, others doubted its effectiveness and/or found it
Rising second and third-year medical students who overly cumbersome. Students expressed concerns about
received milestone-based performance feedback were sur- the quantification of milestone-based feedback and
veyed in September 2015 regarding the assessment system whether “scores” were a valid measure of different perform-
and the portfolio review process. Responses on a five-point ance dimensions.
MEDICAL TEACHER 501

Table 1. Sample of survey results from students using milestones.


Feedback from Medical Students About Competency Milestones
1-strongly disagree, 2-disagree, 3-neutral, 4-agree, S-strongly agree
183 survey respondents
% SD þ D % Neutral % A þ SA Median
I understand the milestones language. 21% 26% 53% 4
Feedback using milestones helps me track my development. 23% 37% 40% 3
Selected Student Comments:
 I think it is good to be cognizant of your strengths and weaknesses, and the portfolio process has helped create an environment in which you strive
to improve as well as to foster humility. 1 think it is wonderful to have a portfolio coach- mine has helped me get in contact with mentors, has good
study tips, is a good point of contact for questions, and 1 think genuinely cares about my success.
 The portfolio process has really allowed me to see how my peers and faculty in nontraditional roles (i.e. not directly providing grades) view my per-
formance. This has been helpful in changing the way 1 interact with groups and growing as a student.
 The milestones and ratings seem very arbitrary and there is not enough variation in the middle ratings. The low scores are very low and almost
unrealistic, it seems like there should be more middle ground.
 The portfolio milestone system has many advantages and a lot of potential. However, it currently struggles from evaluators having “preconceived”
notions of where a student is supposed to fall on the number scale at a current point in his or her education such that growth can be demonstrated.

Standing faculty development meetings were used to In the past, a specific performance issue may have been
collect qualitative feedback from faculty assessor groups. overlooked because there were insufficient data within the
Feasibility concerns included the total length of the assess- context of any given course to prompt remediation. The
ment forms and the time needed for completion. This issue ability to collate measures across settings allows identifica-
improved as assessors became more familiar with the mile- tion of trends in performance indicating a need for
stone language, and as the digital interface was refined. improvement. Performance challenges in professionalism,
Specific milestone anchor language was also a subject of interpersonal communication skills, and systems-based
discussion among assessors. practice have been identified for some students in the pre-
Faculty members noted that the quality of peer feed- clerkship phase, enabling coaching and improvement
back improved compared to prior models at our institution. before these students progressed into clerkships.
The milestone anchors prompted more specific comments Additionally, a subset of clerkship students were identified
regarding both strengths and weaknesses. Some small who benefited from targeted coaching in diagnostic skills.
group facilitators anecdotally noted that students utilized To ensure the equitable use of milestone evidence, the
language from milestone anchors when providing verbal portfolio review process is designed to mitigate against
feedback to peers during group debriefings. major action over a single data point. It is challenging to
For some competencies, patterns in aggregate outcomes determine whether variations in milestones data are due to
data differed from what was expected (see Figure 9). This real differences in student performance, are primarily
triggered additional qualitative analysis in informal focus related to changing contexts, or are artifacts of rater vari-
groups. Stakeholders were presented the quantitative ation. However, this mimics the reality of clinical practice;
results and were provided the milestone language, then students must demonstrate competency across all domains
asked to consider potential contributing factors including: in many different settings and to various supervisors, peers
limitations in the anchor language, limitations of the assess- and patients. Students are instructed to view all feedback
ment setting, and possible true differences in developmen- as valuable information emanating from various perspec-
tal patterns for some competencies. These discussions tives; but through the coaching and review mechanism
determined the appropriate improvement response. Based they can argue that a particular rating is not indicative of
upon evidence, a revised milestone menu was endorsed in their overall performance.
the fall of 2015 (see Supplemental materials). Aggregate milestone outcome data also provide insights
into the performance of the curriculum. Results for a class
of students indicate whether desired competency outcomes
Discussion are generally being attained. This is useful to drive curricu-
Reviewing one medical school’s approach to the design lar improvements and to monitor potential impacts of cur-
and implementation of competency milestones, and per- ricular changes. As an example, our recent curricular
forming analysis of two years of resultant data, is instruct- revision included a shift in the timing of core clerkships
ive regarding the effort needed to ensure a practical, useful from year three to year two. Aggregate student perform-
and valid structure. Preliminary findings indicate that it is ance in history and physical examination at the end of the
feasible to create standardized milestone anchors to be clerkship phase for the first transition class was comparable
applied across settings. Generally, the language of the mile- to historical benchmarks, indicating that the acceleration of
stone anchors is meaningful to students and to faculty the core clinical experience was successful for those
members. Aggregate and individual student data regarding measures.
achievement of these milestones affirm that they distin-
guish relative strengths and weaknesses within the individ-
Future directions
ual and among students.
The critical analysis of milestone performance over two Our initial implementation of the milestone framework was
years’ experience has generated enough confidence to deliberately limited to controlled cohorts of assessors with
include this evidence in formal reviews of student progress. protected time who contributed actively to the
502 K. D. LOMIS ET AL.

Figure 9. Example of comparison of aggregate milestone rating distributions.

development and improvement of the milestones. The util- Additional training for students has been instituted. Many
ity of the milestones for clinical teams assessing students personalized learning goals can be met within the context
within the context of active care delivery remains to be of existing experiences; some may require special learning
determined. Challenges regarding training of assessors and opportunities. We have had anecdotal success with tar-
issues of feasibility given time pressures in the clinical geted remediation plans. High performers must be offered
workplace must be addressed. We have partnered with a the ability to “level up,” aligning advanced activities with
clinical department with a strong track record of providing their competency progression. The high degree of flexibility
meaningful student feedback in order to pilot the imple- and individualization in our revised curriculum puts such
mentation of the milestones language with frontline clinical accommodations within reach. After ensuring the quality of
faculty and resident assessors. longitudinal milestone data, we envision the potential for
Through our analysis, we are increasingly aware of the advancement through our educational program based
importance of the process we established around the util- upon competency rather than time.
ization of the milestones as evidence of performance. The The initial milestone process at our institution has been
structured, faculty-facilitated self-assessment cycles promote successful, but will require critical oversight and ongoing
thoughtful interpretation of the milestone results for each refinement. Extensive, continual training for faculty and stu-
learner; the learner and coach weigh the value of each dents remains critical. Communication with stakeholders to
data point. This helps to control for error in the system, address the practicalities of implementation has promoted
and optimizes the application of this evidence into an sustainability. Critical, data-driven appraisal of our outcomes
actionable learning plan. will help avoid unintended consequences. Confounders,
Ongoing improvements will strengthen the meaningful- such as context and inter-rater variation, must be monitored
ness of the milestones. We plan to focus on how these and considered in any analysis or application of the mile-
data are used to direct future learning. Qualitative results stone data. Most importantly, our team actively acknowl-
indicate there is still work to be done to make the mile- edges and addresses limitations in the model, resulting in a
stone framework most effective for student development. transparent process of continual improvement.
MEDICAL TEACHER 503

Conclusions learning communities in medical education, portfolios, learning goals,


and mentoring.
Based on the experience at one medical school, the imple-
William B. Cutrer, MD, MEd is the assistant dean of undergraduate
mentation of competency milestones for medical students medical education, and associate professor of Pediatrics, Critical Care
is feasible and useful; it is also resource intensive and Medicine at Vanderbilt University School of Medicine. He has published
requires ongoing training, validation and improvement. A and presented widely on how students learn in the workplace and
committed group of thoughtful educators was essential to how to help them more effectively.

development of the program, and collaboration in an envir- Geoffrey M. Fleming, MD FAAP is a Pediatric Intensive Care Unit phys-
onment of mutual respect generated a shared sense of ician with a history of education pursuits including Program Director
for an ACGME accredited fellowship training program and Vice
ownership. Protected time for assessors and coaches has
President for Continuous Professional Development in an ACCME
contributed to the quality of performance evidence and its accredited institution. He has been involved extensively in education
application to future learning. Pilots before full implementa- administration leadership at both the local and national level through
tion, and critical analysis of outcome evidence, have workshop development, course administration, educational design
committees and advocacy.
informed revisions. Significant resources were dedicated to
technological solutions for data capture and display that Cathleen C. Pettepher, PhD, is a Professor of Cancer Biology and
enhance the feasibility and effectiveness of this framework. Medical Education and Administration and Assistant Dean of Medical
Student Assessment at Vanderbilt University School of Medicine. Her
Ongoing quality assurance and agility of thinking enables scholarly interests have focused on designing and implementing cur-
alignment between educational theory and the challenges ricular innovations and the development, implementation and evalua-
of real world implementation. Our results demonstrate that tions of value of peer assessments in professional growth and identify
a competency-based framework can provide actionable formation.
information for learners, thus enabling achievement of Bonnie Miller, MD, MHHC is senior associate dean for health science
desired educational outcomes. education at Vanderbilt University School of Medicine and Executive
Vice-President for Educational Affairs at Vanderbilt University Medical
Center, Professor of Medical Education and Administration and
Acknowledgements Professor of Clinical Surgery. Her scholarly interests include curriculum
innovation and the factors that influence the professional formation of
The authors acknowledge Stan Hamstra, PhD for his helpful review physicians.
and feedback on a draft of this paper.
The content reflects the views the authors and does not necessarily
represent the views of AMA or other participants in this Initiative.
Glossary
Disclosure statement Competency-based assessment (CBA)
Orients the assessment of learner performance to graduate out-
The authors for this article have no interests to report.
come abilities. It is organized around competencies derived
from an analysis of societal and patient needs. Ideally, a major
portion of the assessments should be performed in the context
Funding of the clinical workplace and should be criterion-referenced.
This publication was prepared with financial support from the The assessment process is designed to facilitate the develop-
American Medical Association (AMA) as part of the Accelerating mental progression of competence.
Change in Medical Education Initiative. Adapted from:
1. The role of assessment in competency-based medical educa-
Notes on contributors tion. Eric S. Holmboe, Jonathan Sherbino, Donlin M. Long,
Susan R. Swing, Jason R. Frank, and for the International
Kimberly D. Lomis, MD is associate dean for undergraduate medical CBME Collaborators. Medical Teacher Vol. 32, Iss. 8,2010
education, professor of surgery, and professor of medical education
2. Toward a definition of competency-based education in medi-
and administration at Vanderbilt University School of Medicine. Her
cine: a systematic review of published definitions. Jason R.
scholarly interests focus around curricular innovation, competency-
Frank, Rani Mungroo, Yasmine Ahmad, Mimi Wang, Stefanie
based assessment, and change management.
De Rossi, and Tanya Horsley. Medical Teacher Vol. 32, Iss.
Regina G. Russell, MEd, MA is the director of learning system out- 8,2010
comes for the MD program at Vanderbilt University School of
Milestones
Medicine. Her higher education experience includes student services
administration, sociology and social work education, learner assess- Are discrete points or stages in a developmental progression.
ment and program evaluation. Current work includes measurement of Milestones frameworks in medical education mirror the Dreyfus
medical student outcomes and the impact of organizational environ- model of learner development from novice to expert. Well-
ments on learning. articulated milestones for a specific competency provides a
shared observational approach for assessors and a clear frame
Mario Davidson, PhD, MA (Mathematics Education), MS (Statistics), of reference for learners.
MAS (Mathematics) is a biostatistician for Vanderbilt University Medical
Center. In his career, he has worked in medical and statistical educa- Adapted from:
tion as an instructor and researcher. He is currently the course director 1. Frameworks for learner assessment in medicine: AMEE Guide
for the Statistical Collaborator course presented in the Department of No. 78. Louis Pangaro and Olle ten Cate. Medical Teacher
Biostatistics. Vol. 35, Iss. 6, 2013.
Amy E. Fleming, MD, MHPE is the associate dean for medical student 2. Accreditation Council for Graduate Medical Education
affairs at the Vanderbilt University School of Medicine and associate Milestones implementation.
professor of pediatrics, division of hospital medicine at Monroe Carell
Jr Children's Hospital at Vanderbilt. Her research interests span
504 K. D. LOMIS ET AL.

References Hicks PJ, Schumacher DJ, Benson BJ, Burke AE, Englander R,
Guralnick S, Ludwig S, Carraccio C. 2010b. The pediatrics
Anderson MB. 2012. Aligning education and assessment: improving milestones: conceptual framework, guiding principles, and approach
medical education through assessment. Med Sci Educ. 22:267–269. to development. J Grad Med Educ. 2:410–418.
Baartman LKJ, Bastiaens TJ, Kirschner PA, van der Vleuten CPM. 2007. Hodges B. 2013. Assessment in the post-psychometric era: learning to
Evaluating assessment quality in competence-based education: a love the subjective and collective. Med Teach. 35:564–568.
qualitative comparison of two frameworks. Educ Res Rev. 2:114–129. Holmboe ES, Batalden P. 2015. Achieving the desired transformation:
Boursicot K, Etheridge L, Setna Z, Sturrock A, Ker J, Smee S, thoughts on next steps for outcomes-based medical education.
Sambandam E. 2011. Performance in assessment: consensus state- Acad Med. 90:1215–1223.
ment and recommendations from the Ottawa conference. Med Krupat E, Pelletier SR. 2016. The development of medical student com-
Teach. 33:370–383. petence: tracking its trajectory over time. Med Sci Educ. 26:61–67.
Carraccio CL, Englander R. 2013. From Flexner to competencies: reflec- Linn R, Baker E, Dunbar S. 1991. Complex, performance-based assess-
tions on a decade and the journey ahead. Acad Med. 88:1067–1073. ment: expectations and validation criteria. Educ Res. 20:15–21.
Chertoff J, Wright A, Novak M, Fantone J, Fleming A, Ahmed T, Green Messick S. 1989. Validity. In: Educational measurement. 3rd ed. New
MM, Kalet A, Linsenmeyer M, Jacobs J, et al. 2016. Status of port- York (NY): American Council on Education: Macmillian Publishing
folios in undergraduate medical education in the LCME accredited Company.
US medical school. Med Teach. 38:886–896. Nasca TJ, Philibert I, Brigham T, Flynn TC. 2012. The next GME
Clayton MJ. 1997. Delphi: a technique to harness expert opinion for crit-
accreditation system-rationale and benefits. N Engl J Med.
ical decision-making tasks in education. Educ Psychol. 17:373–386.
366:1051–1056.
Crossley J, Johnson G, Booth J, Wade W. 2011. Good questions, good
Pecaric M, Boutis K, Beckstead J, Pusic M. 2017. A big data and
answers: construct alignment improves the performance of work-
learning analytics approach to process-level feedback in cognitive
place-based assessment scales. Med Educ. 45:560–569.
simulations. Acad Med. 92:175–184.
Downing SM, Haladyna TM. 2004. Validity threats: overcoming interfer-
Regehr G, Ginsburg S, Herold J, Hatala R, Eva K, Oulanova O. 2012.
ence with proposed interpretations of assessment data. Med Educ.
Using “standardized narratives” to explore new ways to represent
38:327–333.
faculty opinions of resident performance. Acad Med. 87:419–427.
Dreyfus SE. 2004. The five-stage model of adult skill acquisition. Bull
Schumacher DJ, Lewis KO, Burke AE, Smith ML, Schumacher JB, Pitman
Sci Technol Soc. 24:177–181.
Englander R, Burke AE, Guralnick S, Benson B, Hicks PJ, Ludwig S, MA, Ludwig S, Hicks PJ, Guralnick S, Englander R, et al. 2013. The
Schumacher D, Johnson L, Carraccio C. 2012. The pediatrics mile- pediatrics milestones: initial evidence for their use as learning road
stones: a continuous quality improvement project is launched-now maps for residents. Acad Pediatr. 13:40–47.
the hard work begins! Acad Pediatr. 12:471–474. Spickard A 3rd, Ahmed T, Lomis K, Johnson K, Miller B. 2016. Changing
Englander R, Cameron T, Ballard AJ, Dodge J, Bull J, Aschenbrener CA. medical school IT to support medical education transformation.
2013. Toward a common taxonomy of competency domains for the Teach Learn Med. 28:80–87.
health professions and competencies for physicians. Acad Med. Swing SR. 2007. The ACGME outcome project: retrospective and pro-
88:1088–1094. spective. Med Teach. 29:648–654.
Frank JR, Mungroo R, Ahmad Y, Wang M, De Rossi S, Horsley T. 2010a. Swing SR, Beeson MS, Carraccio C, Coburn M, Iobst W, Selden NR,
Toward a definition of competency-based education in medicine: a Stern PJ, Vydareny K. 2013. Educational milestone development in
systematic review of published definitions. Med Teach. 32:631–637. the first 7 specialties to enter the next accreditation system. J Grad
Frank JR, Snell LS, Cate OT, Holmboe ES, Carraccio C, Swing SR, Med Educ. 5:98–106.
Harris P, Glasgow NJ, Campbell C, Dath D, et al. 2010b. van der Vleuten CP. 2014. When I say … context specificity. Med
Competency-based medical education: theory to practice. Med Educ. 48:234–235.
Teach. 32:638–645. van der Vleuten CP, Schuwirth LW. 2005. Assessing professional com-
Hicks PJ, Englander R, Schumacher DJ, Burke A, Benson BJ, Guralnick S, petence: from methods to programmes. Med Educ. 39:309–317.
Ludwig S, Carraccio C. 2010a. Pediatrics milestone project: next Whitehead CR, Kuper A, Hodges B, Ellaway R. 2015. Conceptual and
steps toward meaningful outcomes assessment. J Grad Med Educ. practical challenges in the assessment of physician competencies.
2:577–584. Med Teach. 37:245–251.

You might also like