Professional Documents
Culture Documents
Management, 2021.
Citation:
Abstract
Purpose
The purpose of this article is to examine whether the incentivizing type of performance
appraisal (typical of New Public Management) has indeed been superseded by a post-
NPM, developmental type of performance appraisal in European civil services.
Design/methodology/approach
The literature review lead to a unidimensional, twofold typology: incentivizing (NPM)
and developmental (post-NPM) performance appraisal. The empirical basis of the
research is two surveys conducted among top civil servants in 18 European countries.
Findings
Firstly, there are crucial discrepancies between performance appraisal systems in
contemporary European central government administrations, and current theorizing on
performance appraisal. Contrary to our expectations developed on the basis of the latter,
“developmental” and “incentivizing” do not seem to be two distinct types of
performance appraisal; rather, they are two independent dimensions, defining altogether
four different types of performance appraisal systems.
Social Implication
Citizens and communities are direct stakeholders in the development of public service
performance appraisal both as possible or actual employees of public service
organizations and as recipients of public services.
Practical Implications
Our results give orientation to policy makers and public service managers to engaged in
designing or applying performance appraisal systems, in particular by identifying
assailable presumptions underlying many present-time reform trends.
Originality
The paper proposes a new fourfold typology of performance appraisal systems:
incentivizing, developmental, symbolic, and want-it-all.
Sources of data
Performance appraisal systems vary both between and within countries – frequently even
between organizations in one country, or even within organizations (see, e.g., the
separate appraisal systems for senior civil servants). In this research, our primary unit of
analysis is the country level. Within that, we focus on the core of government (ministerial
departments, agencies, and other central state administrative bodies), because it is narrow
enough to enable comparisons across different national public administrations.
The empirical investigation utilizes two datasets. The first one is an expert survey carried
out by the EU Public Administration Network (composed of representatives of public
administration institutions of all EU members, the European Commission, as well as
Norway, Switzerland, and Serbia; hereinafter “EUPAN”). The purpose of the survey
(conducted from July to December 2016) was to provide an overview of individual
performance appraisal systems of civil servants on the central/federal level of
government. This was followed by a second survey (from January to April 2018)
developed and administered by the authors, targeting the same set of respondents: senior
officials delegated to EUPAN working group by the individual member states. These
officials, one per country, were officials in charge of civil service in central governments,
with unique insights into the systemic design features of performance appraisal in their
respective countries. Imperfections frequently attributed to single-respondent surveys
such as ours (for an in-depth discussion in a different but similar field see Krause et al
2018) are nevertheless limited by the fact that the survey results were discussed and
approved both at an EUPAN Working Level Meeting and by EUPAN Directors General
meeting overseeing public administration in the member states. Thus, the survey can be
assumed to reflect performance appraisal systems in functional rather than legalistic
terms with reasonable accuracy.
The data collection instruments used differed in the two surveys. The 2016 survey
reflected the data needs of the ongoing EUPAN project. The nine-page-long instrument
contained approximately 130 items, frequently open-ended, exploring a broad range of
performance appraisal systems’ design features. In contrast, the 2018 follow-up survey
focused more directly on our key ambition here: describing the performance appraisal
systems in place in terms of their incentivizing and developmental features. The design of
the second, follow-up survey reflected our recognition, based on the results of the 2016
survey, that real-life performance appraisal systems are oftentimes either a mixture of
Analytical framework
In order to classify performance appraisal systems into our two-cluster typology, we first
developed, on the basis of conceptual considerations, an analytical framework consisting
of a set of classification criteria. The seven classification criteria enabled us to
differentiate between NPM spirited incentivizing ideal type (hereinafter called “Type I”)
and post-NPM spirited developmental ideal type (hereinafter called “Type II”
performance appraisal systems. The proposed criteria do not constitute a set of sufficient
conditions for designation as Type I or Type II, let alone ensuring that they will actually
operate as Type I or Type II performance appraisal systems. Rather, meeting the given
set of criteria means that a system may reasonably be expected to operate as a Type I or
Type II system (i.e., each one is a necessary but – either individually or jointly -
insufficient conditions). Therefore, failing to meet even a single one of the criteria means
that the given performance appraisal system cannot reasonably be expected to operate as
a Type I or Type II system.
A further consequence of the above methodological approach is that the criteria are not
meant to provide an exhaustive characterization of the respective performance appraisal
system types. Nor can the criteria be interpreted as statistical variables possibly subject to
conventional statistical analytical procedures (such as cluster analysis). Rather, their
purpose is no more (but also no less) than to enable the distinction between the two
types, as well as the classification of existing – oftentimes somewhat inconsistent or
incomplete – performance appraisal systems into our twofold typology. Note that
although, for the sake of brevity, we use numbers (1’s and 0’s) to present our findings,
the analytical approach is entirely qualitative in nature.
Let us now explain and justify the individual criteria.
The main purpose of Type I systems is to increase work performance through providing
external incentives for individuals to increase their efforts. This implies the fulfillment of
the following design features.
[C1:] Work performance has to be evaluated on the basis of a set of well-defined
standards and indicators. Moreover, it needs to be assessed, often expressed in numeric
form (that is, quantitatively) or as a minimum using an ordinal scale. The actual
performance measurement is often conducted in fixed (annual or bi-annual) periods.
[C2:] The system entails monitoring and evaluation of results and past behavior (rather
than, say, focusing on present or future needs or plans alone).
[C3:] Whereas some performance appraisal systems entail a broad range of possible
assessors and sources of assessment data (cf. self-assessment, group assessment, 360-
degree-assessment, and the like), in Type I appraisals, the critical element is the
supervisor’s judgment of performance. Other assessors may contribute to the appraisal
result, but the bulk of the assessment results must be tied to how supervisors judge the
The remaining cases fall into two categories: they either do not meet the minimum
criteria for either the incentivizing or the developmental model, or they meet both. This
finding created a clear need to revise and extend our initial typology, in order to include
most or all of the cases covered in the study.
A central initial assumption of the study – informed by a wealth of authoritative literature
on the topic reviewed earlier – was that, based on pre-existing scholarship and theoretical
considerations, performance appraisal systems (i) can be characterized along a single
conceptual continuum ranging from the “developmental” to the “incentivizing” pole,
and that (ii) these two modes of performance appraisal contradict one another.
Therefore, actual performance appraisal systems were likely to be located close to one
end of the spectrum or the other.
In the light of our above findings, however, both assumptions had to be relaxed. Instead,
we had to admit that “being developmental” and “being incentivizing” did not constitute
two poles of a single dimension, but rather they are two independent dimensions
themselves. Actual systems can feature essential characteristics of both of them, or even
none of them. This empirical finding runs counter existing theorizing and typologies.
The resulting conceptual typology is displayed below.
10
11
e L
tiv t ce is r ary ces le tal ive k TA S G )
try t ita as an urc vis e o
n
t
e en lt p
i ces / en
d
t
s lita ac F E REF ISIN RES type
N
n n p m so er o u u r 6 m e a b O M UO V U (
Co
u
qua C2 / rfor / u p / m seq / m sou C lop ne Qu eed N. LOP EATN. NTI EAT sult
s n e / f E F E F
/ pe C3 C4 co C5 de
v
C7 V C Re
C1 DE IN
Belgium 1 1 1 1 0 0 0 0 4 I.
Bulgaria 1 1 1 1 0 1 1 2 4 I.
Croatia 1 1 1 0 0 0 1 1 3 III.
Czech Republic1 1 1 1 0 1 1 2 4 I.
EU (Commission)
0 1 1 0 1 1 1 3 2 II.
Finland 1 1 1 1 0 1 0 1 4 I.
France 0 1 1 0 0 1 1 2 2 III.
Germany 1 1 1 1 1 1 1 3 4 IV.
Hungary 1 1 1 1 0 1 0 1 4 I.
Ireland 0 1 1 0 1 1 1 3 2 II.
Italy 1 1 1 1 0 0 0 0 4 I.
Luxemburg 1 1 1 0 0 0 1 1 3 III.
Netherlands 1 1 1 1 1 1 1 3 4 IV.
Norway 0 0 1 0 0 1 1 2 1 III.
Poland 1 1 1 0 0 1 1 2 3 III.
Romania 1 1 1 0 0 0 0 0 3 III.
Slovakia 1 1 1 1 0 0 0 0 4 I.
Serbia 1 1 1 1 0 0 1 1 4 I.
Source: the authors
12
13
14
15
References
Adcroft, A. and Willis, R. (2005), “The (un)intended outcome of public sector
performance measurement”, International Journal of Public Sector Management, 18(5), pp.
386-400.
Bauer, M., Peters, B. G., Pierre, J., Yesilkagit, K., & Becker, S. (2021 forthcoming),
Democratic backsliding, populism and public administration, Cambridge University Press:
Cambridge.
Bellé, N. (2015), “Performance‐related pay and the crowding out of motivation in the
public sector: A randomized field experiment”, Public Administration Review, 75(2),
pp.230-241.
16
17
18
19
20