Professional Documents
Culture Documents
Jean-François Henri
To cite this article: Jean-François Henri (2010) The Periodic Review of Performance Indicators:
An Empirical Investigation of the Dynamism of Performance Measurement Systems, European
Accounting Review, 19:1, 73-96, DOI: 10.1080/09638180902863795
JEAN-FRANÇOIS HENRI
Université Laval, Québec, Canada
Introduction
The aim of this study is to examine the association between one attribute of
performance measurement systems (PMS) that is not widely addressed in the
management accounting literature, namely the dynamism of PMS, and organiz-
ational performance. This attribute refers to the periodic review of performance
PMS, more revisions may not always be better. In particular, changes in the
business environment and strategic direction are commonly recognized as
important drivers of change within PMS (e.g. Bititci et al., 2000; Forza and
Salvador, 2000; Lynch and Cross, 1991; Wisner and Fawcett, 1991). Those
external and internal forces promoting changes within PMS reflect the need for
periodic reviews. Therefore, this study examines whether the association
between the dynamism of PMS and organizational performance is contingent
on the level of external and internal change.
Using survey data from a sample of manufacturing firms, the results suggest
that the association between the dynamism of PMS and organizational perform-
ance is a function of the match between the levels of change and the periodic
reviews of performance indicators. Furthermore, it may be more harmful to not
have a dynamic PMS in a context of higher levels of change than to have a
dynamic PMS even if it is not required. Considering the scant attention
devoted to the dynamism of PMS in the management accounting literature
combined with the importance of the periodic review of performance indicators
for practitioners, this paper contributes to the development of knowledge related
to PMS by describing dynamism as an important attribute of PMS, like integra-
tiveness, alignment and measurement quality. The remainder of this paper is
organized as follows. The next section describes the theoretical framework and
presents our research hypothesis. The section after presents the methodology,
including a sample definition, data collection and measurement of constructs.
The results of the analyses are presented, followed by a discussion of the
results and the conclusions of this study.
Theoretical Framework
Definition of Constructs
Dynamism of PMS
The dynamism of PMS is defined as the propensity of the organization to
revise its performance indicators to ensure their relevance and appropriateness
(Kennerley and Neely, 2002, 2003). It refers to an updating process that insti-
tutionalizes the need for continuously changing measures and ensuring evol-
ution in the measurement set (Dixon et al., 1990; Kuwaiti, 2004; Vitale and
Mavrinac, 1995). This update represents the last step in the ongoing process
for developing an effective PMS. It refers to the review necessary to ensure
that PMS are regularly updated and to the periodic re-evaluation of the appro-
priateness of the established performance indicators in view of the current
competitive environment (Forza and Salvador, 2000; Waggoner et al., 1999;
Wisner and Fawcett, 1991). The periodic review of performance measures
comprises the addition of indicators, the deletion of indicators, the changes
in the target and the changes in the definition of the indicators (Bourne
et al., 2000).
76 J.-F. Henri
have a limited lifetime; as a result they tend to decline over time in their ability to
discriminate good from bad performance. Periodic reviews may prevent perform-
ance indicators from reflecting old priorities and inconsistent measures, and
enable PMS to capture a range of performance outcomes (Meyer and Gupta,
1994). In sum, appropriate accounting information supports the effective man-
agement of resources and contributes to organizational performance (Baines
and Langfield-Smith, 2003). More specifically, updated performance indicators
support resource management and performance by reviewing and reprioritizing
internal objectives, deploying changes within the organization, and ensuring
the maintenance of gains through improvement programs (Bititci et al., 2000).
Despite the potentially positive influence of the dynamism of PMS on
organizational performance, periodic reviews do not come without costs. Indeed,
dynamic PMS require human, technological and financial resources to periodically
refine performance indicators. Indeed, several activities have to be performed, such
as the analysis of current performance indicators before deletion or revision, the
identification of the information needs, the development and documentation of
new indicators, the adjustment to the information systems to enable the collection
of new or different indicators, etc. Furthermore, unnecessary changes may create
confusion and a possible lack of motivation among managers and employees, as
well as disrupt the collection of longitudinal data used to analyze trends and to
test causality implicit to the strategy.
The choice of periodically reviewing performance indicators (or not) is the
result of a trade-off between the benefits and costs discussed above. In other
words, there are circumstances whereby the benefits exceed the costs (i.e.
desirability of PMS dynamism) while in other cases the costs exceed the benefits
(i.e. non-desirability of PMS dynamism). The degree of external and internal
changes faced by firms represents one important circumstance that may influence
the need to periodically revise performance indicators. Those changes influence
the level of complexity faced by organizations by increasing the interaction
among numerous factors that must be considered simultaneously. This complexity
contributes to the global uncertainty faced by organizations, i.e. the difference
between the information required and the information that organizations already
possess (Galbraith, 1973).
The periodic review of performance indicators is likely to contribute to organ-
izational performance when the level of global uncertainty is high. In this context,
managers must deal with complex operations and business processes, unstable
product lines and changing technologies (Miller, 1988). As changes in the
frames of reference and central norms are needed, past practices are called into
question, new assumptions about organizations are proposed and changes in
strategic priorities are considered (Argyris and Schön, 1978; Van de Ven, 1986).
Managers need additional information to understand those changing situations
in order to reduce the information gap. Periodic reviews of performance indicators
may help reduce this uncertainty by ensuring an evolution in the measurement set,
improving the information that organizations possess to manage resources, and
The Periodic Review of Performance Indicators 79
Method
Data Collection
The data used in this study were collected in 2002 using a mail-survey approach
and were part of a larger research project related to PMS in manufacturing
firms. The survey implementation followed four steps: (i) pre-notification; (ii)
initial mailing; (iii) first follow up; and (iv) second follow up. The first step con-
sisted of a letter, phone call or e-mail to respondents to generate early interest.
A mail out package including the following three elements was then sent to
every contact name: covering letter, questionnaire, and business-reply envelope.
In some cases, the questionnaire was sent by fax or e-mail. The first follow-up
80 J.-F. Henri
was a postcard reminder three weeks after the initial mailing, while the second was
a phone call or replacement questionnaire three weeks after the first follow-up.
The target population consisted of 2175 top management teams of Canadian
manufacturing firms listed in the Scott’s database with primary and secondary
SIC codes in the range of 21 to 39.3 Furthermore, the firms were large enough
to ensure that organizational variables apply (Miller, 1987) and to ensure that a
formal PMS is in place (Bouwens and Abernethy, 2000). Thus, the firms selected
in the sample respect the following two criteria: (i) sales are at least $20 million
Canadian yearly; (ii) at least 150 people are employed. However, the lack of
contact names in the database in several cases reduced the number of usable
firms in the target population to 1692. Data were collected using a structured
questionnaire sent to the highest member of the ‘corporate’ top management
team (autonomous entity) or ‘local’ top management team (subunit) for which
the identity was revealed in the database.4 A total of 383 usable questionnaires
were received, for a response rate of approximately 24%.5 Appendix 1 presents
the statistics of the respondents in terms of position, experience, size (number
of employees) and industry classification.
To test whether respondents differed from non-respondents, a two-step analy-
sis was conducted. Respondents were first compared with non-respondents in
terms of sample characteristics (size, location, industry). Next, early and late
respondents were compared to detect any difference in the mean score of each
construct.6 Using chi-square statistics, no significant differences were found
between the size, location and industry of respondent firms and non-respondent
firms. A comparison of the means of the variables found no significant differ-
ences between early and late respondents. Hence, the analysis did not reveal
any systematic differences between respondents and non-respondents.
Results
Descriptive Statistics
Table 1 presents the descriptive statistics and the correlation matrix for the main
constructs (Panel A) and specifically for each type of periodic review (Panel B).
Globally, the results suggest that performance indicators are not reviewed exten-
sively by top management teams (mean of 3.69/7). More specifically, while new
performance indicators are added to some extent (mean of 4.49) and new targets
fixed (mean of 4.06), less attention is devoted to discarding old indicators (mean
of 2.43) or to changing the definition of a performance indicator (mean of 3.73).
Interestingly, not only is the difference between the mean score of the deletion
and addition of indicators important (2.43 vs. 4.49) but the median also differs
substantially (2 vs. 5). While 78% of the scores of the deletion item range
between 1 and 3, 74% of the score of the addition item range between 4 and
7. This suggests that more performance indicators are added than deleted. In
other words, as time passes, the number of indicators within the PMS increases.
Larcker (2001) and Ittner et al. (2002). Assuming that at any point in time some
firms will be ‘off-equilibrium’, performance differences are examined (see the
discussion earlier). Basically, this approach assesses in two steps the extent to
which organizational performance is associated with deviations from ‘optimal’
practices. In the first step, we hypothesize and estimate a model for the periodic
reviews of performance indicators. Using a logit model, we estimate the
probability that organizations review extensively their performance indicators
as a function of the level of external and internal changes.10 The residuals for
84
J.-F. Henri
Table 2. Results of OLS regressions examining the association between dynamism of PMS and organizational performance
Periodic reviews Deletion of indicators Addition of indicators Changes in targets Changes in definition
Variable (n ¼ 383) (four types) (type 1) (type 2) (type 3) (type 4)
Intercept 2.032 2.354 2.509 2.263 1.910
Dynamism of PMS 0.197 0.102 0.080 0.124 0.146
PEU 20.085 20.058 20.079 20.057 20.075
Strategic capabilities 0.075 0.081 0.086 0.085 0.085
Industry 0.344 0.333 0.308 0.311 0.356
Size 0.266 0.301 0.278 0.267 0.286
Adjusted R2 0.113 0.058 0.055 0.091 0.112
(p , 0.01) (p , 0.01) (p , 0.01) (p , 0.01) (p , 0.01)
each observation estimate the distance between the practices of the firms and the
‘optimal’ practices as represented by the systematic model. These residuals are
assumed to reflect the over-investment or under-investment of the organization
in the dynamism of PMS. Table 3 presents the results of the logit model
predicting the dynamism of PMS.
Overall, consistent with Kennerley and Neely (2002, 2003), firms facing a
higher level of PEU (proxy for external change) are more likely to make periodic
reviews of performance indicators (0.291; p , 0.05). However, no significant
association is observed between strategic capabilities (proxy for internal
change) and periodic reviews. Some notable differences are observed for each
type of periodic review. First, PEU is only specifically associated with the
addition of performance indicators (p , 0.01). In addition, strategic capabilities
are associated with the addition of performance indicators (p , 0.05) and
changes in definition (p , 0.10). No association is observed with either PEU
or strategic capabilities for the deletion of indicators and changes in targets.
In the second step, OLS regressions are conducted using residuals from the
logit model. Separate variables are computed for positive and negative residuals
to allow for potential difference between over- and under-investment in periodic
reviews of performance indicators.
. The positive residuals from the logit model refer to the firms that review
performance indicators extensively but the predicted probability of review is
less than one. The variable used is based on the value of that positive residual,
or zero if the residual is negative. The expected coefficients are negative as
firms having dynamic PMS but having low predicted probability of review
are expected to be associated with lower performance.
. The negative residuals from the logit model refer to the firms that do not review
performance indicators extensively but the predicted probability of review is
greater than zero. The variable used is based on the value of that negative
86 J.-F. Henri
Table 4 presents the results of the regressions examining the association between
the positive and negative residuals and performance, controlling for PEU, stra-
tegic capabilities, industry, and size. While each regression is statistically signifi-
cant (p , 0.01), it reflects moderate explanatory power (between 0.05 and 0.09).
The results suggest that the association between dynamism of PMS and organiz-
ational performance is a function of the fit between the level of changes and the
periodic reviews of performance indicators. Thus, the results provide support for
hypothesis 1. More specifically, the coefficient of the ‘negative residuals’ vari-
able is positive and significant (1.874; p , 0.01) while the coefficient of the
‘positive residuals’ variable is negative and significant, but to a lesser extent
( – 1.066; p , 0.10). This suggests that lower performance is associated with
firms that (i) do not review extensively performance indicators when the model
predicts they should, and to a lesser extent, (ii) review extensively performance
indicators when the model predicts they should not be doing so. The same results
are observed for each type of performance indicator, except for the changes in
definition whereby no significant association is observed. Lastly, while no associ-
ation is observed between PEU and performance, strategic capabilities, industry
and size are positively associated with performance.
Conclusion
The aim of this study was to examine one attribute of PMS that is not widely
addressed in the management accounting literature, namely the dynamism of
PMS. Following a contingency approach, this study has specifically investigated
the extent to which the association between periodic reviews of performance indi-
cators and organizational performance is contingent on the level of external and
internal changes. Based on a sample of 383 manufacturing firms, the results
suggest three main conclusions.
(i) Despite the fast changes in the current business environment and organiz-
ational requirements, the limited lifetime of performance indicators and
the various claims encouraging the regular update of PMS, it appears that
manufacturing organizations do not update their PMS regularly. While
some effort is being devoted to the addition of new indicators and to the
change of targets, not much attention is directed towards the deletion of
indicators and changes in the definition of indicators.
(ii) Periodic reviews of performance indicators can be beneficial for organizations
by ensuring evolution in the measurement set, improving the information that
organizations possess to manage resources, and continuously maintaining the
Table 4. Results of OLS regressions examining the association between dynamism of PMS prediction model residuals and organizational
performance
Periodic reviews Deletion of indicators Addition of indicators Changes in targets Changes in definition
Variable (n ¼ 383) (four types) (type 1) (type 2) (type 3) (type 4)
87
88 J.-F. Henri
of 12 months. However, changes in PMS may occur over a longer period of time
and not be completely captured by the instrument. Longitudinal and qualitative
approaches will be necessary to provide complementary evidence. Second,
PEU and strategic capabilities are two proxies for external and internal
changes. They may not completely reflect the degree of changes that can
occur. Third, two potential contingency variables (i.e. external and internal
changes) and four types of dynamism (i.e. addition, deletion, targets, and defi-
nition) have been investigated. However, other correlated omitted variables
could influence simultaneously the dynamism of PMS as well as the effect of per-
iodic reviews of performance indicators on performance, such as organizational
factors (e.g. change in leadership, culture, climate, structure, etc) and technologi-
cal factors (e.g. the information systems functionality and capacity). Other types
of dynamism include the change in weights of importance attached to the indi-
cators, a change in the frequency of reporting, and a change in the format of pres-
entation. Fourth, using the survey method to collect data creates a potential for
bias due to common response. Several techniques have been used to reduce or
minimize its effect (Doty and Glick, 1998). In terms of measurement techniques,
the questionnaire reflects different response formats, various scale anchors, and
variations in the wording of items. In terms of data sources, a validation
sample has been developed by seeking the participation of a second respondent
within the firms that have originally returned the questionnaire.
Acknowledgments
The author would like to thank Maurice Gosselin, Sylvie Héroux, Claude Laurin,
Alfred Seaman, and Nicole Thibodeau for their insightful comments and sugges-
tions, as well as the two anonymous reviewers.
Notes
1
Although few accounting studies have explicitly examined the review of performance indi-
cators, many have implicitly addressed this attribute when referring to the feedback process
of control systems. Indeed, the feedback process involves changes in objectives and actions,
as well as changes in the control systems (e.g. Ittner and Larcker, 2001; Otley, 1999). While
not focusing on the review process to ensure that PMS are regularly updated, this framework
suggests that the control systems may also be subject to re-evaluations and modifications.
2
Capabilities are the organizational and strategic routines by which firms synthesize and acquire
knowledge resources, and achieve new resource configurations (Eisenhardt and Martin, 2000;
Kogut and Zander, 1992).
3
In this study, ‘firm’ is a fully autonomous entity or a subunit of a larger firm. In all cases, firms
appeared as separate entities in the database.
4
Following other upper echelon studies (e.g. Carpenter and Fredrickson, 2001), a top manage-
ment team is defined as the top two tiers of an organization’s management team, which
include CEO/general manager, chief operating officer (COO), chief financial officer (CFO),
and the next highest management tier of a firm (senior vice-presidents).
90 J.-F. Henri
5
The response rate was calculated as the percentage of the number of usable returned question-
naires to the number of questionnaires sent, after adjusting for the firms that had closed, ended
manufacturing activities or moved, or for which the contact person had left the organization.
6
Early respondents correspond to the first 10% of all respondents ranked following the reception
date of the completed questionnaire. Late respondents correspond to the last 10% of all
respondents.
7
The results of the second-order CFA suggest that innovativeness and learning orientation are
two first-order constructs that relate to a second-order construct labeled strategic capabilities
(see Appendix 2). Thus, the mean score of innovativeness and learning orientation are com-
bined into one variable (strategic capabilities).
8
The indices used to assess the model are among the most frequently reported, namely NNFI
(non-normed fit index), CFI (comparative fit index), and RMSEA (root mean square error of
approximation). The threshold values recommended are (i) NNFI . 0.90 (Tabachnick and
Fidell, 2001), (ii) CFI . 0.95 (Hu and Bentler, 1995), and (iii) RMSEA , 0.l0 (Browne and
Cudeck, 1993).
9
To assess the inter-rater agreement, an average deviation (AD) index is calculated (Burke et al.,
1999). For all the constructs in the validation sample, the AD is estimated at 0.35 and ranges
from 0.27 to 0.49 for each construct. Compared with the criterion for acceptable inter-rater
agreement and practical significance estimated at 1.2 (Burke and Dunlap, 2002), these
results are quite satisfactory. The criterion is approximated as c/6, where c is the number of
response options for a Likert-type item. A series of t-tests was carried out to determine
whether the mean ratings provided by the first respondent for each construct were significantly
different from the mean ratings from the second respondent. The results do not reflect signifi-
cant differences for any construct (p . 0.05). Moreover, t-tests and chi-square analyses were
carried out to compare the mean ratings and sample characteristics (size, industry, location)
of the firms with two respondents with those of the firms that have one respondent. No signifi-
cant differences were found for all constructs and sample characteristics (p . 0.05) except for
size. Indeed, the firms that have two respondents appear to be smaller.
10
As mentioned by Ittner and Larcker (2001), this model is based on numerous assumptions: (i)
the model is assumed to be the same for each firm, (ii) the model exhibits the correct functional
form, (iii) the model has predictor variables that are measured without errors, and (iv) the model
includes all relevant predictor variables. Furthermore, the choice of logistic regression is three-
fold (Tabachnick and Fidell, 2001). Unlike multiple-regression, logistic regression has no
assumptions about the distributions of the predictor variable and it cannot produce negative pre-
dicted probabilities. In addition, logistic regression is especially useful when the distribution of
responses of the dependent variable is expected to be nonlinear with one or more independent
variable.
References
Amit, R. and Schoemaker, P. (1993) Strategic assets and organizational rent, Strategic Management
Journal, 14(1), pp. 33–46.
Argyris, C. and Schön, D. A. (1978) Organizational Learning (Addison-Wesley ed.) (Reading, MA:
Addison-Wesley).
Baines, A. and Langfield-Smith, K. (2003) Antecedents to management accounting change: a struc-
tural equation approach, Accounting, Organizations and Society, 28(7-8), pp. 675–698.
Banker, R. and Datar, S. (1989) Sensitivity, precision, and linear aggregation of signals for
performance evaluation, Journal of Accounting Research, 27(1), pp. 21–39.
Barney, J. B. (1991) Firms resources and sustained competitive advantage, Journal of Management,
17, pp. 99–120.
Barney, J., Wright, M. and Ketchen, D. J. (1991) The resource-based view of the firm: ten years after,
Journal of Management, 27, pp. 625–641.
The Periodic Review of Performance Indicators 91
Bititci, U. S., Turner, T. and Begemann, C. (2000) Dynamics of performance measurement systems,
International Journal of Operations and Production Management, 20(6), pp. 692–704.
Bourne, M., Mills, J., Wilcox, M., Neely, A. and Platts, K. (2000) Designing, implementing perform-
ance measurement systems, International Journal of Operations and Production Management,
20(7), pp. 754–771.
Bouwens, J. and Abernethy, M. A. (2000) The consequences of customization on management
accounting system design, Accounting, Organizations and Society, 25, pp. 221–241.
Browne, M. W. and Cudeck, R. (1993) Alternative ways of assessing model fit, in: K.A. Bollen and
J.S. Long (Eds) Testing Structural Equation Models (Newbury Park, CA: Sage Publications).
Burke, W. W. (1989) Culture instrument, Columbia University, Working paper.
Burke, M. J. and Dunlap, W. P. (2002) Estimating interrater agreement with the average deviation
index: a user’s guide, Organizational Research Methods, 5(2), pp. 159–172.
Burke, M. J., Finkelstein, L. M. and Dusig, M. S. (1999) On average deviation indices for estimating
interrater agreement, Organizational Research Methods, 2(1), pp. 49–68.
Carpenter, M. A. and Fredrickson, J. W. (2001) Top management teams, global strategic posture, and
the moderating role of uncertainty, Academy of Management Journal, 44(3), pp. 533–545.
Chenhall, R. H. (2005) Integrative strategic performance measurement systems, strategic alignment
of manufacturing, learning and strategic outcomes: an exploratory study, Accounting, Organiz-
ations and Society, 30(5), pp. 395–422.
Courty, P. and Marschke, G. (2003) Dynamics of performance-measurement systems, Oxford
Review of Economic Policy, 19(2), pp. 268.
Damanpour, F. (1991) Organizational innovation: a meta-analysis of effects of determinants and
moderators, Academy of Management Journal, 34(3), pp. 555–590.
Danneels, E. (2002) The dynamics of product innovation and firm competences, Strategic Manage-
ment Journal, 23, pp. 1095–1121.
Dess, G. G. and Robinson, R. B. (1984) Measuring organizational performance in the absence of
objective measures: the case of the privately-held firm and conglomerate business unit, Strategic
Management Journal, 5, pp. 266–273.
Dixon, J. R., Nanni, A. J. and Vollman, T. E. (1990) The New Performance Challenge – Measuring
Operations for World-class Competition (Dow Jones-Irwin ed.) (Homewood, Illinois: Dow
Jones-Irwin).
Donaldson, L. (2001) The Contingency Theory of Organizations (Thousand Oaks, CA: Sage
Publications).
Doty, D. H. and Glick, W. H. (1998) Common methods bias: does common methods variance really
bias results? Organizational Research Methods, 1(4), pp. 374–406.
Dye, R. A. (2004) Strategy selection and performance measurement choice when profit drivers are
uncertain, Management Science, 50(12), pp. 1624.
Eisenhardt, K. M. and Martin, J. A. (2000) Dynamic capabilities: what are they? Strategic Manage-
ment Journal, 21, pp. 1105–1121.
Fiol, C. M. and Lyles, M. A. (1985) Organizational learning, Academy of Management Review, 10(4),
pp. 803–813.
Forza, C. and Salvador, F. (2000) Assessing some distinctive dimensions of performance feedback
information in high performing plants, International Journal of Operations and Production
Management, 20(3), pp. 359–385.
Galbraith, J. (1973) Designing Complex Organizations (Reading, MA: Addison-Wesley).
Govindarajan, V. (1984) Appropriateness of accounting data in performance evaluation: an empirical
examination of environmental uncertainty as an intervening variable, Accounting, Organizations
and Society, 9(2), pp. 125–135.
Hu, L. and Bentler, P. M. (1995) Evaluating model fit, in: R.H. Doyle (Ed.) Structural Equation
Modeling (Thousand Oaks, CA: Sage).
Hult, G. T.M. (1998) Managing the international strategic sourcing process as a market-driven
organizational learning system, Decision Sciences, 29(1), pp. 193–216.
92 J.-F. Henri
Hult, G. T.M. and Ketchen, D. J. (2001) Does market orientation matter? A test of the
relationship between positional advantages and performance, Strategic Management Journal,
22, pp. 899–906.
Hurley, R. F. and Hult, G. T.M. (1998) Innovation, market orientation, and organizational learning: an
integration and empirical examination, Journal of Marketing, 62, pp. 42–54.
Ittner, C. D. and Larcker, D. F. (2001) Assessing empirical research in managerial accounting: a value-
based management perspective, Journal of Accounting and Economics, 32, pp. 349–410.
Ittner, C. D., Lanen, W. N. and Larcker, D. F. (2002) The association between activity-based costing
and manufacturing performance, Journal of Accounting Research, 40(3), pp. 711.
Ittner, C. D., Larcker, D. F. and Meyer, M. W. (2003a) Subjectivity and the weighting of performance
measures: evidence from a balanced scorecard, The Accounting Review, 78(3), pp. 725–758.
Ittner, C. D., Larcker, D. F. and Randall, T. (2003b) Performance implications of strategic perform-
ance measurement in financial services firms, Accounting, Organizations and Society, 28(7-8),
pp. 715–741.
Kennerley, M. and Neely, A. (2002) A framework of the factors affecting the evolution of perform-
ance measurement systems, International Journal of Operations and Production Management,
22(11), pp. 1222–1245.
Kennerley, M. and Neely, A. (2003) Measuring performance in a changing business environment,
International Journal of Operations and Production Management, 23(2), pp. 213–229.
Kogut, B. and Zander, U. (1992) Knowledge of the firm, combinative capabilities, and the replication
of technology, Organization Science, 3, pp. 383–397.
Kuwaiti, M. E. (2004) Performance measurement process: definition and ownership, International
Journal of Operations and Production Management, 24(1), pp. 55–78.
Lawrence, P. R. and Lorsch, J. (1967) Organization and Environment (Boston, MA: Harvard Business
School, Division of Research).
Lingle, J. H. and Schiemann, W. A. (1996). From balanced scorecard to strategic gauges: is measure-
ment worth it? Management Review, March, pp. 56–61.
Lipe, M. G. and Salterio, S. E. (2000) The balanced scorecard: judgemental effects of common and
unique performance measures, The Accounting Review, 75(3), pp. 283–298.
Luft, J. and Shields, M. D. (2003) Mapping management accounting: making structural models from
theory-based empirical research, Accounting, Organizations and Society, 28, pp. 169–249.
Lynch, R. L. and Cross, K. F. (1991) Measure Up – Yardsticks for Continuous Improvement
(Cambridge, MA: Basil Blackwell).
Malina, M. A. and Selto, F. H. (2004) Choice and change of measures in performance measurement
models, 15, pp. 441–469.
McMann, P. and Nanni, A. J. (1994) Is your company really measuring performance? Management
Accounting, 76(5), pp. 55–68.
Meyer, M. W. and Gupta, V. (1994) The performance paradox, in: B. Staw and L.L. Cummings (Eds)
Research in Organizational Behavior (Greenwich, CT: JAI Press).
Milgrom, P. and Roberts, J. (1992) Economics, Organizations and Management (Englewood Cliffs,
NJ: Prentice Hall).
Miller, D. (1987) The structural environmental correlates of business strategy, Strategic Management
Journal, 8(1), pp. 55–76.
Miller, D. (1988) Relating Porter’s business strategies to environment and structure: analysis and
performance implications, Academy of Management Journal, 31(2), pp. 280–308.
Milliken, F. J. (1987) Three types of perceived uncertainty about the environment: state, effect and
response uncertainty, Academy of Management Review, 12, pp. 133–143.
Moers, F. (2005) Discretion and bias in performance evaluation: the impact of diversity and subjec-
tivity, Accounting, Organizations and Society, 30, pp. 67–80.
Moers, F. (2006) Performance measure properties and delegation, The Accounting Review, 81(4),
pp. 897–924.
Nunnally, J. C. (1967) Psychometric Theory (New-York: McGraw-Hill).
The Periodic Review of Performance Indicators 93
Otley, D. (1999) Performance management: a framework for management control systems research,
Management Accounting Research, 10, pp. 363–82.
Porter, M. E. (1985) Competitive Advantage (New York: The Free Press).
Schreyögg, G. and Steinmann, H. (1987) Strategic control: a new perspective, Academy of Manage-
ment Review, 12(1), pp. 91–103.
Schoonhoven, C. B. (1981) Problems with contingency theory: testing assumptions hidden within the
language of contingency ‘theory’, Administrative Science Quarterly, 26(3), pp. 349.
Simons, R. (1990) The role of management control systems in creating competitive advantage: new
perspectives, Accounting, Organizations and Society, 15(1/2), pp. 127–143.
Simons, R. (2000) Performance Measurement and Control Systems for Implementing Strategy (Upper
Saddle River, NJ: Prentice Hall).
Tabachnick, B. G. and Fidell, L. S. (2001) Using Multivariate Statistics, 4th ed. (Needham Heights,
MA: Allyn and Bacon).
Teece, D. J., Pisano, G. and Shuen, A. (1997) Dynamic capabilities and strategic management,
Strategic Management Journal, 18(7), pp. 509–533.
Tymon, W. G., Stout, D. E. and Shaw, K. N. (1998) Critical analysis and recommendations regarding
the role of perceived environmental uncertainty in behavioral accounting research, Behavioral
Research in Accounting, 10, pp. 23–46.
Ullrich, M. J. and Tuttle, B. M. (2004) The effects of comprehensive information reporting systems
and economic inventives on managers’ time-planning decisions, Behavioral Research in
Accounting, 16, pp. 89–105.
Van de Ven, A. H. (1986) Central problems in the management of innovation, Management Science,
32(5), pp. 590–607.
Venkatraman, N. and Ramanujam, V. (1987) Measurement of business economic performance: an
examination of method convergence, Journal of Management, 13(1), pp. 109–122.
Vitale, M. R. and Mavrinac, S. C. (1995) How effective is your performance measurement system?
Management Accounting, 77(2), pp. 43–55.
Waggoner, D. B., Neely, A. D. and Kennerley, M. P. (1999) The forces that shape organisational per-
formance measurement systems: an interdisciplinary review, International Journal of Production
Economicss, 60–61, pp. 53–60.
Ward, P. T., Bickford, D. J. and Leong, G. K. (1996) Configurations of manufacturing strategy,
business strategy, environment and structure, Journal of Management, 22(4), pp. 597–626.
Wernerfelt, B. (1984) A resource-based view of the firm, Strategic Management Journal, 5(2),
pp. 171–180.
Wisner, J. D. and Fawcett, S. E. (1991) Linking firm strategy to operating decisions through perform-
ance measurement, Production and Inventory Management Journal, 32(3), pp. 5–11.
94 J.-F. Henri
Number of Employees %
fewer than 499 66%
Between 500 and 999 18%
Between 1 000 and 4 999 13%
Between 5 000 and 9 999 2%
Between 10 000 and 19 999 1%
Average 796
Industry Classification %
20 Food and kindred products 8,4%
21 Tobacco manufactures 0,3%
22 Textile mill products 3,1%
23 Apparel and other textile products 4,2%
24 Lumber and wood products 10,4%
25 Furniture and fixture 4,2%
26 Paper and allied products 8,1%
27 Printing and publishing 1,8%
28 Chemicals and allied products 4,4%
29 Petroleum and coal products 1,6%
30 Rubber and misc. plastics products 3,9%
31 Leather and leather products 1,3%
32 Stone, clay, glass, and concrete products 3,1%
33 Primary metal industries 6,0%
34 Fabricated metal products 10,4%
35 Industrial machinery and equipment 10,4%
36 Electrical and electronic equipment 7,3%
37 Transportation equipment 7,3%
38 Instrument and related products 2,3%
39 Misc. manufacturing industries 1,3%
The Periodic Review of Performance Indicators 95
2- Strategic capabilities
Please indicate the extent to which the following items describe your organization
Scale: 1 ¼ not descriptive to 7 ¼ very descriptive
(Continued)
96 J.-F. Henri
Continued
Confirmatory factor analysis
Second-
First-order order Cronbach
Items (second order construct) loadings loadings Alpha
Once we quit learning we 1.018
endanger our future
Employee learning is an 1.109
investment, not an expense
Goodness-of-fit of the model: x2 (26) ¼ 72.80 p , 0.001; NNFI ¼ 0.963;
CFI ¼ 0.973; RMSEA ¼ 0.067
Note: Significant at the 0.05 level Significant at the 0.01 level.
Confirmatory Cronbach
Items (first order construct) factor analysis Alpha
Performance indicators were deleted from the 0.726 0.76
measurement system
Performance indicators were added to within the 1.218
measurement system
Changes occurred in performance targets 1.209
Changes occurred in the definition of performance 1.308
indicators
Goodness-of-fit of the model: x2 (2) ¼ 0.687 p . .001;
NNFI ¼ 0.998; CFI¼ 1.0; RMSEA ¼ 0.0001
Note: Significant at the 0.05 level Significant at the 0.01 level.
4- Organizational performance
Please rate the performance of your organization against initial expectations on
each of the following dimensions for the past 12 months.
Scale: 1 ¼ not at all satisfactory to 7 ¼ outstanding