Professional Documents
Culture Documents
Abstract—Pattern classification systems are commonly used in adversarial applications, like biometric authentication, network
intrusion detection, and spam filtering, in which data can be purposely manipulated by humans to undermine their operation. As this
adversarial scenario is not taken into account by classical design methods, pattern classification systems may exhibit vulnerabilities,
whose exploitation may severely affect their performance, and consequently limit their practical utility. Extending pattern classification
theory and design methods to adversarial settings is thus a novel and very relevant research direction, which has not yet been pursued
in a systematic way. In this paper, we address one of the main open issues: evaluating at design phase the security of pattern classifiers,
namely, the performance degradation under potential attacks they may incur during operation. We propose a framework for empirical
arXiv:1709.00609v1 [cs.LG] 2 Sep 2017
evaluation of classifier security that formalizes and generalizes the main ideas proposed in the literature, and give examples of its use
in three real applications. Reported results show that security evaluation can provide a more complete understanding of the classifier’s
behavior in adversarial environments, and lead to better design choices.
Index Terms—Pattern classification, adversarial classification, performance evaluation, security evaluation, robustness evaluation.
and generalize them in our framework (Sect. 3). First, to of an attack, that ranges from targeted to indiscriminate,
pursue security in the context of an arms race it is not depending on whether the attack focuses on a single
sufficient to react to observed attacks, but it is also nec- or few specific samples (e.g., a specific spam email
essary to proactively anticipate the adversary by predicting misclassified as legitimate), or on a wider set of samples.
the most relevant, potential attacks through a what-if
analysis; this allows one to develop suitable countermea- 2.2 Limitations of classical performance evaluation
sures before the attack actually occurs, according to the methods in adversarial classification
principle of security by design. Second, to provide practi-
cal guidelines for simulating realistic attack scenarios, we Classical performance evaluation methods, like k-fold
define a general model of the adversary, in terms of her cross validation and bootstrapping, aim to estimate the
goal, knowledge, and capability, which encompasses and performance that a classifier will exhibit during operation,
generalizes models proposed in previous work. Third, by using data D collected during classifier design.2 These
since the presence of carefully targeted attacks may affect methods are based on the stationarity assumption that the
the distribution of training and testing data separately, data seen during operation follow the same distribution
we propose a model of the data distribution that can as D. Accordingly, they resample D to construct one or
formally characterize this behavior, and that allows us to more pairs of training and testing sets that ideally follow
take into account a large number of potential attacks; we the same distribution as D [9]. However, the presence of
also propose an algorithm for the generation of training an intelligent and adaptive adversary makes the clas-
and testing sets to be used for security evaluation, sification problem highly non-stationary, and makes it
which can naturally accommodate application-specific difficult to predict how many and which kinds of attacks
and heuristic techniques for simulating attacks. a classifier will be subject to during operation, that is,
In Sect. 4 we give three concrete examples of applica- how the data distribution will change. In particular,
tions of our framework in spam filtering, biometric au- the testing data processed by the trained classifier can
thentication, and network intrusion detection. In Sect. 5, be affected by both exploratory and causative attacks,
we discuss how the classical design cycle of pattern while the training data can only be affected by causative
classifiers should be revised to take security into account. attacks, if the classifier is retrained online [11], [14], [15].3
Finally, in Sect. 6, we summarize our contributions, the In both cases, during operation, testing data may follow
limitations of our framework, and some open issues. a different distribution than that of training data, when
the classifier is under attack. Therefore, security evalu-
ation can not be carried out according to the classical
2 BACKGROUND AND PREVIOUS WORK
paradigm of performance evaluation.4
Here we review previous work, highlighting the con-
cepts that will be exploited in our framework.
2.3 Arms race and security by design
2.1 A taxonomy of attacks against pattern classifiers Security problems often lead to a “reactive” arms race
between the adversary and the classifier designer. At
A taxonomy of potential attacks against pattern clas- each step, the adversary analyzes the classifier defenses,
sifiers was proposed in [11], [15], and subsequently and develops an attack strategy to overcome them. The
extended in [14]. We will exploit it in our framework, as designer reacts by analyzing the novel attack samples,
part of the definition of attack scenarios. The taxonomy and, if required, updates the classifier; typically, by re-
is based on two main features: the kind of influence of training it on the new collected samples, and/or adding
attacks on the classifier, and the kind of security violation features that can detect the novel attacks (see Fig. 1, left).
they cause. The influence can be either causative, if it Examples of this arms race can be observed in spam
undermines the learning algorithm to cause subsequent filtering and malware detection, where it has led to a
misclassifications; or exploratory, if it exploits knowl-
edge of the trained classifier to cause misclassifications, 2. By design we refer to the classical steps of [9], which include
without affecting the learning algorithm. Thus, causative feature extraction, model selection, classifier training and performance
evaluation (testing). Operation denotes instead the phase which starts
attacks may influence both training and testing data, when the deployed classifier is set to operate in a real environment.
or only training data, whereas exploratory attacks af- 3. “Training” data refers both to the data used by the learning
fect only testing data. The security violation can be an algorithm during classifier design, coming from D, and to the data
integrity violation, if it allows the adversary to access collected during operation to retrain the classifier through online
learning algorithms. “Testing” data refers both to the data drawn from
the service or resource protected by the classifier; an D to evaluate classifier performance during design, and to the data
availability violation, if it denies legitimate users access classified during operation. The meaning will be clear from the context.
to it; or a privacy violation, if it allows the adversary 4. Classical performance evaluation is generally not suitable for
non-stationary, time-varying environments, where changes in the data
to obtain confidential information from the classifier. distribution can be hardly predicted; e.g., in the case of concept drift
Integrity violations result in misclassifying malicious [20]. Further, the evaluation process has a different goal in this case,
samples as legitimate, while availability violations can i.e., to assess whether the classifier can “recover” quickly after a change
has occurred in the data distribution. Instead, security evaluation aims
also cause legitimate samples to be misclassified as ma- at identifying the most relevant attacks and threats that should be
licious. A third feature of the taxonomy is the specificity countered before deploying the classifier (see Sect. 2.3).
3
Fig. 1. A conceptual representation of the arms race in adversarial classification. Left: the classical “reactive” arms
race. The designer reacts to the attack by analyzing the attack’s effects and developing countermeasures. Right: the
“proactive” arms race advocated in this paper. The designer tries to anticipate the adversary by simulating potential
attacks, evaluating their effects, and developing countermeasures if necessary.
considerable increase in the variability and sophistica- a proactive/security-by-design approach. Attacks were
tion of attacks and countermeasures. simulated by manipulating training and testing samples
To secure a system, a common approach used in according to application-specific criteria only, without
engineering and cryptography is security by obscurity, reference to more general guidelines; consequently, such
that relies on keeping secret some of the system details techniques can not be directly exploited by a system
to the adversary. In contrast, the paradigm of security designer in more general cases.
by design advocates that systems should be designed A few works proposed analytical methods to evaluate
from the ground-up to be secure, without assuming the security of learning algorithms or of some classes
that the adversary may ever find out some important of decision functions (e.g., linear ones), based on more
system details. Accordingly, the system designer should general, application-independent criteria to model the
anticipate the adversary by simulating a “proactive” arms adversary’s behavior (including PAC learning and game
race to (i) figure out the most relevant threats and theory) [10], [12], [14], [16], [17], [37]–[40]. Some of these
attacks, and (ii) devise proper countermeasures, before criteria will be exploited in our framework for empirical
deploying the classifier (see Fig. 1, right). This paradigm security evaluation; in particular, in the definition of
typically improves security by delaying each step of the adversary model described in Sect. 3.1, as high-level
the “reactive” arms race, as it requires the adversary guidelines for simulating attacks.
to spend a greater effort (time, skills, and resources) to
find and exploit vulnerabilities. System security should
2.5 Building on previous work
thus be guaranteed for a longer time, with less frequent
supervision or human intervention. We summarize here the three main concepts more or
The goal of security evaluation is to address issue less explicitly emerged from previous work that will be
(i) above, i.e., to simulate a number of realistic attack exploited in our framework for security evaluation.
scenarios that may be incurred during operation, and 1) Arms race and security by design: since it is not
to assess the impact of the corresponding attacks on the possible to predict how many and which kinds of attacks
targeted classifier to highlight the most critical vulner- a classifier will incur during operation, classifier security
abilities. This amounts to performing a what-if analysis should be proactively evaluated using a what-if analysis,
[21], which is a common practice in security. This ap- by simulating potential attack scenarios.
proach has been implicitly followed in several previous 2) Adversary modeling: effective simulation of attack
works (see Sect. 2.4), but never formalized within a gen- scenarios requires a formal model of the adversary.
eral framework for the empirical evaluation of classifier 3) Data distribution under attack: the distribution of
security. Although security evaluation may also suggest testing data may differ from that of training data, when
specific countermeasures, the design of secure classifiers, the classifier is under attack.
i.e., issue (ii) above, remains a distinct open problem.
3 A FRAMEWORK FOR EMPIRICAL EVALUA -
2.4 Previous work on security evaluation TION OF CLASSIFIER SECURITY
Many authors implicitly performed security evaluation We propose here a framework for the empirical evalu-
as a what-if analysis, based on empirical simulation ation of classifier security in adversarial environments,
methods; however, they mainly focused on a specific that unifies and builds on the three concepts highlighted
application, classifier and attack, and devised ad hoc in Sect. 2.5. Our main goal is to provide a quantitative
security evaluation procedures based on the exploita- and general-purpose basis for the application of the
tion of problem knowledge and heuristic techniques what-if analysis to classifier security evaluation, based on
[1]–[6], [15], [22]–[36]. Their goal was either to point the definition of potential attack scenarios. To this end,
out a previously unknown vulnerability, or to evaluate we propose: (i) a model of the adversary, that allows
security against a known attack. In some cases, spe- us to define any attack scenario; (ii) a corresponding
cific countermeasures were also proposed, according to model of the data distribution; and (iii) a method for
4
generating training and testing sets that are representa- constraints (e.g., correlated features can not be modified
tive of the data distribution, and are used for empirical independently, and the functionality of malicious sam-
performance evaluation. ples can not be compromised [1], [3], [6], [10]).
Attack strategy. One can finally define the optimal
attack strategy, namely, how training and testing data
3.1 Attack scenario and model of the adversary
should be quantitatively modified to optimize the objec-
Although the definition of attack scenarios is ultimately tive function characterizing the adversary’s goal. Such
an application-specific issue, it is possible to give general modifications are defined in terms of: (a.i) how the class
guidelines that can help the designer of a pattern recog- priors are modified; (a.ii) what fraction of samples of
nition system. Here we propose to specify the attack each class is affected by the attack; and (a.iii) how fea-
scenario in terms of a conceptual model of the adversary tures are manipulated by the attack. Detailed examples
that encompasses, unifies, and extends different ideas are given in Sect. 4.
from previous work. Our model is based on the assump- Once the attack scenario is defined in terms of the
tion that the adversary acts rationally to attain a given adversary model and the resulting attack strategy, our
goal, according to her knowledge of the classifier, and her framework proceeds with the definition of the corre-
capability of manipulating data. This allows one to derive sponding data distribution, that is used to construct
the corresponding optimal attack strategy. training and testing sets for security evaluation.
Adversary’s goal. As in [17], it is formulated as the
optimization of an objective function. We propose to
3.2 A model of the data distribution
define this function based on the desired security vio-
lation (integrity, availability, or privacy), and on the attack We consider the standard setting for classifier design
specificity (from targeted to indiscriminate), according to in a problem which consists of discriminating between
the taxonomy in [11], [14] (see Sect. 2.1). For instance, the legitimate (L) and malicious (M) samples: a learning
goal of an indiscriminate integrity violation may be to algorithm and a performance measure have been chosen,
maximize the fraction of misclassified malicious samples a set D of n labelled samples has been collected, and
[6], [10], [14]; the goal of a targeted privacy violation a set of d features have been extracted, so that D =
may be to obtain some specific, confidential information {(xi , yi )}ni=1 , where xi denotes a d-dimensional feature
from the classifier (e.g., the biometric trait of a given user vector, and yi ∈ {L, M} a class label. The pairs (xi , yi )
enrolled in a biometric system) by exploiting the class are assumed to be i.i.d. samples of some unknown distri-
labels assigned to some “query” samples, while mini- bution pD (X, Y ). Since the adversary model in Sect. 3.1
mizing the number of query samples that the adversary requires us to specify how the attack affects the class
has to issue to violate privacy [14], [16], [41]. priors and the features of each class, we consider the
Adversary’s knowledge. Assumptions on the adver- classical generative model pD (X, Y ) = pD (Y )pD (X|Y ).5
sary’s knowledge have only been qualitatively discussed To account for the presence of attacks during operation,
in previous work, mainly depending on the application which may affect either the training or the testing data,
at hand. Here we propose a more systematic scheme or both, we denote the corresponding training and test-
for their definition, with respect to the knowledge of ing distributions as ptr and pts , respectively. We will just
the single components of a pattern classifier: (k.i) the write p when we want to refer to either of them, or both,
training data; (k.ii) the feature set; (k.iii) the learning and the meaning is clear from the context.
algorithm and the kind of decision function (e.g., a linear When a classifier is not under attack, according to
SVM); (k.iv) the classifier’s decision function and its the classical stationarity assumption we have ptr (Y ) =
parameters (e.g., the feature weights of a linear clas- pts (Y ) = pD (Y ), and ptr (X|Y ) = pts (X|Y ) = pD (X|Y ).
sifier); (k.v) the feedback available from the classifier, We extend this assumption to the components of ptr
if any (e.g., the class labels assigned to some “query” and pts that are not affected by the attack (if any), by
samples that the adversary issues to get feedback [14], assuming that they remain identical to the corresponding
[16], [41]). It is worth noting that realistic and minimal distribution pD (e.g., if the attack does not affect the class
assumptions about what can be kept fully secret from the priors, the above equality also holds under attack).
adversary should be done, as discussed in [14]. Examples The distributions p(Y ) and p(X|Y ) that are affected
of adversary’s knowledge are given in Sect. 4. by the attack can be defined as follows, according to the
Adversary’s capability. It refers to the control that definition of the attack strategy, (a.i-iii).
the adversary has on training and testing data. We Class priors. ptr (Y ) and pts (Y ) can be immediately
propose to define it in terms of: (c.i) the attack influence defined based on assumption (a.i).
(either causative or exploratory), as defined in [11], [14]; Class-conditional distributions. ptr (X|Y ) and
(c.ii) whether and to what extent the attack affects the pts (X|Y ) can be defined based on assumptions (a.ii-iii).
class priors; (c.iii) how many and which training and First, to account for the fact that the attack may not
testing samples can be controlled by the adversary in
5. In this paper, for the sake of simplicity, we use the lowercase
each class; (c.iv) which features can be manipulated, and notation p(·) to denote any probability density function, with both
to what extent, taking into account application-specific continuous and discrete random variables.
5
fake biometric trait to the system (e.g., a “gummy” that it is the same for all impostors. We will thus repeat
fingerprint or a photograph of a user’s face). The reason the evaluation twice, considering fingerprint and face
is that, to evade a multimodal system, one expects spoofing, separately. Defining how impostors can manip-
that the adversary should spoof all the corresponding ulate the features (in this case, the matching scores) in a
biometric traits. In this application example, we show spoofing attack is not straightforward. The standard way
how the designer of a multimodal system can verify if of evaluating the security of biometric systems against
this hypothesis holds, before deploying the system, by spoofing attacks consists of fabricating fake biometric
simulating spoofing attacks against each of the matchers. traits and assessing performance when they are submit-
To this end, we partially exploit the analysis in [1], [2]. ted to the system. However, this is a cumbersome task.
We consider a typical multimodal system, made up Here we follow the approach of [1], [2], in which the
of a fingerprint and a face matcher, which operates as impostor is assumed to fabricate “perfect” fake traits,
follows. The design phase includes the enrollment of i.e., fakes that produce the same matching score of the
authorized users (clients): reference templates of their “live” trait of the targeted client. This allows one to use
biometric traits are stored into a database, together with the available genuine scores to simulate the fake scores,
the corresponding identities. During operation, each user avoiding the fabrication of fake traits. On the other hand,
provides the requested biometric traits to the sensors, this assumption may be too pessimistic. This limitation
and claims the identity of a client. Then, each matcher could be overcome by developing more realistic models
compares the submitted trait with the template of the of the distribution of the fake traits; however, this is
claimed identity, and provides a real-valued matching a challenging research issue on its own, that is out of
score: the higher the score, the higher the similarity. We the scope of this work, and part of the authors’ ongoing
denote the score of the fingerprint and the face matcher work [47], [48].
respectively as xfing and xface . Finally, the matching Attack strategy. The above attack strategy modifies only
scores are combined through a proper fusion rule to the testing data, and: (a.i) it does not modify the class
decide whether the claimed identity is the user’s identity priors; (a.ii) it does not affect the genuine class, but all
(genuine user) or not (impostor). impostors; and (a.iii) any impostor spoofs the considered
As in [1], [2], we consider the widely used likelihood biometric trait (either the face or fingerprint) of the
ratio (LLR) score fusion rule [46], and the matching known identity. According to the above assumption, as
scores as its features. Denoting as x the feature vector in [1], [2], any spoofing attack is simulated by replacing
(xfing , xface ), the LLR can be written as: the corresponding impostor score (either the face or
( fingerprint) with the score of the targeted genuine user
L if p(x|Y = L)/p(x|Y = M) ≥ t, (chosen at random from the legitimate samples in the
LLR(x) = (5)
M if p(x|Y = L)/p(x|Y = M) < t, testing set). This can be represented with a function
A(x), that, given an impostor x = (xfing , xface ), returns
where L and M denote respectively the “genuine” and
either a vector (x0fing , xface ) (for fingerprint spoofing) or
“impostor” class, and t is a decision threshold set ac-
(xfing , x0face ) (for face spoofing), where x0fing and x0face are
cording to application requirements. The distributions
the matching scores of the targeted genuine user.
p(xfing , xface |Y ) are usually estimated from training data,
while t is estimated from a validation set. 2) Data model. Since training data is not affected, we
1) Attack scenario. Goal. In this case, each malicious set ptr (X, Y ) = pD (X, Y ). Then, we set (a.i) pts (Y ) =
user (impostor) aims at being accepted as a legitimate pD (Y ); and (a.ii) pts (A = T|Y = L) = 0 and pts (A =
(genuine) one. This corresponds to a targeted integrity T|Y = M) = 1. Thus, pts (X|Y = L) = pD (X|Y = L),
violation, where the adversary’s goal is to maximize the while all the malicious samples in TS have to be ma-
matching score. nipulated by A(x), according to (a.iii). This amounts to
Knowledge. As in [1], [2], we assume that each impostor empirically defining pts (X|Y = M, A = T). As in the spam
knows: (k.i) the identity of the targeted client; and (k.ii) filtering case, pts (X|Y = y, A = F), for y = L, M, will be
the biometric traits used by the system. No knowledge empirically approximated from the available data in D.
of (k.iii) the decision function and (k.iv) its parameters The definition of the above attack scenario and data
is assumed, and (k.v) no feedback is available from the model is summarized in Table 1 (second column).
classifier. 3) Construction of TR and TS. We use the NIST
Capability. We assume that: (c.i) spoofing attacks affect Biometric Score Set, Release 1.10 It contains raw simi-
only testing data (exploratory attack); (c.ii) they do not larity scores obtained on a set of 517 users from two
affect the class priors;9 and, according to (k.i), (c.iii) different face matchers (named ‘G’ and ‘C’), and from
each adversary (impostor) controls her testing malicious one fingerprint matcher using the left and right index
samples. We limit the capability of each impostor by finger. For each user, one genuine score and 516 impostor
assuming that (c.iv) only one specific trait can be spoofed scores are available for each matcher and each modality,
at a time (i.e., only one feature can be manipulated), and for a total of 517 genuine and 266,772 impostor samples.
We use the scores of the ‘G’ face matcher and the ones
9. Further, since the LLR considers only the ratio between the class-
conditional distributions, changing the class priors is irrelevant. 10. http://www.itl.nist.gov/iad/894.03/biometricscores/
10
Here we assume that an anomaly-based IDS is being note that: first, the best case for the adversary is when
designed, using a one-class ν-SVM classifier with a the attack samples exhibit the same histogram of the
radial basis function (RBF) kernel and the feature vector payload’s byte values as the malicious testing samples,
representation proposed in [49]. Each network packet as this intuitively forces the model of the normal traffic
is considered as an individual sample to be labeled as to be as similar as possible to the distribution of the
normal (legitimate) or anomalous (malicious), and is testing malicious samples; and, second, this does not
represented as a 256-dimensional feature vector, defined imply that the attack samples perform any intrusion,
as the histogram of byte frequencies in its payload (this is allowing them to avoid detection. In fact, one may
known as “1-gram” representation in the IDS literature). construct an attack packet which exhibit the same feature
We then focus on the model selection stage. In the values of an intrusive network packet (i.e., a malicious
above setting, it amounts to choosing the values of the ν testing sample), but looses its intrusive functionality, by
parameter of the learning algorithm (which is an upper simply shuffling its payload bytes. Accordingly, for our
bound on the false positive error rate on training data purposes, we directly set the feature values of the attack
[50]), and the γ value of the RBF kernel. For the sake of samples equal to those of the malicious testing samples,
simplicity, we assume that ν is set to 0.01 as suggested without constructing real network packets.
in [24], so that only γ has to be chosen. 2) Data model. Since testing data is not affected by
We show how the IDS designer can select a model the considered causative attack, we set pts (X, Y ) =
(the value of γ) based also on the evaluation of classifier pD (X, Y ). We then encode the attack strategy as fol-
security. We focus on a causative attack similar to the lows. According to (a.ii), ptr (A = T|Y = L) = 0
ones considered in [38], aimed at forcing the learned and ptr (A = T|Y = M) = 1. The former assumption
model of normal traffic to include samples of intrusions implies that ptr (X|Y = L) = pD (X|Y = L). Since the
to be attempted during operation. To this end, the attack training data of one-class classifiers does not contain any
samples should be carefully designed such that they malicious samples (see explanation at the beginning of
include some features of the desired intrusive traffic, this section), i.e., ptr (A = F|Y = M) = 0, the latter
but do not perform any real intrusion (otherwise, the assumption implies that the class prior ptr (Y = M)
collected traffic may be discarded, as explained above). corresponds exactly to the fraction of attack samples in
1) Attack scenario. Goal. This attack aims to cause the training set. Therefore, assumption (a.i) amounts to
an indiscriminate integrity violation by maximizing the setting ptr (Y = M) = pmax . Finally, according to (a.iii),
fraction of malicious testing samples misclassified as the distribution of attack samples ptr (X|Y = M, A =
legitimate. T) will be defined as the empirical distribution of the
M,T M,F
Knowledge. The adversary is assumed to know: (k.ii) malicious testing samples, namely, DTR = DTS . The
the feature set; and (k.iii) that a one-class classifier is distribution ptr (X|Y = L, A = F) = pD (X|L) will also be
L,F
used. No knowledge of (k.i) the training data and (k.iv) empirically defined as the set DTR .
the classifiers’ parameters is available to the adversary, The definition of the attack scenario and data model
as well as (k.v) any feedback from the classifier. is summarized in Table 1 (third column).
Capability. The attack consists of injecting malicious 3) Construction of TR and TS. We use the data set of
samples into the training set. Accordingly, we assume [24]. It consists of 1,699,822 legitimate samples (network
that: (c.i) the adversary can inject malicious samples packets) collected by a web server during five days
into the training data, without manipulating testing data in 2006, and a publicly available set of 205 malicious
(causative attack); (c.ii) she can modify the class priors by samples coming from intrusions which exploit the HTTP
injecting a maximum fraction pmax of malicious samples protocol [51].13 To construct TR and TS, we take into
into the training data; (c.iii) all the injected malicious account the chronological order of network packets as
samples can be manipulated; and (c.iv) the adversary in Sect. 4.1, and the fact that in this application all the
is able to completely control the feature values of the malicious samples in D must be inserted in the testing
malicious attack samples. As in [17], [28]–[30], we repeat set [27], [51]. Accordingly, we set DTR as the first 20,000
the security evaluation for pmax ∈ [0, 0.5], since it is legitimate packets of day one, and DTS as the first 20,000
unrealistic that the adversary can control the majority legitimate samples of day two, plus all the malicious
of the training data. samples. Since this attack does not affect testing samples,
Attack strategy. The goal of maximizing the percentage L,F
we set TS = DTS . We then set DTR = DTR , and
of malicious testing samples misclassified as legitimate, M,F L,T
DTR = DTR = ∅. Since the feature vectors of attack
for any pmax value, can be attained by manipulating the samples are identical to those of the malicious samples,
training data with the following attack strategy: (a.i) the M,T M,F
as above mentioned, we set DTR = DTS (where the
maximum percentage of malicious samples pmax will be latter set clearly includes the malicious samples in DTS ).
injected into the training data; (a.ii) all the injected mali- The size of TR is initially set to 20,000. We then consider
cious samples will be attack samples, while no legitimate different attack scenarios by increasing the number of
samples will be affected; and (a.iii) the feature values of attack samples in the training data, up to 40,000 samples
the attack samples will be equal to that of the malicious
testing samples. To understand the latter assumption, 13. http://www.i-pi.com/HTTP-attacks-JoCN-2006/
12
[28] B. Nelson, M. Barreno, F. J. Chi, A. D. Joseph, B. I. P. Rubinstein, [52] D. Sculley, G. Wachman, and C. E. Brodley, “Spam filtering using
U. Saini, C. Sutton, J. D. Tygar, and K. Xia, “Exploiting machine inexact string matching in explicit feature space with on-line
learning to subvert your spam filter,” in Proc. 1st Workshop on linear classifiers,” in 15th Text Retrieval Conf. NIST, 2006.
Large-Scale Exploits and Emergent Threats. CA, USA: USENIX [53] S. Z. Li and A. K. Jain, Eds., Enc. of Biometrics. Springer US, 2009.
Association, 2008, pp. 1–9. [54] B. Biggio, G. Fumera, and F. Roli, “Adversarial pattern classifica-
[29] B. I. Rubinstein, B. Nelson, L. Huang, A. D. Joseph, S.-h. Lau, tion using multiple classifiers and randomisation,” in Structural,
S. Rao, N. Taft, and J. D. Tygar, “Antidote: understanding and Syntactic, and Statistical Pattern Recognition, ser. LNCS, vol. 5342.
defending against poisoning of anomaly detectors,” in Proc. 9th FL, USA: Springer-Verlag, 2008, pp. 500–509.
ACM SIGCOMM Internet Measurement Conf., ser. IMC ’09. NY,
USA: ACM, 2009, pp. 1–14.
[30] M. Kloft and P. Laskov, “Online anomaly detection under ad- Battista Biggio received the M. Sc. degree in
versarial impact,” in Proc. 13th Int’l Conf. on Artificial Intell. and Electronic Eng., with honors, and the Ph. D. in
Statistics, 2010, pp. 405–412. Electronic Eng. and Computer Science, respec-
[31] O. Dekel, O. Shamir, and L. Xiao, “Learning to classify with tively in 2006 and 2010, from the University of
missing and corrupted features,” Machine Learning, vol. 81, pp. Cagliari, Italy. Since 2007 he has been working
149–178, 2010. for the Dept. of Electrical and Electronic Eng.
[32] B. Biggio, G. Fumera, and F. Roli, “Design of robust classifiers for of the same University, where he holds now a
adversarial environments,” in IEEE Int’l Conf. on Systems, Man, postdoctoral position. From May 12th, 2011 to
and Cybernetics, 2011, pp. 977–982. November 12th, 2011, he visited the University
of Tübingen, Germany, and worked on the secu-
[33] ——, “Multiple classifier systems for robust classifier design in
rity of machine learning algorithms to contami-
adversarial environments,” Int’l Journal of Machine Learning and
nation of training data. His research interests currently include: secure /
Cybernetics, vol. 1, no. 1, pp. 27–41, 2010.
robust machine learning and pattern recognition methods, multiple clas-
[34] B. Biggio, I. Corona, G. Fumera, G. Giacinto, and F. Roli, “Bagging sifier systems, kernel methods, biometric authentication, spam filtering,
classifiers for fighting poisoning attacks in adversarial environ- and computer security. He serves as a reviewer for several international
ments,” in Proc. 10th Int’l Workshop on Multiple Classifier Systems, conferences and journals, including Pattern Recognition and Pattern
ser. LNCS, vol. 6713. Springer-Verlag, 2011, pp. 350–359. Recognition Letters. Dr. Biggio is a member of the IEEE Institute of
[35] B. Biggio, G. Fumera, F. Roli, and L. Didaci, “Poisoning adaptive Electrical and Electronics Engineers (Computer Society), and of the
biometric systems,” in Structural, Syntactic, and Statistical Pattern Italian Group of Italian Researchers in Pattern Recognition (GIRPR),
Recognition, ser. LNCS, vol. 7626. Springer, 2012, pp. 417–425. affiliated to the International Association for Pattern Recognition.
[36] B. Biggio, B. Nelson, and P. Laskov, “Poisoning attacks against
support vector machines,” in Proc. 29th Int’l Conf. on Machine
Learning, 2012. Giorgio Fumera received the M. Sc. degree
in Electronic Eng., with honors, and the Ph.D.
[37] M. Kearns and M. Li, “Learning in the presence of malicious
degree in Electronic Eng. and Computer Sci-
errors,” SIAM J. Comput., vol. 22, no. 4, pp. 807–837, 1993.
ence, respectively in 1997 and 2002, from the
[38] A. A. Cárdenas, J. S. Baras, and K. Seamon, “A framework for the University of Cagliari, Italy. Since February 2010
evaluation of intrusion detection systems,” in Proc. IEEE Symp. on he is Associate Professor of Computer Eng.
Security and Privacy. DC, USA: IEEE CS, 2006, pp. 63–77. at the Dept. of Electrical and Electronic Eng.
[39] B. Biggio, G. Fumera, and F. Roli, “Multiple classifier systems of the same University. His research interests
for adversarial classification tasks,” in Proc. 8th Int’l Workshop on are related to methodologies and applications of
Multiple Classifier Systems, ser. LNCS, vol. 5519. Springer, 2009, statistical pattern recognition, and include mul-
pp. 132–141. tiple classifier systems, classification with the
[40] M. Brückner, C. Kanzow, and T. Scheffer, “Static prediction games reject option, adversarial classification and document categorization.
for adversarial learning problems,” J. Mach. Learn. Res., vol. 13, pp. On these topics he published more than sixty papers in international
2617–2654, 2012. journal and conferences. He acts as reviewer for the main interna-
[41] A. Adler, “Vulnerabilities in biometric encryption systems,” in 5th tional journals in this field, including IEEE Trans. on Pattern Analysis
Int’l Conf. on Audio- and Video-Based Biometric Person Authentication, and Machine Intelligence, Journal of Machine Learning Research, and
ser. LNCS, vol. 3546. NY, USA: Springer, 2005, pp. 1100–1109. Pattern Recognition. Dr. Fumera is a member of the IEEE Institute of
[42] B. Efron and R. J. Tibshirani, An Introduction to the Bootstrap. New Electrical and Electronics Engineers (Computer Society), of the Italian
York: Chapman & Hall, 1993. Association for Artificial Intelligence (AI*IA), and of the Italian Group
[43] H. Drucker, D. Wu, and V. N. Vapnik, “Support vector machines of Italian Researchers in Pattern Recognition (GIRPR), affiliated to the
for spam categorization,” IEEE Trans. on Neural Networks, vol. 10, International Association for Pattern Recognition.
no. 5, pp. 1048–1054, 1999.
[44] F. Sebastiani, “Machine learning in automated text categoriza- Fabio Roli received his M. Sc. degree, with hon-
tion,” ACM Comput. Surv., vol. 34, pp. 1–47, 2002. ors, and Ph. D. degree in Electronic Eng. from
[45] C.-C. Chang and C.-J. Lin, “LibSVM: a library for support vector the University of Genoa, Italy. He was a member
machines,” 2001. [Online]. Available: http://www.csie.ntu.edu. of the research group on Image Processing and
tw/∼cjlin/libsvm/ Understanding of the University of Genoa, Italy,
[46] K. Nandakumar, Y. Chen, S. C. Dass, and A. Jain, “Likelihood from 1988 to 1994. He was adjunct professor
ratio-based biometric score fusion,” IEEE Trans. on Pattern Analysis at the University of Trento, Italy, in 1993 and
and Machine Intell., vol. 30, pp. 342–347, February 2008. 1994. In 1995, he joined the Dept. of Electrical
[47] B. Biggio, Z. Akhtar, G. Fumera, G. Marcialis, and F. Roli, “Robust- and Electronic Eng. of the University of Cagliari,
ness of multi-modal biometric verification systems under realistic Italy, where he is now professor of computer
spoofing attacks,” in Int’l Joint Conf. on Biometrics, 2011, pp. 1–6. engineering and head of the research group on
[48] B. Biggio, Z. Akhtar, G. Fumera, G. L. Marcialis, and F. Roli, pattern recognition and applications. His research activity is focused
“Security evaluation of biometric authentication systems under on the design of pattern recognition systems and their applications
real spoofing attacks,” IET Biometrics, vol. 1, no. 1, pp. 11–24, 2012. to biometric personal identification, multimedia text categorization, and
[49] K. Wang and S. J. Stolfo, “Anomalous payload-based network computer security. On these topics, he has published more than two
intrusion detection,” in RAID, ser. LNCS, vol. 3224. Springer, hundred papers at conferences and on journals. He was a very active
2004, pp. 203–222. organizer of international conferences and workshops, and established
[50] B. Schölkopf, A. J. Smola, R. C. Williamson, and P. L. Bartlett, the popular workshop series on multiple classifier systems. Dr. Roli is
“New support vector algorithms,” Neural Comput., vol. 12, no. 5, a member of the governing boards of the International Association for
pp. 1207–1245, 2000. Pattern Recognition and of the IEEE Systems, Man and Cybernetics
[51] K. Ingham and H. Inoue, “Comparing anomaly detection tech- Society. He is Fellow of the IEEE, and Fellow of the International
niques for http,” in Recent Advances in Intrusion Detection, ser. Association for Pattern Recognition.
LNCS. Springer, 2007, pp. 42–62.