Professional Documents
Culture Documents
Guangdong Key Lab of Information Security, School of Data and Computer Science,
Sun Yat-Sen University, Guangzhou 510006, China
isskxg@mail.sysu.edu.cn
Abstract. During the past decade, image forensics has made rapid
progress due to the growing concern of image content authenticity. In
order to remove or conceal the traces that forensics based on, some far-
sighted forgers take advantage of so-called anti-forensics to make their
forgery more convincing. To rebuild the credibility of forensics, many
countermeasures against anti-forensics have been proposed. This paper
presents a multi-purpose approach to detect various anti-forensics based
on the architecture of Convolutional Neural Networks (CNN), which can
automatically extract features and identify the forged types. Our model
can detect various image anti-forensics both in binary and multi-class
decision effectively. Experimental results show that the proposed method
performs well for multiple well-known image anti-forensic methods.
1 Introduction
With the popularity of digital cameras and powerful image processing software
tools, it is easy to record, edit and share photos. In law enforcement, digital image
content plays an important role in deciding a case. Since the digital image can be
easily altered by the forgers without leaving perceptible artifacts, its authenticity
and reliability must be verified. As a result, image forensics has raised more
and more attention over the last decade [1]. Various digital forensic techniques
[5,7,8,12,13] are proposed to detect the image processing operations, e.g., JPEG
compression, median filtering, resampling and contrast enhancement, which are
considered in our experiments. On the other hand, the farsighted forgers attempt
to fool the investigators by using anti-forensic techniques [2,4,6,9,10,14,15] to
remove or hide the traces that they may leave after certain operations. To restore
authenticity and rebuild the credibility, many countermeasures [3,11,22–24] have
been proposed to detect different anti-forensics.
X. Kang—This work was supported by NSFC (Grant nos. 61379155, U1536204,
61502547, 61332012, 61272453 and NSF of Guangdong province (Grant no.
s2013020012788).
c Springer International Publishing AG 2017
Y.Q. Shi et al. (Eds.): IWDW 2016, LNCS 10082, pp. 3–15, 2017.
DOI: 10.1007/978-3-319-53465-7 1
4 J. Yu et al.
2 Background
In this section, we briefly review the image anti-forensic processing operations
and its traces left.
Fig. 1. DCT coefficient histogram of an image (a) before JPEG compression, (b) after
JPEG compression, (c) after JPEG anti-forensic operation.
traces, the authors [3] proposed a countering JPEG anti-forensic method. In [4],
JPEG blocking artifact is eliminated by median filtering followed by the dither.
This method can significantly disguise the traces left by JPEG compression and
decrease the detection accuracy of the aforementioned work [3]. To detect such
anti-forensic method, another countering method is proposed by estimating the
noise level of the questioned images [22].
3.1 Convolution
In this paper, we recall the major concepts of CNN, which is a deep learning
network that has proved its efficiency in image classification competitions. The
most important operation of CNN is convolution, which is done between the
input image and a filter. There are many filters in a convolutional layer, and
each one leads to a filtered image. Then the second and third steps are applied,
leading to a new image called feature map. Compare to fully connected layer,
convolutional layer shows much better performance in extracting feature because
number of convolutional layers. A convolution from the first layer with the k th
(l)
field leads to a filtered image, denote Ik which also been called feature maps),
such that:
Ik = I (0) ∗ Fkl
(l)
(1)
Generally, a convolutional layer is made of three steps, i.e. convolution, acti-
vation, and pooling. These three consecutive steps can be summarized by exam-
ining the link from the feature map of one layer to that of its previous layer:
l−1
i=K
Ikl Iil−1 ∗ Fk,i
l
(l)
= pool(f (bk + )) (2)
i=1
(l)
with bk being a scalar used for setting a bias to the convolution, f () being the
activation function applied pixel per pixel to the filtered image, and the pooling,
pool(), which pools a local neighborhood, and K l being the number of filters
the lth layer has. Note that the filter kernel (weights), and the bias have to be
learned and are modified by back-propagation of the error.
Since convolution part shows above and anti-forensic operation can both be
presented through a combination of linear and non-linear function, the proposed
convolutional layer is supposed to be able to extract the anti-forensic feature.
In order to see the feature map actually extract useful feature with increasing
depth of convolutional layers, the feature maps of the proposed CNN network
is displayed in Fig. 3. Figure 3(a)–(d) are the feature maps of an original image
obtained from the convolutional layer 1, 2, 3, and 4 respectively. Figure 3(a) and
(b) have 32 feature maps, (c) and (d) have 40 feature maps. The bright area in
a feature map indicates large weights and vice versa.
The feature maps in Fig. 4 are generated by the 4th convolutional layer with
original and anti-forensic images test in our experiments. 40 feature maps are
included in one subgraph. We can see perceptible difference of units in the same
position of certain feature map between original and anti-forensic images. Median
filtering and JPEG operation will cause the image fuzzy phenomenon, and their
anti-forensics both altered by adding corresponding noises, thus it will make
their images details blurred. As seen in Fig. 4(a)–(c), the features of anti-JPEG
and anti-MF are similar and both fuzzy compared to the original image. Anti-CE
8 J. Yu et al.
Fig. 3. Feature maps of original image (a)–(d) are obtained from the convolutional
layer 1, 2, 3, 4 respectively.
is aimed to falsify the forensics and achieve the effect of contrast enhancement.
There is a sharp contrast in the feature maps of anti-CE image compared to
the original. In the operation of resampling, the image will resize according to
the specified scaling factors. Seen in Fig. 4(e)–(f), the feature of the up-sampling
image is just like an amplifier of the original one, while the down-sampling
feature includes more textures than the original. The analysis above shows that
our network can extract useful and decisive features when dealing the counter-
anti-forensic task during the training process.
3.2 Pooling
Pooling layers in CNN summarize the outputs of previous layers and operates
independently on each depth slice of the input, as well as resizing it spatially.
Pooling layer can not only effectively reduce training complexity by eliminating
parameters, but also play an important role in preventing overfitting. Pooling
layers usually use MAX operation. Normally, the neighborhoods summarized by
adjacent pooling units do not overlap, but when it is applied on image anti-
forensics, it is necessary to preserve the adjacent area for better performance.
Based on our experiment, overlapping max pooling layers reduce the error rate
by 2% or more.
3.3 Activation
The Rectified Linear Units (ReLU) is used in our work due to its fast conver-
gence in large datasets [27]. It also increases the nonlinear properties of the
A Multi-purpose Image Counter-anti-forensic Method 9
Fig. 4. The feature maps of (a) Original image, (b) Anti-JPEG image, (c) Anti-MF
image, (d) Anti-CE image, (e) Anti-RES image (up-sampling) and (f) Anti-RES image
(down-sampling) from the last convolutional layer respectively.
decision function and the overall network. The convolutional and non-linearity
transformation can be denoted as:
n
xkj = g( xik−1 ∗ wij
k−1
+ bkj ) (3)
i=1
where ‘*’ denotes convolutional operation; xkj is the jth output feature map in
k−1
the kth layer; wij is the weight at layer k − 1, i, j position; bkj is the bias
parameter, and g denotes non-linearity transformation where ReLU has been
used. ReLU, which has been proved to have better convergence properties than
the sigmoid and tanh functions [17], can be expressed as:
g(x) = max(0, x) (4)
3.4 Dropout
The classification layer consists mainly of two parts: the softmax layer and
the fully connected layer. We adopt a widely used technique called dropout to
avoid overfitting, which randomly sets certain output of fully connected layers to
zero with a probability of 0.5. The softmax layer, trained using a cross-entropy
scheme, is used for classification, which is essentially a nonlinear multinomial
logistic regression. Back propagation algorithm is used to fine-tune weights and
biases for training the parameters in the proposed CNN. Consequently, CNN is
able to use the classification result to adjust the extracted features and build a
better learning mechanism for measuring and classifying anti-forensic methods
automatically.
10 J. Yu et al.
4 Experimental Result
In this section, we first show the network settings of the proposed CNN architec-
ture. Then, the proposed method is tested on detecting six anti-forensic methods
separately in Sect. 4.2. Finally, the detection performance for the multi-class clas-
sification of anti-forensics is given. A total of 10000 original images from BOSS-
base [18] are cropped from the center to a size of 128 × 128 pixels. Without loss
of generality, grayscale images are adopted in all of our experiments. The tam-
pered images are created via different anti-forensics manipulations mentioned in
Sect. 2.
Anti-method Countermeasure
[3] [22] Proposed
Anti-JPEG [2] 80.0 55.78 98.83
Anti-JPEG [4] 58.02 85.87 99.95
Anti-method Countermeasure
[23] Proposed
Anti-MF [14] - 99.38
Anti-MF [15] 98.45 99.35
12 J. Yu et al.
Anti-method Countermeasure
[24] Proposed
Anti-RES [6] 97.67 99.38
method reaches 99.38%, 1.6% higher than the state-of-art specific countermea-
sure [24].
Test/prediction Original Anti-JPEG [2] Anti-JPEG [4] Anti-MF [15] Anti-MF [14] Anti-CE Anti-Res
Original 94.15 * * 0 * 2.35 2.5
Anti-JPEG [2] 1.6 97.1 * 0 0 * *
Anti-JPEG [4] 0 * 99.3 * 0 0 *
Anti-MF [15] 0 * * 99.4 * 0 *
Anti-MF [14] * 0 0 0 99.5 * *
Anti-CE 6.05 * * 0 * 91.75 0
Anti-RES * * * * 1 * 97.35
We also test the proposed method when the database is changed. The new
test database is created as follows: we use 1000 original images from UCID with
A Multi-purpose Image Counter-anti-forensic Method 13
the size of 128 × 128, and create the six corresponding anti-forensic images, 6000
forged images in total. Then we use the model trained with BOSSbase to classify
the database created from UCID. The detection accuracy is 94.7% in this case,
which demonstrates that the proposed method is not strongly related to the
database.
Four different anti-forensic (anti-JPEG [2], anti-MF [15], anti-RES [6], anti-CE
are included) images are used to create 10000 forged images, each type responsi-
ble for 2500. We first train a classifier with 10000 original images from BOSSbase
and the mentioned forged image dataset. Then, we create two test dataset with
the anti-forensic methods (anti-JPEG [4] and anti-MF [14], respectively) that
are not included in the trained dataset. 1000 original images from UCID and
1000 corresponding anti-forensic images are used to create the test dataset. The
accuracy of detecting the forgery of [4] is 98.7%, and the forgery of [14] is 98.65%
respectively, which indicates that the proposed method is able to extend to other
diverse image anti-forensics without much prior information.
5 Conclusion
References
1. Stamm, M.M.C., Wu, M., Liu, K.J.R.: Information forensics: an overview of the
first decade. In: IEEE Access, vol. 1, pp. 167–200, May 2013
2. Stamm, M.M.C., Tjoa, S.K., Lin, W.S., et al.: Anti-forensics of JPEG compression.
In Proceedings of IEEE International Conference on Acoustics, Speech, and Signal
Processing, pp. 1694–1697, March 2010
3. Valenzise, M.G., Tagliasacchi, M., Tubaro, S.: Revealing the traces of JPEG com-
pression anti-forensics. IEEE Trans. Inf. Forensics Secur. 8(2), 335–349 (2013)
4. Stamm, M.M.C., Liu, K.J.R.: Anti-forensics of digital image compression. IEEE
Trans. Inf. Forensics Secur. 6(3), 1050–1065 (2011)
14 J. Yu et al.
5. Popescu, A.A.C., Farid, H.: Exposing digital forgeries by detecting traces of resam-
pling. IEEE Trans. Signal Process. 53(2), 758–767 (2005)
6. Kirchner, M.M., Böhme, R.: Hiding traces of resampling in digital images. IEEE
Trans. Inf. Forensics Secur. 3(4), 582–592 (2008)
7. Stamm, M.M.C., Liu, K.J.R.: Forensic detection of image manipulation using sta-
tistical intrinsic fingerprints. IEEE Trans. Inf. Forensics Secur. 5(3), 492–506 (2010)
8. Stamm, M.M.C., Liu, K.J.R.: Blind forensics of contrast enhancement in digital
images. In: Proceedings of IEEE International Conference on Image Processing,
San Diego, CA, pp. 3112–3115, October 2008
9. Cao, G.G., Zhao, Y., Ni, R., Tian, H.: Anti-forensics of contrast enhancement in
digital images. In: Proceedings of ACM Workshop Multimedia and Security, pp.
25–34 (2010)
10. Kwok, C.-W., Au, O.C., Chui, S.-H.: Alternative anti-forensics method for con-
trast enhancement. In: Shi, Y.Q., Kim, H.-J., Perez-Gonzalez, F. (eds.) IWDW
2011. LNCS, vol. 7128, pp. 398–410. Springer, Heidelberg (2012). doi:10.1007/
978-3-642-32205-1 32
11. Valenzise, G.G., Tagliasacchi, M., Tubaro, S.: Revealing the traces of JPEG com-
pression anti-forensics. IEEE Trans. Inf. Forensics Secur. 8(2), 335–349 (2013)
12. Yuan, H.H.: Blind forensics of median filtering in digital images. IEEE Trans. Inf.
Forensics Secur. 6(4), 1335–1345 (2011)
13. Kirchner, M.M., Fridrich, J.: On detection of median filtering in digital images. In:
Memon, N.D., Dittmann, J., Alattar, A.M., Delp, E.J. (Eds.) Media Forensics and
Security II, part of the IS&T-SPIE Electronic Imaging Symposium, Proceedings,
vol. 7541 of SPIE Proceedings, San Jose, CA, USA, 18–20 January 2010, p. 754110,
SPIE (2010)
14. Dang-Nguyen, D.D.T., Gebru, I.D., Conotter, V., et al.: Counter-forensics of
median filtering. In: 2013 IEEE 15th International Workshop on Multimedia Signal
Processing (MMSP), pp. 260–265, September 2013
15. Wu, Z.Z.H., Stamm, M.C., Liu, K.J.R.: Anti-forensics of median filtering. In:
Proceedings of IEEE International Conference on Acoustic, Speech, and Signal
Processing, pp. 3043–3047, Vancouver, Canada, May 2013
16. Simonyan, K.K., Zisserman, A.: Very deep convolutional networks for large-scale
image recognition. In: Proceedings of ICLR (2015)
17. Krizhevsky, A.A., Sutskever, I., Hinton, G.: Imagenet classification with deep con-
volutional neural networks. In: Advances in Neural Information Processing Sys-
tems, pp. 1097–1105 (2012)
18. Bas, P., Filler, T., Pevný, T.: “Break our steganographic system”: the ins and
outs of organizing BOSS. In: Filler, T., Pevný, T., Craver, S., Ker, A. (eds.)
IH 2011. LNCS, vol. 6958, pp. 59–70. Springer, Heidelberg (2011). doi:10.1007/
978-3-642-24178-9 5
19. Lecun, Y.Y., BoUou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied
to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)
20. Srivastava, N., et al.: Dropout: a simple way to prevent neural networks from
overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014)
21. Nair, V., Hinton, G.E.: Rectified linear units improve restricted boltzmann
machines. In: Proceedings of the 27th International Conference on Machine Learn-
ing (ICML 2010) (2010)
22. Jiang, Y.Y., Zeng, H., Kang, X., et al.: The game of countering JPEG anti-forensics
based on the noise level estimation. In: Proceedings of the Asian-Pacific Signal and
Information Processing Association Annual Submit Conference (APSIPA ASC)
(2013)
A Multi-purpose Image Counter-anti-forensic Method 15
23. Zeng, H.H., Qin, T., Kang, X., et al.: Countering anti-forensics of median filtering.
In: Proceedings of the IEEE International Conference on Acoustics, Speech, and
Signal Processing, pp. 2723–2727 (2014)
24. Peng, A.A., Zeng, H., Kang, X., et al.: Countering anti-forensics of image resam-
pling. In: Proceedings of the 2015 IEEE International Conference on Image Process-
ing (ICIP), pp. 3595–3599, September 2015