Professional Documents
Culture Documents
Relatively few studies have realized the detection of re- performs favorably against state-of-the-art trackers in terms of
gion duplication effectively. Wang et al. propose a divide-and- efficiency, accuracy, and robustness. In this section, we briefly
conquer approach to detect duplication wherein the entire video present the principle and basic steps of the FCT.
is split into subparts, and different types of correlation coeffi- First, assuming that the object location at the (τ − 1)-frame
cients are computed to highlight their similarities [13]. How- (i.e., Iτ −1 ) has been determined and record the current frame
ever, the detection effectiveness is unacceptable if the forged as Iτ . After sampling a set of image patches from the tracking
region is small. In [14], a region duplication forgery detection location at the (τ − 1)-th frame [18], a very sparse random
∗
algorithm is proposed based on Histogram of Oriented Gradi- Gaussian matrix R ∈ Rm s (m << s) is adopted to extract a
∗
ents (HOG) feature matching and video compression properties. low-dimensional measurement vector v ∈ Rm 1 from the high-
∗
This HOG-based algorithm is effective and robust against vari- dimensional Haar-like feature vector x ∈ Rs 1 of these image
ous signal processing manipulations; however, it is not suitable patches, and the extracting process can be defined as:
for long videos because of its unacceptable computational com-
v = Rx (1)
plexity. In [15], a novel copy–move forgery detection scheme
is proposed using adaptive over-segmentation and feature point where each element R(i, j) in R can be defined as:
matching. The algorithm combines a block-based method with ⎧
⎪
⎪ 1 1
with a probability of 2ρ
a feature points-based method and has good performance. How-
√ ⎨
ever, to reduce computational complexity, the test video in [15] R (i, j) = rij = ρ × 0 with a probability of 1 − 1
⎪
⎪
ρ
was processed by the down-sampling, which complicated the ⎩ −1 with a probability of 1
extraction of features and influenced the detection accuracy. 2ρ
(2)
In video forensics, the issues of detection accuracy and com-
where ρ = s/(glog10 (s)) with s = 106 ∼ 1010 and g = 0.4.
putational cost are central to the design of the algorithms, be-
Second, use classifier H(v) to each feature vector v and find
cause a video of even modest length may run into thousands of
the tracking location at the τ -th frame with the maximal classi-
frames. And most of the previous algorithms could not achieve
fier response [19]. H(v) can be calculated with a naive Bayes
both satisfactory detection accuracy and computational effi-
classifier, defined as
ciency, not even mentioning for detecting forged regions with s
mirroring, as reported in [13]–[15]. p(vi |y = 1)p(y = 1)
H (v) = log s i=1
i=1 p(vi |y = 0)p(y = 0)
To address this problem, in this paper, we propose a fast
forgery detection algorithm based on Exponential-Fourier mo-
s
ments for region duplication in videos. In our algorithm, the p(vi |y = 1)
= log (3)
current frame is first split into overlapping blocks, from which i=1
p(vi |y = 0)
the proposed algorithm extracts EFMs features, and performs a
fast match. Then, a post-verification scheme (PVS) is designed where p(y = 1) = p(y = 0) = 0.5, and y ∈ {0, 1} is a binary
to eliminate falsely matched pairs, and locate the altered regions variable that represents the sample label.
in the current frame. Finally, AFCT algorithm is used to track The conditional distributions p(vi |y = 1) and p(vi |y = 0) in
the forged regions in the subsequent frames. the classifier H(v) are assumed to be Gaussian distributed [20]
The main contributions of this paper address the following with four parameters (μ1i , σi1 , μ0i , σi0 ) where:
elements: p(vi |y = 1) ∼ N (μ1i , σi1 ), p(vi |y = 0) ∼ N (μ0i , σi0 ) (4)
1. EFMs are first employed to detect region duplication in
video, realizing the detection for forged regions with Third, sample two sets of image patches from Iτ , extract the
mirroring. Moreover, we optimize the process of EFMs features with these two sets of samples, and update the classifier
extraction and block matching, which improve the com- parameters according to:
putational efficiency greatly. μ1i ← λμ1i + (1 − λ)μ1 (5)
2. Object tracking technique is first introduced into digi-
that are suitable for image analysis and rotation invariant pattern performed:
recognition [21], [22]. Exponent-Fourier moments have a strong
ability to describe an image because the new radial functions ri,j = (c1 j − c2 )2 + (c2 − c1 i)2 (17)
have more zeros, and these zeros are evenly distributed. In this
section, we briefly introduce the Exponential-Fourier Moments c2 − c1 i
θi,j = arctan (18)
as follows: c1 j − c2
In the polar coordinate system, an orthogonal function set
where c1 and c2 can be calculated by (19) and (20):
Pn m (r, θ) is defined:
√
P n m (r, θ) = Qn (r) exp (tmθ) (9) 2
c1 = (19)
N
where exp(tmθ) is the Fourier function, n and m are both N +1
integers, and Qn (r) is the radial function which is defined as: c2 = √ (20)
2N
1 Finally, (13) becomes:
Qn (r) = exp (t2nπr) (10)
r
1
N N
The function set Pn m (r, θ) is orthogonal within 0 ≤ r ≤ 1 EF M sn m = f (i, j)Qn (ri,j ) exp(−tmθi,j )
and 0 ≤ θ ≤ 2π; that is, 2πN 2 i=1 j =1
2π 1 (21)
P n m (r, θ)P k l (r, θ)rdrdθ = 2πδn m k l (11) where f (i, j) represents the pixel value of the image at (i, j).
0 0 The image moment invariants, ||EF M sn m ||, keep good
where δn m k l is the Kronecker symbol. According to the theory invariance under translation, scaling, rotation [22], [23] and
of orthogonal functions, an image f (r, θ) in the polar coordinate mirror with detailed proofs in [23], Low-order moments
systems can be decomposed as: (i.e., ||EF M s(0, 0)||, ||EF M s(1, 0)||, ||EF M s(0, 1)||, and
||EF M s(1, 1)||) retain the low frequency information of the
+∞
+∞
image. Furthermore, the zero order moment ||EF M s(0, 0)||
f (r, θ) = EF M sn m Qn (r) exp(tmθ) (12)
is generally used to represent the “quality” of an im-
n =−∞ m =−∞
age. In this paper, we use ||EF M s(0, 0)||, ||EF M s(1, 0)||,
Therefore, Exponential-Fourier moments are defined using ||EF M s(0, 1)||, and||EF M s(1, 1)|| as the feature vector of
the radial function Qn (r) in the polar system as: each image sub-block, which is defined as:
2π 1
EF M sn m =
1
f (r, θ)Qn (r) exp(−tmθ)rdrdθ V = { EF M sn m |0 ≤ n, m ≤ 1} (22)
2π 0 0
2π 1 III. PROPOSED METHOD
1 1
= f (r, θ) exp(t(2nπr − mθ))rdrdθ (13)
2π 0 0 r Our algorithm primarily consists of the following four stages:
1) feature extraction, 2) block matching, 3) locating the tam-
In the proposed video region duplication detection algorithm
pered areas in the current frame, and 4) tracking the forged
in this paper, Exponential-Fourier moments are calculated in
areas in the subsequent frames.
the Cartesian coordinate systems. The formulas for converting
The first stage of the algorithm extracts EFMs from each block
any point on a two dimensional coordinate plane between polar
using an improved feature extraction method. The second stage
coordinates (r, θ) and Cartesian coordinates (x, y) are given in
applies a new block-matching method to search for potential
(14) and (15):
matching blocks, which significantly reduces the computational
x = r cos θ, y = r sin θ (14) cost. In the third stage, PVS method is designed to eliminate
y falsely matched pairs and locate the tampered areas in the current
rx,y = x2 + y 2 , θx,y = arctan (15) frame. In the final stage, AFCT algorithm is adapted to learn
x
the tampered areas identified in the current frame and track the
The Exponential-Fourier moments expressed in the Cartesian forged regions in subsequent frames.
coordinate systems can be obtained from the above equations:
EF M sn m A. Feature Extraction
In intra-frame region duplication forgery, transformations
1
= f (x, y)Qn (r) exp(−tmθ)dxdy (16) such as scaling, rotation and mirroring (post-processing) are
2π x 2 +y 2 ≤1
often conducted in duplicated regions. Scale Invariant Feature
A digital image of size N × N is an array of pixels f (i, j), Transform (SIFT) [24], [25] and Histograms of Oriented Gra-
where i and j denote rows and columns, respectively. Since dients (HOG) [14], [26], [27] are often used to detect dupli-
EFMs are computed over a unit disk, the center of the image as cation forgeries. However, those features have the following
the origin, and then the following coordinate transformation is shortcomings:
828 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 20, NO. 4, APRIL 2018
TABLE I based on a 640 × 480 pixel image. The results indicate that the
TIME COMPARISON OF TWO METHODS FOR EXTRACTING EFMS
proposed improved algorithm significantly increases the com-
putational efficiency. The information about the computer and
Block size software has been given in Section IV.
Algorithm 8×8 24 × 24 Additionally, the improved algorithm can be expanded to
the extraction of other image moments, such as the Zernike
Conventional method 31.12 s 217.55 s
Improved algorithm 2.51 s 5.13 s
moment, principal component transformation (PCT), and phase
stretch transform (PST). A detailed comparison of EFMs with
these image moments is provided in Section IV-B.
⎡ ⎤ ⎡ ⎤
f (1, 1) , . . . , f (1, N ) Qn r1,1 exp −tmθ1,1 , . . . , Qn r1,N exp −tmθ1,N
⎢. . . ⎥ ⎢. . . ⎥
ME = ⎢
⎣. . .
⎥. ∗ ⎢
⎦ ⎣. . .
⎥
⎦ (24)
f (N, 1) , . . . , f (N, N ) Qn rN ,1 exp −tmθN ,1 , . . . , Qn rN ,N exp −tmθN ,N
SU et al.: FAST FORGERY DETECTION ALGORITHM BASED ON EFMs FOR VIDEO REGION DUPLICATION 829
TABLE II
PSEUDO CODE OF THE IMPROVED BLOCKS-MATCHING ALGORITHM
than 2000 in Im atch and the white patches represent the tam- where α = 1.4, β = 2.3, ui and σi represent the mean and vari-
pered areas. The pseudo code for this step is given in Table IV. ance of the existing targets, while u and σ represent that of the
An example of using PVS to eliminate falsely matched pairs current targets, respectively. The current targets’ extent of de-
is shown in Fig. 4. We observe from Fig. 4(c) that there exist viation from existing targets is judged by differences in mean
SU et al.: FAST FORGERY DETECTION ALGORITHM BASED ON EFMs FOR VIDEO REGION DUPLICATION 831
TABLE V
COMPARISON OF THE SUCCESS RATES
Bolt 74 92
Yarn 65 96
Biker 80 88
David indoor 92 93
Occluded face 97 97
Shaking 76 87
Fig. 6. Experimental results of the FCT algorithm. Fig. 7. Experimental results of the AFCT algorithm.
TABLE VI
COMMON IMAGE MOMENTS AND THEIR RADIAL BASIS FUNCTIONS
⎪ √1
⎨ r ,n = 0
RHFMs [34] Q n (r) = √2 cos(πnr), n is an even number
⎪ r
⎩ √2 sin(π(n + 1)r), n is an odd number
r
PCT [35] Q n (r) = cos(πnr 2 )
PST [35] Q n (r) = sin(πnr 2 )
TABLE VII
COMPARISON OF THE AVERAGE EXTRACTION TIME OF
COMMON IMAGE MOMENTS
Zernike 12 6.32 s
RHFMs 7 5.43 s
PCT 8 5.97 s
PST 9 7.21 s
EFMs 4 2.51 s
Fig. 8. Flow chart of the proposed approach.
versions (60 images in total) from CASIA 2.0.1 We built and TABLE VIII
COMPARISON OF THE AVERAGE EXTRACTION TIME OF EFMS,
analyzed the distribution of the Euclidean distances between all SIFT, AND FISIFT
matching pairs for each image.
In Fig. 9, we illustrated the distribution of Euclidean distances
Algorithm 320∗ 240 640∗ 480 1024∗ 768
between all matching pairs for both the original and forged im-
ages with “Euclidean distance” on the y-axis and “serial number EFMs 0.71 2.51 8.43
of matching pairs” on the x-axis. In this figure, (a) and (c) are the SIFT 1.04 2.76 6.47
FISIFT 1.76 5.87 17.66
distributions of Euclidean distances from two different original
images, while (b) and (d) are the distributions obtained from
their corresponding forged versions. The forged region in (b)
was not subjected to post-processing, while the one in (d) was in Table VI. The EFMs are compared with other image
mirrored. (Note: Euclidean distance values above 0.05 are not moments in terms of their feature dimensions and extraction
shown in Fig. 9) time in Table VII. The comparison experiments were performed
Fig. 8(a) shows that nearly every matching pair’s Euclidean on 50 images at a resolution of 640 × 480 pixels, which were
distance exceeds 0.008, while in Fig. 8(b), some matching pairs divided into blocks of 8 × 8 pixels each. Clearly, the EFMs
exist whose Euclidean distances are distributed in [0, 0.008]. A are superior to the other moments in terms of both the feature
similar phenomenon is evident in Fig. 8(c) and (d). Accordingly, dimension and extraction time.
we inferred that the matching pairs whose Euclidean distances Similarly, we also compared EFMs with SIFT and FISIFT
are below 0.008 are potential matching pairs in the forged im- in terms of the average extraction time. The comparison exper-
ages. Finally, considering the computational complexity and iments were performed on 50 images at resolutions of 320 ×
precision, we fixed the parameter as Tdis = 0.006, which made 240 pixels, 640 × 480 pixels and 1024 × 768 pixels (50 images
our algorithm convenient to use. were used for each resolution). Table VIII shows that FISIFT
requires the highest computation complexity at the various res-
B. Performance Comparison of Different Features olutions, while the EFMs exhibited the best performance at low
resolution. As the image resolution increased, the performance
1) Comparison of the Extraction Time: As described in
of SIFT became superior to that of the EFMs.
Section III-A, we expanded the improved feature extraction al-
2) Comparison of the Accuracy: We also compared the ac-
gorithm to the extraction of other image moments, as shown
curacy of the EFMs with those of other image moment fea-
tures (shown in Table VI) by employing the methods proposed
1Credit for the use of the CASIA Image Tempering Detection Evaluation in Sections III-B and III-C. The experiments were conducted
Database (CAISA TIDE) V2.0 is given to the National Laboratory of Pattern
Recognition, Institute of Automation, Chinese Academy of Science, Corel Im- by selecting 100 forged images from CASIA V2.0. Instead of
age Database and the photographers. http://forensics.idealtest.org the information of original areas, the forged images selected in
834 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 20, NO. 4, APRIL 2018
Fig. 9. The distributions of the Euclidean distances between all matching pairs for both the original and forged images: (a) and (b) are the distributions of
Euclidean distances from two different original images, and (c) and (d) are the distributions from their corresponding forged versions.
Fig. 12. Test video snapshots: (a) the original untampered video clip, (b) the
tampered video clip with intra-frame region duplication, (c) the binary images
of the tampered areas, (d) and (e) the video clips showing the locations of the
Fig. 10. Test video snapshots: (a) the original untampered video clip, (b) the tampered areas determined by the AFCT algorithm, and (f) the tampered areas
tampered video clip with intra-frame region duplication, (c) the binary images in subsequent frames.
of the tampered areas, (d) and (e) the video clips showing the locations of the
tampered areas determined by the AFCT algorithm, and (f) the tampered areas
in subsequent frames. forgery under geometrical transforms accurately, demonstrating
the EFMs’ good invariance under translation, scaling, rotation,
and mirroring.
TABLE X
DETAILS OF SOME OF THE TEST VIDEOS USED IN
SECTION IV-G
Fig. 13. Test video snapshots (a region in the video was copied and pasted
to multiple locations): (a) the original video clip; (b) the tampered video clip Video Length Resolution
with intra-frame region duplication; (c) the binary images of the tampered areas;
(d)–(f) the video clips showing the locations of the tampered areas determined Video 1 100 320∗ 240
by the AFCT algorithm; and (g) the tampered areas in subsequent frames. Video 2 100 640∗ 480
Video 3 150 1280∗ 720
Video 7 100 1024∗ 768
Video 8 150 1024∗ 768
Video 9 150 1024∗ 768
Fig. 17. Test video snapshots: (a) the original untampered video clip, (b) the
tampered video clip with intra-frame region duplication, (c) the binary images
of the tampered areas, (d) and (e) the video clips showing the locations of the
Fig. 16. Test video snapshots (a moving target in different frames is duplicated tampered areas determined by the AFCT algorithm, and (f) the tampered areas
to different locations): (a) the original video clip; (b)–(d) the tampered video in subsequent frames.
clips with intra-frame region duplication; (e) the binary images; (f)–(i) the video
clips showing the locations of the tampered areas determined by the AFCT TABLE XII
algorithm; and (j)–(l) the tampered areas in subsequent frames. PERFORMANCE EVALUATION OF THE
PROPOSED ALGORITHM
TABLE XI
COMPARISON OF THE EXECUTION TIMES OF DIFFERENT ALGORITHMS
Positive Negative
Video Proposed Wang A.V Pun SIFT FISIFT True 96.9% 89.3%
False 3.1% 10.7%
Video 1 42 68 206 711 52 80
Video 2 96 222 785 >1000 112 157
Video 3 431 671 >1000 >1000 444 582
Video 7 192 464 >1000 >1000 249 308
Video 8 243 550 >1000 >1000 312 401 F N indicates that a forged frame is detected as authentic.
Video 9 224 538 >1000 >1000 280 396 Theoretically, if the algorithm used to detect frame duplica-
tion achieves higher precision and recall, its detection rate is
considered better. (T P + T N ) represents the total number of
detections, and (T P + T N + F P + F N ) represents the total
and FISIFT, the time required for block matching is higher than
number of frames in the experiments. Therefore, DA is the per-
that needed by EFMs, resulting in increases in the total time.
centage of correct detection. It is assumed that a higher DA
corresponds to a better detection rate using the proposed algo-
G. Comparison of the Detection Accuracy
rithm. Fig. 17 and Table XII show the performance evaluations
To evaluate the capability of the proposed algorithm, we con- of the proposed algorithm.
sider three performance indices, i.e., precision rate (P R), recall We compared the proposed algorithm results with those of
rate (RR), and detection accuracy (DA), which are defined as other methods in terms of the accuracy and mirror invariance.
follows: The comparison experiments were performed on the experi-
TP TP mental dataset of 50 videos downloaded from SULFA (23,000
PR = RR = (32) frames in total). Region duplication was simulated by selecting
TP + FP TP + FN
an area in a frame and duplicating it to another non-overlapping
TP + TN
DA = (33) position in the same frame and several subsequent frames.
TP + TN + FP + FN The experimental dataset was divided into two parts: the 35
where T P indicates that an authentic frame is detected as au- videos (13,000 frames in total) in which region duplication
thentic, T N indicates that a forged frame is detected as a forgery, was simulated without post-processing and the remaining 15
F P indicates that an authentic frame is detected as a forgery, and videos (10,000 frames in total) in which duplicated regions were
838 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 20, NO. 4, APRIL 2018
TABLE XIII
COMPARISON OF THE DETECTION ACCURACY AMONG ALGORITHMS
Pun 91.4% NO
A.V. 88.3% NO
Wang 69.7% NO
SIFT 87.9% NO
FISIFT 92.2% YES
Proposed 93.1% YES
TABLE XIV
COMPARISON OF THE DETECTION ACCURACY AMONG ALGORITHMS FOR
DETECTING DUPLICATED REGIONS SUBJECTED TO GEOMETRICAL
TRANSFORMATION
TABLE XV [10] W. Chen and Y. Shi, “Detection of double MPEG compression based on
PROOF OF THE JUDGING CRITERION DESCRIBED IN SECTION III-B first digit statistics,” in International Workshop on Digital Watermarking
(Lecture Notes in Computer Science), vol. 5450. New York, NY, USA:
Springer, 2009, pp. 16–30.
[11] S.-Y. Liao and T.-Q. Huang, “Video copy-move forgery detection and
localization based on Tamura texture features,” in Proc. 2013 6th Int.
Congr. Image Signal Process., 2013, pp. 864–868.
[12] J. Yang, T. Huang, and L. Su, “Using similarity analysis to detect frame
duplication forgery in videos,” Multimedia Tools Appl., vol. 75, pp. 1793–
1811, 2014.
[13] W. Wang and H. Farid, “Exposing digital forgeries in video by de-
tecting duplication,” in Proc. 9th workshop Multimedia Security, 2007,
pp. 35–42.
[14] A. V. Subramanyam and S. Emmanuel, “Video forgery detection using
HOG features and compression properties,” in Proc. IEEE Int. Workshop
Multimedia Signal Process., 2012, pp. 89–94.
[15] C.-M. Pun, X.-C. Yuan, and X.-L. Bi, “Image forgery detection using
adaptive oversegmentation and feature point matching,” IEEE Trans. Inf.
Forensics Security, vol. 10, no. 8, pp. 1705–1716, Aug. 2015.
[16] Z. Kaihua, Z. Lei, and Y. Ming-Hsuan, “Fast compressive tracking,” IEEE
Trans. Pattern Anal. Mach. Intell., vol. 36, no. 10, pp. 2002–2015, Oct.
2014.
[17] K. Zhang, L. Zhang, and M.-H. Yang, “Real-time compressive tracking,”
in Proc. Eur. Conf. Comput. Vis., 2012, pp. 864–877.
[18] E. J. Candes and T. Tao, “Near optimal signal recovery from random pro-
jections: Universal encoding strategies,” IEEE Trans. Inf. Theory, vol. 52,
no. 12, pp. 5406–5425, Dec. 2006.
[19] A. Jordan, “On discriminative vs. generative classifiers: A comparison
of logistic regression and naive bayes,” Adv. Neural Inf. Process. Syst.,
vol. 14, pp. 841–842, 2002.
[20] P. Diaconis and D. Freedman, “Asymptotics of graphical projection pur-
suit,” Ann. Statist., vol. 12, pp. 793–815, 1984.
[21] Z. Ping, “FFT algorithm of complex exponent moments and its application
in image recognition,” in Proc. SPIE, Int. Soc. Opt. Eng., 2014, vol. 9159,
pp. 4177–4180.
[22] H.-T. Hu, Y.-D. Zhang, C. Shao, and Q. Ju, “Orthogonal moments based
on exponent functions: Exponent-Fourier moments,” Pattern Recognit.,
vol. 47, pp. 2596–2606, 2014.
[23] Y. Jiang, “Exponent moments and its application in pattern recognition,”
In future research, we will extend our method to detect more Ph.D. dissertation, Beijing Univ. Posts Telecommun., Beijing, China,
challenging types of video forgery. 2011.
[24] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,”
Int. J. Comput. Vis., vol. 60, pp. 91–110, 2004.
APPENDIX [25] I. Amerini, L. Ballan, R. Caldelli, A. Del Bimbo, and G. Serra, “A SIFT-
based forensic method for copy–move attack detection and transformation
Mathematical proof of the judging criterion described in recovery,” IEEE Trans. Inf. Forensics Security, vol. 6, no. 3, pp. 1099–
Section III-B is shown in Table XV. 1110, Sep. 2011.
[26] N. Dalal and B. Triggs, “Histograms of oriented gradients for human de-
tection,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit.,
REFERENCES 2005, pp. 886–893.
[27] J. C. Lee, C. P. Chang, and W. K. Chen, “Detection of copy–move im-
[1] H. Yin, W. Hui, H. Li, C. Lin, and W. Zhu, “A novel large-scale digital age forgery using histogram of orientated gradients,” Inf. Sci., vol. 321,
forensics service platform for internet videos,” IEEE Trans. Multimedia, pp. 250–262, 2015.
vol. 14, no. 1, pp. 178–186, Feb. 2012. [28] W. L. Zhao and C. W. Ngo, “Flip-invariant sift for copy and object de-
[2] T. Stütz, F. Autrusseau, and A. Uhl, “Non-blind structure-preserving sub- tection,” IEEE Trans. Image Process., vol. 22, no. 3, pp. 980–991, Mar.
stitution watermarking of H.264/CAVLC inter-frames,” IEEE Trans. Mul- 2013.
timedia, vol. 16, no. 5, pp. 1337–1349, Aug. 2014. [29] Y. Li, “Image copy-move forgery detection based on polar cosine trans-
[3] M. Kobayashi, T. Okabe, and Y. Sato, “Detecting video forgeries based form and approximate nearest neighbor searching,” Forensic Sci. Int.,
on noise characteristics,” in Proc. 3rd Pacific Rim Symp. Adv. Image Video vol. 224, pp. 59–67, 2013.
Technol., Tokyo, Japan, 2009, pp. 306–317. [30] Y. Lai and T. Huang, “Image region copy-move of forgery detection based
[4] S. Milani et al., “An overview on video forensics,” APSIPA Trans. Signal on exponential-Fourier moments,” J. Image Graph., vol. 20, pp. 1212–
Inf. Process., vol. 1, pp. 1229–1233, 2012. 1221, 2015.
[5] X. Feng, I. J. Cox, and G. Doerr, “Normalized energy density-based [31] Fast Compressive Tracking. 2013. [Online]. Available: http://www4.
forensic detection of resampled images,” IEEE Trans. Multimedia, comp.polyu.edu.hk/∼cslzhang/FCT/FCT.htm
vol. 14, no. 3, pp. 536–545, Jun. 2012. [32] G. Qadir, S. Yahaya, and A. T. Ho, “Surrey university library for forensic
[6] S. A. H. Tabatabaei, O. Ur-Rehman, N. Zivic, and C. Ruland, “Secure and analysis (SULFA) of video content,” in Proc. IET Conf. Image Process.,
robust two-phase image authentication,” IEEE Trans. Multimedia, vol. 17, 2012, pp. 1–6.
no. 7, pp. 945–956, Jul. 2015. [33] S. J. Ryu, M. Kirchner, M. J. Lee, and H. K. Lee, “Rotation invariant
[7] M. Kobayashi, T. Okabe, and Y. Sato, “Detecting video forgeries based on localization of duplicated image regions based on Zernike moments,”
noise characteristics,” in Advances in Image and Video Technology. New IEEE Trans. Inf. Forensics Security, vol. 8, no. 8, pp. 1355–1370, Aug.
York, NY, USA: Springer, 2009, pp. 306–317. 2013.
[8] C.-C. Hsu, T.-Y. Hung, C.-W. Lin, and C.-T. Hsu, “Video forgery detec- [34] H. Ren, Z. Ping, W. Bo, W. Wu, and Y. Sheng, “Multidistortion-invariant
tion using correlation of noise residue,” in Proc. IEEE 10th Workshop image recognition with radial harmonic Fourier moments,” J. Opt. Soc.
Multimedia Signal Process., 2008, pp. 170–174. Amer. A, vol. 20, pp. 631–637, 2003.
[9] W. Wang and H. Farid, “Exposing digital forgeries in video by detecting [35] P. T. Yap, X. Jiang, and A. Chichung Kot, “Two-dimensional polar har-
double quantization,” in Proc. 11th ACM Workshop Multimedia Security, monic transforms for invariant image representation,” IEEE Trans. Pattern
2009, pp. 39–48. Anal. Mach. Intell., vol. 32, no. 7, pp. 1259–1270, Jul. 2010.
840 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 20, NO. 4, APRIL 2018
Lichao Su received the B.E. and M.E. degrees in Yuecong Lai was born in Ganzhou, China, in 1991.
computer science and technology from Fujian Nor- He received the Bachelor’s Degree of Science in
mal University, Fuzhou, China, in 2011 and 2014, information and computing sciences from Henan
respectively. He is currently working toward the University of Engineering, Zhengzhou, China, in
Doctoral degree in computer science at Xiamen Uni- 2013, and the Master’s Degree of Science in software
versity, Xiamen, China. His research interests include engineering from Fujian Normal University, Fuzhou,
video processing, multimedia forensics, and informa- China, in 2016. His research interests include image
tion security. processing and multimedia forensics.