iii - universiti teknologi malaysiaeprints.utm.my/id/eprint/81700/1/omarismaelibrahimpfc... ·...
TRANSCRIPT
-
iii
To the Source of Kindness Who Provided Me with Ambition,
My Parents “Assoc. Prof Dr. Ismael and Mrs. Asma”.
To My Beloved Wife “Amina”, for her Love and Support this Research
Would Have Never Been Completed.
To My Brother “Ibrahim” and My Sisters “Noor” and “Marwa” for
being patience, supportive, and understanding.
To all my family, my relatives and friends receives my deepest gratitude
and love for their patience and support during the years of my study.
-
iv
ACKNOWLEDGEMENT
Thanks to Allah SWT for everything I was able to achieve and for everything
I tried but I was not able to achieve.
First of all, I would like to take this opportunity to gratefully acknowledge the
wholehearted supervision of Professor Dr. Ghazali bin Sulong during this work. His
dedication, skillful guidance, helpful suggestions and constant encouragement made it
possible for me to deliver a dissertation of appreciable quality and standard.
I am forever indebted to my parents for their patience and understanding,
alleviating my family responsibilities and encouraging me to concentrate on my study.
Finally, and most importantly, I would like to express special thanks to my
brothers, sisters and friends for their support when it was most required. Without their
help and encouragement, this study would not have been completed.
In addition, I want to thank my beloved friends Dr. Mohammed Al-Ali, Dr.
Ahmed Abdullah Ahmed, Dr. Akeel Shebeeb, Dr. Mohamed Shaban, Dr. Saeed Al
Hameed, and Omar Alzaide for their constant support through good and bad times.
Thank you for being the friend that I always wanted, I always needed and I always
deserved Finally, I would like to offer my special thanks to the faculty of computing
staff, your help will never be forgotten.
-
v
ABSTRACT
Sophisticated digital video editing tools has made it easier to tamper real videos
and create perceptually indistinguishable fake ones. Even worse, some post-processing
effects, which include object insertion and deletion in order to mimic or hide a specific
event in the video frames, are also prevalent. Many attempts have been made to detect
such as video copy-move forgery to date; however, the accuracy rates are still
inadequate and rooms for improvement are wide-open and its effectiveness is confined
to the detection of frame tampering and not localization of the tampered regions. Thus,
a new detection scheme was developed to detect forgery and improve accuracy. The
scheme involves seven main steps. First, it converts the red, green and blue (RGB)
video into greyscale frames and treats them as images. Second, it partitions each frame
into non-overlapping blocks of sized 8x8 pixels each. Third, for each two successive
frames (S2F), it tracks every block’s duplicate using the proposed two-tier detection
technique involving Diamond search and Slantlet transform to locate the duplicated
blocks. Fourth, for each pair of the duplicated blocks of the S2F, it calculates a
displacement using optical flow concept. Fifth, based on the displacement values and
empirically calculated threshold, the scheme detects existence of any deleted objects
found in the frames. Once completed, it then extracts the moving object using the same
threshold-based approach. Sixth, a frame-by-frame displacement tracking is
performed to trace the object movement and find a displacement path of the moving
object. The process is repeated for another group of frames to find the next
displacement path of the second moving object until all the frames are exhausted.
Finally, the displacement paths are compared between each other using Dynamic Time
Warping (DTW) matching algorithm to detect the cloning object. If any pair of the
displacement paths are perfectly matched then a clone is found. To validate the
process, a series of experiments based on datasets from Surrey University Library for
Forensic Analysis (SULFA) and Video Tampering Dataset (VTD) were performed to
gauge the performance of the proposed scheme. The experimental results of the
detection scheme were very encouraging with an accuracy rate of 96.86%, which
markedly outperformed the state-of-the-art methods by as much as 3.14%.
-
vi
ABSTRAK
Kecanggihan alat penyuntingan video digital telah membuat lebih mudah
untuk mengubah video gangguan sebenar dan memalsukannya supaya ia tidak dapat
dibezakan. Lebih buruk lagi, terdapat beberapa kesan pasca-pemprosesan, yang
berleluasa termasuk sisipan dan penghapusan objek untuk meniru atau
menyembunyikan peristiwa tertentu dalam bingkai video. Pelbagai cubaan telah dibuat
untuk mengesan pemalsuan video salinan-langkah sehingga kini; walau
bagaimanapun, kadar ketepatan masih tidak mencukupi dan ruang untuk
penambahbaikan adalah terbuka luas dan keberkesanannya terbatas kepada
pengesanan bingkai dan kawasan setempat yang tidak diganggu. Oleh itu, satu skim
pengesanan baru telah dibangunkan untuk mengesan pemalsuan dan meningkatkan
ketepatan. Skim ini melibatkan tujuh langkah utama. Pertama, ia menukarkan video
merah, hijau dan biru (RGB) ke dalam bingkai skala kelabu dan menganggap mereka
sebagai imej. Kedua, ia menyekat setiap bingkai ke dalam blok bukan pertindihan
piksel setiap satu bersaiz 8x8. Ketiga, bagi setiap dua bingkai berturut-turut (S2F), ia
menjejaki dua salinan setiap blok dengan menggunakan teknik pengesanan dua
peringkat yang dicadangkan dan melibatkan carian Berlian dan Pengubah Slantlet
untuk mencari blok pendua. Keempat, bagi setiap pasangan blok pendua daripada S2F,
ia mengira anjakan menggunakan konsep aliran optik. Kelima, berdasarkan nilai-nilai
anjakan dan ambang pengiraan empirikal, skim ini mengesan kewujudan sebarang
objek terpadam yang dijumpai di dalam bingkai. Setelah selesai, ia kemudiannya
mengekstrak objek yang bergerak menggunakan pendekatan berasaskan ambang-
sama. Keenam, pengesanan anjakan bingkai demi bingkai dilakukan untuk mengesan
pergerakan objek dan mencari laluan anjakan objek yang bergerak. Proses ini diulangi
untuk satu lagi kumpulan bingkai bagi mencari jalan anjakan seterusnya untuk objek
kedua yang bergerak sehingga kesemua bingkai habis. Akhirnya, laluan anjakan
dibandingkan antara satu sama lain dengan menggunakan algoritma sepadan
Lengkungan Masa Dinamik (DTW) yang hampir sama untuk mengesan objek
pengklonan. Jika mana-mana pasangan daripada laluan anjakan merupakan bandingan
yang sempurna maka pengklonan akan ditemui. Bagi mengesahkan proses ini, satu siri
eksperimen berdasarkan set data dari Perpustakaan Universiti Untuk Analisis Forensik
(SULFA) dan Video Mengganggu Set Data (VTD) telah dijalankan untuk mengukur
prestasi skim yang dicadangkan. Keputusan eksperimen skim pengesanan sangat
menggalakkan dengan kadar ketepatan 96.86%, yang ketara mengatasi kaedah terkini
sebanyak 3.14%.
-
vii
TABLE OF CONTENTS
CHAPTER TITLE PAGE
DECLARATION ii
DEDICATION iii
ACKNOWLEDGEMENT iv
ABSTRACT v
ABSTRAK vi
TABLE OF CONTENTS vii
LIST OF TABLES xii
LIST OF FIGURES xiv
LIST OF ABBREVIATIONS xxiv
LIST OF ALGORITHMS xxv
1 INTRODUCTION 1
1.1 Overview 1
1.2 Background of Research 3
1.2.1 Pixel-Based Approach 6
1.2.2 Format-Based Approach 6
1.2.3 Camera-Based Approach 7
1.2.4 Geometric-Based Approach 8
1.3 Problem Statements 9
1.4 Research Goal 11
1.5 Research Objectives 12
1.6 Research Scope 12
1.7 Significance of the Study 13
-
viii
1.8 Thesis Outline 13
2 LITERATURE REVIEW 15
2.1 Introduction 15
2.2 Digital Video Forgery Detection 16
2.2.1 Source Identification 19
2.2.2 Splicing 22
2.2.3 Copy-Move Forgery 23
2.3 Framework Processing in Video Forgery
Detection 26
2.4 Tampering of Video Content in Passive
Approaches 28
2.4.1 Tampering in Spatial Domain 31
2.4.1.1 Block Level Attack 33
2.4.1.2 Pixel Level Attack 33
2.4.1.3 Object Removal Attack 35
2.4.1.4 Object Addition Attack 35
2.4.1.5 Object Modification Attack 36
2.4.2 Tampering in Temporal Domain 37
2.4.2.1 Frames Addition Attack 39
2.4.2.2 Frames Removal Attack 40
2.4.2.3 Frame Swapping Attack 40
2.4.3 Tampering in Spatial-Temporal Domain 42
2.5 Related Work 45
2.6 Video Compression 51
2.7 Summary 52
3 RESEARCH METHODOLOGY 55
3.1 Introduction 55
3.2 Research Framework 56
3.2.1 Phase A: Problem Formulation 58
3.2.2 Phase B: Design 58
3.2.3 Phase C: Development 59
-
ix
3.2.4 Phase D: Implementation 59
3.2.5 Phase E: Validation of the Proposed
Methods 62
3.2.5.1 Qualitative Evaluation 63
3.2.5.2 Quantitative Evaluation
Approach
63
3.3 Benchmarking 66
3.4 Datasets in Video Forgery Detection 67
3.4.1 SULFA Dataset 67
3.4.1.1 Video Forgery Process 68
3.4.1.2 Manipulating Digital Video
Forgery 69
3.4.2 The Video Tampering Dataset (VTD) 72
3.4.2.1 Video Tampering Domain 72
3.4.2.2 Framework of Video Forgery
Process 78
3.4.2.3 Ground Truth 80
3.8 Summary 81
4 DETAILED PROPOSED METHODS OF VIDEO
COPY-MOVE FORGERY DETECTION
82
4.1 Introduction 82
4.2 Convert Videos into Image Frames 85
4.3 Computation of Optical Flow 87
4.3.1 Convert RGB Colour Frames to
Greyscale 91
4.3.2 Partition Frame into Non-Overlapping
Blocks 93
4.3.3 Diamond Search Algorithm 96
4.3.4 Slantlet Transform-based Matching
Function 100
4.3.5 Optical Flow Computation for Each
Two Successive Frames 114
-
x
4.4 Optical Flow in Copy-Move Forgery Detection 118
4.4.1 Detection of Deleted Object in Video
Forgery
119
4.4.2 Cloning Object Detection 127
4.4.2.1 Extraction of Moving Object 132
4.4.2.2 Represent Moving Object 139
4.4.2.3 Compute Similarity Matching
and Forgery Detection 148
4.5 Summary 157
5 RESULTS AND DISCUSSION 158
5.1 Introduction 158
5.2 Performance Evaluation of Deleting Object
Detection 160
5.2.1 Qualitative Evaluation of Deleting
Object Detection 160
5.2.1.1 Experimental Results for
SULFA Dataset 161
5.2.1.2 Experimental Results for
VTD Dataset 167
5.2.2 Quantitative Evaluation of Deleting
Object Detection 178
5.2.2.1 Frame-based Detection
Evaluation 178
5.2.2.1 Block-based Detection
Evaluation 183
5.3 Performance Evaluation of Cloning Object
Detection 188
5.3.1 Qualitative Evaluation of Cloning
Object Detection 189
5.3.2 Quantitative Evaluation of Cloning
Object Detection 198
5.3.2.1 Frame-based Detection 199
-
xi
Evaluation
5.3.2.2 Block-based Cloning Object
Detection Evaluation 201
5.4 Benchmarking 204
5.5 Summary 205
6 CONCLUSIONS AND FURTHER OUTLOOK 206
6.1 Contributions 208
6.2 Further Works 212
REFERENCES 213
-
xii
LIST OF TABLES
TABLE NO. TITLE PAGE
2.1 Multi frequent happenings in spatio-temporal visual
videos, where parameters are chosen to simulate the
real cases 45
2.2 Summary of video forgery detection method and
classifier extracted features with dataset by previous
work 46
3.1 Summary of operational research design for Phase D 60
3.2 Performance Measures 62
3.3 The meaning of statistical measures, TP, TN, FN,
and FP 65
3.4 Current methods for benchmarking 66
4.1 The parameters of the function gi (n) with different
pixels numbers 104
4.2 The parameters of both filters fi(n)and hi(n)with
different number of pixels 106
4.3 Coefficients of fi(n) and hi(n) functionsusing the
Algorithm 4.7 with N=8 107
5.1 Frame-based deleting object detection using TP, TN,
FP, and FN. 179
5.2 Result of Frame-based deleting object detection 181
5.3 Jaccard and Dice scores for the deleting object
localisation 185
-
xiii
5.4 Evaluation result of cloning object detection using
SULFA Dataset
199
5.5 Jaccard and Dice scores for the cloning object
localisation 201
5.6 The proposed method versus the state-of-the-art
methods in terms of accuracy rate 204
-
xiv
LIST OF FIGURES
FIGURE NO. TITLE PAGE
1.1 Text sub-images (blocks) Example of video Copy-
Move forgery (a) Original video(b) Tampered video
from SULFA dataset 3
1.2 Sample of video copy-move forgery (a) Original
video (b) Tampering video from VTD dataset 4
1.3 Example of video copy-move forgery 10
2.1 Classification of video forgery detection approaches 18
2.2 Forged videos recorded realistic scenes (a)
Wrapped box on a bookshelf (b) Bicycle in front of
a building 21
2.3 Video copy-move tampering (a) Original video (b)
Delete/add object (c) Delete/add sequence of frames
(d) Tampering both copy-move 23
2.4 Example of a forged video copy-move 24
2.5 Example of (a) Original and (b) Forged of video 25
2.6 General forgery detection 27
2.7 (a) Original video; (b) Spatially tampered video;
(c) Temporally tampered video; and (d) Spatio-
temporal tampered video 29
2.8 Number of papers related to digital video passive
approaches 30
2.9 Slice and macro-blocks Structure 32
2.10 An example of (a) Frame Drop, (b) Frame
-
xv
Swapping, and (c) Frame Copying where source is
the video sequence presented
37
2.11 An example of temporal domain: (a) Original
frames; (b) Frame Copying; (c) Frame Drop; and
(d) Frame Swapping 41
2.12 Spatio temporal connotation in passive tampering
detection 44
2.13 Arrows show prediction dependencies between
frames 52
3.1 Block diagram of the research design 57
3.2 The Evaluation of localisation of the video forged
block detection using Jaccard (J) coefficient and
Dice (D) coefficient 66
3.3 A block diagram of the general forgery process 69
3.4 Manipulating video; (a) actual Video, (b) Tampered
video and (c) Subtracted ground-truth frame of (a) 70
3.5 Structures of SULFA Datasets 71
3.6 (a) Original video; (b) Spatially tampered video;
(c)Temporally tampered video; and (d) Spatio-
temporal tampered video. 73
3.7 The process of copy-move forgery 75
3.8 Splicing tampering in video frames 76
3.9 An example of temporal domain: (a) Original
frames; (b) Frame Copying; (c) Frame Drop; and
(d) Frame Swapping 77
3.10 Swapping frames (drop and add) 78
3.11 Framework of the video tampering process 79
3.12 Deducted ground-truth frame of figures (a) and (b) 80
4.1 The proposed methodology framework 84
4.2 Example of conversion of video into frames of the
video named: “Forgery camera_demo” of VTD
dataset 86
4.3 Calculation of Oflow between Frame#68 and
-
xvi
Frame#69 of “Forgery Dahua_HDCVI” video of
VTD dataset
89
4.4 Displacement velocity of a vector movement 90
4.5 Video frame conversion: Group of RGB video
frames and its converted greyscale of “Forgery
white car” video of VTD dataset 93
4.6 Video frame partitioning into non-overlapping
blocks of 8×8 pixels of video 6_(Elephant) of
SULFA Dataset 95
4.7 Full Search algorithms within thirteen crosses show
all possible checking-point positions within the
circle 97
4.8 Two search patterns are employed in the Diamond
Search algorithm; (a) Large diamond search (LDS)
(b) Small diamond search (SDS) 98
4.9 Search window path examples of three cases of
checking-point. (a) The corner point LDS LDS,
(b) The edge point LDS LDS, and (c) The corner
point LDS SDS 99
4.10 Diamond Search for block movement estimation 100
4.11 The SLTBlock matrix operation 110
4.12 An example of a SLTBlock creation 112
4.13 Calculate movement vector of displacement in each
S2F 115
4.14 The 5 × 5 Gaussian smoothing filter 115
4.15 The optical flow in three temporally-consecutive
frames using the proposed method from video
named: “Real man _street” of VTD dataset 117
4.16 X and Y elements values of the displacement
movement of blocks representing the Optical flow
of human movement in a videonamed “Real man
_street” of VTD dataset 118
4.17 Example of (a) Original (b) Forged duplicate region
-
xvii
of video frames to hide a specific event, extracted
from video 11_(person) of SULFA Dataset
120
4.18 Example of (a) Original (b) forged optical flow
vector plot of video frames to hide a specific event,
extracted from video 11_(person) of SULFA
Dataset 121
4.19 Minimum movements in video frames of SULFA
dataset 123
4.20 A flowchart for detection of deleted of object using
minimum movement in videos frames 125
4.21 Result of the Optical Flow matrix of the frame 138
extracted from 11_ forgery (person) video of
SULFA Dataset 126
4.22 The identical object movement in video named
“02_forgery (red ball)” of SULFA Dataset 129
4.23 Example of (a) Original and (b) Forged optical flow
vector plot of video frames to copy an identical
object movement in video named “2_forgery (red
ball)” of SULFA Dataset 131
4.24 Maximum movements in videos frames of SULFA
dataset 134
4.25 The optical flow matrix of the frame 66 extracted
from “2_forgery (red ball)” video of SULFA dataset 136
4.26 Example of (a) Original and (b) Forged optical flow
vector plot of moving object detection from video
named “2_forgery (red ball)” of SULFA Dataset 138
4.27 8-connectivity to link and label all the movement
values 141
4.28 Example of(a) Forged frames 102 and 103, (b)
Vector plot of optical flow and (c) movement
extraction from video named “4_forgery (red ball)”
of SULFA Dataset 142
4.29 The center point Cp of moving objects from frame
-
xviii
66 to 72 video named “2_forgery (red ball)” of
SULFA dataset
145
4.30 Example of displacement path of each object
movement in video (a) Blue Object movement
labeled 1from frame 62 to 94, (b) Black Object
movement labeled 2 from frame 123 to 152 and (c)
Red Object movement labeled 3 from frame 223 to
255 of connected moving objects from video named
“2_forgery (red ball)” SULFA Dataset 147
4.31 The difference between Euclidean and DTW
matching similarity measure: (a) Euclidean and (b)
DTW frame 62 to 94 and frame 123 to 152 of
connected object movement from video named
“2_forgery (red ball)” of SULFA Dataset 149
4.32 Initialization of the alignment matrix 151
4.33 (a) to (d) Filling of the alignment matrix 152
4.34 Trackback of the optimal complete alignment: (a)
Alignment matrix with the initial point and (b)
Alignment matrix with the optimal path 153
4.35 Overall process of computing the similarity of two
displacement paths using DTW matching 155
5.1 The general framework to achieving the proposed
objectives 159
5.2 Detection result of deleting object obtained from
frame 153 of “11_forgery (person)” video (a)
Original frame (b) Tampered frame (c) Ground-
truth (d) Optical flow (e) The detection result with
forged area (red) 161
5.3 Detection result of deleting object obtained from
frame 292 of “11_forgery (person)” video (a)
Original frame (b) Tampered frame (c) Ground-
truth (d) Optical flow (e) The detection result with
forged area (red)
162
-
xix
5.4 Detection result of deleting object obtained from
frame 103 of “13_forgery (person walk)” video (a)
Original frame (b) Tampered frame (c) Ground-
truth (d) Optical flow (e) The detection result with
forged area (red) 163
5.5 Detection result of deleting object obtained from
frame 131 of “13_forgery (person walk)” video (a)
Original frame (b) Tampered frame (c) Ground-
truth (d) Optical flow (e) The detection result with
forged area (red) 164
5.6 Detection result of deleting object obtained from
frame 38 of “14_forgery (person walk)” video (a)
Original frame (b) Tampered frame (c) Ground-
truth (d) Optical flow (e) The detection result with
forged area (red) 165
5.7 Detection result of deleting object obtained from
frame 175 of “14_forgery (person walk)” video (a)
Original frame (b) Tampered frame (c) Ground-
truth (d) Optical flow (e) The detection result with
forged area (red) 166
5.8 Detection result of deleting object obtained from
frame 65 of “Camera_Demo” video (a) Original
frame (b) Tampered frame (c) Ground-truth (d)
Optical flow (e) The detection result with forged
area (red) 167
5.9 Detection result of deleting object obtained from
frame 130 of “Camera_Demo” video (a) Original
frame (b) Tampered frame (c) Ground-truth (d)
Optical flow (e) The detection result with forged
area (red)
168
5.10 Detection result of deleting object obtained from
frame 400 of “CCTV_London_Str” video (a)
Original frame (b) Tampered frame (c) Ground-
-
xx
truth (d) Optical flow (e) The detection result with
forged area (red)
169
5.11 Detection result of deleting object obtained from
frame 416 of “CCTV_London_Str” video (a)
Original frame (b) Tampered frame (c) Ground-
truth (d) Optical flow (e) The detection result with
forged area (red) 170
5.12 Detection result of deleting object obtained from
frame 400 of “Clarity_Sample” video (a) Original
frame (b) Tampered frame (c) Ground-truth (d)
Optical flow (e) The detection result with forged
area (red) 171
5.13 Detection result of deleting object obtained from
frame 430 of “Clarity_Sample” video (a) Original
frame (b) Tampered frame (c) Ground-truth (d)
Optical flow (e) The detection result with forged
area (red) 172
5.14 Detection result of deleting object obtained from
frame 185 of “Dahua_HDCVI” video (a) Original
frame (b) Tampered frame (c) Ground-truth (d)
Optical flow (e) The detection result with forged
area (red) 173
5.15 Detection result of deleting object obtained from
frame 1 of “man _street” video (a) Original frame
(b) Tampered frame (c) Ground-truth (d) Optical
flow (e) The detection result with forged area (red)
174
5.16 Detection result of deleting object obtained from
frame 90 of “man _street” video (a) Original frame
(b) Tampered frame (c) Ground-truth (d) Optical
flow (e) The detection result with forged area (red) 175
5.17 Detection result of deleting object obtained from
frame 285 of “man _street” video (a) Original frame
(b) Tampered frame (c) Ground-truth (d) Optical
-
xxi
flow (e) The detection result with forged area (red)
176
5.18 Detection result of deleting object obtained from
frame 315 of “White Car” video (a) Original frame
(b) Tampered frame (c) Ground-truth (d) Optical
flow (e) The detection result with forged area (red) 177
5.19 Performance measures of the proposed detection
method implemented on a series of videos from
both SULFA and VTD datasets 182
5.20 The deleting object blocks detection in video
images using Jaccard coefficient and Dice
coefficient 183
5.21 The Jaccard and Dice coefficient detection for of
each video in SULFA and VTD datasets 186
5.22 Incomplete tampered region localization detected
by the proposed method on SULFA dataset, video
named 11_forgery (person) (a) Detection result
after using the proposed method (b) Ground-truth
(c) Incomplete detected the full frame247 187
5.23 Incomplete tampered region localization detected
by the proposed method on SULFA
dataset,13_forgery (person walk) (a) Detection
result after using the proposed method (b) Ground-
truth (c) Incomplete detected full frame 116
187
5.24 Detection result of cloning object obtained from
“2_forgery” video (a) Frame 67 (b) Tampered
frame 128 (c) Optical flow 67 (d) Optical flow
tampered 128 (e) Ground-truth (f) The detection
result of processing by red colour 190
5.25 Detection result of cloning object obtained from
“3_forgery (red ball)” (a) Frame 59 (b) Tampered
frame 244 (c) Optical flow 59 (d) Optical flow
tampered 128 (e) Ground-truth (f) The detection
-
xxii
result of processing by red colour 191
5.26 Detection result of cloning object obtained from
“4_forgery (red ball)” (a) Frame108 (b) Tampered
frame 74 (c) Optical flow 108 (d) Optical flow
tampered 74 (e) Ground-truth (f) The detection
result of processing by red colour 192
5.27 Detection result of cloning object obtained from
“4_forgery (red ball)” (a) Frame 137 (b) Tampered
frame 103 (c) Optical flow 137 (d) Optical flow
tampered 103 (e) Ground-truth (f) The detection
result of processing by red colour 193
5.28 Detection result of cloning object obtained from
“6_forgery (red ball)” (a) Frame 24 (b) Tampered
frame 226 (c) Optical flow 24 (d) Optical flow
tampered 226 (e) Ground-truth (f) The detection
result of processing by red colour 194
5.29 Detection result of cloning object obtained from
“7_forgery (red ball)” (a) Frame 377 (b) Tampered
frame 102 (c) Optical flow 377 (d) Optical flow
tampered102 (e) Ground-truth (f) The detection
result of processing by red colour
195
5.30 Detection result of cloning object obtained from
“8_forgery (red ball)” (a) Frame 232 (b) Tampered
frame 145 (c) Optical flow 232 (d) Optical flow
tampered 145 (e) Ground-truth (f) The detection
result of processing by red colour
196
5.31 Detection result of cloning object obtained from
“9_forgery (red ball)” (a) Frame 535 (b) Tampered
frame 340 (c) Optical flow 535 (d) Optical flow
tampered 340 (e) Ground-truth (f) The detection
result of processing by red colour 197
5.32 Performance results obtained by the proposed
method using a series of videos from SULFA
-
xxiii
dataset 200
5.33 The Jaccard and Dice coefficients for each video of
SULFA dataset 202
5.34 Examples of incomplete tampered region
localization detected by the proposed method on
SULFA dataset, 2_forgery (red ball) (a) Detected
cloning object region (b) Ground-truth (c)
Incomplete detected frame 130 203
5.35 Examples of incomplete tampered region
localization detected by the proposed method on
SULFA dataset, 7_forgery (red car) (a) Detected
cloning object region (b) Ground-truth (c)
Incomplete detected frame 98 203
-
xxiv
LIST OF ABBREVIATIONS
GOP - Group of Pictures
SULFA - Surrey University Library for Forensic Analysis
VTD - Video Tampering Dataset
NLF - Noise Level Function
CCN - Circular Correlation Norm
HHT - Hilbert-Huang transform
DCT - Discrete Cosine Transform
DS - Diamond Search
SLT - Slantlet Transform
DTW - Dynamic Time Warping
Th1 - Threshold
Th2 - Threshold
TP - True Positive
TN - True Negative
FN - False Negative
FP - False Positive
S2F - Two Successive Frame
Oflow - optical flow
LDS - Large Diamond Search
SDS - Small Diamond Search
MAD - Mean Absolute Difference
MAE - Mean Absolute Error
DWT - Discrete Wavelet Transform
GS - Gaussian Smoothing
-
xxv
LIST OF ALGORITHMS
ALGORITHM TITLE PAGE
4.1 Conversion of videos into image frames 86
4.2 Conversion of video frames to greyscale 92
4.3 Frames divided into non-overlapping blocks 95
4.4 Calculation of parameter of gi(n) function 104
4.5 Calculation of parameters of fi(n) and hi(n)
filters 106
4.6 Calculation of coefficients of fi (n)and
hi (n) functions 107
4.7 Creation of SLT filter matrix 108
4.8 Transposed operation of SLTfilter matrix 109
4.9 Calculation of SLT Block matrix 110
4.10 Proposed MAE Block Matching function 113
4.11 Minimum optical flow movement value in
each S2F 123
4.12 Detection of the motionless area in each S2F
using the Th1 127
4.13 Maximum optical flow movement value in
each frame 135
4.14 Detection of the moving object area in S2F
using the Th2 136
4.15 Extraction of connected moving object 141
4.16 Computing displacement path based on center
point method 146
-
xxvi
4.17 DTW matching alignment process 154
4.18 Similarity measure between two displacement
paths 156
-
CHAPTER 1
INTRODUCTION
1.1 Overview
In the last of few years, with the advent of digital media such as video, images
and audio through internet, the means and the incentive to create digital forgeries have
also multiplied with it. As a matter of fact, powerful tamper media or editing software
and tools allow the creation of perceptually persuasive digital forgeries techniques.
Evolutions in visual digital video technologies such as digital transmission,
compression, storage, and video-conferencing have supported society in many ways.
Compared with digital image, the tampering of digital video is often more
sophisticated and time-consuming (Richao et al., 2014) although it is becoming easier
with the popularity of video editing tools, such as Video Editor and Adobe Photoshop.
The video processing software is commonly used to delete or incorporate
moving objects and change the forged regions with information garnered from their
neighbouring areas (Su et al., 2015). In this background video authentication refers to
a process that confirms the authenticity of a specific video as captured by camera
through searching and detecting various forensic types as to the tampering method. In
this regard, a video sequence can be modified through several forensic methods like
-
2
modification, combination or create of new video contents. The aim behind video
manipulation is to tamper, doctor or fake an authentic video. The real videos may be
utilized as sources of their tampered counterparts, and such tampering can be
conducted on a single video source or on many sources (Upadhyay and Singh, 2012).
Video forgeries mainly fall into two types of techniques that can be used for
video tampering detection: active forgeries and passive forgeries. In active forgeries,
(Di Martino and Sessa, 2012; Ram et al., 2009), the tampered region can be extracted
using a pre-embedded information such as watermark and fingerprint. However, this
scheme must have source files to embed the watermark first otherwise the detection
process will fail (Ng et al., 2006).
In passive approaches, techniques can be divided into three categories (C.-S.
Lin and Tsay, 2013) namely, source identification (sensor type of camera such as
noise), splicing techniques (multiple video-based forgery), and copy-move detection
techniques (single video-based forgery).
Source camera identification (C.-C. Hsu et al., 2008; Kang et al., 2012) is a
crucial issue that focuses on many issues that are linked to camera that is concerned
with identifying the source of a digital device; for example, mobile phones,
camcorders, and cameras. On the other hand, in splicing techniques forgery (Wahab
et al., 2014), two videos are combined to create one tampered video or a composite of
two or more videos are combined to create a fake video. Furthermore, splicing
tampering becomes difficult if the directions and lighting conditions are different
during recording with a dynamic camera (He et al., 2012; Y.-F. Hsu and Chang, 2007).
One of the major challenges that are faced by digital forensics is video copy-
move forgery or digital content tampering (Li and Huang, 2014). More recently, there
have been various types of forgery methods developed by hackers, with operation
duplication operation or copy-move on top of the list. In the context of cloning, the
-
3
main objective is to hide or incorporate an object from the same video scene to develop
a new scene. This process has become widely used as a malicious way of hiding
evidence (Qadir et al., 2012).
(a) (b)
Figure 1.1 Example of video Copy-Move forgery (a) Original video (b) Tampered
video from VTD dataset
Figure 1.1 illustrates an example of video copy-move forgery, where (a) the
original video and (b) the tampered video. For example, the ducks are recorded in the
video twice by taking part of the video (e.g., white car) and pasting it in another region
within the same video. It is challenging to detect this type of video forgery if the copy-
move procedure is carefully and actually carried out. Therefore, it is necessary to have
reliable and efficient methods to detect copy-move forgery for applications in law
enforcement, particularly forensics (Milani et al., 2012).
1.2 Background of Research
Dynamic developments in digital technologies and extensively utilized digital
video recording systems along with sophisticated video editing software, high quality
-
4
processing tools and algorithms, and low cost accessibility as well as easy to operate
digital multimedia devices have led to the increased video tampering and the challenge
of video authentication.
Video copy-move forgery has been identified as a vital form of forgery as it
utilizes the same video frames sources and destination-apart from this, the video frame
is copied and pasted on another part of the same video. In fact there are more subtle
cases found in the standard dataset video copy-move forgery whereby copied frames
are pasted on several places on the same video (many-to-many). Figure 1.2 illustrates
the above cases.
Figure 1.2 Sample of video copy-move forgery (a) Original video (b) Tampering
video from VTD dataset
In the beginning, the problem of copy-move lies in appointing video
authentication that plays a key role in detecting and determining region duplication,
frame duplication or object duplication of video forgery, and locating the factors that
affect video forgery (Richao, et al., 2014).
-
5
Pioneering methods depend on intrinsic features such as pixel value and
statistical features as well as video files characteristics. Among the many methods that
are considered for video forgery detection are those that are based on the identification
of acquisition device and detection of whether or not two video clips stem from the
one source. In relation to this, (Kot and Cao, 2013) stated that owing to the statistical
source features sensitive nature towards tampering and modifications, it is suitable to
be used in addressing tampering.
A majority of the previous works (Pathak and Patil, 2014; D'Amiano, et al., 2015;
Bestagini et al., 2013) have used Local Binary Patterns (LBP)-based features to identify
duplicated image regions. Though LBP is effective against distortions, scaling, JPEG
compression, blurring and noise adding, it however becomes ineffective when forged
areas are small. This failure may lead to inadvertent errors in the subsequent important
processes such as detection of moving objects and deleting objects. This is to
emphasise that precise duplicated region detection is utmost important in video copy-
move forgery detection. Failure which will result in low detection accuracy.
Beside the duplicated region, detection of moving objects, which traverse from
frame to frame, is also important. Most of the previous works (Pathak and Patil, 2014;
D'Amiano, et al., 2015; Bestagini et al., 2013) used LESH (Local Energy based Shape
Histogram) along with lexicographical sorting to determine objects’ trajectory similarity. The
method is similar to shape-based image retrieval approach. The method has a good efficiency
in detecting copy-move. However, the efficiency falls as the quality of the video frames
degrades. Video tampering refers to the generation of faked videos by adding, deleting
or altering new video object. It usually consists of detection/tracking, video
manipulation, video in-painting and video layer fusion (Kot and Cao, 2013).
The detection methods in passive blind video copy-move forgeries can be
categorized into four and they are pixel-based approach, format-based approach,
camera-based approach and geometric-based approach (Lin and Tsay, 2014).
-
6
1.2.1 Pixel-Based Approach
According to (Wang and Farid, 2007), pixel-based approaches make use of
high correlation between authentic and forged areas in video frames for the detection
of copy-paste forgery. The drawback lies in the fact that high correlation between
frames is common in normal videos rendering the method useless if the copied regions
are in use from other video frames. In (G.-S. Lin et al., 2011) proposed a video
detection method known as a coarse-to-fine grained method that uses the variation in
colour histograms of adjacent frames that are similar spatially and temporally – it
makes use of the macro-block based correlation algorithm to identify duplication.
However, their method is not able to determine region forgery.
In a similar study, (Zhang et al., 2009) brought forward a method that uses
ghost shadow artefacts presented by the consistencies in painting in order to detect
forged moving object region. Their approach differentiates static background from
moving foreground through block matching that is sensitive to the noise property
(illumination alterations), but due to its inaccuracy, the tampered region in each frame
cannot be identified by their method.
1.2.2 Format-Based Approach
An appearance of format-based approach in the video makes up part of the
forgery chain-of-evidence. A study of high MPEG is usually initialized with
compression video coding standard since some of the research in MPEG video
forgeries focus on the properties of the frames compression efficiency and how it is
affected when a video is tampered.
-
7
In this background, (Wang and Farid, 2006), proposed spatio-temporal domain
artefacts of doubly compressed MPEG video frames in type of I-frame that is like a
frame sequence of JPEG compressed image although there is considerable correlation
among frames in GOP (Group of Pictures). In relation to this, the determination of
predictive coded I-frame double compression is akin to double compression detection,
and in the case of GOP, insertion or removal of frames will increase the error of motion
estimation. Their method effectively works in detecting frame manipulation but not
in locating tampered object regions. In (Luo et al., 2008) study, the authors brought
forward a new method using the temporal blocking artefacts patterns to detect whether
deletion or insertion has been done on MPEG video prior to recompression with
different GOP structures. Their approaches are effective in frame-level forgery but
not in region-level tampering and localization.
1.2.3 Camera-Based Approach
The main steps of this approach is extracting different types of fingerprints
based on a set of videos then applying pattern recognition techniques in order to detect
forgery. Some fingerprints recognition, which can be used in these techniques include
noise patterns, lens distortion, and inconsistence-related artifacts (Kancherla and
Mukkamala, 2012).
As aforementioned, another set of video source camera identification method
is based on the extraction and measurement of noise characteristics that stem from
camera sensors. Noise is generally random, unwanted and variation of pixel values in
digital file (e.g., videos and images) that are sensors-generated. According to (Bayram
et al., 2005), the noise patterns are mostly utilized as a part of identification process
source owing to their deterministic properties that stem from CCD sensors.
-
8
(Kobayashi et al., 2010) proposed photon approach shot noise to detect
tampered level regions by using noise characteristics. Their method exploited the
inconsistencies of photon shot noise caused by various video source cameras to detect
between the original and forgery regions. But, their methods approach can merely
detect forgery regions in static scene videos and not suspicious level regions in videos
captured by a moving camera.
1.2.4 Geometric-Based Approach
Geometric-based approaches make use of measurements of objects in the video
and their positions relative to the moving video camera. Based on (Conotter et al.,
2011) study, a forgery technique approach to detect physically inconsistent
implausible trajectory of objects in video frame sequences was employed. However,
their technique can detect manipulation regions in a video sequence and limits the form
of the video frame (i.e., de-interlaced or interlaced).
The research questions regarding automatic video copy-move forgery
detection that are answered in this thesis are:
1. What different techniques for video forgery copy-move detection have been
proposed to date and where does state-of-the-art methods stand today in terms
of detection rates?
2. Which type of features can be extracted to characterizing the video frames?
3. Can a new video forgery copy-move detection technique, which could achieve
better performances in comparison to existing techniques in terms of detection
rate, be proposed?
-
9
4. How can the exact location of forged area in video forgery copy-move
detection be located?
1.3 Problem Statement
Malicious manipulations and video tampering without any evidence being left
behind has become very affordable and highly used, due to existence of extremely
powerful editing tools such as Adobe Photoshop and video editing software.
Therefore, there has been a swift augmentation of the digitally altered videos on the
Internet and mainstream media. This tendency depicts grave vulnerabilities as well as
minimizes the reliability of digital video. For such reasons, upcoming techniques in
verification of the authenticity and integrity of digital video have been regarded as
being significant, more so when putting into consideration the videos presented as
news items, as evidence in forensic investigation, such as murder surveillance, or part
of video forensics. From such a perspective, the principal goal in video forensics is to
determine and detect video forgery forensics (Amerini et al., 2013).
There entail a myriad of challenges faced by passive technique of detecting
video forgeries and equally have their constraints and setbacks. One of the fascinating
challenges facing the current scholars and researchers in this field is the reduction of
counterfeit positive rate of such approaches, in establishing effusive automatic system
with the capacity to identify image falsification from a wide perspective of video
formats. Additionally, the system detecting such falsification is made to increase its
dependability, robustness, and competence of operation. The key setback realized in
passive approaches is their requirement for several initial videos to approximate the
internal traces, whereas in capable situations there entails nothing else rather than the
video in query (Chen Moet al., 2008). Additional studies regarding this analysis can
be accessed in (Lanjewar et al., 2014).
-
10
(c) Forged video
(b) Target part of video
Figure 1.3 Example of video copy-move forgery (Qadir, et al., 2012)
Figure 1.3 shows an example of video copy-move forgery in which a region
from one image is copied and pasted within the same video (Muhammad Ghulam et
al., 2014). For instance, there has been a continuous problem in identifying copy-
move areas that have been rotated and scaled from different angles (Lanjewar et al.,
2014).
Against this backdrop, this research therefore concludes that existing video
copy-move detection methods still suffer many drawbacks, which include, among
others:
i. Their performance is mediocre when it comes to video compression, and
in GOP, where addition or deletion of frame increases the estimation of
motion error (Dong et al., 2012; Li and Huang, 2014; Shanableh, 2013).
Their method’s effectiveness is confined to the detection of frame
tampering and not localization of tampered regions.
+ (a) Original video
-
11
ii. The presence of homogeneous regions in the tampering video further
complicates the video copy-move forgery detection, which normally
increases the false positive and the accuracy rate is far from satisfactory
(Hyun et al., 2013; Su, et al., 2015; Subramanyam and Emmanuel, 2013).
iii. Notwithstanding the achievements realized by prior studies entailing high
correlations between original and forged regions in copy-move forgery
detection (e.g., Hsu et al., 2008; Thakur, 2013; Wang et al., 2014), high
correlation is known to be common in natural videos and the methods
proposed are not effective if the copied regions are within-frame object
tampering calling for more enhancements.
Thus, the remaining issues and drawbacks of the previous works, which have
been mentioned above, compel the author to pursue the research to seek a new
approach to improve the detection rate of video copy-move forgery. With that in mind,
a specific and focused research goal along with its objectives and scope are articulated
and given in the following sub-sections.
1.4 Research Goal
The study aims to design and develop a new video copy-move forgery
detection scheme with high accuracy based on optical flow methods.
-
12
1.5 Research Objectives
In order to achieve the goal, this study aims to fulfil the following main
objectives:
i. To develop a new method to trace duplicated blocks from each pair of
successive frames of a video using a two-tier approach comprising
Diamond search and Slantlet transform.
ii. To propose a new method to detect and localise both deleted objects and
moving objects using block displacements.
iii. To propose a new cloning object detection method based on displacement
paths of moving objects using Dynamic Time Warping (DTW) matching
algorithm.
1.6 Research Scope
The scope of this research is limited to the following:
i. Datasets: two sets of datasets namely, SULFA (Surrey University Library
for Forensic Analysis) (Qadir, et al., 2012) and VTD (Video Tampering
Dataset) are employed to evaluate the performance of the proposed scheme.
The former is a standard dataset obtained from
http://sulfa.cs.surrey.ac.uk/forged.php. On the other hand, the latter is a self
-created dataset which can be found at:
https://www.youtube.com/channel/UCZuuu-iyZvPptbIUHT9tMrA.
-
13
ii. Performance evaluation: This study’s only concern is the accuracy rate,
while the computational complexity is beyond its domain.
iii. Type of forgery: This study only focus on copy-move or copy-paste-move
forgery, while other kinds are out of scope of this study.
1.7 Significance of the Study
It is strongly believed that several applications like video copy-move forgery
detection investigations of digital video for forensic investigation such as in the case
of video surveillance, and presenting video evidence in courts of law need more
advanced detection and authentication techniques to prove the trustworthiness of
digital video. In light of the above mentioned issues, the results of this research are
expected to contribute to what is currently known about video copy-move forgery
detection. Nonetheless, the significance of this study is not only limited to forgery
detection, but also to the development of a new method that can be used in the future
in many applications in the field of computer vision.
1.8 Thesis Outline
The organization of this thesis is given in this section. The rest of the chapters
in this thesis begin with brief sections that highlight the aims of each section of the
chapter, and sums up with a short conclusion. Chapter 1, provides an overview of the
research problem and a brief background. The objectives of the research are also
described in this chapter.
-
14
In Chapter 2, an in-depth review of the existing literature on authentication of
video digital processing on the whole, as well as passive methods in attaining the
study’s objective, specifically are presented. The currently employed approaches and
criteria within the context of counterfeit digital video recognition are highly defined in
this chapter. A review of a current study is conducted inclusive of recent techniques
and methods employed in sensing video forgery tampering. This study area is
moderately novel hence the meagre sources relative to the study topic. Therefore, the
reviewed and availed approaches relate to extensive processing of digital videos, but
only somewhat related to forgery detection in videos forensic. All the chapters are
independent but a flow and coherency of ideas throughout the entire thesis are ensured.
Chapter 3, presents a clear roadmap of this study to guide the reader to achieve
a quick grasp of the detailed research framework. The advantages of using the popular
dataset in the newly developed methods are emphasised. The layout of the entire
research framework, strategies, and procedures are highlighted. This is followed by
Chapter 4, where a detailed design of the proposed method is provided along with the
step-by-step processes and the proposed algorithms employed in it. The chapter also
provides a discussion of the proposed methods entailing the proposed method’s
implementation on video copy-move forgery detection.
The next chapter (Chapter 5), provides the results of the proposed method used
on two datasets SULFA and the VTD dataset of video copy-move forgery detection,
along with the experimental results, detailed analyses, and discussion. It explains the
qualitative and quantitative measurements that are carried out for the performance
evaluations and implementation of the method for every phase with the inclusion of
the detection of the tampered videos and localisation of the forgery region. The
qualitative measurements are based on visual human inspection, while the quantitative
measurements are performed using standard approaches. Lastly, Chapter 6 concludes
the study by emphasizing the major contributions, enumerating the major
achievements achieved and providing recommendations for future studies.
-
215
Conotter, V., O'Brien, J. F. and Farid, H. (2012). Exposing digital forgeries in ballistic
motion. Information Forensics and Security, IEEE Transactions on, 7(1), 283-
296.
Dalal, N. and Triggs, B. (2005). Histograms of oriented gradients for human detection.
Proceedings of the 2005 Computer Vision and Pattern Recognition, 2005. CVPR
2005. IEEE Computer Society Conference on, 886-893.
D'Amiano, L., Cozzolino, D., Poggi, G. and Verdoliva, L. (2015). Video forgery
detection and localization based on 3D patchmatch. Proceedings of the 2015
Multimedia & Expo Workshops (ICMEW), 2015 IEEE International Conference
on, 1-6.
Davarzani, R., Yaghmaie, K., Mozaffari, S. and Tapak, M. (2013). Copy-move forgery
detection using multiresolution local binary patterns. Forensic science
international, 231(1), 61-72.
Di Martino, F. and Sessa, S. (2012). Fragile watermarking tamper detection with
images compressed by fuzzy transform. Information Sciences, 195, 62-90.
Dice, L. R. (1945). Measures of the amount of ecologic association between species.
Ecology, 26(3), 297-302.
Dong, Q., Yang, G. and Zhu, N. (2012). A MCEA based passive forensics scheme for
detecting frame-based video tampering. Digital Investigation, 9(2), 151-159.
Elnagar, A. and Basu, A. (1995). Motion detection using background constraints.
Pattern Recognition, 28(10), 1537-1554.
Esmaeilani, R. (2014). Source identification of captured video using photo response
non-uniformity noise pattern and svm classifiers. Master thesis, Universiti
Teknologi Malaysia.
Fan, X. and Tjahjadi, T. (2015). A spatial-temporal framework based on histogram of
gradients and optical flow for facial expression recognition in video sequences.
Pattern Recognition, 48(11), 3407-3416.
Farid, H. (2003). A picture tells a thousand lies. New Scientist, (2411), 38-41.
Farid, H. (2009). Image forgery detection. Signal Processing Magazine, IEEE, 26(2),
16-25.
Fridrich, A. J., Soukal, B. D. and Lukáš, A. J. (2003). Detection of copy-move forgery
in digital images. Proceedings of the 2003 in Proceedings of Digital Forensic
Research Workshop, 1-10.
-
216
Fu, A. W.-C., Keogh, E., Lau, L. Y., Ratanamahatana, C. A. and Wong, R. C.-W.
(2008). Scaling and time warping in time series querying. The VLDB Journal—
The International Journal on Very Large Data Bases, 17(4), 899-921.
Garg, R., Varna, A. L. and Wu, M. (2011). Seeing ENF: natural time stamp for digital
video via optical sensing and signal processing. Proceedings of the 2011
Proceedings of the 19th ACM international conference on Multimedia, 23-32.
Ghanbari, M. (1990). The cross-search algorithm for motion estimation. IEEE
Transactions on Communications, 38(7), 950-953.
Ghanem, K. (2013). Towards more accurate clustering method by using dynamic time
warping. International Journal of Data Mining & Knowledge Management
Process (IJDKP), 3(2), 107-118.
Gironi, A., Fontani, M., Bianchi, T., Piva, A. and Barni, M. (2014). A video forensic
technique for detecting frame deletion and insertion. Proceedings of the 2014
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International
Conference on, 6226-6230.
Goldman, D. B., Gonterman, C., Curless, B., Salesin, D. and Seitz, S. M. (2008). Video
object annotation, navigation, and composition. Proceedings of the 2008
Proceedings of the 21st annual ACM symposium on User interface software and
technology, 3-12.
He, Z., Lu, W., Sun, W. and Huang, J. (2012). Digital image splicing detection based
on Markov features in DCT and DWT domain. Pattern Recognition, 45(12),
4292-4299.
Hikmat N, A. and Safa'a AA, A. (2010). Implementation of 8-point Slantlet transform
based polynomial cancellation coding-OFDM system using FPGA. Proceedings
of the 2010 Systems Signals and Devices (SSD), 2010 7th International Multi-
Conference on, 1-6.
Hsu, C.-C., Hung, T.-Y., Lin, C.-W. and Hsu, C.-T. (2008). Video forgery detection
using correlation of noise residue. Proceedings of the 2008 Multimedia Signal
Processing, 2008 IEEE 10th Workshop on, 170-174.
Hsu, Y.-F. and Chang, S.-F. (2007). Image splicing detection using camera response
function consistency and automatic segmentation. Proceedings of the 2007
Multimedia and Expo, 2007 IEEE International Conference on, 28-31.
-
217
Hyun, D.-K., Lee, M.-J., Ryu, S.-J., Lee, H.-Y. and Lee, H.-K. (2013). Forgery
detection for surveillance video. The Era of Interactive Media (pp. 25-36):
Springer.
Ince, S. and Konrad, J. (2008). Occlusion-aware optical flow estimation. IEEE
Transactions on Image Processing, 17(8), 1443-1451.
Jaccard, P. (1912). The distribution of the flora in the alpine zone. New phytologist,
11(2), 37-50.
Jaiswal, S. and Dhavale, S. (2013). Video Forensics in Temporal Domain using
Machine Learning Techniques. International Journal of Computer Network and
Information Security, 5(9), 58.
Jing, G., Rajan, D. and Siong, C. E. (2005). Motion detection with adaptive
background and dynamic thresholds. Proceedings of the 2005 2005 5th
International Conference on Information Communications & Signal Processing,
41-45.
Johnson, M. K. and Farid, H. (2006). Exposing digital forgeries through chromatic
aberration. Proceedings of the 2006 Proceedings of the 8th workshop on
Multimedia and security, 48-55.
Johnson, M. K. and Farid, H. (2007). Exposing digital forgeries through specular
highlights on the eye. Proceedings of the 2007 Information Hiding, 311-325.
Kancherla, K. and Mukkamala, S. (2012). Novel blind video forgery detection using
markov models on motion residue. Intelligent Information and Database
Systems (pp. 308-315): Springer.
Kang, X., Li, Y., Qu, Z. and Huang, J. (2012). Enhancing source camera identification
performance with a camera reference phase sensor pattern noise. Information
Forensics and Security, IEEE Transactions on, 7(2), 393-402.
Kattoush, A. H. (2012). A radon slantlet transforms based OFDM system design and
performance simulation under different channel conditions. ISRN
Communications and Networking, 2012, 1-10.
Kim, C. and Vasudev, B. (2005). Spatiotemporal sequence matching for efficient video
copy detection. Circuits and Systems for Video Technology, IEEE Transactions
on, 15(1), 127-132.
-
218
Kitagawa, G. (1994). The two-filter formula for smoothing and an implementation of
the Gaussian-sum smoother. Annals of the Institute of Statistical Mathematics,
46(4), 605-623.
Kobayashi, M., Okabe, T. and Sato, Y. (2009). Detecting video forgeries based on
noise characteristics. Advances in Image and Video Technology (pp. 306-317):
Springer.
Kobayashi, M., Okabe, T. and Sato, Y. (2010). Detecting forgery from static-scene
video based on inconsistency in noise level functions. IEEE Transactions on,
Information Forensics and Security, 5(4), 883-892.
Koga, T. (1981). Motion-compensated interframe coding for video conferencing.
Proceedings of the 1981 Proc. NTC 81, (pp. C9-6).
Kopparapu, S. and Satish, M. (2014). Optimal Gaussian Filter for Effective Noise
Filtering. Computer Science, 12(2), 1-7.
Kot, A. C. and Cao, H. (2013). Image and Video Source Class Identification. Digital
Image Forensics (pp. 157-178): Springer.
Kroeger, T., Timofte, R., Dai, D. and Van Gool, L. (2016). Fast Optical Flow using
Dense Inverse Search. Computer Vision and Pattern Recognition, 1(3), 1-25.
Kurosawa, K., Kuroki, K. and Saitoh, N. (1999). CCD fingerprint method-
identification of a video camera from videotaped images. IEEE International
Conference on Image Processing, vol.3, 537-540.
Lanjewar, S. B., P. A. Khaire, and R. Meshram, (2014). An approach towards image
forgery detection. International Journal of Advanced Research in Computer
Science. vol.5, no.8, 205-209.
Le Gall, D. (1991). MPEG: A video compression standard for multimedia applications.
Communications of the ACM, 34(4), 46-58.
Li, F. and Huang, T. (2014). Video copy-move forgery detection and localization
based on structural similarity. Proceedings of the 2014 Proceedings of the 3rd
International Conference on Multimedia Technology (ICMT 2013), 63-76.
Li, R., Yu, S. and Yang, X. (2007). Efficient spatio-temporal segmentation for
extracting moving objects in video sequences. IEEE Transactions on Consumer
Electronics, 53(3), 1161-1167.
Li, X., Jing, T. and Li, X. (2010). Image splicing detection based on moment features
and Hilbert-Huang Transform. Proceedings of the 2010 Information Theory and
-
219
Information Security (ICITIS), 2010 IEEE International Conference on, 1127-
1130.
Liao, S.-Y. and Huang, T.-Q. (2013). Video copy-move forgery detection and
localization based on Tamura texture features. Proceedings of the 2013 Image
and Signal Processing (CISP), 2013 6th International Congress on, 864-868.
Lin, C.-S. and Tsay, J.-J. (2013). Passive approach for video forgery detection and
localization. Proceedings of the 2013 The Second International Conference on
Cyber Security, Cyber Peacefare and Digital Forensic (CyberSec2013), 107-
112.
Lin, C.-S. and Tsay, J.-J. (2014). A passive approach for effective detection and
localization of region-level video forgery with spatio-temporal coherence
analysis. Digital Investigation, 11(2), 120-140.
Lin, G.-S., Chang, J.-F. and Chuang, C.-H. (2011). Detecting frame duplication based
on spatial and temporal analyses. Proceedings of the 2011 Computer Science &
Education (ICCSE), 2011 6th International Conference on, 1396-1399.
Lucas, B. D. and Kanade, T. (1981). An iterative image registration technique with an
application to stereo vision. Proceedings of the 1981 International Joint
Conference on Artificial Intelligence, 674-679.
Luo, W., Wu, M. and Huang, J. (2008). MPEG recompression detection based on
block artifacts. Proceedings of the 2008 Electronic Imaging 2008, 68190X-
68190X-68112.
Mahajan, A., Sharma, K. and Kumar, D. (2014). Comparative Analysis of Pixel Based
Motion Estimation with Block Based Motion Estimation. International Journal
of Recent Research Aspects, 1(2), 78-83.
Malviya, A. V. and Ladhake, S. A. (2016). Pixel Based Image Forensic Technique for
Copy-move Forgery Detection Using Auto Color Correlogram. Procedia
Computer Science, 79, 383-390.
Manzanera, A. and Richefeu, J. C. (2007). A new motion detection algorithm based
on Σ–Δ background estimation. Pattern Recognition Letters, 28(3), 320-328.
Metkar, S. and Talbar, S. (2013). Performance Evaluation of Block Matching
Algorithms for Video Coding Motion Estimation Techniques for Digital Video
Coding (pp. 13-31): Springer.
-
220
Milani, S., Bestagini, P., Tagliasacchi, M. and Tubaro, S. (2012). Multiple
compression detection for video sequences. Proceedings of the 2012 Multimedia
Signal Processing (MMSP), 2012 IEEE 14th International Workshop on, 112-
117.
Milani, S., Fontani, M., Bestagini, P., Barni, M., Piva, A., Tagliasacchi, M., et al.
(2012). An overview on video forensics. APSIPA Transactions on Signal and
Information Processing, 1, 1-18.
Mondaini, N., Caldelli, R., Piva, A., Barni, M. and Cappellini, V. (2007). Detection of
malevolent changes in digital video for forensic applications. Proceedings of the
2007 Electronic Imaging 2007, 65050T-65050T-65012.
Muhammad, Ghulam, Munner H. Al-Hammadi, Muhammad Hussain, and George
Bebis, (2014). Image Forgery Detection Using Steerable Pyramid Transform and
Local Binary Pattern. Machine Vision and Applications. vol.25, no.4, pp.985-
995.
Mutt, S. and Kumar, S. (2009). Secure image Steganography based on Slantlet
transform. Proceedings of the 2009 Methods and Models in Computer Science,
2009. ICM2CS 2009. Proceeding of International Conference on, 1-7.
Ng, T.-T., Chang, S.-F., Lin, C.-Y. and Sun, Q. (2006). Passive-blind image forensics.
Multimedia Security Technologies for Digital Rights, 15, 383-412.
Pathak, A. and Patil, D. (2014). Video Forgery Detection Based on Variance in
Luminance and Signal to Noise Ratio using LESH Features and Bispectral
Analysis. International Journal of Computer Science and Mobile Computing,
3(7), 318-327.
Peng, F., Nie, Y.-y. and Long, M. (2011). A complete passive blind image copy-move
forensics scheme based on compound statistics features. Forensic science
international, 212(1), e21-e25.
Podilchuk, C. I. and Zeng, W. (1998). Image-adaptive watermarking using visual
models. Selected Areas in Communications, IEEE Journal on, 16(4), 525-539.
Popescu, A. C. and Farid, H. (2004). Statistical tools for digital forensics. Proceedings
of the 2004 Information Hiding, 128-147.
Popescu, A. C. and Farid, H. (2005). Exposing digital forgeries in color filter array
interpolated images. IEEE Transactions on, Signal Processing, 53(10), 3948-
3959.
-
221
Qadir, G., Yahaya, S. and Ho, A. T. (2012). Surrey university library for forensic
analysis (SULFA) of video content. IET Conference on, Proceedings of the 2012
Image Processing (IPR 2012), 1-6.
Qin, Y.-l., Sun, G.-l., Wang, S.-z. and ZHANG, X.-p. (2010). Blind detection of video
sequence montage based on GOP abnormality. Act Electronica Sinica, 38(7),
1597-1602.
Ram, S., Bischof, H. and Birchbauer, J. (2009). Active fingerprint ridge orientation
models Advances in Biometrics (pp. 534-543): Springer.
Ratanamahatana, C. A. and Keogh, E. (2004). Everything you know about dynamic
time warping is wrong. Proceedings of the 2004 Third Workshop on Mining
Temporal and Sequential Data, 53-63.
Redi, J. A., Taktak, W. and Dugelay, J.-L. (2011). Digital image forensics: a booklet
for beginners. Multimedia Tools and Applications, 51(1), 133-162.
Ren, Y., Chua, C.-S. and Ho, Y.-K. (2003). Motion detection with no stationary
background. Machine Vision and Applications, 13(5-6), 332-343.
Richao, C., Gaobo, Y. and Ningbo, Z. (2014). Detection of object-based manipulation
by the statistical features of object contour. Forensic science international, 236,
164-169.
Richardson, I. E. (2004). H. 264 and MPEG-4 video compression: video coding for
next-generation multimedia: John Wiley & Sons.
Schneider, M. and Chang, S.-F. (1996). A robust content based digital signature for
image authentication. Proceedings of the 1996 Image Processing, 1996.
Proceedings., International Conference on, 227-230.
Selesnick, I. W. (1999). The slantlet transform. IEEE Transactions on Signal
Processing, 47(5), 1304-1313.
Shaid, S. (2009). Estimating optimal block size of copy-move attack detection on
highly textured image. Universiti Teknologi Malaysia.
Shanableh, T. (2013). Detection of frame deletion for digital video forensics. Digital
Investigation, 10(4), 350-360.
Shih, T. K., Tang, N. C. and Hwang, J.-N. (2007). Ghost shadow removal in multi-
layered video inpaintinga. Proceedings of the 2007 Multimedia and Expo, 2007
IEEE International Conference on, 1471-1474.
-
222
Shih, T. K., Tang, N. C., Tsai, J. C. and Hwang, J.-N. (2011). Video motion
interpolation for special effect applications. Systems, Man, and Cybernetics,
Part C: Applications and Reviews, IEEE Transactions on, 41(5), 720-732.
Shivakumar, B. and Baboo, L. D. S. S. (2010). Detecting copy-move forgery in digital
images: a survey and analysis of current methods. Global Journal of Computer
Science and Technology, 10(7).
Sowmya, K. and Chennamma, H. (2015). A Survey on video forgery detection.
International Journal of Computer Engineering and Applications, 9(2), 17-27.
Srinivasan, R. and Rao, K. (1985). Predictive coding based on efficient motion
estimation. IEEE Transactions on Communications, 33(8), 888-896.
Su, L., Huang, T. and Yang, J. (2015). A video forgery detection algorithm based on
compressive sensing. Multimedia Tools and Applications, 74(17), 6641-6656.
Su, Y., Zhang, J. and Liu, J. (2009). Exposing digital video forgery by detecting
motion-compensated edge artifact. Proceedings of the 2009 Computational
Intelligence and Software Engineering, 2009. CiSE 2009. International
Conference on, 1-4.
Subramanyam, A. and Emmanuel, S. (2012). Video forgery detection using HOG
features and compression properties. Proceedings of the 2012 Multimedia Signal
Processing (MMSP), 2012 IEEE 14th International Workshop on, 89-94.
Subramanyam, A. and Emmanuel, S. (2013). Pixel estimation based video forgery
detection. Proceedings of the 2013 Acoustics, Speech and Signal Processing
(ICASSP), 2013 IEEE International Conference on, 3038-3042.
Sun, T., Wang, W. and Jiang, X. (2012). Exposing video forgeries by detecting MPEG
double compression. Proceedings of the 2012 Acoustics, Speech and Signal
Processing (ICASSP), 2012 IEEE International Conference on, 1389-1392.
Thakur, M. K. (2013). Tampered videos: detection and quality assessment. PhD thesis,
Jaypee Institute of Information Technology.
Upadhyay, S. and Singh, S. K. (2011). Learning based video authentication using
statistical local information. Proceedings of the 2011 Image Information
Processing (ICIIP), 2011 International Conference on, 1-6.
Upadhyay, S. and Singh, S. K. (2012). Video Authentication: Issues and Challenges.
International Journal of Computer Science Issues, 9(1), 409-418.
-
223
Vazquez-Padin, D., Fontani, M., Bianchi, T., Comesaña, P., Piva, A. and Barni, M.
(2012). Detection of video double encoding with GOP size estimation. 2012
IEEE International Workshop on, Proceedings of the Information Forensics and
Security (WIFS), 151-156.
Wahab, A. W. A., Bagiwa, M. A., Idris, M. Y. I., Khan, S., Razak, Z. and Ariffin, M.
R. K. (2014). Passive video forgery detection techniques: a survey. 10th
International Conference on, Proceedings of the 2014 Information Assurance
and Security (IAS), 29-34.
Wang, Q., Li, Z., Zhang, Z. and Ma, Q. (2014). Video inter-frame forgery
identification based on consistency of correlation coefficients of gray values.
Journal of Computer and Communications, 2(04), 51.
Wang, W. (2009). Digital video forensics. PhD thesis, Dartmouth College, Hanover,
New Hampshire.
Wang, W. and Farid, H. (2006). Exposing digital forgeries in video by detecting double
MPEG compression. Proceedings of the 2006 Proceedings of the 8th workshop
on Multimedia and security, 37-47.
Wang, W. and Farid, H. (2007a). Exposing digital forgeries in interlaced and
deinterlaced video. Information Forensics and Security, IEEE Transactions on,
2(3), 438-449.
Wang, W. and Farid, H. (2007b). Exposing digital forgeries in video by detecting
duplication. Proceedings of the 2007b Proceedings of the 9th workshop on
Multimedia & security, 35-42.
Wang, W. and Farid, H. (2009). Exposing digital forgeries in video by detecting double
quantization. Proceedings of the 2009 Proceedings of the 11th ACM workshop
on Multimedia and security, 39-48.
Wang, W., Jiang, X., Wang, S., Wan, M. and Sun, T. (2013). Identifying video forgery
process using optical flow Digital-Forensics and Watermarking (pp. 244-257):
Springer.
Wang, Z. (2006). Time series matching: a multi-filter approach. Courant Institute of
Mathematical Sciences New York.
Wolf, S. and Pinson, M. (2009). A no reference (NR) and reduced reference (RR)
metric for detecting dropped video frames. Proceedings of the 2009 Fourth
-
224
International Workshop on Video Processing and Quality Metrics for Consumer
Electronics (VPQM-09).,
Wu, X., Li, J., Zhang, Y. and Tang, S. (2008). Spatio-temporal visual consistency for
video copy detection. Proceedings of the 2008 Visual Information Engineering,
2008. VIE 2008. 5th International Conference on, 414-419.
Wu, Y., Jiang, X., Sun, T. and Wang, W. (2014). Exposing video inter-frame forgery
based on velocity field consistency. Proceedings of the 2014 Acoustics, Speech
and Signal Processing (ICASSP), 2014 IEEE International Conference on,
2674-2678.
Xiaoling, C. and Huimin, Z. (2012). A Novel Video Tamper Detection Algorithm
Based on Semi-fragile Watermarking Advances in Information Technology and
Industry Applications (pp. 489-497): Springer.
Yu, D., Yu, X., Hu, Q., Liu, J. and Wu, A. (2011). Dynamic time warping constraint
learning for large margin nearest neighbor classification. Information Sciences,
181(13), 2787-2796.
Yu, J. and Srinath, M. D. (2001). An efficient method for scene cut detection. Pattern
Recognition Letters, 22(13), 1379-1391.
Zeng, F.-z., LU, Y.-s. and ZHOU, Y. (2011). The video tamper detection algorithm
based on semi-fragile watermark with compressive sensing [J]. Journal of
Circuits and Systems, 4, 016.
Zhang, J., Su, Y. and Zhang, M. (2009). Exposing digital video forgery by ghost
shadow artifact. Proceedings of the 2009 Proceedings of the First ACM
workshop on Multimedia in forensics, 49-54.
Zheng, J., Li, B., Zhou, B. and Li, W. (2005). Fast motion detection based on
accumulative optical flow and double background model. Proceedings of the
2005 International Conference on Computational and Information Science, 291-
296.
Zhu, C., Lin, X., Chau, L.-P., Ang, H.-A. and Ong, C.-Y. (2004). Efficient inner search
for faster diamond search. Signal processing, 84(3), 527-533.
Zhu, S. and Ma, K.-K. (2000). A new diamond search algorithm for fast block-
matching motion estimation. Image Processing, IEEE Transactions on, 9(2),
287-290.
-
Table of Corrections
Required Corrections Page No.
The Amendment Made Page No.
Title Video Copy-Move Forgery Detection Scheme
Based On Displacement Paths
Abstract Amended
Chapter 1
Typographical errors including English
language and UTM Formatting
Checked and amended 2, 9 and
11
Background of Research 4 Added two paragraphs about motivation of
previous works 5
Problem statement 8 Added a paragraph about motivation of this
study 9
Figure 1.3 adding (a), (b) and (c)
10 Added figure 1.3 the link motivation of this
study 10
Adding reference at paragraph 1 line 2
10 Added (Muhammad Ghulam et al., 2014) 10
Adding paragraphs about motivation of this study
10 Added two paragraph in page 10 to link a
motivation of Problem Statement 11
Chapter 2
Typographical errors including English
language and UTM Formatting
Checked and amended 16, 26, 37 and
51
Give critical analysis (short coming, advantage
of other works) 22
Critical analyses are given in four separate pages
24, 25, 32 and
50
Chapter 3
Typographical errors including English
language and UTM Formatting
Checked and amended 59 and
62
Research framework should be contributions
and link to objective 59
Research framework is amended to include the contributions and link to objectives
60 and 61
Adding words 78 Add “see in Figure 3.11” words 78
Chapter 4
Typographical errors including English
language and UTM Formatting
Checked and amended 89, 94,
98, 100, and 106
-
Chapter title need to be changed
81 Changed to “ PROPOSED VIDEO COPY-MOVE
FORGERY DETECTION SCHEME” 82
Need to tie up with the research problem
81 Added paragraph about the proposed
method’s main components and link with research problem and objectives
82
Insert small caption in Fig. 4.1
83 Changed caption of Figure 4.1: “Proposed video
copy-move forgery detection scheme” 84
Chapter 5
Typographical errors including English
language and UTM Formatting
Checked and amended 160,193, 194, 195 and 196
Indicating the figure with a symbol in the relevant
figures and tables 180
Each video in Table 5.2 indicated to the name of each video in both datasets SULA and VTD
181
Indicating the figure with a symbol in the relevant
figures and tables 184
Each video in Table 5.3 indicated to the name of each video in both datasets SULA and VTD
185
Changing words in Figure 5.28
193 Changing words from “ (red ball) “ to
“(Elephant) “ words in Figure 5.28 194
Changing words in Figure 5.29
194 Changing words from “ (red ball) “ to “(red
car)“ words in Figure 5.29 195
Changing words in Figure 5.30
195 Changing words from “ (red ball) “ to “(white
ball) “ words in Figure 5.30 196
Changing words in Figure 5.31
196 Changing words from “ (red ball) “ to “(white
car)“ words in Figure 5.31 197
Indicating the figure with a symbol in the relevant
figures and tables 198
Each video in Table 5.4 indicated to the name of each video in dataset SULA
199
Indicating the figure with a symbol in the relevant
figures and tables 200
Each video in Table 5.5 indicated to the name of each video in dataset SULA
201
Chapter 6
Typographical errors including English
language and UTM Formatting
Checked
second paragraph is a revisit to problem
background, objectives 205
A new paragraph (2nd. Paragraph) is added to address the issue
206
References
Reference has been deleted form References
212
Al-Sanjary, O. I., Ahmed, A. A. and Sulong, G. (2016). Development of a video tampering dataset for forensic investigation. Forensic
Science International, 266, 565-572.
213