robust hashing for image authe

9
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 8, NO. 1, JANUARY 2013 55 Robust Hashing for Image Authentication Using Zernike Moments and Local Features Yan Zhao, Shuozhong Wang, Xinpeng Zhang,and Heng Yao, Member, IEEE Abstract—A robust hashing method is developed for detecting image forgery including removal, insertion, and replacement of ob- jects, and abnormal color modication, and for locating the forged area. Both global and local features are used in forming the hash sequence. The global features are based on Zernike moments rep- resenting luminance and chrominance characteristics of the image as a whole. The local features include position and texture infor- mation of salient regions in the image. Secret keys are introduced in feature extraction and hash construction. While being robust against content-preserving image processing, the hash is sensitive to malicious tampering and, therefore, applicable to image authen- tication. The hash of a test image is compared with that of a refer- ence image. When the hash distance is greater than a threshold and less than , the received image is judged as a fake. By decom- posing the hashes, the type of image forgery and location of forged areas can be determined. Probability of collision between hashes of different images approaches zero. Experimental results are pre- sented to show effectiveness of the method. Index Terms—Forgery detection, image hash, perceptual robust- ness, saliency, Zernike moments. I. INTRODUCTION W ITH the widespread use of image editing software, en- suring credibility of the image contents has become an important issue. Image hashing is a technique that extracts a short sequence from the image to represent its contents, and therefore can be used for image authentication. If the image is maliciously modied, the hash must be changed signicantly. Meanwhile, unlike hash functions in cryptography such as MD5 and SHA-1 that are extremely sensitive to slight changes in the input data, the image hash should be robust against normal image processing. In general, a good image hash should be rea- sonably short, robust to ordinary image manipulations, and sen- sitive to tampering. It should also be unique in the sense that Manuscript received April 16, 2012; revised September 20, 2012; accepted September 25, 2012. Date of publication October 09, 2012; date of current version December 26, 2012. This work was supported by the National Sci- ence Foundation of China (60832010, 61071187). The associate editor coor- dinating the review of this manuscript and approving it for publication was Prof. Chiou-Ting Hsu. Y. Zhao is with School of Communication and Information Engineering, Shanghai University, Shanghai 200072, China, and also with the School of Computer and Information Engineering, Shanghai University of Electric Power, Shanghai 200090, China (e-mail: [email protected]). S. Wang and X. Zhang are with School of Communication and Informa- tion Engineering, Shanghai University, Shanghai 200072, China (e-mail: [email protected]; [email protected]). H. Yao was with the School of Communication and Information Engineering, Shanghai University, Shanghai 200072, China. He is now with University of Shanghai for Science and Technology, Shanghai 200093, China (e-mail: [email protected]). Color versions of one or more of the gures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identier 10.1109/TIFS.2012.2223680 different images have signicantly different hash values, and secure so that any unauthorized party cannot break the key and coin the hash. To meet all the requirements simultaneously, es- pecially perceptual robustness and sensitivity to tampering, is a challenging task. Various image hashing methods have been proposed. Monga et al. [1] develop a two-step framework that includes feature extraction (intermediate hash) and coding of the intermediate result to form the nal hash. That has become a routine practice in many image hashing methods. Many previous schemes are either based on global [2]–[5] or local [6]–[11] features. Global features are generally short but insensitive to changes of small areas in the image, while local features can reect regional mod- ications but usually produce longer hashes. In [2], Xiang et al. propose a method using invariance of the image histogram to geometric deformations. It is robust to geometric attacks, but cannot distinguish images with similar histograms but different contents. Tang et al. [3] develop a global method using nonnegative matrix factorization (NMF). The image is rst converted into a xed-sized pixel array. A secondary image is obtained by rearranging pixels and applying NMF to produce a feature-bearing coefcient matrix, which is then coarsely quantized. The obtained binary string is scrambled to generate the image hash. Swaminathan et al. [4] propose an image hash method based on rotation invariance of Fourier-Mellin transform and present a new framework to study the security issues of existing image hashing schemes. Their method is robust to geometric distortions, ltering oper- ations, and various content-preserving manipulations. In [5], Lei et al. calculate DFT of the invariant moments of signicant Radon transform coefcients, and normalize/quantize the DFT coefcients to form the image hash for content authentication. Kheliet al. [6] propose a robust and secure hashing scheme based on virtual watermark detection. The method is robust against normal image processing operations and geometric transformation, and can detect content changes in relatively large areas. In another work, Monga et al. [7] apply NMF to pseudo-randomly selected subimages. They construct a secondary image, and obtain low-rank matrix approximation of the secondary image with NMF again. The matrix entries are concatenated to form an NMF-NMF vector. The inner prod- ucts of the NMF-NMF vector and a set of weight vectors are calculated. Because the nal hash comes from the secondary image with NMF, their method cannot locate forged regions. In analyzing the NMF-NMF method, Fouad et al. [8] point out that, among the three keys it uses, the rst one for pseudo-ran- domly selecting several subimages is crucial. However, it can be accurately estimated based on the observation of image hash pairs when reused several times on different images. A 1556-6013/$31.00 © 2012 IEEE

Upload: raghul-ramasamy

Post on 30-Dec-2015

144 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Robust Hashing for Image Authe

IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 8, NO. 1, JANUARY 2013 55

Robust Hashing for Image Authentication UsingZernike Moments and Local FeaturesYan Zhao, Shuozhong Wang, Xinpeng Zhang, and Heng Yao, Member, IEEE

Abstract—A robust hashing method is developed for detectingimage forgery including removal, insertion, and replacement of ob-jects, and abnormal color modification, and for locating the forgedarea. Both global and local features are used in forming the hashsequence. The global features are based on Zernike moments rep-resenting luminance and chrominance characteristics of the imageas a whole. The local features include position and texture infor-mation of salient regions in the image. Secret keys are introducedin feature extraction and hash construction. While being robustagainst content-preserving image processing, the hash is sensitiveto malicious tampering and, therefore, applicable to image authen-tication. The hash of a test image is compared with that of a refer-ence image. When the hash distance is greater than a thresholdand less than , the received image is judged as a fake. By decom-posing the hashes, the type of image forgery and location of forgedareas can be determined. Probability of collision between hashesof different images approaches zero. Experimental results are pre-sented to show effectiveness of the method.

Index Terms—Forgery detection, image hash, perceptual robust-ness, saliency, Zernike moments.

I. INTRODUCTION

W ITH the widespread use of image editing software, en-suring credibility of the image contents has become an

important issue. Image hashing is a technique that extracts ashort sequence from the image to represent its contents, andtherefore can be used for image authentication. If the image ismaliciously modified, the hash must be changed significantly.Meanwhile, unlike hash functions in cryptography such asMD5and SHA-1 that are extremely sensitive to slight changes inthe input data, the image hash should be robust against normalimage processing. In general, a good image hash should be rea-sonably short, robust to ordinary image manipulations, and sen-sitive to tampering. It should also be unique in the sense that

Manuscript received April 16, 2012; revised September 20, 2012; acceptedSeptember 25, 2012. Date of publication October 09, 2012; date of currentversion December 26, 2012. This work was supported by the National Sci-ence Foundation of China (60832010, 61071187). The associate editor coor-dinating the review of this manuscript and approving it for publication wasProf. Chiou-Ting Hsu.Y. Zhao is with School of Communication and Information Engineering,

Shanghai University, Shanghai 200072, China, and also with the Schoolof Computer and Information Engineering, Shanghai University of ElectricPower, Shanghai 200090, China (e-mail: [email protected]).S. Wang and X. Zhang are with School of Communication and Informa-

tion Engineering, Shanghai University, Shanghai 200072, China (e-mail:[email protected]; [email protected]).H. Yao was with the School of Communication and Information Engineering,

Shanghai University, Shanghai 200072, China. He is now with Universityof Shanghai for Science and Technology, Shanghai 200093, China (e-mail:[email protected]).Color versions of one or more of the figures in this paper are available online

at http://ieeexplore.ieee.org.Digital Object Identifier 10.1109/TIFS.2012.2223680

different images have significantly different hash values, andsecure so that any unauthorized party cannot break the key andcoin the hash. To meet all the requirements simultaneously, es-pecially perceptual robustness and sensitivity to tampering, is achallenging task.Various image hashing methods have been proposed. Monga

et al. [1] develop a two-step framework that includes featureextraction (intermediate hash) and coding of the intermediateresult to form the final hash. That has become a routine practicein many image hashing methods. Many previous schemes areeither based on global [2]–[5] or local [6]–[11] features. Globalfeatures are generally short but insensitive to changes of smallareas in the image, while local features can reflect regional mod-ifications but usually produce longer hashes.In [2], Xiang et al. propose a method using invariance of

the image histogram to geometric deformations. It is robust togeometric attacks, but cannot distinguish images with similarhistograms but different contents. Tang et al. [3] develop aglobal method using nonnegative matrix factorization (NMF).The image is first converted into a fixed-sized pixel array.A secondary image is obtained by rearranging pixels andapplying NMF to produce a feature-bearing coefficient matrix,which is then coarsely quantized. The obtained binary string isscrambled to generate the image hash. Swaminathan et al. [4]propose an image hash method based on rotation invarianceof Fourier-Mellin transform and present a new framework tostudy the security issues of existing image hashing schemes.Their method is robust to geometric distortions, filtering oper-ations, and various content-preserving manipulations. In [5],Lei et al. calculate DFT of the invariant moments of significantRadon transform coefficients, and normalize/quantize the DFTcoefficients to form the image hash for content authentication.Khelifi et al. [6] propose a robust and secure hashing scheme

based on virtual watermark detection. The method is robustagainst normal image processing operations and geometrictransformation, and can detect content changes in relativelylarge areas. In another work, Monga et al. [7] apply NMFto pseudo-randomly selected subimages. They construct asecondary image, and obtain low-rank matrix approximation ofthe secondary image with NMF again. The matrix entries areconcatenated to form an NMF-NMF vector. The inner prod-ucts of the NMF-NMF vector and a set of weight vectors arecalculated. Because the final hash comes from the secondaryimage with NMF, their method cannot locate forged regions.In analyzing the NMF-NMF method, Fouad et al. [8] point outthat, among the three keys it uses, the first one for pseudo-ran-domly selecting several subimages is crucial. However, it canbe accurately estimated based on the observation of imagehash pairs when reused several times on different images. A

1556-6013/$31.00 © 2012 IEEE

Page 2: Robust Hashing for Image Authe

56 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 8, NO. 1, JANUARY 2013

lexicographical image hashing scheme has been proposed [9],in which a dictionary is constructed based on a large number ofimage feature vectors called words taken from various imageblocks. Words are used to represent image blocks in generatingthe hash. In [10], a wavelet-based image hashing method isdeveloped. The input image is partitioned into nonoverlappingblocks, and the pixels of each block are modulated using apermutation sequence. The image undergoes pixel shufflingand then wavelet transform. The subband wavelet coefficientsare used to form an intermediate hash, which is permuted againto generate the hash sequence. This method is robust to mostcontent-preserving operations and can detect tampered areas.Lv et al. [11] propose a SIFT-Harris detector to identify themost stable SIFT key points under various content-preservingoperations. The extracted local features are embedded intoshape-context-based descriptors to generate an image hash.The method is robust against geometric attacks and can be usedto detect image tampering. The performance is degraded whenthe detected key points from the test image do not coincidewith that of the original.Besides the aforementioned methods, a concept of forensic

hash for information assurance is proposed in [12]. A compactforensic hash is based on Radon transform and the scale spacetheory, and used as side information in image authentication. Itis aimed to address a wider range of issues than simply decidingwhether an image is a fake. These include telling the historyof image manipulations and estimating parameters of geomet-rical transformation. The geometrical transformation parame-ters allow image registration to be achieved without resortingto the original image so that the forged areas can be located.In another work of forensic hashing [13], SIFT features are en-coded into compact visual words to estimate geometric transfor-mations, and block-based features are used to detect and localizeimage tampering.In the present paper, we propose a method combining ad-

vantages of both global and local features. The objective is toprovide a reasonably short image hash with good performance,i.e., being perceptually robust while capable of detecting andlocating content forgery. We use Zernike moments of the lu-minance/chrominance components to reflect the image’s globalcharacteristics, and extract local texture features from salientregions in the image to represent contents in the correspondingareas. Distance metrics indicating the degree of similarity be-tween two hashes are defined to measure the hash performance.Two thresholds are used to decide whether a given imageis an original/normally-processed or maliciously doctoredversion of a reference image, or is simply a different image.The method can be used to locate tampered areas and tell thenature of tampering, e.g., replacement of objects or abnormalmodification of colors. Compared with some other methodsusing global features or local features alone, the proposedmethod has better overall performance in major specifications,especially the ability of distinguishing regional tampering fromcontent-preserving processing.In Section II, Zernike moments, salient region detection and

texture features are briefly introduced. Section III presents theproposed image hashing scheme, and describes the procedureof image authentication. Section IV gives experimental results

and analyzes performance of the method. Section V concludesthe paper.

II. BRIEF DESCRIPTION OF USEFUL TOOLS AND CONCEPTS

A. Zernike Moments

Zernike moments (ZM) of order and repetition of a dig-ital image are defined as [14]–[16]:

(1)

where is a Zernike polynomial of order and repe-tition [14]:

(2)

in which is even, andare real-valued radial polynomials. Suppose is a ro-

tation angle, and and the ZM of the original androtated images respectively:

(3)

Thus, the magnitude of ZM is rotation invariant, while the phasechanges with the angle:

(4)

B. Salient Region Detection

A salient region in an image is one that attracts visual atten-tion. According to [17], information in an image can be viewedas a sum of two parts: that of innovation and that of prior knowl-edge. The former is new and the latter redundant. The informa-tion of saliency is obtained when the redundant part is removed.Log spectrum of an image, , is used to represent general

information of the image. Because log spectra of different im-ages are similar, there exists redundant information in . Let

denote the redundant information defined as convolutionbetween and an low-pass kernel :

(5)

Spectral residual representing novelty of the image, , canbe obtained by subtracting from , which is then in-versely Fourier transformed to give a saliency map :

(6)

We choose a threshold equal to three times of the mean ofto determine the salient regions. For example, Fig. 1(a)

is an original image, (b) its saliency map, (c) the salient regionsafter thresholding, and (d) the image marked with four circum-scribed rectangles of the largest connected salient regions. Wewill extract the image’s local features from these rectangles.

C. Texture Features

Texture is an important feature to human visual perception. In[18] and [19], the authors propose six texture features relating tovisual perception: coarseness, contrast, directionality, line-like-ness, regularity and roughness. In this work, we use coarseness

Page 3: Robust Hashing for Image Authe

ZHAO et al.: ROBUST HASHING FOR IMAGE AUTHENTICATION USING ZERNIKE MOMENTS AND LOCAL FEATURES 57

Fig. 1. Salient region detection: (a) Original image. (b) Saliency map.(c) Salient region. (d) Four rectangles.

and contrast as defined below, plus skewness and kur-tosis, to describe the texture properties.To evaluate coarseness around a pixel at , the pixels in

its neighborhood sized are averaged:

(7)

where is the gray-level of pixel . At each point, differences between pairs of the average values of

nonoverlapping neighborhoods on opposite sides of the pixelin horizontal and vertical directions are:

(8)

For that point, find the size that leads to the highest differencevalue and call it .

(9)

Take average of over a region, and call it coarseness of thatregion, .Contrast describes the degree of image brightness variation,

calculated from variance and the fourth-order moment ofthe gray values within the region:

(10)

III. PROPOSED HASHING SCHEME

In this section, we describe the proposed image hashingscheme and the procedure of image authentication using thehash. The hash is formed from Zernike moments to representglobal properties of the image, and the texture features insalient regions to reflect local properties.

A. Image Hash Construction

The image hash generation procedure includes the followingsteps, referring to Fig. 2.1) Preprocessing: The image is first rescaled to a fixed size

with bilinear interpolation, and converted from RGB tothe YCbCr representation. Y and are used as lumi-nance and chrominance components of the image to generate thehash. The aim of rescaling is to ensure that the generated imagehash has a fixed length and the same computation complexity.

Fig. 2. Block diagram of the proposed image hashing method.

TABLE IZERNIKE MOMENTS OF DIFFERENT ORDERS

Small leads to loss of fine details, while large results inhigh computation complexity. We choose as an ap-propriate trade-off.2) Global Feature Extraction: Zernike moments of Y and

are calculated. Because shape features can be obtainedfrom a small number of low frequency coefficients, the orderdoes not need to be large. We choose . Further, since

, only is needed. We do notuse as it represents the average intensity. Table I lists theZernike moment features from order 1 to order 5. Thus we have

Zernike moments in total. Magnitudes of theZernike moments are rounded and used to form a global vector,

. Each element in is no more than 255. A secretkey is used to randomly generate a row vector with 22random integers in [0, 255]. The encrypted global vector isobtained as .3) Local Feature Extraction: largest salient regions are

detected from the luminance image Y. The coordinates of topleft corner, and width/height of each circumscribed rectangleare used to form a -element vector , rep-resenting the position and size of each salient region.From the statistics based on 3 000 images as shown in Fig. 3,

more than 97% of the images have no more than 6 salient re-gions. For those having more than 6 salient regions, the totalarea of the 7th and smaller salient regions are less than 1.5%of the image area. With a larger , fewer salient regions aremissing but will lead to a longer image hash. For example, thepercentage of images with no more than 7 salient regions is99.5%. We choose as a reasonable trade-off.

Page 4: Robust Hashing for Image Authe

58 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 8, NO. 1, JANUARY 2013

Fig. 3. Statistics of salient region numbers based on 3000 images.

TABLE IICONSTITUTION OF IMAGE HASH

Local texture features of each salient region including coarse-ness and contrast , skewness, and kurtosis are computedand rounded to give a 6-element vector .If an image has less than 6 salient regions, the positions andtexture features of the missing ones are set to zero. The posi-tion/size and texture vectors of all salient regions together forma local feature vector ,which contains 48 integers. A secret key is used to ran-domly generate a row vector containing 48 random inte-gers in [0, 255]. An encrypted local vector is then obtained by

.4) Hash Construction: The global and salient local vectors

are concatenated to form an intermediate hash, namely, which is then pseudo-randomly scrambled based on a

secret key to produce the final hash sequence . Table IIgives the constitution of an image hash of 70 integers. Sinceall integers are in the range of [0, 255], the hash is

bits long.

B. Image Authentication

In image authentication, the hash of a trusted image, , isavailable and called the reference hash. The hash of a receivedimage to be tested, , is extracted using the above method.These two hashes are compared to determine whether the testimage has the same contents as the trusted one or has been ma-liciously tampered, or is simply a different image. Here, two im-ages having the same contents (visual appearance) do not needto have identical pixel values. One of them, or both, may havebeen modified in normal image processing such as contrast en-hancement and lossy compression. In this case, we say the twoimages are perceptually the same, or similar. The image authen-tication process is performed in the following way.

1) Feature Extraction: Pass the test image through thesteps as described in Section A to obtain the intermediate hashwithout encryption, namely .2) Hash Decomposition: With the secret keys and, restore the intermediate hash from the reference hash to ob-

tain , which is a concatenated feature se-quence of the trusted image. Decompose it into global and localfeatures.3) Salient Region Matching: Check if the salient regions

found in of the test image match those in of the trustedimage. If the matched areas of a pair of regions are large enough,the two regions are considered as being matched. Reshuffle thetexture vectors by moving the matched components in each ofthe texture vector pair to the left-most and, for notational sim-plicity, still call them and . For example, if there are threesalient regions in the reference image and two in the test image,

The first two pairs of subvectors in and may either bematched or unmatched. The vectors and are reshuffledaccordingly.4) Distance Calculation and Judgment: We use a distance

between hashes of an image pair as a metric to judge simi-larity/dissimilarity of the two images. To define the hash dis-tance, a feature vector is formed by concatenating the globalfeature vector and the reshuffled texture feature vector ,namely . The vector does not contribute to thedistance calculation but will be used to locate forged regions.The hash distance between the test image and the reference isthe Euclidean distance between and :

(11)

For a pair of similar images, texture features in the corre-sponding salient regions are close to each other. However, sinceno currently available method of saliency detection is perfect,the salient regions obtained from an image after content-pre-serving processing may not always precisely match that of theoriginal. If this happens, difference between the test image andthe original will be exaggerated. In practice, the global struc-ture of an image represented by Zernike moments is sufficientto distinguish similar from dissimilar. To minimize the adverseinfluence of saliency detection inaccuracy, we omit in calcu-lating the hash distance for similar images:

(12)

The issue of saliency detection will be discussed later. Havingdefined the hash distance, we can use it first to distinguish sim-ilar and dissimilar images according to a threshold . If, the two images are said to be similar, otherwise the im-

ages are dissimilar. We then need to further determine whetherthe test image is a tampered version of the reference image, orsimply a different one. To do so, compare the distance with asecond thresholds . The test image is judged as tampered if

. Otherwise it is a completely different image.

Page 5: Robust Hashing for Image Authe

ZHAO et al.: ROBUST HASHING FOR IMAGE AUTHENTICATION USING ZERNIKE MOMENTS AND LOCAL FEATURES 59

C. Determination of Thresholds

To determine a threshold for differentiating two sets of data,A and B, we need to know the probability distribution functions(PDF) of samples taken from these data sets. The chi-square test[20] is used for the purpose. Assume that the data satisfy one ofseveral common distributions: Poisson, lognormal, and normal,and apply the chi-square test to find which is the closest. Thestatistic is calculated as

(13)

where is the number of trials, the frequency of a distancebeing , and is probability of the distancebeing calculated from the tested PDF. The PDF is the one thatproduces the smallest value. For simplicity, we choose thehorizontal coordinate of the two PDF curves’ intersection as thethreshold in the present work irrespective of the particular PDFshapes and the costs of different types of errors.To perform the chi-square test, a sufficiently large number

of test images are needed. For this purpose, we construct anoriginal image database, a similar image database, and aforged image database. The original image database contains1 000 different images downloaded from the internet andcaptured with digital cameras. A total of 10 000 similar imagesare obtained with ten content-preserving operations on eachoriginal image. The operations include gamma correction with

and 1.15, JPEG compression with and 80,zero-mean Gaussian noise contamination with and0.001, scaling with factors 0.5 and 1.5, and rotation by 1 and5 . The forged image database is generated by pasting a foreignblock into each image. The pasted area is 10% of the host. Theforged image database contains 7 000 images with various suchpasted blocks. These image databases will also be used forexperimental verification of the proposed method in Section IV.We now calculate hash distances of 1) 24 000 similar image

pairs taken from the original and similar image databases, 2) thesame number of forged images and their original versions, and3) 499 500 pairs of different images taken from the originalimage database. Chi-square tests show that the distance valuesbetween similar images follow the exponential distribution witha mean of 1.88. The distances between forged images and theiroriginal versions follow the lognormal distribution with a meanof 3.46 and standard deviation 0.88. Those between differentimages follow the gamma distribution with a mean of 10.3 andstandard deviation 8.16. Fig. 4 shows three curves of these prob-ability distribution functions. The curves of exponential and log-normal distributions intersect at (6.76, 0.014), and those of log-normal and gamma at (50.16, 0.0079). Therefore we choose

and as the thresholds in this work to differ-entiate similar, forged, and different images.

D. Forgery Classification and Localization

Having found that a test image is a fake, the next job is to lo-cate the forged region and tell the nature of forgery. Four types

Fig. 4. Distribution of hash distance.

Fig. 5. Forgery classification and localization when .

of image forgery can be identified: removal, insertion and re-placement of objects, and unusual color changes. Forgery classi-fication and localization are performed as follows, and schemat-ically illustrated in Fig. 5.Decode and into components representing global and

local features, and find the number of matched salient regionsand the numbers of salient regions in the reference and test

images, and .1) If , some objects have been removed fromthe received test image. Positions of the missing objects arelocated by comparing the saliency indices.

2) If , the test image contains some addi-tional objects whose positions are located by comparingthe saliency indices.

3) If , check the luminance and chrominancecomponents in the Zernike moments and calculate the fol-lowing distances:

If is greater than by a threshold , the test imagecontains substantial color changes with respect to the ref-erence image while the luminance changes are consider-ably smaller. Thus the test image is judged as being tam-pered with unusual color modifications. The thresholdis chosen as 5 based on experiments of 50 image pairs.

4) If and is less than , thetest image contains replaced objects because in this caseluminance changes are dominant. Calculate the distance

Page 6: Robust Hashing for Image Authe

60 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 8, NO. 1, JANUARY 2013

TABLE IIICOMPARISON OF HASH PERFORMANCE

Fig. 6. Distances of 1000 hash pairs of similar images.

between the texture feature vectors of each salient region,. The -th salient region

having maximal is recognized as being replaced.5) If and , some of the salient regions arenot matched. Mark the mismatched salient regions in thetest image as being tampered.

IV. PERFORMANCE STUDIES

A. Robustness and Anti-Collision Test

Fig. 6 shows hash distances between 1 000 different imagesand the images after normal processing manipulations includinggamma correction with , JPEG coding with ,zero mean Gaussian noise addition with , rotation by5 , scaling with factors 0.5 and 1.5, and slight cropping with 2%of the image width/height removed. We observe that more than98.5% of all distances are less than . The exceptions cor-respond to gamma correction, zero mean Gaussian noise addi-tion, scaling of 1.5, and cropping. In these cases image intensitychanges may have affected the saliency map.If two different images have a hash distance less than , col-

lision occurs. Most published methods have good anticollisionperformance. In our case, integration of the gamma distributionPDF from negative infinity to gives a very low collision prob-ability of . Anticollision performance is important,

in particular, to the application of content-based image retrieval(CBIR).

B. Forgery Detection Capability

As shown in Fig. 4, the proposed method can differentiatesimilar (i.e., perceptually the same), forged, and different im-ages. A qualitative comparison between the proposed methodand [3], [6], [7], and [10] is given in Table III, showing goodoverall performance of the proposed method, which generateshashes with the third shortest length, is robust against severalnormal processing manipulations, and can detect and locatesmall area tampering. Not included in the table is the abilityof resisting large area cropping and large-angle rotation sincethe proposed method is not designed to be robust against suchoperations. Cropping that changes the image’s geometricalcenter will cause significant differences in the Zernike mo-ments, and large-angle rotation will affect the saliency-relatedlocal features. These are limitations of the method. Nonethe-less, removal of a few lines (no more than 2% of the imagewidth/height) and small-angle rotation are tolerable as will beshown in the ROC performance below. As far as image rotationis concerned, a picture will clearly appear abnormal from thephotographical point of view if it is rotated by, say, more than5 . Such rotation may be viewed as a destructive modification,viz., tampering.Performance of the proposed method is basically due to the

combination of global and local features. The former reflect theoverall structure of the image, and the latter, based on contentsaliency, are used to identify and characterize changes in visu-ally important regions. In [3], features are extracted from a sec-ondary image composed of image blocks, while in [10] eachhash element corresponds to a nonoverlapping block. They arenot robust against rotation and cropping since these operationswill change the blocks. The hash of [7] is robust against slightcropping but not to rotation even if the angle is small becauseit is based on pseudo-randomly selected subimages, which ischanged after rotation so that the hash will also be changed. Themethods of [3] and [7] are not designed to localize forgery. Themethod proposed in [6] is robust against geometric attacks as thefeatures are extracted from overlapped blocks of a high-passedversion of the image, but it cannot detect small area forgery.We now evaluate hash performance in terms of forgery detec-

tion errors. It is based on experiments on 1 000 uncompressed

Page 7: Robust Hashing for Image Authe

ZHAO et al.: ROBUST HASHING FOR IMAGE AUTHENTICATION USING ZERNIKE MOMENTS AND LOCAL FEATURES 61

Fig. 7. ROC performance. (a) ROC curve. (b) Zoomed-in view of the ROCcurve.

images taken from the original image database, 1 000 imagesprocessed with content-preserving operations taken from thesimilar image database, and 1 000 images from the forged imagedatabase. Fig. 7(a) shows curves of receiver operating charac-teristics (ROC) corresponding to six types of content-preservingprocessing: gamma correction, JPEG coding, Gaussian noiseaddition, rotation, scaling, and slight cropping. Fig. 7(b) is azoomed-in view of the circle-marked part of (a). It is observedthat the hash has good ability in distinguishing these operationsfrom regional forgery. Note that, in calculating ROC, the falsenegative (FN) and false positive (FP) errors are errors in differ-entiating between similar and forged images rather than betweensimilar and different images. The error probabilities are definedas in (14) and (15), shown at the bottom of the page.Fig. 8 compares the ROC curves of the proposed method and

those of [3] and [7], obtained using the same set of images asin Fig. 7. The proposed method has shown stronger ability todistinguish content-preserving operations from regional forgerythan the other two methods. All three methods can distinguishforgery from normal operations such as JPEG coding, additivenoise and scaling because these only cause tiny changes to theimage contents. As stated in the above, the method of [3] cannotresist rotation and cropping, and [7] is not robust against small-angle rotation. Comparable ROC curves of [6] and [10] are notavailable as [6] provides examples of tampering in relativelylarge areas, and the hash distance in [10] is given in a matrixform rather than a scalar quantity.

C. Forgery Localization

We have tested 150 image pairs, with the original imagesdownloaded from the Internet and the forged ones producedmanually using Photoshop. The forged images are all correctlydetected.Without considering forgery classification, the successrate of forgery localization is 96%. The success rate drops to87% when both localization and classification are considered.The latter is lower because saliency detection is imperfect sothat accuracy of forgery classification may be affected.

Fig. 8. Comparison of ROC curves with the methods of [3] and [7] in termsof various image processing operations: (a) Noise addition and JPEG coding.(b) Gamma correction and rotation. (c) Scaling and cropping.

A few examples of forgery detection are shown in Table IVand Fig. 9. The table gives original and forged images, salientregions and detection results. In the third column from the left,the solid purple rectangles indicate matched salient regions; thedashed blue rectangles are salient regions that only exist in theforged images but not in the original; and the dash-dotted yellowrectangles show regions where salient regions exist in the orig-inal but disappear after forgery. The rectangles in the fourthcolumn indicate the detected forged regions. In the table, thefirst two rows are examples of object removal, the third andfourth rows object insertion, and the last two rows object re-placement.Fig. 9 shows examples of unusual color changes, in which

(a) and (c) are original images, and (b) and (d) forged ones.In the first example, hash distance , indicating thetest image is tampered. As is much smaller than

, there exists abnormal color change. Similarly, inthe second example, hash distance is large enoughto be judged as being tampered, and is significantlysmaller than , indicating unusual color change.

D. Computation Complexity

To compare computation complexity of the differentmethods, we consider average time consumed in calculatingimage hashes on a desktop computer with Dual Core 2.8-GHzCPU and 2 GB RAM, running Matlab. The average time of theproposed method is 2.7 s, and those of [3] and [7] are 2.98 sand 1.43 s respectively. These are not significantly different.

(14)

(15)

Page 8: Robust Hashing for Image Authe

62 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 8, NO. 1, JANUARY 2013

TABLE IVDISTANCE BETWEEN ORIGINAL AND FORGED IMAGES

Fig. 9. Examples of unusual color changes. (a) Original. (b) Forged. (c) Orig-inal. (d) Forged.

According to the experimental data given in [6], their methodis much faster than [7].Computation complexity is not mentioned in [10]. We may

conjecture that the wavelet transform does not take much time,and the major computation load is in determining the decisionthreshold based on the maximum error matrix by consideringvarious image processing operations.

V. CONCLUSION

In this paper, an image hashing method is developed usingboth global and local features. The global features are based onZernike moments representing the luminance and chrominancecharacteristics of the image as a whole. The local features in-clude position and texture information of salient regions in theimage.Hashes produced with the proposed method are robust

against common image processing operations including bright-ness adjustment, scaling, small angle rotation, JPEG codingand noise contamination. Collision probability between hashesof different images is very low. The proposed scheme has areasonably short hash length and good ROC performance.The method described in this paper is aimed at image authen-

tication. The hash can be used to differentiate similar, forged,and different images. At the same time, it can also identify thetype of forgery and locate fake regions containing salient con-tents. In the image authentication, a hash of a test image is gen-erated and compared with a reference hash previously extractedfrom a trusted image. When the hash distance is greater than thethreshold but less than , the received image is judged as afake. By decomposing the hashes, the nature of image forgeryand locations of forged areas can be determined.It should be stressed that the success of image authentica-

tion using the proposed scheme depends to a large extent on the

Page 9: Robust Hashing for Image Authe

ZHAO et al.: ROBUST HASHING FOR IMAGE AUTHENTICATION USING ZERNIKE MOMENTS AND LOCAL FEATURES 63

accuracy of saliency detection. The method proposed in [17]is used due to its acceptable accuracy and computation com-plexity. One can always incorporate a better saliency detectionscheme, whenever it is available, into the algorithm for im-proved performance. Further study is desired to find featuresthat better represent the image contents so as to enhance thehash’s sensitivity to small area tampering while maintainingshort hash length and good robustness against normal imageprocessing.

REFERENCES[1] V. Monga, A. Banerjee, and B. L. Evans, “A clustering based approach

to perceptual image hashing,” IEEE Trans. Inf. Forensics Security, vol.1, no. 1, pp. 68–79, Mar. 2006.

[2] S. Xiang, H. J. Kim, and J. Huang, “Histogram-based image hashingscheme robust against geometric deformations,” in Proc. ACM Multi-media and Security Workshop, New York, 2007, pp. 121–128.

[3] Z. Tang, S.Wang, X. Zhang,W.Wei, and S. Su, “Robust image hashingfor tamper detection using non-negative matrix factorization,” J. Ubiq-uitous Convergence Technol., vol. 2, no. 1, pp. 18–26, May 2008.

[4] A. Swaminathan, Y. Mao, and M. Wu, “Robust and secure imagehashing,” IEEE Trans. Inf. Forensics Security, vol. 1, no. 2, pp.215–230, Jun. 2006.

[5] Y. Lei, Y.Wang, and J. Huang, “Robust image hash in Radon transformdomain for authentication,” Signal Process.: Image Commun., vol. 26,no. 6, pp. 280–288, 2011.

[6] F. Khelifi and J. Jiang, “Perceptual image hashing based on virtual wa-termark detection,” IEEE Trans. Image Process., vol. 19, no. 4, pp.981–994, Apr. 2010.

[7] V. Monga and M. K. Mihcak, “Robust and secure image hashing vianon-negative matrix factorizations,” IEEE Trans. Inf. Forensics Secu-rity, vol. 2, no. 3, pp. 376–390, Sep. 2007.

[8] K. Fouad and J. Jianmin, “Analysis of the security of perceptual imagehashing based on non-negative matrix factorization,” IEEE SignalProcess. Lett., vol. 17, no. 1, pp. 43–46, Jan. 2010.

[9] Z. Tang, S. Wang, X. Zhang, W. Wei, and Y. Zhao, “Lexicograph-ical framework for image hashing with implementation based on DCTand NMF,”Multimedia Tools Applicat., vol. 52, no. 2–3, pp. 325–345,2011.

[10] F. Ahmed, M. Y. Siyal, and V. U. Abbas, “A secure and robust hash-based scheme for image authentication,” Signal Process., vol. 90, no.5, pp. 1456–1470, 2010.

[11] X. Lv and Z. J. Wang, “Perceptual image hashing based on shape con-texts and local feature points,” IEEE Trans. Inf. Forensics Security, vol.7, no. 3, pp. 1081–1093, Jun. 2012.

[12] W. Lu, A. L. Varna, and M. Wu, “Forensic hash for multimedia infor-mation,” in Proc. SPIE,Media Forensics and Security II, San Jose, CA,Jan. 2010, 7541.

[13] W. Lu and M. Wu, “Multimedia forensic hash based on visual words,”in Proc. IEEE Conf. on Image Processing, Hong Kong, 2010, pp.989–992.

[14] H. Lin, J. Si, and G. P. Abousleman, “Orthogonal rotation-invariantmoments for digital image processing,” IEEE Trans. Image Process.,vol. 17, no. 3, pp. 272–282, Jan. 2008.

[15] S. Li, M. C. Lee, and C. M. Pun, “Complex Zernike moments featuresfor shape-based image retrieval,” IEEE Trans. Syst., Man, Cybern. A,Syst. Humans, vol. 39, no. 1, pp. 227–237, Jan. 2009.

[16] Z. Chen and S. K. Sun, “A Zernike moment phase based descriptorfor local image representation and matching,” IEEE Trans. ImageProcess., vol. 19, no. 1, pp. 205–219, Jan. 2010.

[17] X. Hou and L. Zhang, “Saliency detection: A spectral residual ap-proach,” in Proc. IEEE Int. Conf. Computer Vision and Pattern Recog-nition, Minneapolis, MN, 2007, pp. 1–8.

[18] T. Deselaers, D. Keysers, and H. Ney, “Features for image retrieval: Aquantitative comparison,” in Lecture Notes in Computer Science, 2004,vol. 3175, pp. 228–236, Springer.

[19] H. Tamura, S. Mori, and T. Yamawaki, “Textural features corre-sponding to visual perception,” IEEE Trans. Syst., Man, Cybern., vol.8, no. 6, pp. 460–472, Jun. 1978.

[20] E. Lehmann and J. Romano, Testing Statistical Hypotheses, 3rd ed.New York: Springer, 2005, pp. 590–599.

Yan Zhao received the B.S. degree from Xi’an Jiao-tong University, China, in 1999, and the M.E. degreefrom Shanghai Jiaotong University, China, in 2005.She is currently working toward the Ph.D. degree atShanghai University, China.She is also a lecturer with Shanghai University of

Electric Power, China. Her research interests includeimage processing and digital forensics.

Shuozhong Wang received the B.S. degree fromPeking University, Beijing, China, in 1966, andthe Ph.D. degree from University of Birmingham,Birmingham, U.K., in 1982.Currently, he is a Professor with Shanghai Univer-

sity, Shanghai, China. His research interests includeimage processing, audio processing, and informationhiding.

Xinpeng Zhang received the B.S. degree in compu-tational mathematics from Jilin University, China, in1995, and the M.E. and Ph.D. degrees in communica-tion and information system from Shanghai Univer-sity, China, in 2001 and 2004, respectively.Since 2004, he has been with the faculty of the

School of Communication and Information Engi-neering, Shanghai University, where he is currentlya Professor. He was with the State University ofNew York at Binghamton as a visiting scholar fromJanuary 2010 to January 2011, and with University

of Konstanz as an experienced researcher sponsored by the Alexander vonHumboldt Foundation from May 2011 to August 2012. His research interestsinclude multimedia security, image processing, and digital forensics. He haspublished more than 170 papers in these areas.

Heng Yao (S’11–M’12) received the B.S. degreefrom Hefei University of Technology, China, in2004, the M.E. degree from Shanghai NormalUniversity, China, in 2008, and the Ph.D. degreefrom Shanghai University, China, in 2012.Currently, he is with School of Optical-Electrical

and Computer Engineering, University of Shanghaifor Science and Technology, China. His researchinterests include digital forensics, image processing,and pattern recognition.