 Research
 Open access
 Published:
A contextadaptive SPN predictor for trustworthy source camera identification
EURASIP Journal on Image and Video Processing volumeÂ 2014, ArticleÂ number:Â 19 (2014)
Abstract
Sensor pattern noise (SPN) has been recognized as a reliable device fingerprint for camera source identification (CSI) and image origin verification. However, the SPN extracted from a single image can be contaminated largely by image content details from scene because, for example, an image edge can be much stronger than SPN and hard to be separated. So, the identification performance is heavily dependent upon the purity of the estimated SPN. In this paper, we propose an effective SPN predictor based on eightneighbor contextadaptive interpolation algorithm to suppress the effect of image scene and propose a source camera identification method with it to enhance the receiver operating characteristic (ROC) performance of CSI. Experimental results on different image databases and on different sizes of images show that our proposed method has the best ROC performance among all of the existing CSI schemes, as well as the best performance in resisting mild JPEG compression, especially when the falsepositive rate is held low. Because trustworthy CSI must often be performed at low falsepositive rates, these results demonstrate that our proposed technique is better suited for use in realworld scenarios than existing techniques. However, our proposed method needs many such as not less than 100 original images to create camera fingerprint; the advantage of the proposed method decreases when the camera fingerprint is created with less original images.
1. Introduction
Digital images are easy to modify and edit via imageediting software. Image content becomes unbelievable. Using this kind of forged image should be avoided as evidence in a court of law, as news, as part of a medical record, or as financial documents. There are some works focused on image component forensics in recent years [1â€“3]. The work in [3] first proposed using the imaging sensor pattern noise (SPN) to trace back the imaging device and solve the camera source identification (CSI) problem. They extracted SPN from wavelet highfrequency coefficients using the waveletbased denoising filter [4]. A camera reference SPN is built by averaging residual noise from multiple images taken by the same camera. In [5], an innovative and recently introduced denoising filter, namely, a sparse 3D transformdomain collaborative filtering (BM3D) [6], is used to extract the SPN. This filter is based on an enhanced sparse representation in a transform domain. A maximum likelihood method is proposed in [7] to estimate the camera reference SPN. It will be named the MLE CSI method for short in this paper. Later, [8] proposed a more stable detection statistic, the peaktocorrelation energy measure (PCE), to suppress periodic noise contamination and enhance CSI performance. The authors of [9] proposed a forgerydetection method using SPN to determine if an image is tampered. Li [10] demonstrated that the SPN extracted from a single image can be contaminated by image scene details and proposed some models to attenuate the strong signal component of noise residue. However, attenuating strong components from scene details may also attenuate the useful SPN components [11]. Kang et al. [11] proposed a detection statistic correlation over circular correlation norm (CCN) to lower the falsepositive rate and a whitecamera reference SPN to enhance the ROC performance [12]. The noise residues extracted from the original images are whitened first and then averaged to generate the whitecamera phase reference SPN. We call it the phase CSI method for short in the rest of this paper.
Although there have been some prior studies dedicated to improving the performance of CSI based on SPN in recent years, an effective method to eliminate the contamination of the image scene details is still lacking. In order to reduce the impact of scene details while preserving SPN at the same time, an edgeadaptive SPN predictor based on a fourneighbor contextadaptive interpolation (PCAI4) [13] was proposed and has been proved to have improvement on CSI performance via extensive experiments. This paper is an extension work of our conference paper [13]. Because the method PCAI4 only predicts the center pixel from its fourneighboring pixels, in this paper, we will extend this method by making use of all the eightneighboring pixels and propose an edgeadaptive SPN predictor based on eightneighbor contextadaptive interpolating prediction, as well as a CSI method with this advanced predictor. We have also conducted extensive experiments on different image datasets and reported new results in this paper. Thanks to its adaptability to image edge and context, the predicted SPN is much purer and performs better for CSI. The experimental results on different image databases show that our proposed method can achieve the best ROC performance among all of the existing CSI schemes on different sizes of images and has the best performance in resisting mild JPEG compression.
The rest of this paper is organized as follows. In Section II, we will first introduce our contextadaptive interpolating prediction algorithm. Then, an eightneighbor SPN predictor is proposed to improve the CSI performance. In Section III, we evaluate the performance of our proposed algorithm and compare its performance with stateoftheart CSI methods on different image databases. The conclusion of this paper is made in Section IV.
2. Advance SPN predictor based on adaptive interpolation
2.1 Contextadaptive interpolator
The contextadaptive interpolation (CAI) method predicts a center pixel from its fourneighbor pixels. We will call it the â€˜CAI4â€™ in this paper. The SPN predictor using CAI4 [13] is based on the CAI [14] interpolation algorithm which is adapted from the gradientadaptive predictor (GAP) [15]. In the CAI4 method, the local regions are classified into four types: smooth, horizontally edged, vertically edged, and other. A mean filter is used to estimate the centerpixel value in smooth region; in edged regions, the center pixel is predicted along the edge. In other regions, a median filter is applied. Taking p to be a centerpixel value to be predicted, and tâ€‰=â€‰[n,â€‰s,â€‰e,â€‰w]^{T} to be a vector of its fourneighboring pixels as in Figure 1, the predicted pixel value \stackrel{\xcc\u201a}{\mathit{p}} using CAI4 method can be formulated as
In (1), a smooth region will never be estimated as the edged region, and the interpolation prediction in the edged regions are adapted from the GAP [15]. The center pixel is predicted according to different types of edge regions, which is classified by the fourneighboring pixel values with an empirical threshold. The threshold has little impact on the experimental results and set to be 20 according to the former work [15].
2.2 Extending CAI4 to CAI8
The CAI4 method only predicts the center pixel from its fourneighbor pixels because it is proposed as an adaptive interpolation algorithm and is not aware of the other four diagonal pixels. As we are using it to predict SPN knowing all the neighbor pixels in Figure 1, we can extend and enhance the CAI4 method by making use of all the eightneighboring pixels. We call this method â€˜CAI8â€™ in short form.
In CAI8 method, the local regions are classified into six types: smooth, horizontally edged, vertically edged, leftdiagonal edge, rightdiagonal edge, and others. In the smooth region, a mean filter is used to estimate the center pixel from the eightneighboring pixels; in the horizontal and vertical edge regions, the centerpixel value is predicted along the edge as the same as CAI4. In the diagonaledge region, the centerpixel value is also estimated along the corresponding edge; in other regions, a median filter is applied. Taking pâ€² to be the centerpixel value to be predicted by CAI8, tâ€²â€‰=â€‰[n,â€‰s,â€‰e,â€‰w,â€‰en,â€‰es,â€‰wn,â€‰ws]^{T} to be a vector of its eightneighboring pixels as shown in Figure 1, then the predicted pixel value {\hat{\mathit{p}}}^{\text{'}} using the CAI8 method can be formulated as follows:
In (2), the centerpixel value is predicted along different directions of the edge, including in the diagonally edged region which is ignored by CAI4. So, the predicted result can suppress the interference of image edge better and has less prediction error.
2.3 Source camera identification with SPN predictor based on CAI8
SPN can be contaminated largely by the image scene, especially in the texture regions. Method CAI8 can predict a centerpixel value accurately in allusion to different local regions because it is adaptive to image edge and local context. So, the difference between the predicted value and actual value can suppress the impact of image edge better while preserving the SPN components at the same time.
Let yâ€‰=â€‰{y_{ i }  iâ€‰=â€‰0, 1, â€¦, N1} be the camera reference SPN, and xâ€‰=â€‰{x_{ i }} be the noise residue extracted from a test image. For the null hypothesis, y is not the correct camera reference SPN of the noise residue x extracted from a test image, i. e., the test image is not taken by the reference camera. In other words, x is a negative sample for y. For the affirmative hypothesis, y is the correct camera reference SPN of the noise residue x extracted from a test image, i.e., the test image is taken by the reference camera. In other words, x is a positive sample for y.
In the following, we will propose a contextadaptive SPN predictor based on CAI8, which is called PCAI8 in short form, and a source camera identification method with PCAI8.

(1)
Firstly, we take the difference D of the predicted value and actual value,
\mathbf{D}=\mathbf{I}\mathrm{CAI}\left(\mathbf{I}\right),(3)
where CAI(â‹…) means the pixelwise CAI8 prediction as shown in Equation 2.

(2)
In order to further eliminate the impact of the image scene and extract a more accurate camera reference SPN, we then perform a pixelwise adaptive Wiener filter based on the statistics estimated from the neighborhood of each pixel, assuming that the SPN is a white Gaussian signal corrupted by image content. For each pixel (i, j), the optimal predictor for the estimated SPN is
\mathbf{W}\left(\mathit{i},\mathit{j}\right)=\mathbf{D}\left(\mathit{i},\mathit{j}\right)\frac{{\mathit{\xcf\u0192}}_{0}^{2}}{{\widehat{\mathit{\xcf\u0192}}}^{2}\left(\mathit{i},\mathit{j}\right)+{\mathit{\xcf\u0192}}_{0}^{2}},(4)
where {\hat{\mathrm{\xcf\u0192}}}^{2} represents the estimated local variance for the original noisefree image, and {\mathrm{\xcf\u0192}}_{0}^{2} represents the overall variance of the additive white Gaussian noise (AWGN) signal, i.e., the SPN here. To a large extent, the performance of the predictor depends on the accuracy of the estimated local variance. We use the maximum a posteriori probability (MAP) estimation to estimate the local variance as following:
where m is the size of a neighborhood N_{ m } for each pixel. Here, we take mâ€‰=â€‰3. The overall variance of the SPN {\mathrm{\xcf\u0192}}_{0}^{2} is also unknown. The detailed discussion of the choice of the parameter {\mathrm{\xcf\u0192}}_{0}^{2} can be found in [3]; the authors of [3] found that the choice of the parameter {\mathrm{\xcf\u0192}}_{0}^{2} has little impact on the experimental results, and our experiments also verified this point. We follow the work in [3] and use {\mathrm{\xcf\u0192}}_{0}^{2}=9 in all experiments to make sure that the predictor extracts a relatively consistent level of the SPN.
Our proposed SPN predictor PCAI8 is adaptive to different image edge regions according to all eightneighbor pixels, and the PCAI8 method is more accurate than PCAI4 in classifying edge's area, so it is expected that the predicted SPN has less scene noise from the original image than PCAI4 and other denoising filters.

(3)
The estimated camera reference SPN y' is obtained by averaging all the residual noise W _{ k } {W _{ k }(i, j)} (the estimated SPN from each image) extracted from the same camera as follows:
{\mathbf{y}}^{\mathbf{\xe2\u20ac\xb2}}=\frac{{\displaystyle \underset{\mathit{k}=0}{\overset{\mathit{L}1}{\mathrm{\xe2\u02c6\u2018}}}{\mathbf{W}}_{\mathit{k}}}}{\mathit{L}},(6)
where L denotes the total number of images used for the extraction of camera reference SPN. The residual noise W_{ k }(i, j) is extracted pixelwise according to Equation 4.

(4)
In order to further suppress the unwanted artifacts caused by camera processing operations such as color interpolation and JPEG compression blocking artifacts, we adopt two preprocessing operations proposed in [7] to enhance the estimated SPN before it is used for identification. So, the final estimated camera reference SPN y can be expressed as
\mathbf{y}=\mathit{WF}\phantom{\rule{0.12em}{0ex}}\left(\mathit{ZM}\left(\mathit{y}\text{'}\right)\right),(7)
where the ZM(â‹…) operation makes y' to have zero mean in every row and column, and the WF(â‹…) operation makes ZM(y') to have a flat frequency spectrum using the Wiener filter in Fourier domain.

(5)
Finally, calculate the detection statistic c(x, y) between the camera reference SPN y and the noise residue x extracted from a test image with Equation 4. We use the detection statistic CCN to measure the similarity between the image noise residue x and a camera's reference SPN y. We use CCN instead of PCE [8] because it can lower the falsepositive rate at the same truepositive rate (please refer to [11] for details). The CCN value c(x, y) is defined as:
\mathit{c}\left(\mathbf{x}\mathbf{,}\mathbf{y}\right)=\frac{\mathbf{xy}/\mathit{N}}{\sqrt{\frac{1}{\mathit{N}\left\mathbf{A}\right}{\displaystyle \underset{\mathit{m}\xe2\u02c6\u2030\mathbf{A}}{\mathrm{\xe2\u02c6\u2018}}{\mathit{r}}_{\mathbf{xy}}^{2}\left(\mathit{m}\right)}}}=\frac{{\mathit{r}}_{\mathbf{xy}}\left(0\right)}{\sqrt{\frac{1}{\mathit{N}\left\mathbf{A}\right}{\displaystyle \underset{\mathit{m}\xe2\u02c6\u2030\mathbf{A}}{\mathrm{\xe2\u02c6\u2018}}{\mathit{r}}_{\mathbf{xy}}^{2}\left(\mathit{m}\right)}}}(8)
where A is a small neighbor area around zero where {\mathit{r}}_{\mathbf{xy}}\left(0\right)=\frac{1}{\mathit{N}}\mathbf{xy}=\frac{1}{\mathit{N}}{\displaystyle \underset{\mathit{i}=0}{\overset{\mathit{N}1}{\mathrm{\xe2\u02c6\u2018}}}{\mathit{x}}_{\mathit{i}}}{\mathit{y}}_{\mathit{i}}, and Î‘ is the size of A. The size of A is chosen to be a block of 11â€‰Ã—â€‰11 pixels. The circular shift vector y_{ m }â€‰=â€‰{y_{iâŠ•m}}, where the operationâ€‰âŠ•â€‰is modulo N addition in â„¤_{ N }. The circular crosscorrelation r_{ xy }(m) is defined as
In the next section, we will evaluate the CSI performance of our proposed method.
3. Experimental results
In this section, we will compare the CSI performance of the proposed PCAI8 method with the existing stateoftheart methods on two different image databases. In â€˜Part Aâ€™ section, an image database built by ourselves is used. In this database, blue sky images can be used to extract more accurate reference patterns. In â€˜Part Bâ€™ section, we use a public image database, the â€˜Dresden Image Databaseâ€™ (DID) [16], which can be downloaded from the internet [17]. Cameras in this image database cover different camera brands or models and different devices of the same camera model. We choose two of Li's models, â€˜model 3â€™ and â€˜model 5', in our experimental comparison because they show better results according to Li's work [10]. Furthermore, all model parameters are chosen the same as those in Li's work, and we use model 3 or model 5 to denote the image noise residue attenuated by model 3 or model 5 in our results. As a result, we compare our PCAI8 method with the MLE method from [7], BM3D method [5], PCAI4 method [13], phase method [11], and Li's method [10] (i.e., model 3 and model 5).
The CSI experiments are performed on the image block with different sizes cropped from the center of the fullsize images. Our experiments are performed in the luminance channel of all images because the luminance channel contains information of all the three RGB channels. In fact, experiments in the other channel are also performed and have similar results.
The detection statistic CCN is used to measure the similarity between the image noise residue x and a camera's reference SPN y for all methods. In order to make a fair comparison, before the calculation of detection statistic, for all four methods, we performed the same preprocessing operations as shown in (7) on the estimated reference PRNU/SPN y before the calculation of detection statistic. The experiments on different image databases demonstrate that our method always has the best performance among all existing methods regardless of using CCN, PCE, or correlation as a detection statistic. So, we report the experimental results with detection statistic CCN to measure the similarity between the image noise residue x and a camera's reference SPN y for all methods.
3.1 Part A
On the first image database, we use seven different cameras in our experiments. Table 1 shows the image format, native resolution, and imaging sensor property of the cameras (PS means PowerShot). All images are in JPEG format with the highest JPEG quality factor provided by the cameras, except in raw data format for the Nikon D40 (Shanghai, China) and Minolta A2 (Konica, Tokyo, Japan). For each camera, we have two subimage datasets which are the test image dataset and original image dataset, respectively. The original image dataset is used for camera reference SPN extraction. It has been proved that a more accurate camera reference SPN can be extracted by using blue sky images [7]. So, the original images are taken on a sunny day of the blue sky whose content is flat or near flat. The test images are taken under a variety of environments, from indoor furniture to outdoor sight. The images in the test image dataset are used as test samples for CSI. The CSI experiment is performed on the image block with different sizes from 128â€‰Ã—â€‰128 to 512â€‰Ã—â€‰512. The image block is cropped from the center of a fullsize photo.
For each chosen camera, we extract the camera reference SPN using Lâ€‰=â€‰100 images from the original image dataset, 200 test images of this camera are selected as the positive samples, and 1,200 test images of the other six cameras (each camera is responsible for 200) are selected as the negative samples. All the test images are chosen randomly from the test image dataset. Totally, we get 200 positive and 1,200 negative samples of CCN values for each chosen camera.
To obtain the overall ROC curve, for a given detection threshold, we count the number of truepositive decisions and the number of falsepositive decisions for each camera and then sum them up to obtain the total number of truepositive decisions and falsepositive decisions. Then, the total truepositive rate (TPR) and total falsepositive rate (FPR) are calculated to draw the overall ROC curve.
The overall ROC curve performances of our proposed PCAI method compared with other SPN CSI methods are shown in Figures 2, 3 and 4. In practical applications, it is often necessary to ensure a sufficiently low FPR; therefore, the ROC performance in low FPR case is more critical. So, the horizontal axis of all the ROC curves in this paper is in logarithmic scale, in order to show the detail of the ROC curves with a low FPR.
The experimental results show that the proposed PCAI8 method outperforms the others and enhances the ROC performance of CSI for images of different sizes. The proposed PCAI8 method, the PCAI4 method, and the phase SPN method can achieve a 100% TPR at a low FPR on an image block of 512â€‰Ã—â€‰512 pixels in our experimental environment. From Figures 2, 3 and 4, we also notice that both PCAI methods, including PCAI4 and PCAI8, achieve better ROC performance than other methods because of the SPN predictor PCAI has less scene noise residue. Compared to PCAI4, PCAI8 always achieves better performance than PCAI4, which means that the PCAI8method can suppress the scene noise better than PCAI4.
Table 2 shows the TPR of the different methods at a low FPR of 10^{3}. From the table data, we find that the TPR of the proposed method is always the largest regardless of the image size. The experimental results indicate that the proposed method raises the TPR prominently in the case of trustworthy identification which is with a low FPR. For example, on small image block size of 256â€‰Ã—â€‰256, the TPR of our proposed PCAI8 method is 99.3%, the TPR of the MLE, phase, PCAI4, BM3D, model 3, and model 5 methods is 96.8%, 98%, 98.6%, 92.2%, 97.4%, and 96.9%, respectively. The improvement is 2.5%, 1.3%, 0.7%, 7.1%, 1.9%, and 2.5%, respectively.
When an image is JPEGcompressed, the SPN is impaired at the same time, so it becomes more difficult to use SPN for CSI. Figure 5 shows the overall ROC curves performance on JPEGcompressed images of 512â€‰Ã—â€‰512 pixels, with a quality factor (QF) of 90%. The number of test images is the same as that mentioned above. The results with the other sizes are not shown here because they are also similar. The experimental results show that the proposed PCAI8 method also has the best performance in resisting mild JPEG compression and achieves perfect detection.
Although camera fingerprint can be created with as much as possible original images, sometime we cannot have as much as 100 original images for camera fingerprint creation. So, we also investigate the performance when camera fingerprint is extracted using less than 100, e.g., 30, original images from the original image dataset; the other setup is the same as Figure 3. It is observed from Figure 6 that the advantage of the proposed PCAI decreases when the camera fingerprint is extracted using only 30 original images, but it still achieves similar performance as the stateoftheart MLE method.
3.2 Part B
In this part, we report the experimental results on 3,320 images of 17 cameras from the Dresden Image Database. This image database contains some images with some special shooting environment and setting, such as a high ISO value which results in high shooting noises. It makes the CSI challenging on this image database. The 17 camera devices belong to four camera brands or models. Each camera model has 3 to 5 different camera devices. The different camera devices with the same camera model have the same incamera processing, such as JPEG compress and color filter array (CFA) interpolation. Table 3 shows the information of each device. Device ID is the unique identification for each camera device. Image no. denotes the number of images in the camera devices, and the resolution is the native resolution of the camera devices.
Most settings of the experiments in this part are similar with the ones in â€˜Part Aâ€™ section. We use the luminance channel of all the images to extract sensor pattern noises of test images and reference SPN of each camera device. All the image blocks are of three sizes (i.e., 128â€‰Ã—â€‰128, 256â€‰Ã—â€‰256, and 512â€‰Ã—â€‰512 pixels) and are all cropped from the center of fullsize images. In this image database, exactly blue sky images are not available. All the images are ordinary scene pictures in daily life. There are about 200 images of each camera device (Table 3).
In our experiments, we use the fivefold crossvalidation method. Assume that one database contains Nâ€‰Ã—â€‰K images taken by N cameras; each camera is responsible for K images. Firstly, we divide the images of each camera device into five groups averagely. In each fold, we randomly choose one group as the test image dataset (about K/5 images for each camera), and the other four groups as original images dataset (about Kâ€‰Ã—â€‰4/5 images for each camera). The original image dataset is used for extracting the camera reference SPN, and images from the test image dataset may be used as positive test samples or negative test samples. For each chosen camera, we extract the camera reference SPN using its original image dataset; the test images (about K/5 images) of this camera are selected as the positive samples, and the test images of the other Nâ€‰â€‰1 cameras (each camera is responsible for K/5 images) are selected as the negative samples. So, we get K/5 CCN values of positive samples and K/5â€‰Ã—â€‰(Nâ€‰â€‰1) CCN values of negative samples for each chosen camera. After five folds, totally, we get K CCN values of positive samples and Kâ€‰Ã—â€‰(Nâ€‰â€‰1) CCN values of negative samples for each camera. At last, the overall ROC curve is obtained in a similar way as mentioned in â€˜Part Aâ€™ section.
An obvious characteristic of this database is that some camera devices belong to the same brand. Most of the previous works, including the experiments in â€˜Part Aâ€™ section, only considered different camera brands. It might lead to a problem that we cannot make a clear division of camera source identification and camera model identification because the extracted SPN might contain part of camera model noises, which could be regarded as fingerprints of a special camera model. These noises play different roles in experiments dependent on the models of tested cameras. So, if all the tested cameras come from different camera brands, the SPN with more camera model noises might give a better performance than the more accurate one which is with less camera model noises in it. And, the results of such experiments are not very reliable when different camera devices of the same camera model are considered.
In order to make the experiments more convincing, we first compare the performance between our method and other methods in the same camera brand. Figure 7 shows the overall ROC curve performance on images of five camera devices (device ID: C0 to C4) in Casio_EXZ150. We use the fivefold cross validation method in this experiment. Only the results on 512â€‰Ã—â€‰512 sizes are showed since the results in other sizes are similar.
The experimental results show that our proposed method has the best performance in identifying the source of images taken by the same camera brand and model. The proposed method can achieve a high TPR of 97% at a low FPR of 10^{3} for images with size of 512â€‰Ã—â€‰512, which means that only few images are misjudged.
In the following, we report the CSI experimental results on the whole DID database. In plotting the overall ROC curves on all the images in the DID database, we totally get 3,320 CCN values of positive samples and 53,120 CCN values of negative samples. The results with three different image sizes are shown in Figures 8, 9 and 10.
The experimental results also show that both PCAI8 and PCAI4 have better performance than the other methods in identifying images of different source camera models regardless of different image sizes. Table 4 shows the TPR of the different methods at a low FPR of 10^{3}. It shows that the TPR of the proposed PCAI8 method is always the largest at a low FPR. For example, on small image block of size 128â€‰Ã—â€‰128, the TPR of the PCAI8 method is 46.2%, the TPR of the MLE, phase, PCAI4, BM3D, model 3, and model 5 methods is 39.1%, 37.7%, 42.3%, 37.7%, 32.2%, and 26.8%, respectively. The improvement is 7.1%, 9.5%, 3.9%, 9.5%, 14.0%, and 19.4%, respectively. The performance of PCIA8 achieves little better than that of PCAI4.
The experimental results in both â€˜Part Aâ€™ and â€˜Part Bâ€™ sections show that the propose method achieves better performance for CSI whether the influence of camera model is considered or not. In â€˜Part Aâ€™ section, we compare all methods on seven cameras with different camera models in our image database. In â€˜Part Bâ€™ section, we test all methods on five camera devices with the same model and also test all methods on 17 camera devices with the same model or different models. All the experiments on images with different sizes show that our proposed method has the best ROC performance among all of the existing CSI schemes.
The computation time to get the noise residue x from a test image of each method with IntelÂ® (Santa Clara, CA, USA) XeonÂ®CPU E52603 1.80 GHz and Matlab (MathWorks, Bangalore, India) is shown in Table 5. It is observed that both PCAI4 and PCAI8 methods have the best efficiency.
4 Conclusion
In this paper, we propose a source camera identification scheme based on an eightneighbor contextadaptive SPN predictor to enhance the ROC performance of CSI. The SPN predictor can suppress the effect of image content better and lead to a more accurate SPN estimation because of its adaptability of different image edge regions. Extensive experiment results on different image databases and on different sizes of images show that our proposed PCAI method achieves the best ROC performance among all of the stateoftheart CSI schemes and also has the best performance in resisting mild JPEG compression (e.g., with a quality factor of 90%) simultaneously, especially when the falsepositive rate is held low (e.g., P_{fp}â€‰=â€‰10^{3}). Because trustworthy CSI must often be performed at low falsepositive rates, these results demonstrate that our proposed technique is better suited for use in realworld scenarios than existing techniques. However, our proposed method needs many such as not less than 100 original images to create a camera fingerprint; the advantage of the proposed method decreases when the camera fingerprint is created with less original images.
References
Swaminathan A, Wu M, Liu KJR: Nonintrusive component forensics of visual sensors using output images. IEEE. T. Inf. Foren. Sec. 2007, 2(1):91106.
Swaminathan A, Wu M, Liu KJR: Digital image forensics via intrinsic fingerprints. IEEE. T. Inf. Foren. Sec. 2008, 3(1):101117.
LukÃ¡Å¡ J, Fridrich J, Goljan M: Digital camera identification from sensor pattern noise. IEEE. T. Inf. Foren. Sec. 2006, 1(2):205214. 10.1109/TIFS.2006.873602
Mihcak MK, Kozintsev I, Ramchandran K: Spatially adaptive statistical modeling of wavelet image coefficients and its application to denoising. Paper presented at the IEEE international conference on acoustics, speech, and signal processing, vol. 6. Phoenix, AR, USA; May1999:32533256.
Cortiana A, Conotter V, Boato G, de Natale FGB: Performance comparison of denoising filters for source camera identification. Paper presented at the SPIE conference on media watermarking, security, and forensics III, vol. 7880. San Jose, CA, USA; Jan. 2011:778007.
Dabov K, Foi A, Katkovnik V, Egiazarian K: Image denoising by sparse 3D transformdomain collaborative filtering. IEEE Trans. Image Process 2007, 16(8):20802095.
Chen M, Fridrich J, Goljan M, LukÃ¡Å¡ J: Determining image origin and integrity using sensor noise. IEEE. T. Inf. Foren. Sec. 2008, 3(1):7490.
Goljan M: Digital camera identification from images  estimating false acceptance probability. Paper presented at the international workshop on digitalforensics and watermarking. Busan, Korea: LNCS 5450; Dec. 2008:454468.
Fridrich J, Chen M, Goljan M: Imaging sensor noise as digital Xray for revealing forgeries. Paper presented at the 9th international workshop on information hiding. Saint Malo, France; July 2007:342358.
Li CT: Source camera identification using enhanced sensor pattern noise. IEEE. T. Inf. Foren. Sec. 2010, 5(2):280287.
Kang X, Li Y, Qu Z, Huang J: Enhancing source camera identification performance with a camera reference phase sensor pattern noise. IEEE. T. Inf. Foren. Sec. 2012, 7(2):393402.
Kang X, Li Y, Qu Z, Huang J: Enhancing ROC performance of trustworthy camera source identification. Paper presented at the SPIE conference on electronic imagingmedia watermarking, security, forensics XIII, vol. 7880. San Francisco, CA, USA; Jan. 2011:78800178800109.
Wu G, Kang X, Liu KJR: A context adaptive predictor of sensor pattern noise for camera source identification. Paper presented at the 19th international conference on image processing. Orlando, FL, USA; 15â€“20 Sept. 2012:237240.
Liu W, Zeng W, Dong L, Yao Q: Efficient compression of encrypted grayscale images. IEEE Trans. Image Process 2010, 19(4):10971102.
Wu X, Memon N: Contextbased adaptive lossless image coding. IEEE. T. Commun 1997, 45(4):437444. 10.1109/26.585919
Gloe T, BÃ¶hme R: Proceedings of the 25th Symposium on Applied Computing, vol.2. Springer, New York; 2010:15851591.
Dresden Image Database Dresden: Technische Universitaet Dresden; 2009â€“2014.http://forensics.inf.tudresden.de/ddimgdb . Accessed 3 May 2013
Acknowledgements
This work was supported by NSFC (grant nos. 61379155 and U1135001), 973 Program (grant no. 2011CB302204), the Research Fund for the Doctoral Program of Higher Education of China (grant no. 20110171110042), NSF of Guangdong Province (grant no.s2013020012788), and the National Science & Technology Pillar Program (grant no. 2012BAK16B06).
Author information
Authors and Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that we have no competing interests.
Authorsâ€™ original submitted files for images
Below are the links to the authorsâ€™ original submitted files for images.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Kang, X., Chen, J., Lin, K. et al. A contextadaptive SPN predictor for trustworthy source camera identification. J Image Video Proc 2014, 19 (2014). https://doi.org/10.1186/16875281201419
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/16875281201419