
Citation: Gang Cao, Antao Zhou, Xianglin Huang, Gege Song, Lifang Yang, Yonggui Zhu. Resampling detection of recompressed images via dual-stream convolutional neural network[J]. Mathematical Biosciences and Engineering, 2019, 16(5): 5022-5040. doi: 10.3934/mbe.2019253
[1] | Bin Zhang, Linkun Sun, Yingjie Song, Weiping Shao, Yan Guo, Fang Yuan . DeepFireNet: A real-time video fire detection method based on multi-feature fusion. Mathematical Biosciences and Engineering, 2020, 17(6): 7804-7818. doi: 10.3934/mbe.2020397 |
[2] | Pengcheng Cao, Weiwei Liu, Guangjie Liu, Jiangtao Zhai, Xiaopeng Ji, Yuewei Dai . Steganographic coding scheme based on dither convolutional trellis under resampling mechanism. Mathematical Biosciences and Engineering, 2019, 16(5): 6015-6033. doi: 10.3934/mbe.2019301 |
[3] | Jun Gao, Qian Jiang, Bo Zhou, Daozheng Chen . Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview. Mathematical Biosciences and Engineering, 2019, 16(6): 6536-6561. doi: 10.3934/mbe.2019326 |
[4] | Xinyi Wang, He Wang, Shaozhang Niu, Jiwei Zhang . Detection and localization of image forgeries using improved mask regional convolutional neural network. Mathematical Biosciences and Engineering, 2019, 16(5): 4581-4593. doi: 10.3934/mbe.2019229 |
[5] | Yufeng Li, Chengcheng Liu, Weiping Zhao, Yufeng Huang . Multi-spectral remote sensing images feature coverage classification based on improved convolutional neural network. Mathematical Biosciences and Engineering, 2020, 17(5): 4443-4456. doi: 10.3934/mbe.2020245 |
[6] | Guanghua Fu, Qingjuan Wei, Yongsheng Yang . Bearing fault diagnosis with parallel CNN and LSTM. Mathematical Biosciences and Engineering, 2024, 21(2): 2385-2406. doi: 10.3934/mbe.2024105 |
[7] | Tao Zhang, Hao Zhang, Ran Wang, Yunda Wu . A new JPEG image steganalysis technique combining rich model features and convolutional neural networks. Mathematical Biosciences and Engineering, 2019, 16(5): 4069-4081. doi: 10.3934/mbe.2019201 |
[8] | Bo Wang, Yabin Li, Xue Sui, Ming Li, Yanqing Guo . Joint statistics matching for camera model identification of recompressed images. Mathematical Biosciences and Engineering, 2019, 16(5): 5041-5061. doi: 10.3934/mbe.2019254 |
[9] | Qi Cui, Ruohan Meng, Zhili Zhou, Xingming Sun, Kaiwen Zhu . An anti-forensic scheme on computer graphic images and natural images using generative adversarial networks. Mathematical Biosciences and Engineering, 2019, 16(5): 4923-4935. doi: 10.3934/mbe.2019248 |
[10] | Xi Lu, Zejun You, Miaomiao Sun, Jing Wu, Zhihong Zhang . Breast cancer mitotic cell detection using cascade convolutional neural network with U-Net. Mathematical Biosciences and Engineering, 2021, 18(1): 673-695. doi: 10.3934/mbe.2021036 |
Digital images play an important part in delivering information, leading public opinion and taking evidence in legal proceedings and criminal investigations. Photo editing and operation tools make progress in creating forged images without leaving any perceptible artifacts. The malicious image manipulation may lead to serious ethical and legal problems. Therefore, it is significant to detect the image manipulation blindly [1].
Image manipulation generally destroys some inherent consistency within original images, or leaves new unique operation trail. Such clues could be exploited to detect manipulation blindly by forensic analysts. In the last decade, numerous methods have been developed to detect various types of image manipulations, such as JPEG compression [2,3], contrast enhancement [4,5,6], sharpening [7], median filtering [8], image splicing [9,10,11,12,13,14,15] and computer graphic rendering [16].
Blind detection of image resampling has received extensive attention. Existing image resampling detection methods operate by identifying operation traces from spatial [17,18,19,20,21,22] and frequency domains [23,24]. Popescu et al. propose to detect the statistical correlation of interpolated pixels by expectation maximization (EM) algorithm [17]. Mahdian and Saic propose to detect pixel interpolation traces by capturing the statistical changes on signal covariance structure [21]. In [19], subspace decomposition and random matrix theory are used to identify resampling traces in upscaled images. Feng et al. exploit normalized energy density to derive a 19-dimensional feature vector, which is fed into a SVM classifier [23]. In addition, Kirchner et al. [24] model the detectable resampling artifacts in spatial and frequency domains by means of the variance of prediction residue. A semi-intrusive forensic method is proposed to identify the category of resampling operators in [25]. Pasquini et al. [26] analyze the detectability of downsampling in theory.
However, such prior resampling detection methods are rather fragile with post-JPEG compression, even for the quality factor Q around 95 [15,32]. The periodic pattern introduced by resampling is disturbed by JPEG blocking artifacts. In order to detect resampling in JPEG images, the periodicity in the second derivative of interpolated images is employed as evidence of resampling detection [27]. Furthermore, Nataraj et al. [28] propose to suppress JPEG periodic patterns by adding Gaussian noise to post-compressed images. However, such prior methods are ineffective in low quality JPEG images. Kirchner and Gloe [29] demonstrate that the previous JPEG compression before resampling can well detect resampling in recompressed images. Although such an approach can detect resampling in the cases of low and moderate quality primary compression, it is invalidated for the high quality case. In [30], the operation chain including JPEG compression and resampling is detected by the transformed block artifacts extracted from pixel and discrete cosine transforms (DCT) domains. Li et al. propose a data-driven resampling detection method based on improved spatial rich model (SRM) features [31].
Convolutional neural network (CNN) is an important deep learning algorithm widely used in computer vision. A constrained convolutional neural network has been proposed for detecting image resampling in the prior works [32,33]. Such a method achieves the best performance in resampling detection compared with previous methods.
In this paper, we propose a new CNN framework to detect resampling in recompressed images. A noise extraction layer consisting of high pass filters is inserted into the CNN architecture. An input image can be converted into noise residuals by such filters for suppressing the influence of image contents. Simultaneously, a novel dual-stream CNN is presented to extract the high level feature of resampling traces. Note that resampling is often enforced locally to adjust region dimensions for creating realistic spliced images. Additionally, the proposed dual-stream CNN method is applied to the detection of such resampling-involved image splicing. Extensive experimental results verify that the proposed resampling detection method can achieve the state-of-the-art performance.
In this work, we have made the following major contributions:
● A noise extraction layer is constructed by using two low-order high pass filters to suppress image content and highlight operational features.
● A dual-stream convolutional neural network is proposed based on the methodology of two streams feature fusion, which could automatically learn the resampling features in images.
The rest of the paper is organized as follows. Section 2 describes the proposed resampling detection scheme in detail. Experimental results and discussions are presented in Section 3, followed by the conclusion drawn in Section 4.
In this section, the proposed image resampling detection scheme is presented in detail. Firstly, we describe the overall framework of the network in subsection 2.1, and then introduce the design motivation and function of each module in subsections 2.2–2.4, respectively. Finally, the specific functions of each layer in the dual-stream CNN framework are analyzed in subsection 2.5.
We address resampling detection as a pattern classification problem, which would be resolved by deep learning methodology. Figure 1 depicts the overall design of our proposed dual-stream CNN architecture for resampling detection. It consists of three main components. First, in order to capture different dimensional features, the adjacent pixels differences from both horizontal and vertical directions are extracted by a noise extraction layer. Second, the high level representation of image manipulation features are generated by horizontal and vertical streams. In order to capture the correlation of both directions, horizontal and vertical streams are interleaved and concatenated. Lastly, the concatenated features of horizontal and vertical streams are fed into the sigmoid/softmax layer. Note that the sigmoid layer is used for binary classification, and the softmax layer is used for multiple classification. Below we present a detailed introduction of such three components and the different layers used in our CNN architecture.
Resampling detection is typically interfered by image content which should be suppressed. It has been found that the resampling trace generally exists in the redundant domain of spatial-domain images and is irrelevant to image content [14,34,35,36,37,38]. In our approach, noise is modeled by the residual between a pixel and its estimation yielded by interpolating only neighboring pixels. To accomplish this, a special convolution layer is set as the first layer of the network, namely noise extraction layer. As shown in Figure 2, two low-order highpass filters are chosen as convolution kernels of noise extraction layers, which would not need to be trained cautiously. The adjacent pixels differences in horizontal and vertical directions can be captured by the filters shown in Figures 2(a) and (b), respectively. In particular, a grayscale image of 256×256 pixels is first convolved with 1×3 and 3×1 filters with stride 1 and padding 1. Such filters could learn noise features from windows through the correlation of local pixels. As a result, the noise extraction layers would yield a noise map of prediction residuals with the dimension of 256×256×1.
Noise from horizontal and vertical directions is extracted through the noise extraction layer. Then high-level representation of resampling trace would be extracted from such noise. Experimental observations show that a part of evidence would be lost and the overall resampling detection performance would degrade if a single, instead of double, direction(s) of noise gradient was used. Therefore, two identical pipelines (horizontal stream and vertical stream) are designed to extract features independently from different directions, and the weights of two pipelines are not shared. As shown in Figure 1, the horizontal and vertical streams are composed of five similar groups. Each group consists of four layers including the convolution, batch normalization, activation and pooling ones. The fifth group receives additional features from the interleaved stream. Finally, the features generated by the horizontal/vertical and interleaved streams are concatenated.
In order to determine the resampling behavior, it requires to fuse the two directional features for making a decision together. Therefore, we propose a novel feature fusion strategy of interleaved stream, which consists of four similar groups. The first group consists of a convolutional layer, a batch normalization layer and an activation layer. The remaining groups have an additional pooling layer. The features from the first group in horizontal and vertical streams are concatenated, and then enforced by 1×1 convolutional kernels with stride 1. Such a type of convolution kernel weights each position of the two feature graphs linearly for fusion. Then the remaining three groups further extract the high-level representation of fused features. Finally, the feature map output from interleaved stream is interpolated back to the horizontal and vertical streams. As cooperative learning, the horizontal and vertical streams are aware of each other without affecting the feature extraction in each direction.
The final features learned from the previous layers are fed into the classifier, which is a fully connected layer using softmax or sigmoid functions. Through such a classifier, the probability that the feature belongs to each category is obtained, and the most probable category is the result of the classifier. The two involved functions are specified by
P(y=1|x)=11+e−z | (1) |
P(y=j|x)=ezj∑Kk=0ezk | (2) |
where Eq. (1) is a sigmoid function used in binary classification and z is the value of neurons in the full connected layer. P(y=1|x) is the probability that x belongs to the positive category. Eq. (2) is a softmax function used in multiple classification, where zj is the value of the jth neuron in the full connected layer. P(y=j|x) belongs to the probability that x is the jth category.
The convolution layer in CNN is used to extract features. The input feature maps in convolution layers are computed as
F(n)j=K∑k=0F(n−1)k∗ω(n)kj+b(n)j | (3) |
where ∗ indicates a 2d convolution operation, F(n−1)k signifies the kth feature map generated in the (n−1)th layer, ω(n)kj indicates the k channel of the jth convolution kernel in the nth layer, b(n)j is the jth bias in the nth layer and F(n)j is the jth feature map generated in the nth layer.
Three convolution kernels with different sizes (1×1, 3×3 and 5×5) are used in the proposed model. In horizontal/vertical stream, the convolution layers "H/V1~5 " are with 64, 64, 64, 64,128 filters of size 5×5×1, 3×3×64, 5×5×64, 3×3×64 and 3×3×128, respectively. The output dimensions of such convolutional layers are 256×256×64, 127×127×64, 63×63×64, 31×31×64, 15×15×64, respectively. In interleaved stream, the convolution layers "I1~4" are with 64 filters of size 1×1×128, 5×5×64, 5×5×64, 3×3×64, respectively. Their output dimensions are 127×127×64, 127×127×64, 63×63×64, 31×31×64, respectively. The stride is set as 1 in all convolution layers.
In order to solve the change of the data distribution in the middle layer during the training of CNN, we need to normalize the feature maps generated by the convolutional layers. To do this, a batch normalization layer is used between the convolution layer and the activation layer. The batch normalization operation within the CNN architecture is given from Eqs. (4) to (7).
First, the mean and variance of all data in a batch are calculated as
μ=1mm∑i=0xi | (4) |
σ2=1mm∑i=0(xi−μ)2 | (5) |
where m denotes the number of data in a batch, xi signifies the ith data in a batch, μ and σ2 indicate the mean and variance, respectively. Then each data is normalized to generate a new data ˆxi with a mean of 0 and a variance of 1. That is,
ˆxi=xi−μ√σ2+ϵ | (6) |
where ϵ is small floating-point number greater than 0 to prevent dividing by zero errors. Finally, all data is scaled and shifted as
yi=γˆxi+β | (7) |
where yi indicates the ith output of the batch normalization layer. γ and β are the parameters learned by a network.
The newly generated data can make better use of the non-linear function of the activation function. Huge change in back layer parameters caused by small change in front layer parameters is prevented by batch normalization layer.
In order to compensate for the expressive deficiency of linear models, the convolution layer is generally followed by an activation layer containing a non-linear function called activation function. The features generated by convolution layer are transformed into another space by activation function, and the data can be better classified. Note that TanH, Sigmoid, and ReLu are widely applied activation functions. The data is squashed to [-1, 1] by TanH function. TanH has a good performance for the features with significant difference, because it can enlarge continuously the effect of features. Besides, TanH is preferred to Sigmoid in practical application due to TanH's mean value is 0. Although the training speed of ReLu function is faster than that of TanH, it is fragile in training and can not use a larger learning rate. Hence, activation layers are equipped with the TanH function in our network.
The pooling layer aims to downsample the feature map. It reduces the number of elements in feature maps, and imports hierarchical structure by increasing the observation window of continuous convolution layers. In our method, max pooling and global average pooling are used. The max pooling which outputs the maximum value in each window of the input feature map is applied to all pooling layers except the fifth one in horizontal/vertical streams. Max pooling layers use kernels of size 3×3 and stride of 2, which is the smallest one to capture the notion of left/right, up/down and center. Global average pooling downsamples the feature maps to 1 via the average pooling and it can replace full connection layer to reduce model parameters. Note that global average pooling is only used in the last pooling layer of horizontal and vertical streams.
In this section, we first introduce the image datasets and the image processing tool used in our experiments. Then the related parameter settings of our proposed model are described in detail. Finally, extensive experiments are conducted to evaluate the effectiveness and efficacy of our proposed resampling detection scheme. Source code of our algorithm is publicly available on the Github *.
*https://github.com/zhouantao/Resampling/
In following experiments, the image database published by ALASKA steganography challenge is used as testing data †. Such a dataset contains 50000 RAW images of size 3154×5286 taken by several different cameras. The main image contents include animals, plants, landscapes and buildings, etc.. We use the first 5000 images taken by Canon-EOS100D camera. Each image is divided into subimages of 512×512 pixels, and seven central subimages are retained. Then JPEG compression is performed using Q = 95-97 for simulating the default compression setting in cameras. As such, 35000 subimages are generated as our original image dataset. The datasets used in all sub-experiments would be generated from such an original dataset. All operations are performed on RGB color images for simulating real image processing. The operations are implemented by the built-in image processing toolbox of MATLAB software.
To measure the performance of resampling detection methods, accuracy is employed as evaluation metrics. Our model yields a category label for each image and compares it with the real one. The accuracy is defined as the ratio of correctly classified images' number to the total number of images.
All models in our experiments are implemented using Keras with a single NVIDIA Titan GPU. The network is supervised training using labeled data, and the network weight is initialized by Xavier. The green channel images are fed into the network. We train CNN model using stochastic gradient descent with a batch size of 32 images, momentum [39] of 0.9, and weight decay of 1e-5. The learning rate is initialized at 0.01. Besides, step decay of learning rate is used in our method. We gradually reduce the learning rate during training. We divide the learning rate by 10 when the validation loss stop improving for three epochs. The number of training epochs is 30 epochs.
In this subsection, we set up two different sets of experiments on the detection of resampling in the recompressed image. JPEG is a lossy compression algorithm which can destroy the resampling trace. Figure 3 shows the results when an uncompressed image has undergone JPEG compression with Q = 90, 70, 50, respectively. As shown in Figure 3, more image details including the resampling trace are lost and destroyed with the decrease of Q. For Q = 50, the blocking artifacts becomes rather obvious. We first evaluate the performance of our model by specifying the resampling parameter and JPEG quality factors. Then we challenge a difficult task that the resampling parameters and Q are chosen randomly. The performance of the proposed method on different sizes of images is also evaluated.
In this subsection, we test our proposed method in the recompressed images with different scaling factors and JPEG compression quality factors. We first create a dataset without resampling. We crop the 256×256 central area of each image in the original dataset, and then compress it using Q = 50-90 with interval of 5. Second, the images in the original dataset are bilinearly scaled by 50%, 120% and 150%, and their central 256×256 block image is cut and recompressed with different Qs. As such, 27 sub-databases are generated in terms of the scaling and Qs, and each consists of 35000 unresampled images and corresponding resampled versions. Each database is divided into three parts: 50000 images for training, 10000 images for validation and the rest for testing.
In order to estimate the performance of the proposed model, such a model is compared with constrained convolutional neural networks [32,33] using both the same training and testing datasets. The comparison results are illustrated in Figure 4. The proposed CNN is superior to constrained convolutional neural networks. Our scheme achieves higher detection accuracy in the scenarios of mild upscaling and low quality compression. It can be noted from Figure 4 that our method is similar to Bayar's method [32,33] in performance with high quality post-compression. However, under the post-compression of medium and low quality factors, our method shows a clear advantage. In the case of a QF of 50 and an upsampling of 120, our method improves performance by nearly 4% compared to Bayar's method; when the downsampling is 50, the performance of our method is also improved by nearly 4%.
In the previous experiments, one certain scaling factor and quality factor are only included in each of databases. Next, we assess the performance of our approach at performing image resampling detection in a more complex scenario where Q and the resampling parameters are chosen arbitrarily. Additionally, the effectiveness of our method is also evaluated under different sizes of images. The original images are first scaled bilinearly with a random scaling factor. The range of upscaling and downscaling factors is respectively from 110% to 200% and from 50% to 90%, and the interval is 10. Then we retain the central blocks of the resampled images and their original version with different dimensions, i.e., 256×256, 128×128, 64×64. After this, each of the image blocks is compressed using the Q selected arbitrarily in the range of 50-100 with the interval of 10.
Finally, three databases with different dimensions are created, where each database corresponds to three sub databases, i.e., upsampling-JPEG, downsampling-JPEG and unaltered-JPEG. Each sub database consists of 35,000 images in which 25,000 images are used for training, 5,000 images are used for validation, and the rest part is used for testing. We compare the results of our proposed method and Bayar's method [32,33] under images with different dimensions, as shown in Table 1. The experimental results demonstrate that our model can accurately extract resampling features from images with different resampling factors. Our approach increases the overall classification rate from 88.46% to 99.07%. Although the accuracy of our method in 64×64 downscaled images is only 88.46%, it is still far higher than that of Bayar's method. We can notice that the Bayar's method performance degrades severely when reducing the dimension of image.
Image size | 256×256 | 128×128 | 64×64 | |||
Resampling type | Upscaling | Downscaling | Upscaling | Downscaling | Upscaling | Downscaling |
Bayar's method [32,33] | 97.29 | 97.24 | 95.19 | 91.93 | 88.19 | 80.03 |
Proposed method | 99.07 | 98.45 | 97.06 | 95.47 | 92.49 | 88.46 |
In order to verify whether our model can maintain excellent performance in more complex scenarios, we used our CNN to detect image resampling when one manipulation is applied for resampled images, followed by JPEG recompression. We design another set of experiments where each image is first resampled, then it is edited by one manipulations, and finally compressed using different Qs. Five image processing methods and parameters are displayed in Table 2, and they are Gamma correction (GC), Mean filtering (MeanF), Gaussian filtering (GF), Median filtering (MedF) and Wiener filtering(WF).
Operation type | Parameters |
Gamma correction(GC) | γ: 0.5, 0.6, 0.7, 0.8, 0.9, 1.2, 1.4, 1.6, 1.8, 2.0 |
Mean filtering(MeanF) | window size: 3×3, 5×5, 7×7 |
Gaussian filtering(GF) | window size: 3×3, 5×5, 7×7, σ: 0.8-1.6 |
Median filtering(MedF) | window size: 3×3, 5×5, 7×7 |
Wiener filtering(WF) | window size: 3×3, 5×5, 7×7 |
The original images are first resampled with a random scaling factor within 110%-200%, 50%-90% for upscaling and downscaling, respectively. The range of upsampling and downsampling factors is respectively from 110% to 200% and from 50% to 90%, and the interval is 10. Then the resampled image blocks and their corresponding unaltered versions are processed by five operations listed in Table 2, and the operation parameters are selected randomly and uniformly from the set. After this, each manipulated image block is compressed using the selected Q arbitrarily in the range of 50-100 with the interval of 10 and then the central 256×256 block is retained. Finally, five databases are built which are GC, MeanF, GF, MedF and WF.
Table 3 depicts the detection accuracy for the proposed method and Bayar's method [32,33] when one manipulation is applied for resampled images. From Table 3, the performances of our and Bayar's methods have different extent of reduction against different post-processing operations. Noticeably, our proposed method is superior to Bayar's method. For upsampled images, median filtering has a greater impact on the detection results of our model with only 87.90% accuracy, while gamma correction has a smaller impact and with an accuracy rate of 97.85%. Both values have significantly higher than 81.89% and 96.50% of Bayar's method. Although the influence of post-processing operation on downsampling is greater than that on upsampling, our proposed method results are much better than the results of Bayar's method. The detection rates of our approach for each type of manipulation are typically greater than 77% except for the MeanF operation images which is detected with an accuracy of 73.88%. Such experimental results show that our scheme can still detect whether the JPEG image with post-processing has been resampled.
Post-processing type | GC | MeanF | GF | MedF | WF | |||||
Resampling type | Up | Down | Up | Down | Up | Down | Up | Down | Up | Down |
Bayar's method [32,33] | 96.50 | 95.46 | 81.89 | 67.03 | 85.32 | 73.41 | 82.39 | 67.91 | 84.81 | 71.14 |
Proposed method | 97.85 | 97.37 | 86.69 | 73.88 | 89.14 | 84.26 | 87.90 | 77.12 | 90.45 | 84.46 |
In the above experiments, the proposed method can effectively detect the resampling features in the recompressed images with different scaling factors and JPEG quality factors. In this subsection, in order to test the detection capability via resampling parameters, a multiclass classifier is performed for the blind estimation of scaling factors.
Similarly to the previous set of experiments, 35000 image blocks of size 512×512 from the original dataset are used to build the database. The images are respectively sampled using bilinear interpolation with fifteen different scaling factors, i.e., 50%, 60%, 70%, 80%, 110%, 120%, 130%, 140%, 150%, 160%, 170%, 180%, 190% and 200%. And then the central 256×256 blocks of the sampled images are retained to form the final database. Next the generated blocks of size 256×256 pixels are compressed using the Q randomly selected in the range of 50–100 at the interval of 10. Eventually, 15 different databases are created. 25000, 5000, 5000 image samples are used for the model training, validation and testing, respectively. Table 4 shows the confusion matrix achieved by the proposed CNN. Actuals belong on the side of the confusion matrix and predictions are across the top. The accuracies for the proposed CNN with different scaling factors are greater than 91% and the scaling factor of 200% performs the best results of 99.58%. The images with some scaling factors, such as 60% and 70%, seem to be difficult to distinguish from those with the scaling factor 50%.
Tru╲Pre | 50% | 60% | 70% | 80% | 90% | 110% | 120% | 130% | 140% | 150% | 160% | 170% | 180% | 190% | 200% |
50% | 91.56 | 4.32 | 2.54 | 0.76 | 0.4 | 0.14 | 0.02 | 0.1 | 0.02 | 0 | 0.04 | 0 | 0.04 | 0.06 | 0 |
60% | 4.04 | 93.2 | 0.96 | 0.9 | 0.26 | 0.18 | 0.08 | 0.04 | 0.04 | 0.12 | 0.02 | 0.04 | 0.04 | 0.08 | 0 |
70% | 2.3 | 1.5 | 93.72 | 1.28 | 0.7 | 0.1 | 0.1 | 0.06 | 0.08 | 0.02 | 0.02 | 0.02 | 0.06 | 0.02 | 0.02 |
80% | 0.62 | 0.86 | 1.5 | 95.46 | 0.98 | 0.32 | 0.02 | 0.02 | 0.1 | 0.04 | 0.02 | 0.04 | 0 | 0.02 | 0 |
90% | 0.36 | 0.3 | 0.84 | 0.88 | 96.9 | 0.26 | 0.12 | 0.12 | 0.08 | 0.02 | 0.02 | 0.02 | 0.02 | 0.06 | 0 |
110% | 0.08 | 0.3 | 0.16 | 0.46 | 0.46 | 97.12 | 0.56 | 0.26 | 0.14 | 0.14 | 0.02 | 0.06 | 0.06 | 0.14 | 0.04 |
120% | 0 | 0 | 0.1 | 0.14 | 0.06 | 0.58 | 98.38 | 0.38 | 0.06 | 0.1 | 0.04 | 0.06 | 0.04 | 0.04 | 0.02 |
130% | 0.08 | 0.02 | 0.02 | 0 | 0.06 | 0.14 | 0.24 | 98.86 | 0.1 | 0.16 | 0.08 | 0.08 | 0.1 | 0.04 | 0.02 |
140% | 0.06 | 0.02 | 0.08 | 0.18 | 0.08 | 0.18 | 0.12 | 0.42 | 97.58 | 0.46 | 0.24 | 0.34 | 0.1 | 0.06 | 0.08 |
150% | 0.04 | 0.08 | 0.02 | 0 | 0.06 | 0.08 | 0.08 | 0.1 | 0.24 | 98.64 | 0.08 | 0.32 | 0.06 | 0.14 | 0.06 |
160% | 0 | 0 | 0 | 0.02 | 0 | 0.02 | 0.06 | 0.06 | 0.14 | 0.14 | 98.92 | 0.06 | 0.36 | 0.12 | 0.1 |
170% | 0.04 | 0.02 | 0.02 | 0.02 | 0 | 0.06 | 0.02 | 0.02 | 0.08 | 0.18 | 0.12 | 98.96 | 0.12 | 0.24 | 0.1 |
180% | 0.02 | 0.02 | 0 | 0.04 | 0.04 | 0 | 0.06 | 0.04 | 0.06 | 0 | 0.34 | 0.06 | 99 | 0.18 | 0.14 |
190% | 0.02 | 0 | 0 | 0.02 | 0.02 | 0.02 | 0.02 | 0.08 | 0.1 | 0.08 | 0.08 | 0.12 | 0.16 | 98.98 | 0.3 |
200% | 0.02 | 0 | 0 | 0 | 0.02 | 0 | 0.02 | 0 | 0 | 0.02 | 0.02 | 0.08 | 0.06 | 0.18 | 99.58 |
Resampling is a common operation used in image splicing. Here we evaluate the feasibility of our model in image splicing detection. In this experiment, 64×64 and 96×96 are chosen as a small size to detect resampling. Large size images can be used with 96×96 pixels model, otherwise 64×64 pixels. In order to train the variant CNN, we create a dataset using the original dataset. These image blocks are first cut to 64×64 and 96×96 pixels subblocks, and then 300,000 subblocks are randomly selected, followed by resampling and JPEG recompression, respectively. Scaling factors and quantity factors are randomly chosen in the range of 0.5 to 2 at the interval of 0.01 and the range of 50 to 100 at the interval of 1. Finally, we create two datasets of different image sizes, each consists of 900,000 images (300,000 upsampled images, 300,000 downsampled images and 300,000 unresampled images). The datasets are divided into three parts: 80% of the images are used for training, 10% are used for validation, and the rest part is used for testing. Two binary classifier are trained to deal with upsampling and downsampling respectively.
The resampling detection results are shown in Table 5. From Table 5, we can observe when the image size is 96×96, the detection accuracy of upsampling is 92%, and that of downsampling is 83%. When the image size is reduced to 64×64, the accuracy of upsampling classification reaches 89%, which is already an acceptable result. However, the accuracy of downsampling is only 77%.
Image size | Upscaling | Downscaling |
64×64 | 89% | 77% |
96×96 | 92% | 83% |
The wild forensics dataset [13] is applied to verify the effectiveness of the proposed approach using 64×64 pixels in image splicing detection. The tampered areas of the test images are resampled by unknown parameters and compressed by JPEG. Figure 5 shows the results of our and four prior splicing detection methods, which include DCT [9], CFA1 [10], CFA2 [11], ADQ [12]. Our method successfully detects and locates the forged region in some images that are not well detected by the other methods. This indicates that the resampling feature is helpful for image splicing detection. However, such a method can only detect the resampled part of the forged image. If the forged part has not been resampled or the image has been processed by a variety of post-processing such as sharpening and filtering, our method will not work.
Generalization ability, which is the performance of trained models on unknown data, is an important evaluation index for CNN. Here we test the model trained by ALASKA image database on another new dataset. The models trained in Section 3.3.1 (150% scaling factor, Q = 65, 75, 85) and Section 3.3.2 (upsampling operation, image size 256×256 and 128×128) are tested. The Dersern database [40], which is a benchmark dataset in image forensics research community, is used as the new test dataset. It contains 16960 JPEG images taken by 27 cameras. We process the Dersern dataset according to the same image processing method as that in the subsection 3.3.1. Finally, 14 datasets are created and each contains 33,920 images, half of which are resampled ones.
As shown in Table 6, the results discover that the detection accuracy on Dersern keeps above 80% but lower than that on ALASKA, especially for low Qs. For example, the accuracy reduces from 99.94% to 90.30% at Q = 85, and drops from 99.46% to 80.46% at Q = 65. This may attribute to the too single source of the training images.
Dataset | upscaling 150% | upscaling 110–200% | |||
Q = 85 | Q = 75 | Q = 65 | 256×256 | 128×128 | |
ALASKA | 99.94% | 99.84% | 99.46% | 99.07% | 98.45% |
Dersern | 90.30% | 86.03% | 80.40% | 85.25% | 83.41% |
The structure of CNN is related to the final detection accuracy. In this subsection, we use a series of experiments to discuss the rationality and effectiveness of our model structure. Three parts of the proposed model have been considered, including the noise extraction layers, horizontal stream or vertical stream and the activation functions. The sub-database (gamma correction, down-sampling) generated in Section 3.4 is utilized to accomplish the experiment here.
As described in Section 2.2, two low-order highpass filters are used to extract noise maps in our proposed method. A set of experiments are conducted to prove the effectiveness of noise extraction layers in this section. We consider the three different cases about the noise extraction layer, including the use of two high-order highpass filters from Figure 6, two low-order highpass filters that used in our method and removing the noise extraction layer. The experimental results are shown in Figure 7. We can observe that low-order highpass filter achieves the best performance, and its accuracy is distinctively higher than other methods. The accuracy of high-order highpass filter is slightly higher than removing noise extraction layer.
In order to assess the performance of horizontal stream and vertical stream, four cases are studied, which are only using horizontal stream, only using vertical stream, using horizontal and vertical streams without interleaved stream, and using horizontal and vertical streams with interleaved stream. Table 7 depicts that the performance of our model is superior to other model, and the accuracy is 97.37%. The accuracies of horizontal and vertical streams models are 95.52% and 96.25%, respectively. Both methods are significantly lower than that in the other two methods. The primary reason is that only one directional feature is considered. The accuracy rate typically greater than 97% when the horizontal and vertical features are considered simultaneously. Performance of the model with interleaved stream is better than that without interleaved stream, since the directional correlation is further strengthened. Thus, the proposed dual-stream architecture and the interleaved stream strategy could improve the resampling detection performance.
Models | Accuracy |
Horizontal stream | 95.52% |
Vertical stream | 96.25% |
Two streams without interleaved stream | 97.04% |
Two streams with interleaved stream | 97.37% |
The selection of suitable activation function is rather important for CNNs. In order to evaluate the effectiveness of the proposed model with different activation functions, two commonly used activation functions including TanH and ReLu are addressed. We can see from Figure 8 that TanH is more suitable for our model, since performance of the network with TanH is better than that with ReLu. Furthermore, TanH owns better stability and faster convergence than ReLu.
In this paper, we presented a new deep learning-based method to detect resampling in recompressed images. A noise extraction layer is used to extract noise residual and a dual-stream CNN is proposed to extract resampling features from noise maps in different directions. The experimental results show that the proposed method could not only detect the resampling of recompressed images, but also achieve excellent robustness against some additional post-processing operations. Moreover, we extensively apply the global resampling detection method to resampling parameter estimation and image splicing detection. It should be noted that the proposed method can not detect resampling traces if the image has undergone complex post-processing or anti-forensic manipulations. In the future, we would extend our work to identify more types of resampling in the existence of complex operation chain and anti-forensics attacks.
This work was supported in part by the National Natural Science Foundation of China (grant nos. 61401408, 61772539, 11571325) and the Fundamental Research Funds for the Central Universities (grant nos. CUC2019B021, 3132017XNG1715).
All authors declare no conflicts of interest in this paper.
[1] | M. Stamm, M. Wu and K. J. Liu, Information forensics: An overview of the first decade, IEEE Access, 1 (2013), 167–200. |
[2] | H. Farid, Exposing digital forgeries from JPEG ghosts, IEEE T. Inf. Foren. Sec., 4 (2009), 154– 160. |
[3] | W. Luo, J. Huang and G. Qiu, JPEG error analysis and its applications to digital image forensics, IEEE T. Inf. Foren. Sec., 5 (2010), 480–491. |
[4] | G. Cao, Y. Zhao, R. Ni, et al., Contrast enhancement-based forensics in digital images, IEEE T. Inf. Foren. Sec., 9 (2014), 515–525. |
[5] | G. Cao, Y. Zhao, R. Ni, et al., Attacking contrast enhancement forensics in digital images, Sci. China Inform. Sci., 57 (2014), 1–13. |
[6] | G. Cao, L. Huang, H. Tian, et al., Contrast enhancement of brightness-distorted images by improved adaptive gamma correction, Comput. Electr. Eng., 66 (2018), 569–582. |
[7] | F. Ding, G. Zhu, W. Dong, et al., An efficient weak sharpening detection method for image forensics, J. Vis. Commun. Image R., 50 (2018), 93–99. |
[8] | X. Kang, M. Stamm, A. Peng, et al., Robust median filtering forensics using an autoregressive model, IEEE T. Inf. Foren. Sec., 8 (2013), 1456–1468. |
[9] | S. Ye, Q. Sun and E. C. Chang, Detecting digital image forgeries by measuring inconsistencies of blocking artifact, IEEE International Conference on Multimedia and Expo, (2007), 12–15. |
[10] | A. Dirik and N. Memon, Image tamper detection based on demosaicing artifacts, IEEE International Conference on Image Processing, (2009), 1497–1500. |
[11] | P. Ferrara, T. Bianchi, A. De Rosa, et al., Image forgery localization via fine-grained analysis of CFA artifacts, IEEE T. Inf. Foren. Sec., 7 (2012), 1566–1577. |
[12] | I. Amerini, R. Becarelli, R. Caldelli, et al., Splicing forgeries localization through the use of first digit features, IEEE International Workshop on Info. Forensics and Security, (2014), 143–148. |
[13] | M. Huh, A. Liu, A. Owens, et al., Fighting fake news: image splice detection via learned self-consistency, European Conference on Computer Vision, (2018), 101–117. |
[14] | P. Zhou, X. Han, V. Morariu, et al., Learning rich features for image manipulation detection, IEEE Conference on Computer Vision and Pattern Recognition, (2018), 1053–1061. |
[15] | J. Bunk, J. Bappy and T. Mohammed, Detection and localization of image forgeries using resampling features and deep learning, IEEE Conference on Computer Vision and Pattern Recognition Workshops, (2017), 1881–1889. |
[16] | J. Wang, T. Li, X. Luo, et al., Identifying computer generated images based on quaternion central moments in color quaternion wavelet domain, IEEE T. Circ. Syst. Vid., (2018), 1. |
[17] | A. Popescu and H. Farid, Exposing digital forgeries by detecting traces of resampling, IEEE T. Signal Proces., 53 (2005), 758–767. |
[18] | D.Padín, P.ComesanaandF.González, AnSVDapproachtoforensicimageresamplingdetection, European Signal Processing Conference, (2015), 2067–2071. |
[19] | D. Padín, F. González and P. Comesana, A random matrix approach to the forensic analysis of upscaled images, IEEE T. Inf. Foren. Sec., 12 (2017), 2115–2130. |
[20] | Y. Kao, H. Lin, C. Wang, et al., Effective detection for linear up-sampling by factor of fraction, IEEE T. Image Process., 21 (2012), 3443–3453. |
[21] | B. Mahdian and S. Saic, Blind authentication using periodic properties of interpolation, IEEE T. Inf. Foren. Sec., 3 (2008), 529–538. |
[22] | T. Qiao, R. Shi, X. Luo, et al., Statistical model-based detector via texture weight map: application in re-sampling authentication, IEEE T. Multimedia, 21 (2019), 1077–1092. |
[23] | X. Feng, I. J. Cox and G. Doerr, An energy-based method for the forensic detection of re-sampled images, IEEE International Conference on Multimedia and Expo, (2011), 1–6. |
[24] | M. Kirchner, Fast and reliable resampling detection by spectral analysis of fixed linear predictor residue, ACM Workshop on Multimedia and Security, (2008), 11–20. |
[25] | G. Cao, Y. Zhao and R. Ni, Forensic identification of resampling operators: a semi non-intrusive approach, Foren. Sci. Int., 216 (2012), 29–36. |
[26] | C. Pasquini and R. Bohme, Information-theoretic bounds for the forensic detection of downscaled signals, IEEE T. Inf. Foren. Sec., 14 (2019), 1928–1943. |
[27] | A. C. Gallagher, Detection of linear and cubic interpolation in JPEG compressed images, Canadian Conference on Computer and Robot Vision, (2005), 65–72. |
[28] | L. Nataraj, A. Sarkar and B. Manjunath, Adding gaussian noise to "denoise" JPEG for detecting image resizing, IEEE International Conference on Image Processing, (2009), 1477–1480. |
[29] | M. Kirchner and T. Gloe, On resampling detection in re-compressed images, IEEE International Workshop on Information Forensics and Security, (2009), 21–25. |
[30] | Z. Chen, Y. Zhao and R. Ni, Detection of operation chain: JPEG-resampling-JPEG, Signal Process.-Image, 57 (2017), 8–20. |
[31] | H. Li, W. Luo, X. Qiu, et al., Identification of various image operations using residual-based features, IEEE T. Circ. Syst. Vid., 28 (2018), 31–35. |
[32] | B. Bayar and M. Stamm, On the robustness of constrained convolutional neural networks to JPEG post-compression for image resampling detection, IEEE International Conference on Acoustics, Speech and Signal Processing, (2017), 2152–2156. |
[33] | B. Bayar and M. Stamm, Constrained convolutional neural networks: a new approach towards general purpose image manipulation detection, IEEE T. Inf. Foren. Sec., 13 (2018), 2691–2706. |
[34] | J. Fridrich and J. Kodovsky, Rich models for steganalysis of digital images, IEEE T. Inf. Foren. Sec., 7 (2012), 868–882. |
[35] | Y. Ma, X. Luo, X. Li, et al., Selection of rich model steganalysis features based on decision rough set α-positive region reduction, IEEE T. Circ. Syst. Vid., 29 (2019), 336–350. |
[36] | Y. Zhang, C. Qin, W. Zhang, et al., On the fault-tolerant performance for a class of robust image steganography, Signal Process., 146 (2018), 99–111. |
[37] | X. Luo, X. Song, X. Li, et al., Steganalysis of HUGO steganography based on parameter recognition of syndrome-trellis-codes, Multimed. Tools Appl., 75 (2016), 13557–13583. |
[38] | B. Chen, H. Li and W. Luo, Image processing operations identification via convolutional neural network, preprint, arXiv:1709.02908. |
[39] | I. Sutskever, J. Martens, G. E. Dahl, et al., On the importance of initialization and momentum in deep learning, International Conference on Machine Learning, (2013), 1139–1147. |
[40] | T. Gloe and R. Bohme, The Dresden image database for benchmarking digital image forensics, J. Digit. Foren. Pract., 3 (2010), 1584–1590. |
1. | Shangjun Luo, Jiarui Liu, Wenbo Xu, Wei Lu, Yanmei Fang, Jinhua Zeng, Shaopei Shi, Upscaling factor estimation on pre-JPEG compressed images based on difference histogram of spectral peaks, 2021, 94, 09235965, 116223, 10.1016/j.image.2021.116223 | |
2. | Shangjun Luo, Junwei Luo, Wei Lu, Yanmei Fang, Jinhua Zeng, Shaopei Shi, Yue Zhang, Resampling Factor Estimation via Dual-stream Convolutional Neural Network, 2020, 66, 1546-2226, 647, 10.32604/cmc.2020.012869 | |
3. | Xiaolong Wei, Xianglin Huang, LiFang Yang, Gang Cao, Zhulin Tao, Bing Wang, Jing An, Hierarchical RNNs-Based transformers MADDPG for mixed cooperative-competitive environments, 2022, 43, 10641246, 1011, 10.3233/JIFS-212795 | |
4. | Xiaogang Zhu, Shuaiqi Liu, Bing Fan, Xiangjun Li, Yiping Zhu, Haozheng Yu, Forensics Analysis of Resampling via ConvNeXt Block, 2023, 32, 0218-1266, 10.1142/S0218126623500330 | |
5. | Viacheslav Pavlovich Meshchaninov, Ivan Andreevich Molodetskikh, Dmitriy Sergeevich Vatolin, Alexey Gennadievich Voloboy, Combining contrastive and supervised learning for video super-resolution detection, 2022, 20712898, 1, 10.20948/prepr-2022-80 | |
6. | Sunen Chakraborty, Paramita Dey, Kingshuk Chatterjee, 2023, Chapter 17, 978-981-99-2709-8, 197, 10.1007/978-981-99-2710-4_17 | |
7. | Vijayakumar Kadha, Venkatesh Vakamullu, Santos Kumar Das, Madhusudhan Mishra, Joyatri Bora, 2023, Chapter 15, 978-3-031-25193-1, 183, 10.1007/978-3-031-25194-8_15 | |
8. | Jingyun Hu, Ru Xue, Guofeng Teng, Shiming Niu, Danyang Jin, Image splicing manipulation location by multi-scale dual-channel supervision, 2023, 83, 1573-7721, 31759, 10.1007/s11042-023-16705-y |
Image size | 256×256 | 128×128 | 64×64 | |||
Resampling type | Upscaling | Downscaling | Upscaling | Downscaling | Upscaling | Downscaling |
Bayar's method [32,33] | 97.29 | 97.24 | 95.19 | 91.93 | 88.19 | 80.03 |
Proposed method | 99.07 | 98.45 | 97.06 | 95.47 | 92.49 | 88.46 |
Operation type | Parameters |
Gamma correction(GC) | γ: 0.5, 0.6, 0.7, 0.8, 0.9, 1.2, 1.4, 1.6, 1.8, 2.0 |
Mean filtering(MeanF) | window size: 3×3, 5×5, 7×7 |
Gaussian filtering(GF) | window size: 3×3, 5×5, 7×7, σ: 0.8-1.6 |
Median filtering(MedF) | window size: 3×3, 5×5, 7×7 |
Wiener filtering(WF) | window size: 3×3, 5×5, 7×7 |
Post-processing type | GC | MeanF | GF | MedF | WF | |||||
Resampling type | Up | Down | Up | Down | Up | Down | Up | Down | Up | Down |
Bayar's method [32,33] | 96.50 | 95.46 | 81.89 | 67.03 | 85.32 | 73.41 | 82.39 | 67.91 | 84.81 | 71.14 |
Proposed method | 97.85 | 97.37 | 86.69 | 73.88 | 89.14 | 84.26 | 87.90 | 77.12 | 90.45 | 84.46 |
Tru╲Pre | 50% | 60% | 70% | 80% | 90% | 110% | 120% | 130% | 140% | 150% | 160% | 170% | 180% | 190% | 200% |
50% | 91.56 | 4.32 | 2.54 | 0.76 | 0.4 | 0.14 | 0.02 | 0.1 | 0.02 | 0 | 0.04 | 0 | 0.04 | 0.06 | 0 |
60% | 4.04 | 93.2 | 0.96 | 0.9 | 0.26 | 0.18 | 0.08 | 0.04 | 0.04 | 0.12 | 0.02 | 0.04 | 0.04 | 0.08 | 0 |
70% | 2.3 | 1.5 | 93.72 | 1.28 | 0.7 | 0.1 | 0.1 | 0.06 | 0.08 | 0.02 | 0.02 | 0.02 | 0.06 | 0.02 | 0.02 |
80% | 0.62 | 0.86 | 1.5 | 95.46 | 0.98 | 0.32 | 0.02 | 0.02 | 0.1 | 0.04 | 0.02 | 0.04 | 0 | 0.02 | 0 |
90% | 0.36 | 0.3 | 0.84 | 0.88 | 96.9 | 0.26 | 0.12 | 0.12 | 0.08 | 0.02 | 0.02 | 0.02 | 0.02 | 0.06 | 0 |
110% | 0.08 | 0.3 | 0.16 | 0.46 | 0.46 | 97.12 | 0.56 | 0.26 | 0.14 | 0.14 | 0.02 | 0.06 | 0.06 | 0.14 | 0.04 |
120% | 0 | 0 | 0.1 | 0.14 | 0.06 | 0.58 | 98.38 | 0.38 | 0.06 | 0.1 | 0.04 | 0.06 | 0.04 | 0.04 | 0.02 |
130% | 0.08 | 0.02 | 0.02 | 0 | 0.06 | 0.14 | 0.24 | 98.86 | 0.1 | 0.16 | 0.08 | 0.08 | 0.1 | 0.04 | 0.02 |
140% | 0.06 | 0.02 | 0.08 | 0.18 | 0.08 | 0.18 | 0.12 | 0.42 | 97.58 | 0.46 | 0.24 | 0.34 | 0.1 | 0.06 | 0.08 |
150% | 0.04 | 0.08 | 0.02 | 0 | 0.06 | 0.08 | 0.08 | 0.1 | 0.24 | 98.64 | 0.08 | 0.32 | 0.06 | 0.14 | 0.06 |
160% | 0 | 0 | 0 | 0.02 | 0 | 0.02 | 0.06 | 0.06 | 0.14 | 0.14 | 98.92 | 0.06 | 0.36 | 0.12 | 0.1 |
170% | 0.04 | 0.02 | 0.02 | 0.02 | 0 | 0.06 | 0.02 | 0.02 | 0.08 | 0.18 | 0.12 | 98.96 | 0.12 | 0.24 | 0.1 |
180% | 0.02 | 0.02 | 0 | 0.04 | 0.04 | 0 | 0.06 | 0.04 | 0.06 | 0 | 0.34 | 0.06 | 99 | 0.18 | 0.14 |
190% | 0.02 | 0 | 0 | 0.02 | 0.02 | 0.02 | 0.02 | 0.08 | 0.1 | 0.08 | 0.08 | 0.12 | 0.16 | 98.98 | 0.3 |
200% | 0.02 | 0 | 0 | 0 | 0.02 | 0 | 0.02 | 0 | 0 | 0.02 | 0.02 | 0.08 | 0.06 | 0.18 | 99.58 |
Image size | Upscaling | Downscaling |
64×64 | 89% | 77% |
96×96 | 92% | 83% |
Dataset | upscaling 150% | upscaling 110–200% | |||
Q = 85 | Q = 75 | Q = 65 | 256×256 | 128×128 | |
ALASKA | 99.94% | 99.84% | 99.46% | 99.07% | 98.45% |
Dersern | 90.30% | 86.03% | 80.40% | 85.25% | 83.41% |
Models | Accuracy |
Horizontal stream | 95.52% |
Vertical stream | 96.25% |
Two streams without interleaved stream | 97.04% |
Two streams with interleaved stream | 97.37% |
Image size | 256×256 | 128×128 | 64×64 | |||
Resampling type | Upscaling | Downscaling | Upscaling | Downscaling | Upscaling | Downscaling |
Bayar's method [32,33] | 97.29 | 97.24 | 95.19 | 91.93 | 88.19 | 80.03 |
Proposed method | 99.07 | 98.45 | 97.06 | 95.47 | 92.49 | 88.46 |
Operation type | Parameters |
Gamma correction(GC) | γ: 0.5, 0.6, 0.7, 0.8, 0.9, 1.2, 1.4, 1.6, 1.8, 2.0 |
Mean filtering(MeanF) | window size: 3×3, 5×5, 7×7 |
Gaussian filtering(GF) | window size: 3×3, 5×5, 7×7, σ: 0.8-1.6 |
Median filtering(MedF) | window size: 3×3, 5×5, 7×7 |
Wiener filtering(WF) | window size: 3×3, 5×5, 7×7 |
Post-processing type | GC | MeanF | GF | MedF | WF | |||||
Resampling type | Up | Down | Up | Down | Up | Down | Up | Down | Up | Down |
Bayar's method [32,33] | 96.50 | 95.46 | 81.89 | 67.03 | 85.32 | 73.41 | 82.39 | 67.91 | 84.81 | 71.14 |
Proposed method | 97.85 | 97.37 | 86.69 | 73.88 | 89.14 | 84.26 | 87.90 | 77.12 | 90.45 | 84.46 |
Tru╲Pre | 50% | 60% | 70% | 80% | 90% | 110% | 120% | 130% | 140% | 150% | 160% | 170% | 180% | 190% | 200% |
50% | 91.56 | 4.32 | 2.54 | 0.76 | 0.4 | 0.14 | 0.02 | 0.1 | 0.02 | 0 | 0.04 | 0 | 0.04 | 0.06 | 0 |
60% | 4.04 | 93.2 | 0.96 | 0.9 | 0.26 | 0.18 | 0.08 | 0.04 | 0.04 | 0.12 | 0.02 | 0.04 | 0.04 | 0.08 | 0 |
70% | 2.3 | 1.5 | 93.72 | 1.28 | 0.7 | 0.1 | 0.1 | 0.06 | 0.08 | 0.02 | 0.02 | 0.02 | 0.06 | 0.02 | 0.02 |
80% | 0.62 | 0.86 | 1.5 | 95.46 | 0.98 | 0.32 | 0.02 | 0.02 | 0.1 | 0.04 | 0.02 | 0.04 | 0 | 0.02 | 0 |
90% | 0.36 | 0.3 | 0.84 | 0.88 | 96.9 | 0.26 | 0.12 | 0.12 | 0.08 | 0.02 | 0.02 | 0.02 | 0.02 | 0.06 | 0 |
110% | 0.08 | 0.3 | 0.16 | 0.46 | 0.46 | 97.12 | 0.56 | 0.26 | 0.14 | 0.14 | 0.02 | 0.06 | 0.06 | 0.14 | 0.04 |
120% | 0 | 0 | 0.1 | 0.14 | 0.06 | 0.58 | 98.38 | 0.38 | 0.06 | 0.1 | 0.04 | 0.06 | 0.04 | 0.04 | 0.02 |
130% | 0.08 | 0.02 | 0.02 | 0 | 0.06 | 0.14 | 0.24 | 98.86 | 0.1 | 0.16 | 0.08 | 0.08 | 0.1 | 0.04 | 0.02 |
140% | 0.06 | 0.02 | 0.08 | 0.18 | 0.08 | 0.18 | 0.12 | 0.42 | 97.58 | 0.46 | 0.24 | 0.34 | 0.1 | 0.06 | 0.08 |
150% | 0.04 | 0.08 | 0.02 | 0 | 0.06 | 0.08 | 0.08 | 0.1 | 0.24 | 98.64 | 0.08 | 0.32 | 0.06 | 0.14 | 0.06 |
160% | 0 | 0 | 0 | 0.02 | 0 | 0.02 | 0.06 | 0.06 | 0.14 | 0.14 | 98.92 | 0.06 | 0.36 | 0.12 | 0.1 |
170% | 0.04 | 0.02 | 0.02 | 0.02 | 0 | 0.06 | 0.02 | 0.02 | 0.08 | 0.18 | 0.12 | 98.96 | 0.12 | 0.24 | 0.1 |
180% | 0.02 | 0.02 | 0 | 0.04 | 0.04 | 0 | 0.06 | 0.04 | 0.06 | 0 | 0.34 | 0.06 | 99 | 0.18 | 0.14 |
190% | 0.02 | 0 | 0 | 0.02 | 0.02 | 0.02 | 0.02 | 0.08 | 0.1 | 0.08 | 0.08 | 0.12 | 0.16 | 98.98 | 0.3 |
200% | 0.02 | 0 | 0 | 0 | 0.02 | 0 | 0.02 | 0 | 0 | 0.02 | 0.02 | 0.08 | 0.06 | 0.18 | 99.58 |
Image size | Upscaling | Downscaling |
64×64 | 89% | 77% |
96×96 | 92% | 83% |
Dataset | upscaling 150% | upscaling 110–200% | |||
Q = 85 | Q = 75 | Q = 65 | 256×256 | 128×128 | |
ALASKA | 99.94% | 99.84% | 99.46% | 99.07% | 98.45% |
Dersern | 90.30% | 86.03% | 80.40% | 85.25% | 83.41% |
Models | Accuracy |
Horizontal stream | 95.52% |
Vertical stream | 96.25% |
Two streams without interleaved stream | 97.04% |
Two streams with interleaved stream | 97.37% |