
The image super-resolution reconstruction method can improve the image quality in the Internet of Things (IoT). It improves the data transmission efficiency, and is of great significance to data transmission encryption. Aiming at the problem of low image quality in image super-resolution using neural networks, a self-attention-based image reconstruction method is proposed for secure data transmission in IoT environment. The network model is improved, and the residual network structure and sub-pixel convolution are used to extract the feature of the image. The self-attention module is used extract detailed information in the image. Using generative confrontation method and image feature perception method to improve the image reconstruction effect. The experimental results on the public data set show that the improved network model improves the quality of the reconstructed image and can effectively restore the details of the image.
Citation: Hongan Li, Qiaoxue Zheng, Wenjing Yan, Ruolin Tao, Xin Qi, Zheng Wen. Image super-resolution reconstruction for secure data transmission in Internet of Things environment[J]. Mathematical Biosciences and Engineering, 2021, 18(5): 6652-6671. doi: 10.3934/mbe.2021330
[1] | Zhijing Xu, Jingjing Su, Kan Huang . A-RetinaNet: A novel RetinaNet with an asymmetric attention fusion mechanism for dim and small drone detection in infrared images. Mathematical Biosciences and Engineering, 2023, 20(4): 6630-6651. doi: 10.3934/mbe.2023285 |
[2] | Huqing Wang, Zhixin Sun . Research on multi decision making security performance of IoT identity resolution server based on AHP. Mathematical Biosciences and Engineering, 2021, 18(4): 3977-3992. doi: 10.3934/mbe.2021199 |
[3] | Dengzhi Liu, Zhimin Li, Chen Wang, Yongjun Ren . Enabling secure mutual authentication and storage checking in cloud-assisted IoT. Mathematical Biosciences and Engineering, 2022, 19(11): 11034-11046. doi: 10.3934/mbe.2022514 |
[4] | Shuaiyu Bu, Yuanyuan Li, Wenting Ren, Guoqiang Liu . ARU-DGAN: A dual generative adversarial network based on attention residual U-Net for magneto-acousto-electrical image denoising. Mathematical Biosciences and Engineering, 2023, 20(11): 19661-19685. doi: 10.3934/mbe.2023871 |
[5] | Jimin Yu, Jiajun Yin, Shangbo Zhou, Saiao Huang, Xianzhong Xie . An image super-resolution reconstruction model based on fractional-order anisotropic diffusion equation. Mathematical Biosciences and Engineering, 2021, 18(5): 6581-6607. doi: 10.3934/mbe.2021326 |
[6] | Jiushuang Wang, Ying Liu, Huifen Feng . IFACNN: efficient DDoS attack detection based on improved firefly algorithm to optimize convolutional neural networks. Mathematical Biosciences and Engineering, 2022, 19(2): 1280-1303. doi: 10.3934/mbe.2022059 |
[7] | Zilong Liu, Jingbing Li, Jing Liu . Encrypted face recognition algorithm based on Ridgelet-DCT transform and THM chaos. Mathematical Biosciences and Engineering, 2022, 19(2): 1373-1387. doi: 10.3934/mbe.2022063 |
[8] | Hui Yao, Yuhan Wu, Shuo Liu, Yanhao Liu, Hua Xie . A pavement crack synthesis method based on conditional generative adversarial networks. Mathematical Biosciences and Engineering, 2024, 21(1): 903-923. doi: 10.3934/mbe.2024038 |
[9] | Shitharth Selvarajan, Hariprasath Manoharan, Celestine Iwendi, Taher Al-Shehari, Muna Al-Razgan, Taha Alfakih . SCBC: Smart city monitoring with blockchain using Internet of Things for and neuro fuzzy procedures. Mathematical Biosciences and Engineering, 2023, 20(12): 20828-20851. doi: 10.3934/mbe.2023922 |
[10] | Huqing Wang, Feng Xiang, Wenbing Zhao, Zhixin Sun . ONS resolution prediction based on Rasch model. Mathematical Biosciences and Engineering, 2019, 16(6): 6683-6695. doi: 10.3934/mbe.2019333 |
The image super-resolution reconstruction method can improve the image quality in the Internet of Things (IoT). It improves the data transmission efficiency, and is of great significance to data transmission encryption. Aiming at the problem of low image quality in image super-resolution using neural networks, a self-attention-based image reconstruction method is proposed for secure data transmission in IoT environment. The network model is improved, and the residual network structure and sub-pixel convolution are used to extract the feature of the image. The self-attention module is used extract detailed information in the image. Using generative confrontation method and image feature perception method to improve the image reconstruction effect. The experimental results on the public data set show that the improved network model improves the quality of the reconstructed image and can effectively restore the details of the image.
With the improvement of people's living standard and quality, the use of smart devices in the Internet of Things[1,2,3,4] has become more common, and it has a very broad range of application scenarios in smart grids, telemedicine, and smart transportation[5,6,37]. The intelligent service of the IoT needs data as the basis. How to effectively obtain different types of data resources and analyze them becomes particularly important. In addition, data transmission and data encryption among IoT devices have also become a major challenge for IoT security[7,8,9,11]. Image as an important carrier of IoT data, can more clearly and intuitively reflect the information content. Image super-resolution reconstruction can turn a low-resolution image into a high-resolution image[12], which improves the transmission efficiency of data in the IoT and reduces the huge overhead of data encryption and key exchange during data transmission[10,13,22,23].
In recent years, the rapid development of global deep learning has made the research on image super-resolution reconstruction methods become the focus of the public, which promotes the development of image super-resolution reconstruction [58,59]. In the research dealing with computer vision, many tasks can not only be solved by deep learning, but the effects of deep learning-based methods are also better than traditional methods[35]. The neural network is applied to image super-resolution reconstruction, and the mapping relationship between input low-quality images and output high-quality images is established through learning, so as to improve the image quality. This method has great potential value in the application of Internet of things, such as medical images, high-resolution images can help more accurate judgment of the disease, and can also improve the user's visual experience in daily life.
Low resolution images limit the accuracy of information transmission to some extent, thus reducing the efficiency of information transmission[14,38]. Image super-resolution reconstruction technology can improve the resolution of the image[15], thereby obtaining richer image details. The existing image super-resolution reconstruction methods are roughly divided into three categories : interpolation-based methods, reconstruction-based methods and learning-based methods[16].
Image super-resolution method based on interpolation. Interpolation-based method is mainly based on the image interpolation technology, also known as image scaling, which is to infer the value of unknown pixels by using the known pixels through the interpolation function or interpolation kernel. In this method, each known pixel on the image is regarded as a point on the image plane, and the value of unknown pixels to be interpolated is determined by using each known data for interpolation calculation. Finally, the digital image restoration technology is used to remove blur and reduce image noise. Commonly used interpolation methods mainly include nearest neighbor interpolation[19,20,25], bilinear interpolation[27] and bicubic interpolation[28,51].
Although the method based on interpolation is simple and fast, there are also some obvious technical defects. The processing effect of image edge (i.e. pixel mutation) is poor, and it is prone to sawtooth and block effect, which can not reach the ideal super-resolution.
Image super-resolution method based on reconstruction.The imaging process of the image is inverted through the reconstruction method, and the local or global prior model of the image is introduced to establish the observation model between the LR image and the HR image. Finally, the optimization model of data fidelity term and image prior regularization is established to solve it. Tsai[29] proposed a super-resolution reconstruction method for recovering additional high frequency information from multiple images in the Fourier transform domain for the first time. However, this method assumes that there is no motion blur and observation noise in the image, and ignores the point spread function of the optical system, so it is only suitable for the ideal image degradation model. The improved method uses recursive least squares, spatial blur, relative motion, discrete DCT transform and wavelet transform to remove the image observation noise, which effectively improves the visual effect of the reconstructed image[30,31].
The spatial domain method model the spatial factors (such as optical blur, motion blur, etc.) that affect the image imaging effect, so it is easier to approach the actual application. Common spatial super-resolution reconstruction methods mainly include non-uniform sampling interpolation (NUI), maximum a posteriori probability (MAP), iterative back projection, convex set projection[12,32,34]. The method based on reconfiguration has relatively good effect on image processing, but requiring the image to have better prior knowledge, so it is not fully applicable to directly reconstruct the images of the IoT with large magnification.
Image super-resolution method based on learning. Since machine learning has good application effect in the field of computer vision, the learning-based image super-resolution reconstruction method has obtained fruitful research results. Dong et al.[33] applied the convolutional neural network to super-resolution reconstruction of a single image, and its reconstruction effect was far superior to traditional methods. Kim et al.[36] proposed a VDSR method, which effectively improves reconstruction effect by expanding the network depth. Talab et al.[42] proposed ESPCN, which enlarges a low-resolution image into a high-resolution image through an interpolation method, and then uses a convolutional network to get a high-resolution image to make the texture of the generated image clearer. For making the reconstructed image meet the signal-to-noise ratio requirements, Ledig et al.[43] proposed an SRGAN method that uses perceptual loss to enhance the effect of network training, thereby recovering rich high-frequency detail information. Lim et al.[45] improved the EDSR method from the perspective of data processing, replacing the batch normalization in the residual module, making the network more versatile[39]. A single-frame character image super-resolution reconstruction method based on wavelet neural network is proposed, which improves the anti-interference performance of the model[48].
Although the image super-resolution method based on deep learning can improve the quality of restored images, it can not meet the quality requirements of data transmission in the Internet of Things. For actual IoT usage scenarios, domain knowledge, prior knowledge, and deep learning frameworks are not fully utilized, and they cannot complement each other in practical applications, so it can not pay more attention to the useful domain prior knowledge in image feature extraction, which leads to the poor quality of high-resolution image. Therefore, this paper proposes an image super-resolution method with self-attention. The main contributions of the work presented in this paper are as follows:
1) Introducing the self-attention into an image super-resolution model, which improves the model's attention to detail in IoT images.
2) Improving network model, extracting feature information from IoT images using residual network structure and subpixel convolution.
The rest of this paper is organized as follows. The section 2 briefly introduces the related work of neural networks and loss functions. Network structure and problem formulation is discussed in section 3. The section 4 displays and analyzes the experimental results. The paper is concluded in section 5.
For the high magnification factor, the image super-resolution reconstruction method is ill-posed[53]. The reconstructed super-resolution (SR) images have blurred texture details, and the reconstruction effect is not ideal. Generative Adversarial Networks (GAN)[50] has strong image generation capabilities, and well solve the above ill-posed problems. Therefore, the image super-resolution method based on GAN is proposed[51]. This method learns the mapping relationship of the network through large-scale IoT image training[15], and the GAN structure is shown in Figure 1.
GAN is composed of generator network G and discriminant network D. G is responsible for generating the corresponding high-resolution image by feeding the low-resolution image x. D is responsible for identifying its input image, judging from real data as true, judging from generated data from generator G as false. Therefore, the image super-resolution model based on GAN can be expressed as:
L(G,D)=Ex,y[lnD(x,y)]+Ex,y[ln(1−D(y,G(x)))]. | (1) |
Here, x represents the input low-resolution image, y represents the real high-resolution image. By minimizing the generator network loss and maximizing the discriminator loss, the overall network model is continuously trained through the game process between the G and the D. When GAN is successfully trained and converged, the Nash equilibrium between the generator and the discriminator is achieved, and the ability of the model to generate high-resolution images is improved.
Ledig et al.[43] introduced GAN into the field of image super-resolution and proposed SRGAN. Ledig combined the deep residual network with the sub-pixel convolution model, and used the residual network to extract the image feature and eliminate its noise. Up-sampling the extracted image features using the sub-pixel convolution model to gradually restore the detailed information in the image[48]. SRGAN makes the whole network pay more attention to the semantic feature difference between the generated super-resolution image and the real image, rather than the color brightness difference between pixels. However, this method performs super-resolution restoration of the image from a global perspective, ignoring the details of the image. Therefore, we proposed an improved method based on its work, adding self-attention to produce high-quality super-resolution image.
Subpixel convolution is a super-resolution method that uses image features for image up sampling [41,49,68]. It is composed of a set of expansion filters, which continuously adjust the parameters in the filter through learning. By establishing an image mapping relationship, the low-resolution input image is filtered to output an enlarged high-resolution image. For each feature image, a more complex amplification filter is trained and a sub-pixel convolutional layer is formed. Using sub-pixel convolutional layers to replace commonly used image scale expansion methods (such as bilinear interpolation and deconvolution operations, etc.) can improve the accuracy of image pixel amplification while reducing the computational complexity in reconstruction operations. Compared with the traditional methods, the parameters of the image magnification filter in subpixel convolution can be changed adaptively through training and learning. Therefore, using this method to enlarge the image can greatly reduce the risk and get a high-quality image with clearer structure and texture. The sub-pixel convolution process is shown in the Figure 2.
Set the zoom factor of the image to k. Subpixel convolutional layer receives low-resolution feature images, and after convolution of two hidden layers, the feature image with the same size as the input image is obtained. The number of feature channels is k2. Then rearrange all pixels in the channel features. By combining the single pixels on the multi-channel feature map into the pixel values on a feature map, the image is rearranged into one image, that is, the high-resolution image is a linear combination of k2 feature pixels.
The loss function is also known as the objective function, it's often used to evaluate the inconsistency between the actual output and the expected output of the neural network [17,18]. In the process of training, the neural network optimizes the model by minimizing the loss function, narrowing the gap between the predicted image and the real image and improving the ability of model generation.
When optimizing the parameters of the neural network, loss functions such as absolute value loss and mean square error loss are usually used to calculate the target difference value. However, these loss functions are difficult to deal with the uncertainty of high-frequency details such as textures, which leads to over-smooth reconstruction images and cannot produce clear image boundaries. Perception loss can be judged from the perspective of overall image information to increase the global perception of the model. Mualfah et al.[54] use gray level co-occurrence matrix to extract image feature information, and encrypt the asymmetric key through a single layer perceptron. Johnson et al.[55] used the perceptual loss function to train the feedforward network of image conversion task to generate high-quality images. Leading et al.[43] added the perceptual loss function in the production resolution model, making the entire network pay more attention to differences in high-level semantic information. The generated image and real image are input into the pre-training network model, and the perceptual loss of the two images is obtained by extracting the image feature information and calculating[44]:
Lp=‖φ(ˆy)−φ(y)‖2. | (2) |
Here, ϕ represents the pre-training network, ˆy and y epresent the generated high-resolution and real high-resolution images, respectively. By calculating the perceptual loss between images, the expression ability of network model on semantic information can be improved.
The image super-resolution model maps the image according to the feature information provided, and the restoration result of the model depends on the texture detail information of the image[21]. The content of the image includes brightness information, color information, structure information and texture information. By comparing the content information of the generated and real high-resolution images, the network can be better optimized and the quality of image generation can be improved. Cheon et al.[57] used discrete cosine transform coefficient loss and differential content loss to solve the trade-off problem. Lee et al.[60] used the loss of new content generated by the network to overcome the loss of detailed information caused by the pooling operation. By calculating the mean square error per pixel of fake image G(x) and real image y, the difference between the two images is compared. The content loss formula is as follows:
Lc=1WHW∑i=1H∑j=1(yi,j−G(x)i,j)2. | (3) |
Here, W and H respectively represent the width and height of images, (i,j) represents a pixel in images[24], x represents the input low-resolution image, y represents the ground truth, and G represents the generator network.
Image super-resolution reconstruction can be seen as a process of restoring high-frequency information of low-resolution images, that's why so much low-frequency information can be transmitted to the final output high-resolution image directly [40,46,47]. In the process of feature extraction, low-frequency information and high-frequency information are not distinguished while learning, which results in the network can not make full use of the high-frequency information in LR images. By adding the self-attention, he network model can focus more on the important information in the image and ignore the irrelevant information, which can improve these problems effectively.
Attention in neural networks has the same function as the way the human eye observes objects, and information with different characteristics is treated in the form of a network module differently. The self-attention structure is visualized as shown in Figure 3. The self-attention consists of three parts: K, Q, V[44]. Q represents the input query of the model, K represents the reference with high similarity to Q, and V represents the output content information corresponding to K. According to different processing tasks, the information stored in each part is different. In image processing tasks, the inherent information of image features is used for attention interaction, and the original feature map is usually mapped to Q, K, V. We calculate the correlation weight coefficient between Q and K and normalizing, and finally superimposed the coefficient into V.
Most of the research work on the combination of deep learning[65,66,67] and visual attention focuses on the use of masks to form the attention[61]. However, this method cannot dynamically adjust the network's attention to feature images, so we use the self-attention to filter image feature information, and add a larger attention coefficient to the information of interest. The self-attention is also called the internal attention. It is mainly an attention that calculates the correlation between different positions in a single sequence. Vaswani et al.[62] use the full attention on machine translation tasks, it proves that the network structure based on the attention can run in parallel and requires less training time. The self-attention mechanism is used to pay attention to the different information in the image, and the feature map is convolved to obtain Q, K, V. The self-attention formula is as follows:
Attention(Q,K,V)=softmax(QKT√dk)V. | (4) |
Here, dk represents the output dimension of the feature image.Multiplying the self-attention feature map with the dynamic adjustment coefficient and superimposing it on the original feature input can realize the amplification of the detailed information of interest and suppress the irrelevant noise information in the image.
The self-attention-based image super-resolution network model proposed in this paper includes two parts: a generator network and a discriminator network. The generator network consists of 16 deep residual modules, 3 convolution modules, and 2 sub-pixel convolutions, and 2 self-attention modules. The specific network structure is shown in Figure 4.
The low-resolution image is input into the network, the features of the image are extracted through the convolution operation, and the high-resolution image is finally output. In order to achieve the purpose of enabling the generator to capture the content information of the image greatly, this paper uses 16 depth residual modules to extract the feature information. Each residual block consists of 2 convolutional layers, 2 batch normalization layers and 1 activation layer. The activation layer uses the PReLU to perform the activation operation, so that during the network training process, the slope of the negative part can be dynamically adjusted according to the input image data, so as to improve the generalization and avoid over-fitting of the model. In order to enable the network to extract detailed information in the image, this paper adds self-attention after the first and second convolution blocks. Self attention is used to adjust the information of interest in the feature map to improve the attention of texture details and image content, and reduce the interference of noise data and irrelevant information. The feature map after self-attention calculation is input into the sub-pixel convolution layer, and the resolution of the image is enlarged by 4 times by using two up-sampling convolutions, and finally the high-resolution image is obtained through the convolution operation. All parameters of the network are described in detail in Table 1.
Layer name | Output size | Parameter | Numbers |
Conv_1 | 24 × 24 | [9 × 9, 64] | × 1 |
SA_1 | 24 × 24 | [1 × 1, 8] [1×1, 8] [1×1, 64] | × 1 |
ResConv | 24 × 24 | [3×3, 64] [3×3, 64] | × 16 |
Conv_2 | 24 × 24 | [3 × 3, 64] | × 1 |
SA_2 | 24 × 24 | [1 × 1, 8] [1 × 1, 8] [1 × 1, 64] | × 1 |
SubpixelConv | 96 × 96 | [3 × 3, 256] | × 2 |
Conv_3 | 96 × 96 | [9 × 9, 3] | × 1 |
In order to improve the ability of the network to recognize the global features of the image, a larger convolution kernel is used in the input and output parts of the generator to improve the receptive field of vision. At the same time, in order to reduce the number of parameters and prevent the model from being too large, a smaller convolution kernel is used in the model for feature extraction. The discriminator model judges the real high-resolution images and the generated high-resolution images[56]. Its essence is a classifier model, which judges the real images as true and the generated images as false.The discriminator network structure is shown in Figure 5.
The discriminator model is composed of 8 convolutional blocks, which are used to extract the deep feature information of the image, and judge the generated image and the real image on the high-level semantic information. To adapting to different input image sizes, an adaptive average pooling operation is used to pool the obtained feature images, and finally the classification results are output through a multilayer linear perceptron. The network parameters of the discriminator are described in Table 2.
Layer name | Output size | Parameter |
Conv_1 | 96 × 96 | [3 × 3, 64] |
Conv_2 | 48 × 48 | [3 × 3, 64] |
Conv_3 | 48 × 48 | [3 × 3, 128] |
Conv_4 | 24 × 24 | [3 × 3, 128] |
Conv_5 | 24 × 24 | [3 × 3, 256] |
Conv_6 | 12 × 12 | [3 × 3, 256] |
Conv_7 | 12 × 12 | [3 × 3, 512] |
Conv_8 | 6 × 6 | [3 × 3, 512] |
AdaptiveAvgPool | 6 × 6 | - |
FC_1 | 1024 | - |
FC_2 | 1 | - |
The image super-resolution reconstruction model designed in this paper can reconstruct the image of any resolution. When the discriminator is used to judge the image, the adaptive pooling operation is used to adapt to different resolutions. The convoluted feature image is pooled to get a uniform output size, and finally the true and false images are distinguished.
Use loss function to optimize image super-resolution network based on neural network. The quality of the image super-resolution model depends largely on the overall optimization goal. In this paper, the loss function of the self-attention-based image super-resolution reconstruction model is composed of three parts, namely, generative adversarial loss, perceptual loss, and content loss. In order to improve the overall perception ability of the model, this paper uses the pre-trained VGG-19 network to calculate the perception loss of the model. Input the real high-resolution image and generated high-resolution image to the VGG-19 network respectively, obtain the corresponding feature map through feature extraction, and obtain the perceptual loss by calculating the difference between the feature maps. The specific calculation is as follows:
Lp=1CC∑j=1‖φj(y)−φj(G(x))‖2. | (5) |
Here, C represents the number of channels of the feature map, φj represents the feature image at the j level, y represents the ground turth image, x represents the low-resolution image, and G represents the generator. The generative adversarial loss is used to improve the generation ability and discrimination ability of generators and discriminators. The content loss function mainly constrains the content information and compares content differences. The overall loss of the model can be expressed as
Ltotal=βL(G,D)+Lp+Lc. | (6) |
Here, β represents the generative adversarial loss coefficient, and the perceptual loss and content loss are respectively represented as Lp and Lc. The perceptual loss function and content loss function are used to judge the generated image and the real image. Figure 6 describes all the loss function calculation processes.
In order to evaluate the superiority of the method more comprehensively and accurately, subjective evaluation indicators and objective evaluation indicators are used to evaluate and analyze the reconstruction effect of the self-attention-based image super-resolution model proposed in this paper. The subjective evaluation mainly evaluates and scores the satisfaction of the reconstructed image from the perspective of visual effects, while the objective evaluation uses quantitative evaluation indicators to compare the similarity between the reconstructed image and the real one.
Average subjective opinion score usually evaluates the image quality based on the subjective feelings[63]. In this paper, MOS is used as a subjective evaluation metric. The quality of the image is scored by a sufficient number of observers. MOS evaluation metric are shown in Table 3.
Leval | Inpairment | Score |
1 | Very annoying | 1-2 |
2 | Annoying | 3-4 |
3 | Slightly annoying | 5-6 |
4 | Perceptible, but not annoying | 7-8 |
5 | Imperceptible | 9-10 |
MOS evaluation metric shows the human visual evaluation of the image is good or bad, with simple operation, intuitive results. The evaluation method based on the user's subjective intentions has a total of 50 participants, and used a random algorithm to select 4 pairs of images from the Bicubic model, SRGAN model and the high-resolution images generated by the model in this article to experiment. Each participant is shown reconstructed images generated by the three models respectively for the same input image within 5 seconds. All participants are not informed of the image generation model after making their own judgment. However, it is difficult to ensure the objective fairness of data, because of the evaluation of image quality depends entirely on the aesthetic preference.
Different from the MOS, the objective evaluation uses a series of method formulas to analyze the image more rigorously. PSNR and SSIM are the most basic indicators for measuring the quality of compressed reconstructed images in image super-resolution tasks[64]. Therefore, we use these two indicators to evaluate the improved method proposed in this article and compare it with other models. The larger the value of PSNR and SSIM, the better the quality of the image, that is, the better the performance of the image super-resolution method[52].
PSNR. PSNR is the ratio of the maximum power of the signal to the noise power of the signal[56]. It is used to measure the quality of the reconstructed image and is usually expressed in decibels (dB). The higher the score obtained by the PSNR indicator, the smaller the pixel difference between the generated image and the real image. Its calculation formula is expressed as follows:
MSE=1H×WH∑i=1W∑j=1[X(i,j)−Y(i,j)]2. | (7) |
PSNR=10log10(2n−1)2MSE. | (8) |
Here, X and Y respectively denote the generated super-resolution images and the real super-resolution images, W and H respectively denotes the width and height of images, (i,j) representing each pixel, which n is the number of bits of the pixel.
SSIM. PSNR also has its limitations, it cannot fully reflect the consistency of the image quality and the human visual effect, so SSIM is used for further comparison. SSIM is structural similarity, which is an evaluation index for evaluating the similarity of images in structure. Its calculation formula is as follows:
SSIM=(2μxμy+c1)(2σxy+c2)(μ2x+μ2y+c1)(σ2x+σ2y+c2). | (9) |
Here, μx and μy respectively represents the average value of the real image and the generated image, σ2x and σ2y respectively represents the variance of the real image and the generated image, σxy represents the covariance of the real image and the generated image. c1=(k1,L)2, c2=(k2,L)2 are constants to maintain stability and L is the dynamic range of pixel values. k1=0.01, k2=0.03.
All the experiments in this paper are done on the same lab environment, and the network model proposed in this article is implemented using Python 3.6 and Pytorch 1.4.0 software development environment. It runs on a 64-bit Windows 10 operating system, the processor is Intel(R) Core(TM) i9-10900X CPU @ 3.70GHz 3.70 GHz, and the graphics card is NVIDIA GeForce RTX 2080 Ti and CUDA 10.2.
The experiment uses COCO2014 data to train the network model proposed in this paper. In order to restore the complex image environment in the data transmission of Internet of things, SET5, SET14 and BSD1003 public data sets are used for model testing. From the BSD100 dataset, different kinds of images such as airplane, vase, racing car, human and animal are selected for testing; Different types of image data, such as bird, head, baby, butterfly and woman, are selected in SET5 data set; Zebra, pepper, flower, bridge and man are selected in SET14 dataset. Some of the test images in the dataset are shown in Figure 7.
Firstly, we pre-processed the training dataset. The HR images are scaled using the bilinear down-sampling method, and the resolution of the images are reduced by 4 times to obtain the LR images. In order to reduce the interference in the image information, so we de-averaging all LR images, and normalizing all pixels to [-1, 1].
In the training phase, the pre-processed LR images are input into the generative network, and the HR images are used as references to improve the ability of the generation through adversarial iteration. After 50 times of training, the final image super-resolution reconstruction model is obtained.
In the testing phase, the trained network model is used for testing. We use bilinear interpolation and SRGAN model as a comparisons.
Under the same experimental conditions, we compared the reconstruction effects of the bicubic method, the super-resolution method utilizing the generative adversarial network, and our improved method integrated with the self-attention. The experimental results of the three models are shown in Figure 8.
All the low-resolution images used in this paper are obtained by down sampling. The reconstructed images obtained by traditional methods and deep learning methods can not recover the same details as the original image, but in contrast, the reconstruction effect obtained by our method is better. It can be seen from the Figure 3 that the reconstructed images of Bicubic can only retain the high-frequency information such as the contour and color of the images, but reconstruction of details and texture is poor. Compared with the super-resolution image obtained by Bicubic, the effect of the SRGAN model has been significantly improved, and there are better results in terms of texture and contour. Because self-attention improves the attention ability of the network, it better retains the detailed feature information. It can be seen from Figure 3(c) that the improved method proposed in this paper produces higher image quality and reconstructs the details of the image. The color of the reconstructed image is richer and more natural than other methods.
In order to verify the effectiveness of this method, we quantitatively evaluate the super-resolution images generated by different methods. For the 50 participants who participated in the subjective evaluation, without telling them the corresponding method of each reconstructed image, each group of images was scored and the average value was calculated. MOS is shown in Table 4.
Method | Group1 | Group2 | Group3 | Group4 |
Bicubic | 4.200 | 4.230 | 4.166 | 4.067 |
SRGAN | 6.833 | 7.166 | 6.833 | 6.710 |
SA+SRGAN(Ours) | 7.066 | 7.533 | 7.250 | 6.900 |
It shows that the Bicubic interpolation method has the lowest recognition for the effect of image reconstruction, because the interpolation method only enlarges the LR image, and there are a lot of zero-filling operations in the reconstruction process. There is no effective extraction of image feature information. The SRGAN model and the image super-resolution reconstruction method with the self-attention get higher scores. Compared with the reconstruction method that only uses the generative confrontation network, because our method pays attention to the detailed features, it has a better effect on the boundary and detail processing of images while retaining the basic texture features. The best evaluation result was obtained on the group's score. The results show that the method proposed in this article is superior than the other two methods.
In order to compare the effects of different methods more objectively, we further used quantitative indicators such as PSNR and SSIM to quantitatively compare Set5, Set14 and BSD100. The quantitative evaluation results are shown in Table 5.
Method | Set5 | Set14 | BSD100 | |||
PSNR | SSIM | PSNR | SSIM | PSNR | SSIM | |
Bicubic | 27.067 | 0.749 | 24.502 | 0.682 | 23.924 | 0.626 |
SRGAN | 28.072 | 0.843 | 26.057 | 0.731 | 25.490 | 0.686 |
SA+SRGAN(Ours) | 28.745 | 0.854 | 26.656 | 0.745 | 26.037 | 0.697 |
The results are summarized in Table 4 confirmed that when the scaling factor is 4, the super-resolution reconstruction method based on self-attention we proposed has a certain improvement in PSNR and SSIM compared with other methods. On the Set5 data set, PSNR and SSIM increased by 0.673 and 0.011; on the Set14 data set, the PSNR and SSIM increased by 0.599 and 0.014, respectively; on the BSD100 data set, the PSNR increased by 0.547, and the SSIM increased by 0.011. The objective evaluation indicators show that our method can better recover the detailed information and improve the quality of the super-resolution.
In order to improve the transmission efficiency of image data in the Internet and reduce the huge overhead of data encryption during data transmission, we have propoed an image super-resolution method based on the self-attention. By introducing the attention, the details of the images are preserved during the reconstruction, and the images are more photo-realistic. Our method can effectively reduce the huge overhead of image transmission and data encryption of the Internet, and improve the efficiency of data transmission and the security of data encryption.
At present, most super-resolution reconstruction methods simulate the image degradation process under natural conditions by down-sampling high-resolution images to obtain low-resolution images. However, low-resolution images in practical applications usually have many problems such as motion distortion, optical blur, and noise pollution. How to solve the problem of image degradation caused by down-sampling is also a direction worth exploring for the image super-resolution method of the Internet of Things.
This work was supported in part by the Japan Society for the Promotion of Science (JSPS) Grants-in-Aid for Scientific Research (KAKENHI) under Grant 21K17737.
All authors declare no conflict of interest in this paper.
[1] |
J. Zhang, K. Yu, Z. Wen, X. Qi, A. K. Paul, 3D Reconstruction for Motion Blurred Images Using Deep Learning-Based Intelligent Systems, CMC Comput. Mater. Continua, 66 (2021), 2087-2104. doi: 10.32604/cmc.2020.014220
![]() |
[2] | W. Wang, H. Xu, M. Alazab, T. R. Gadekallu, Z. Han, C. Su, Blockchain-Based Reliable and Efficient Certificateless Signature for IIoT Devices, IEEE Trans. Ind. Inf., 2021. Available from: https://ieeexplore.ieee.org/document/9444140. |
[3] |
L. Tan, H. Xiao, K. Yu, M. Aloqaily, Y. Jararweh, A blockchain-empowered crowdsourcing system for 5g-enabled smart cities, Comput. Stand. Interfaces, 76 (2021), 103517. doi: 10.1016/j.csi.2021.103517
![]() |
[4] |
L. Zhen, Y. Zhang, K. Yu, N. Kumar, A. Barnawi, Y. Xie, Early Collision Detection for Massive Random Access in Satellite-Based Internet of Things, IEEE Trans. Veh. Technol., 70 (2021), 5184-5189. doi: 10.1109/TVT.2021.3076015
![]() |
[5] |
B. C. Chifor, I. Bica, V. V. Patriciu, F. Pop, A security authorization scheme for smart home Internet of Things devices, Future Gener. Comput. Syst., 86 (2018), 740-749. doi: 10.1016/j.future.2017.05.048
![]() |
[6] | L. Zhang, Z. Zhang, W. Wang, Z. Jin, Y. Su, H. Chen, Research on a covert communication model realized by using smart contracts in blockchain environment, IEEE Syst. J., 2021 (2021), 1-12. |
[7] |
B. B. Zarpelão, R. S. Miani, S. Rodrigo, C. T. Kawakani, Miani, S. C. de Alvarenga, A survey of intrusion detection in Internet of Things, J. Network Comput. Appl., 84 (2017), 25-37. doi: 10.1016/j.jnca.2017.02.009
![]() |
[8] | L. Zhen, A. K. Bashir, K. Yu, Y. D. Al-Otaibi, C. H. Foh, P. Xiao, Energy-efficient random access for LEO satellite-assisted 6G Internet of remote things, IEEE Internet Things J., 8 (2020), 5114-5128. |
[9] |
C. Feng, K. Yu, A. K. Bashir, Y. D. Al-Otaibi, Y. Lu, S. Chen, D. Zhang, Efficient and secure data sharing for 5G flying drones: a blockchain-enabled approach, IEEE Network, 35 (2021), 130-137. doi: 10.1109/MNET.011.2000223
![]() |
[10] | C. Feng, K. Yu, M. Aloqaily, M. Alazab, Z. Lv, S. Mumtaz, Attribute-based encryption with parallel outsourced decryption for edge intelligent IoV, IEEE Trans. Veh. Technol., 69 (2020), 213784-13795. |
[11] | H. Li, K. Yu, B. Liu, C. Feng, Z. Qin and G. Srivastava, An Efficient ciphertext-policy weighted attribute-based encryption for the internet of health things, IEEE J. Biomed. Health Inf., 2021. Available from: https://ieeexplore.ieee.org/document/9416735. |
[12] | D. Qiu, L. Zheng, S. Zhang, Y. Liu, An Image Super-resolution Reconstruction Method by Using of Deep Learning, in 2019 IEEE 4th International Conference on Image, Vision and Computing (ICIVC), (2019), 213-217. |
[13] | Y. Yang, H. Cai, Z. Wei, H. Lu, K. K. R. Choo, Towards lightweight anonymous entity authentication for IoT applications, in Australasian conference on information security and privacy, Springer, Cham, (2016), 265-280. |
[14] |
C. Sun, J. Lv, J. Li, R. Qiu, A rapid and accurate infrared image super-resolution method based on zoom mechanism, Infrared Phys. Technol., 88 (2018), 228-238. doi: 10.1016/j.infrared.2017.11.033
![]() |
[15] |
X. Feng, J. Li, Z. Hua, Guided filter-based multi-scale super-resolution reconstruction, CAAI Trans. Intell. Technol., 5 (2020), 128-140. doi: 10.1049/trit.2019.0065
![]() |
[16] |
Z. Huang, C. Jing, Super-resolution reconstruction method of remote sensing image based on multi-feature fusion, IEEE Access, 8 (2020), 18764-18771. doi: 10.1109/ACCESS.2020.2967804
![]() |
[17] | N. Shi, L. Tan, W. Li, X. Qi, K. Yu, A Blockchain-Empowered AAA Scheme in the Large-Scale HetNet, Digital Commun. Networks, 2021. Available from: https://doi.org/10.1016/j.dcan.2020.10.002. |
[18] | Z. Guo, A. K. Bashir, K. Yu, J. C. Lin, Y. Shen, Graph Embedding-based Intelligent Industrial Decision for Complex Sewage Treatment Processes, Int. J. Intell. Syst., 2020. Available from: https://doi.org/10.1002/int.22540. |
[19] |
G. T. Reddy, M. P. K. Reddy, K. Lakshmanna, R. Kaluri, D. S. Rajput, G. Srivastava, T. Baker, SAnalysis of dimensionality reduction techniques on big data, IEEE Access, 8 (2020), 54776-54788. doi: 10.1109/ACCESS.2020.2980942
![]() |
[20] |
L. Zhang, Y. Zou, W. Wang, Z. Jin, Y. Su, H. Chen, Resource allocation and trust computing for blockchain-enabled edge computing system, Comput. Secur., 105 (2021), 102249. doi: 10.1016/j.cose.2021.102249
![]() |
[21] |
X. Yao, Q. Wu, P. Zhang, F. X. Bao, Adaptive rational fractal interpolation function for image super-resolution via local fractal analysis, Image Vision Comput., 82 (2019), 39-49. doi: 10.1016/j.imavis.2019.02.002
![]() |
[22] | J. Song, Q. Zhong, W. Wang, C. Su, Z. Tan, Y. Liu, FPDP: Flexible privacy-preserving data publishing scheme for smart agriculture, IEEE Sens. J., 2020. Available from: https://ieeexplore.ieee.org/document/9170612. |
[23] | W. Wang, H. Huang, L. Zhang, C. Su, Secure and efficient mutual authentication protocol for smart grid under blockchain, Peer Peer Networking Appl., 2020 (2020), 1-13. |
[24] | L. Wang, S. Yang, J. Jia, A super-resolution reconstruction algorithm based on feature fusion, 2020 39th Chinese Control Conference (CCC), (2020), 3060-30605. |
[25] |
R. R. Schultz, R. L. Stevenson, A Bayesian approach to image expansion for improved definition, IEEE Trans. Image Process., 3 (1994), 233-242. doi: 10.1109/83.287017
![]() |
[26] | M. Yu, H. Wang, M. Liu, P. Li, Overview of Research on Image Super-Resolution Reconstruction, 2021 IEEE International Conference on Information Communication and Software Engineering (ICICSE), (2011), 131-135. |
[27] | K. T. Gribbon, D. G. Bailey, A novel approach to real-time bilinear interpolation, in Proceedings. DELTA 2004. Second IEEE International Workshop on Electronic Design, Test and Applications, (2004), 126-131. |
[28] |
R. Keys, Cubic convolution interpolation for digital image processing, IEEE Trans. Acoust. Speech Signal Process., 29 (1981), 1153-1160. doi: 10.1109/TASSP.1981.1163711
![]() |
[29] | R. Tsai, Multiframe image restoration and registration, Adv. Comput. Visual Image Process., 1 (1984), 317-339. |
[30] |
Y. Abe, Y. J. Iiguni, Image restoration from a downsampled image by using the DCT, Signal Process., 87 (2007), 2370-2380. doi: 10.1016/j.sigpro.2007.03.010
![]() |
[31] | P. Liu, H. Zhang, K. Zhang, L. Lin, W. Zuo, Multi-level wavelet-CNN for image restoration, Proceedings of the IEEE conference on computer vision and pattern recognition workshops, (2018), 773-782. |
[32] |
S. W. Jung, T. H. Kim, S. J. Ko, A novel multiple image deblurring technique using fuzzy projection onto convex sets, IEEE Signal Process. Lett., 16 (2009), 192-195. doi: 10.1109/LSP.2008.2012227
![]() |
[33] | C. Dong, C. C. G. Loy, K. M. He, X. O. Tang, Learning a deep convolutional network for image super-resolution, European conference on computer vision, (2014), 184-199. |
[34] |
D. Sun, Q. Gao, Y. Lu, Z. Huang, T. Li, A novel image denoising algorithm using linear Bayesian MAP estimation based on sparse representation, Signal Process., 100 (2014), 132-145. doi: 10.1016/j.sigpro.2014.01.022
![]() |
[35] | H. Li, Q. Zheng, J. Zhang, Z. Du, Z. Li, B. Kang, Pix2Pix-Based Grayscale Image Coloring Method, J. Comput.-Aided Comput. Graphics, 33 (2021), 929-938. |
[36] | J. Kim, J. K. Lee, K. M. Lee, Accurate image super-resolution using very deep convolutional networks, Proceedings of the IEEE conference on computer vision and pattern recognition, (2016), 1646-1654. |
[37] | K. Yu, L. Lin, M. Alazab, L. Tan, B. Gu, Deep learning-based traffic safety solution for a mixture of autonomous and manual vehicles in a 5G-enabled intelligent transportation system, IEEE Trans. Intell. Transp. Syst., 22 (2020), 4337-4347. |
[38] |
K. Yu, L. Tan, M. Aloqaily, H. Yang, Y. Jararweh, Blockchain-enhanced data sharing with traceable and direct revocation in IIoT, IEEE Trans. Ind. Inf., 17 (2021), 7669-7678. doi: 10.1109/TII.2021.3049141
![]() |
[39] | C. Y. Ma, J. W. Zhu, Y. J. Li, J. R. Li, Y. Jiang, X. Li, Single image super resolution via wavelet transform fusion and SRFeat network, J. Ambient Intell. Hum. Comput., (2020), 1-9. |
[40] |
K. Yu, M. Arifuzzaman, Z. Wen, D. Zhang, T. Sato, A Key Management Scheme for Secure Communications of Information Centric Advanced Metering Infrastructure in Smart Grid, IEEE Trans. Instrum. Meas., 64 (2015), 2072-2085. doi: 10.1109/TIM.2015.2444238
![]() |
[41] | L. Tan, K. Yu, A. K. Bashir, X. Cheng, F. Ming, L. Zhao, et al., Towards Real-time and Efficient Cardiovascular Monitoring for COVID-19 Patients by 5G-Enabled Wearable Medical Devices: A Deep Learning Approach, Neural Compu. Appl., 2021. Available from: https://doi.org/10.1007/s00521-021-06219-9. |
[42] | M. A. Talab, S. Awang, S. A. M. Najim, Super-low resolution face recognition using integrated efficient sub-pixel convolutional neural network (ESPCN) and convolutional neural network (CNN), 2019 IEEE International Conference on Automatic Control and Intelligent Systems (I2CACIS), (2019), 331-335. |
[43] | C. Ledig, L. Theis, F. Huszár, J. Caballero, A. Cunningham, A. Acosta, et al., Photo-realistic single image super-resolution using a generative adversarial network, Proceedings of the IEEE conference on computer vision and pattern recognition, (2017), 4681-4690. |
[44] |
A. Wang, Z. Fang, Y. Gao, X. Jiang, S. Ma, Depth estimation of video sequences with perceptual losses, IEEE Access, 6 (2018), 30536-30546. doi: 10.1109/ACCESS.2018.2846546
![]() |
[45] | B. Lim, S. Son, H. Kim, S. Nah, K. Mu Lee, Enhanced deep residual networks for single image super-resolution, Proceedings of the IEEE conference on computer vision and pattern recognition workshops, (2017), 136-144. |
[46] | L. Tan, N. Shi, K. Yu, M. Aloqaily, Y. Jararweh, A Blockchain-Empowered Access Control Framework for Smart Devices in Green Internet of Things, ACM Trans. Internet Technol., 21 (2021), 1-20. |
[47] | Z. Guo, K. Yu, A. Jolfaei, A. K. Bashir, A. O. Almagrabi, N. Kumar, A Fuzzy Detection System for Rumors through Explainable Adaptive Learning, IEEE Trans. Fuzzy Syst., 2021. Available from: https://doi.org/10.1109/TFUZZ.2021.3052109. |
[48] |
L. Guo, M. Woźniak, An image super-resolution reconstruction method with single frame character based on wavelet neural network in internet of things, Mobile Networks Appl., 26 (2021), 390-403. doi: 10.1007/s11036-020-01681-6
![]() |
[49] | W. Shi, J. Caballero, F. Huszár, J. Totz, A. P. Aitken, R. Bishop, et al., Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network, Proceedings of the IEEE conference on computer vision and pattern recognition, (2016), 1874-1883. |
[50] |
A. Creswell, T. White, V. Dumoulin, K. Arulkumaran, V. Sengupta, A. A. Bharath, Generative adversarial networks: An overview, IEEE Signal Process. Mag., 35 (2018), 53-65. doi: 10.1109/MSP.2017.2765202
![]() |
[51] | M. Yu, H. Wang, M. Liu, P. Li, Overview of Research on Image Super-Resolution Reconstruction, 2021 IEEE International Conference on Information Communication and Software Engineering (ICICSE), (2021), 131-135. |
[52] | S. Lei, X. Liao, Z. Tao, Content-aware Upsampling for Single Image Super-resolution, 2020 Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), (2020), 213-217. |
[53] | S. E. El-Khamy, M. M. Hadboud, M. I. Dessouky, B. M. Salam, F. E. A. El-Samie, A new super-resolution image reconstruction algorithm based on wavelet fusion, Proceedings of the Twenty-Second National Radio Science Conference, 2005. NRSC 2005., (2005), 195-204. |
[54] |
D. Mualfah, Y. Fatma, R. Ramadhan, Anti-forensics: The image asymmetry key and single layer perceptron for digital data security, Journal of Physics: Conference Series, 1517 (2020), 012106. doi: 10.1088/1742-6596/1517/1/012106
![]() |
[55] | J. Johnson, A. Alahi, L. Fei-Fei, Perceptual losses for real-time style transfer and super-resolution, European conference on computer vision, (2016), 694-711. |
[56] |
K. Fu, J Peng, H. Zhang, X. Wang, J. Frank, Image super-resolution based on generative adversarial networks: a brief review, Comput. Mater. Continua, 64 (2020), 1977-1997. doi: 10.32604/cmc.2020.09882
![]() |
[57] | M. Heon, J. H. Kim, J. H. Choi, J. S. Lee, Generative adversarial network-based image super-resolution using perceptual content losses, Proceedings of the European Conference on Computer Vision (ECCV) Workshops, 2018. |
[58] |
Z. Guo, L. Tang, T. Guo, K. Yu, M. Alazab, A. Shalaginov, Deep graph neural network-based spammer detection under the perspective of heterogeneous cyberspace, Future Gener. Comput. Syst., 117 (2021), 205-218. doi: 10.1016/j.future.2020.11.028
![]() |
[59] | K. Yu, L. Tan, X. Shang, J. Huang, G. Srivastav, P. Chatterjee, Efficient and Privacy-Preserving Medical Research Support Platform Against COVID-19: A Blockchain-Based Approach, IEEE Consum. Electron. Mag., 10 (2021), 111-120. |
[60] |
D. Lee, S. Lee, H. Lee, K. Lee, H. J. Lee, Resolution-preserving generative adversarial networks for image enhancement, IEEE Access, 7 (2019), 110344-110357. doi: 10.1109/ACCESS.2019.2934320
![]() |
[61] | C. F. Song, Y. Huang, W. L. Ouyang, L. Wang, Mask-guided contrastive attention model for person re-identification, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (2018), 1179-1188. |
[62] | A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, et al., Attention is all you need, in Advances in neural information processing systems, (2017), 5998-6008. |
[63] | A. N. Moldovan, I. Ghergulescu, C. H. Muntean, A novel methodology for mapping objective video quality metrics to the subjective MOS scale, 2014 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting, (2014), 1-7. |
[64] |
Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli, Image quality assessment: from error visibility to structural similarity, IEEE Trans. Image Process., 13 (2004), 600-612. doi: 10.1109/TIP.2003.819861
![]() |
[65] |
K. Yu, L. Tan, L. Lin, X. Cheng, Z. Yi, T. Sato, Deep Learning Empowered Breast Cancer Auxiliary Diagnosis for 5GB Remote E-Health, IEEE Wireless Commun., 28 (2021), 54-61. doi: 10.1109/MWC.001.2000374
![]() |
[66] | L. Tan, K. Yu, F. Ming, X. Cheng, G. Srivastava, Secure and Resilient Artificial Intelligence of Things: a HoneyNet Approach for Threat Detection and Situational Awareness, IEEE Consum. Electron. Mag., 2021. Available from: https://doi.org/10.1109/MCE.2021.3081874. |
[67] | Z. Guo, K. Yu, Y. Li, G. Srivastava, J. C. W. Lin, Deep Learning-Embedded Social Internet of Things for Ambiguity-Aware Social Recommendations, IEEE Trans. Network Sci. Eng., 2021. Available from: https://doi.org/10.1109/TNSE.2021.3049262. |
[68] | K. Yu, Z. Guo, Y. Shen, W. Wang, J. C. Lin, T. Sato, Secure Artificial Intelligence of Things for Implicit Group Recommendations, IEEE Internet Things J., 2021. Available from: http://dx.doi.org/10.1109/JIOT.2021.3079574. |
1. | Noman Zahid, Ali Hassan Sodhro, Usman Rauf Kamboh, Ahmed Alkhayyat, Lei Wang, AI-driven adaptive reliable and sustainable approach for internet of things enabled healthcare system, 2022, 19, 1551-0018, 3953, 10.3934/mbe.2022182 | |
2. | Hongan Li, Guanyi Wang, Qiaozhi Hua, Zheng Wen, Zhanli Li, Ting Lei, An image watermark removal method for secure internet of things applications based on federated learning, 2022, 0266-4720, 10.1111/exsy.13036 | |
3. | Hong’an Li, Jiangwen Fan, Qiaozhi Hua, Xinpeng Li, Zheng Wen, Meng Yang, Biomedical sensor image segmentation algorithm based on improved fully convolutional network, 2022, 197, 02632241, 111307, 10.1016/j.measurement.2022.111307 | |
4. | Xiaonan Shi, Jian Huang, Bo Huang, An Underground Abnormal Behavior Recognition Method Based on an Optimized Alphapose-ST-GCN, 2022, 31, 0218-1266, 10.1142/S0218126622502140 | |
5. | Xiao Chen, Ruyun Jing, Chaowen Sun, Attention mechanism feedback network for image super-resolution, 2022, 31, 1017-9909, 10.1117/1.JEI.31.4.043006 | |
6. | Qingfang Liu, Baosheng Kang, Qiaozhi Hua, Zheng Wen, Haipeng Li, G. Thippa Reddy, Visual Attention and Motion Estimation-Based Video Retargeting for Medical Data Security, 2022, 2022, 1939-0122, 1, 10.1155/2022/1343766 | |
7. | Youzhong Ma, Qiaozhi Hua, Zheng Wen, Ruiling Zhang, Yongxin Zhang, Haipeng Li, Thippa Reddy G, k Nearest Neighbor Similarity Join Algorithm on High-Dimensional Data Using Novel Partitioning Strategy, 2022, 2022, 1939-0122, 1, 10.1155/2022/1249393 | |
8. | Yunyan Xu, Fu-Sheng Tsai, The Strategy of How to Deeply Integrate Technology and Finance in the Internet Environment, 2022, 2022, 1687-9813, 1, 10.1155/2022/5018160 | |
9. | Khushboo Singla, Rajoo Pandey, Umesh Ghanekar, A review on Single Image Super Resolution techniques using generative adversarial network, 2022, 266, 00304026, 169607, 10.1016/j.ijleo.2022.169607 | |
10. | Hong-an Li, Guanyi Wang, Kun Gao, Haipeng Li, A Gated Convolution and Self-Attention-Based Pyramid Image Inpainting Network, 2022, 31, 0218-1266, 10.1142/S0218126622502085 | |
11. | Peng Zhao, Yongxin Zhang, Qiaozhi Hua, Haipeng Li, Zheng Wen, Bio-Inspired Optimal Dispatching of Wind Power Consumption Considering Multi-Time Scale Demand Response and High-Energy Load Participation, 2023, 134, 1526-1506, 957, 10.32604/cmes.2022.021783 | |
12. | Zhi-Hua zhao, Li Chen, Bifurcation Fusion Network for RGB-D Salient Object Detection, 2022, 31, 0218-1266, 10.1142/S0218126622502152 | |
13. | Ya Zhou, Fu-Sheng Tsai, The Optimization Strategy of College Students’ Ideological and Political Management under the Internet+ Environment, 2022, 2022, 1687-9813, 1, 10.1155/2022/2556976 | |
14. | Hong-an Li, Liuqing Hu, Qiaozhi Hua, Meng Yang, Xinpeng Li, Image Inpainting Based on Contextual Coherent Attention GAN, 2022, 31, 0218-1266, 10.1142/S0218126622502097 | |
15. | Mei Gao, Baosheng Kang, Blind Image Inpainting Using Low-Dimensional Manifold Regularization, 2022, 31, 0218-1266, 10.1142/S0218126622502115 | |
16. | Zengle Li, Bin Zhi, Enlong Liu, Analysis of the Deformation Characteristics of Loess Based on Thermal–Mechanical Coupling under an Energy Internet, 2022, 31, 0218-1266, 10.1142/S021812662250116X | |
17. | Qianqian Liu, Xiaoyan Zhang, Qiaozhi Hua, Zheng Wen, Haipeng Li, Yan Huo, Adaptive Differential Evolution Algorithm with Simulated Annealing for Security of IoT Ecosystems, 2022, 2022, 1530-8677, 1, 10.1155/2022/6951849 | |
18. | Hong-an Li, Min Zhang, Zhenhua Yu, Zhanli Li, Na Li, An improved pix2pix model based on Gabor filter for robust color image rendering, 2022, 19, 1551-0018, 86, 10.3934/mbe.2022004 | |
19. | Peng Zhao, Dan Zhang, Shengling Geng, Mingquan Zhou, Two-Stage Generator Network for High-Quality Image Inpainting in Future Internet, 2023, 12, 2079-9292, 1490, 10.3390/electronics12061490 | |
20. | Hong-an Li, Lanye Wang, Jun Liu, Application of multi-level adaptive neural network based on optimization algorithm in image style transfer, 2024, 83, 1573-7721, 73127, 10.1007/s11042-024-18451-1 | |
21. | Hong'an Li, Man Liu, Jiangwen Fan, Qingfang Liu, Biomedical image segmentation algorithm based on dense atrous convolution, 2024, 21, 1551-0018, 4351, 10.3934/mbe.2024192 | |
22. | Cai Fujun, Liu Zhenliang, RETRACTED ARTICLE: Application of image optical processing technology based on computer vision in image simulation of oil painting teaching, 2024, 56, 1572-817X, 10.1007/s11082-023-06138-0 | |
23. | Hong-an Li, Diao Wang, Min Zhang, Jun Liu, Image color rendering based on frequency channel attention GAN, 2024, 18, 1863-1703, 3179, 10.1007/s11760-023-02980-7 | |
24. | Xiao Sun, Song Wang, Jucheng Yang, Feng Wei, Yuan Wang, Two-Stage Deep Single-Image Super-Resolution With Multiple Blur Kernels for Internet of Things, 2023, 10, 2327-4662, 16440, 10.1109/JIOT.2023.3268285 | |
25. | Shushu Yin, Hefan Li, Yu Sang, Tianjiao Ma, Tie Li, Mei Jia, 2023, Chapter 12, 978-981-99-5967-9, 165, 10.1007/978-981-99-5968-6_12 | |
26. | Bogdan Kovalenko, Vladimir Lukin, 2023, BPG-based compression of Poisson noisy images, 979-8-3503-9611-9, 1, 10.1109/DESSERT61349.2023.10416544 | |
27. | Xiaoyan Zhang, Mengmeng He, Hongan Li, DAU-Net: A medical image segmentation network combining the Hadamard product and dual scale attention gate, 2024, 21, 1551-0018, 2753, 10.3934/mbe.2024122 | |
28. | Marwan Farag, Friedhelm Schwenker, 2023, Chapter 9, 978-3-031-31326-4, 154, 10.1007/978-3-031-31327-1_9 | |
29. | Hongan Li, Lizhi Cheng, Jun Liu, A new degradation model and an improved SRGAN for multi-image super-resolution reconstruction, 2024, 1368-2199, 1, 10.1080/13682199.2024.2331813 |
Layer name | Output size | Parameter | Numbers |
Conv_1 | 24 × 24 | [9 × 9, 64] | × 1 |
SA_1 | 24 × 24 | [1 × 1, 8] [1×1, 8] [1×1, 64] | × 1 |
ResConv | 24 × 24 | [3×3, 64] [3×3, 64] | × 16 |
Conv_2 | 24 × 24 | [3 × 3, 64] | × 1 |
SA_2 | 24 × 24 | [1 × 1, 8] [1 × 1, 8] [1 × 1, 64] | × 1 |
SubpixelConv | 96 × 96 | [3 × 3, 256] | × 2 |
Conv_3 | 96 × 96 | [9 × 9, 3] | × 1 |
Layer name | Output size | Parameter |
Conv_1 | 96 × 96 | [3 × 3, 64] |
Conv_2 | 48 × 48 | [3 × 3, 64] |
Conv_3 | 48 × 48 | [3 × 3, 128] |
Conv_4 | 24 × 24 | [3 × 3, 128] |
Conv_5 | 24 × 24 | [3 × 3, 256] |
Conv_6 | 12 × 12 | [3 × 3, 256] |
Conv_7 | 12 × 12 | [3 × 3, 512] |
Conv_8 | 6 × 6 | [3 × 3, 512] |
AdaptiveAvgPool | 6 × 6 | - |
FC_1 | 1024 | - |
FC_2 | 1 | - |
Leval | Inpairment | Score |
1 | Very annoying | 1-2 |
2 | Annoying | 3-4 |
3 | Slightly annoying | 5-6 |
4 | Perceptible, but not annoying | 7-8 |
5 | Imperceptible | 9-10 |
Method | Group1 | Group2 | Group3 | Group4 |
Bicubic | 4.200 | 4.230 | 4.166 | 4.067 |
SRGAN | 6.833 | 7.166 | 6.833 | 6.710 |
SA+SRGAN(Ours) | 7.066 | 7.533 | 7.250 | 6.900 |
Method | Set5 | Set14 | BSD100 | |||
PSNR | SSIM | PSNR | SSIM | PSNR | SSIM | |
Bicubic | 27.067 | 0.749 | 24.502 | 0.682 | 23.924 | 0.626 |
SRGAN | 28.072 | 0.843 | 26.057 | 0.731 | 25.490 | 0.686 |
SA+SRGAN(Ours) | 28.745 | 0.854 | 26.656 | 0.745 | 26.037 | 0.697 |
Layer name | Output size | Parameter | Numbers |
Conv_1 | 24 × 24 | [9 × 9, 64] | × 1 |
SA_1 | 24 × 24 | [1 × 1, 8] [1×1, 8] [1×1, 64] | × 1 |
ResConv | 24 × 24 | [3×3, 64] [3×3, 64] | × 16 |
Conv_2 | 24 × 24 | [3 × 3, 64] | × 1 |
SA_2 | 24 × 24 | [1 × 1, 8] [1 × 1, 8] [1 × 1, 64] | × 1 |
SubpixelConv | 96 × 96 | [3 × 3, 256] | × 2 |
Conv_3 | 96 × 96 | [9 × 9, 3] | × 1 |
Layer name | Output size | Parameter |
Conv_1 | 96 × 96 | [3 × 3, 64] |
Conv_2 | 48 × 48 | [3 × 3, 64] |
Conv_3 | 48 × 48 | [3 × 3, 128] |
Conv_4 | 24 × 24 | [3 × 3, 128] |
Conv_5 | 24 × 24 | [3 × 3, 256] |
Conv_6 | 12 × 12 | [3 × 3, 256] |
Conv_7 | 12 × 12 | [3 × 3, 512] |
Conv_8 | 6 × 6 | [3 × 3, 512] |
AdaptiveAvgPool | 6 × 6 | - |
FC_1 | 1024 | - |
FC_2 | 1 | - |
Leval | Inpairment | Score |
1 | Very annoying | 1-2 |
2 | Annoying | 3-4 |
3 | Slightly annoying | 5-6 |
4 | Perceptible, but not annoying | 7-8 |
5 | Imperceptible | 9-10 |
Method | Group1 | Group2 | Group3 | Group4 |
Bicubic | 4.200 | 4.230 | 4.166 | 4.067 |
SRGAN | 6.833 | 7.166 | 6.833 | 6.710 |
SA+SRGAN(Ours) | 7.066 | 7.533 | 7.250 | 6.900 |
Method | Set5 | Set14 | BSD100 | |||
PSNR | SSIM | PSNR | SSIM | PSNR | SSIM | |
Bicubic | 27.067 | 0.749 | 24.502 | 0.682 | 23.924 | 0.626 |
SRGAN | 28.072 | 0.843 | 26.057 | 0.731 | 25.490 | 0.686 |
SA+SRGAN(Ours) | 28.745 | 0.854 | 26.656 | 0.745 | 26.037 | 0.697 |