This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
The main reason of motion blur is that there is rapid relative motion between the camera and the captured object during the exposure time. The blurring of images will reduce the perceptual quality of human beings. It also has a negative impact on advanced visual tasks such as object detection and semantic understanding. Image deblurring is a common and important problem in the field of image processing and computer vision. However, due to the complexity of motion blur processing, most existing methods may not produce satisfactory results when the blur kernel is complex, and the details of the required clear image are abundant. In addition, because the infrared (IR) imaging system is more complex than the natural imaging system, the degradation degree of infrared images is relatively high, such as Gaussian blur, motion blur, and noise pollution. Therefore, infrared image deblurring plays an important role in the IR imaging system. Some researchers are dedicated to hardware-based research for infrared image deblurring. In literature [1], the fluttered shutter is used to solve the problem of infrared image deblurring. Literature [2] uses an ordinary inertial measurement unit (IMU) to estimate the trajectory of the camera movement during the exposure time. Oswald-Tranta et al. [3] used the parameterized Wiener filter method to blur the infrared images obtained from the infrared detector of the microbolometer. Oswald-Tranta also committed to obtaining accurate temperature measurements by deblurring infrared images [4]. Wang et al. [5] used the iterative Wiener filter to estimate the PSF filter of motion blur in infrared images. The deblurring method based on infrared imaging hardware equipment is more expensive. Therefore, the algorithm-based deblurring of infrared images is more widely used. Luo et al. [6] developed a new infrared blurred image restoration model based on the principle of nonuniform exposure. In order to eliminate the motion blur of the image and restore the image, Jing et al. [7] proposed an infrared target motion deblurring method based on the Haar wavelet transform. Liua et al. [8] proposed a method of using Lp-quasi-linear norm and the overlapping sparse total variation method to blur infrared images.
Inspired by the great progress of traditional blind deblurring methods and learning-based blind deblurring methods recently, we propose a method based on GAN and channel prior discrimination. Specifically, the innovation of this article is summarized as follows:
(i) A channel-based inverse prior discrimination is proposed. And this method is built into a new framework of the GAN. It improves the blind deblurring performance of infrared images.
(ii) Different blur types are caused by the motion of the camera or object. In view of this situation, two different methods were used to synthesize two kinds of blurred datasets.
(iii) In the experimental stage, we conducted extensive experiments which were carried out on two different datasets. The method proposed in this article is compared with the other four advanced methods qualitatively and quantitatively.
2. Related Work
2.1. Image Deblurring
The solutions to deblurring problems are mainly divided into two types: blind deblurring and nonblind deblurring. The early related work is mainly nonblind deblurring, that is, the ambiguity function is assumed to be known. Most of these algorithms rely on the Lucy–Richardson algorithm and Wiener or Tikhonov filter which are sensitive to noise to perform deconvolution operation and obtain IS estimation. However, in reality, ambiguity functions are often uncertain. It is unrealistic to find the ambiguity function for each pixel. Therefore, a lot of recent works are focused on blind deblurring. The first modern bold attempt was Fergus et al.’s [9] variational Bayesian method to eliminate uniform camera shake. In the past decade, many methods [10–20] have solved the blur caused by camera shake by considering the uniform blur on the image. This kind of algorithm first estimates camera motion according to the induced blur kernel and then reverses the effect by performing deconvolution operation. Unfortunately, these algorithms are usually unable to eliminate nonuniform motion blur.
In fact, due to the camera rotation, radial camera motion, depth of field change, or rapid movement of objects, images taken in the field may experience more complex heterogeneous blur. Therefore, most existing nonuniform blind deblurring methods [21–26] are based on specific motion models. For example, Gupta et al. [27] proposed to model camera motion as a motion density function. The blurring kernel of spatial variables can be derived directly from it. By specifying a prior of sparsity and compactness in density, an optimization problem is formulated, and the density function and deblurred image can be solved iteratively. A new projection motion path model is proposed in [28, 29]. Another method to eliminate spatial variation ambiguity is to estimate through block-by-block blurring kernel [30–32]. Segmented blurring estimation [24, 33] also considers the spatial variation blur caused by the object movement.
In recent years, some methods based on the convolutional neural network (CNN) have appeared [23, 34–42]. Schuler et al. [39] made the first heuristic attempt, focusing on unified blind deblurring, including modules for feature extraction, blurring kernel estimation, and clear image estimation. Sun et al. [40] used the CNN to estimate the blurring kernel. Chakrabarti [43] put forward another advanced method. This method learnt to predict the plural Fourier coefficients of the deconvolution filter of the input patches of blurred images and then used the traditional optimization strategy to estimate the global blurring kernel from the restored patches. And Gong et al. [34] used the fully convolved network movement flow to estimate. All these methods use the CNN to estimate unknown ambiguity functions. Recently, Noroozi et al. [23] and Nah et al. [44] adopted the kernel-free end-to-end method, using the multiscale CNN to directly remove images. Tao et al.’s latest work [42] expands the multiscale CNN from [37] to scale the recursive CNN to realize image deblurring, and the effect is impressive. Ramakrishnan et al. [38] used a combination of pix2pix framework [45] and densely connected convolution network [46] to perform blind kernel-free image deblurring. These methods can deal with different sources of blur. Since Ramakrishnan et al., the success of GAN in image restoration has also affected the deblurring of single image. Ramakrishnan et al. [38] firstly solved the problem of image deblurring by referring to the idea of image translation [45]. Recently, Kupyn et al. [36] introduced DeblurGAN; it is developed by Wasserstein GAN [47] with gradient penalty and perceived loss.
2.2. GAN
Generative adversarial network, commonly known as GAN, was proposed by Goodfellow [48] and inspired by the zero-sum game in game theory. This game has achieved many exciting results in image restoration [49]. After style conversion [45, 50, 51], it can even be used in other fields. The system includes a generator G and a discriminator D; they constitute a minmax game for two. The generator tries to capture the potential actual data distribution and outputs new data samples, while the discriminator tries to distinguish whether the input data come from the real data distribution. The minmax game with the value function V(G, D) is represented by the following formula [1]. Both generator and discriminator can be constructed based on the CNN and trained based on the above ideas.
GAN is known for its ability to preserve textural details in images, create solutions that are close to the real image, and be perceptually persuasive. Literature [51] was further developed; it is based on the conditional GAN [52] and trains a cyclic consistency goal. This target generates a more realistic image in the task of image migration. Inspired by this idea, Isola [45] put forward the earliest idea of image deblurring based on the GAN. Recently, great progress has been made in the related fields of image super-resolution [53] and image restoration [54] by applying the GAN.
2.3. Dark Channel Prior Algorithm
He et al. [55] proposed a defogging algorithm (DCP) based on the dark channel prior. DCP is based on the assumption that most nonsky patches of outdoor fog-free images contain some pixels. These pixels have very low intensity in at least one color channel. For any image I, its dark channel Idark (x) is given by the following formula:
Many methods use dark channels and bright channels to complete image defogging [55, 56], and they are also used to estimate the blurring kernel in conventional blind image deblurring [15, 57]. In [15], Pan et al. proposed to use the regularization term based on L0 additionally on the dark channel image to improve the gradient-based L0-minimization blind deblurring method [11]. In [57], Yan et al. further combined and used L0-based regularization in both dark and bright channel images.
3. Method
In this work, the purpose of the infrared image deblurring model is to restore a clear image when only the blurred infrared image is given. In this paper, the architecture, proposed in [51], is used to build two sets of GAN models. The generators are GB2S: IB⟶IS and GS2B: IS⟶IB. GB2S restores clear images from blurred images, while GS2B generates blurred images from clear images. The discriminators are DB and DS. DB tries to distinguish whether the input is a blurred image, while DS tries to distinguish whether the input is sharp. The architecture of the proposed method is shown in Figure 1. The input in the method is the blurred image and clear image. The clear image is sent to the generator GS2B to generate the corresponding blurred image. The generated blurred image is sent to the generator GB2S to generate a deblurred image. The generated deblurred image and the real clear image are sent to the discriminator DS together to identify true and fake. The real blurred image is input into the generator GB2S to generate a deblurred image. The generated deblurred image is sent to the generator GS2B to synthesize the blurred image. The synthesized blurred image and the real blurred image are sent to the discriminator DB to determine the authenticity. Through continuous iteration, the generator can generate more realistic deblurred images. The algorithm flow is summarized as Algorithm 1.
[figure omitted; refer to PDF]
As we all know, both BN and IN layers use a batch of mean and variance to normalize features during training and use the estimated mean and variance of the whole training dataset during testing. One of the potential motivations for applying BN or IN is to accelerate the training of deep neural networks (DNNs). However, recent work [58] on single-image super-resolution points out that the BN layer will bring artifacts in training and testing stages. Especially, these artifacts are more likely to occur with the deepening of the network and training under the framework of the GAN. When turned to blind deblurring, the above empirical discussion shows that the IN layer will bring similar artifacts, that is, irregular block color shift. Therefore, no IN or BN layer is introduced in the residual block, as shown in Figure 3. The network configuration of the generator and discriminator is shown in Tables 1 and 2.
[figure omitted; refer to PDF]
Table 1
The layer structure and parameters of the generator.
Layer (type) | Output shape | Parameters |
ReflectionPad2d-1 | [−1, 3, 262, 262] | 0 |
Conv2d-2 | [−1, 64, 256, 256] | 1,792 |
InstanceNorm2d-3 | [−1, 64, 256, 256] | 0 |
ReLU-4 | [−1, 64, 256, 256] | 0 |
Conv2d-5 | [−1, 128, 128, 128] | 73,856 |
InstanceNorm2d-6 | [−1, 128, 128, 128] | 0 |
ReLU-7 | [−1, 128, 128, 128] | 0 |
Conv2d-8 | [−1, 256, 64, 64] | 295,168 |
InstanceNorm2d-9 | [−1, 256, 64, 64] | 0 |
ReLU-10 | [−1, 256, 64, 64] | 0 |
ReflectionPad2d-11 | [−1, 256, 66, 66] | 0 |
Conv2d-12 | [−1, 256, 64, 64] | 590,080 |
ReLU-13 | [−1, 256, 64, 64] | 0 |
ReflectionPad2d-14 | [−1, 256, 66, 66] | 0 |
Conv2d-15 | [−1, 256, 64, 64] | 590,080 |
ResidualBlock-16 | [−1, 256, 64, 64] | 0 |
ConvTranspose2d-65 | [−1, 128, 128, 128] | 295,040 |
InstanceNorm2d-66 | [−1, 128, 128, 128] | 0 |
ReLU-67 | [−1, 128, 128, 128] | 0 |
ConvTranspose2d-68 | [−1, 64, 256, 256] | 73,792 |
InstanceNorm2d-69 | [−1, 64, 256, 256] | 0 |
ReLU-70 | [−1, 64, 256, 256] | 0 |
ReflectionPad2d-71 | [−1, 64, 262, 262] | 0 |
Conv2d-72 | [−1, 3, 256, 256] | 1,731 |
Tanh-73 | [−1, 3, 256, 256] | 0 |
Table 2
The layer structure and parameters of the discriminator.
Layer (type) | Output shape | Parameters |
Conv2d-1 | [−1, 64, 128, 128] | 1,792 |
LeakyReLU-2 | [−1, 64, 128, 128] | 0 |
Conv2d-3 | [−1, 128, 64, 64] | 73,856 |
InstanceNorm2d-4 | [−1, 128, 64, 64] | 0 |
LeakyReLU-5 | [−1, 128, 64, 64] | 0 |
Conv2d-6 | [−1, 256, 32, 32] | 295,168 |
InstanceNorm2d-7 | [−1, 256, 32, 32] | 0 |
LeakyReLU-8 | [−1, 256, 32, 32] | 0 |
Conv2d-9 | [−1, 512, 31, 31] | 1,180,160 |
InstanceNorm2d-10 | [−1, 512, 31, 31] | 0 |
LeakyReLU-11 | [−1, 512, 31, 31] | 0 |
Conv2d-12 | [−1, 1, 30, 30] | 4,609 |
3.2. Loss Function
3.2.1. Adversarial Loss
Adversarial loss includes generator adversarial loss and discriminator adversarial loss, where generator adversarial loss is defined as follows:
Among them, the first item is the adversarial loss between the reconstructed blurred image
Among them, the first term is the loss function of the discriminator DB error identification, and the second term is the loss function of the discriminator DS error identification.
3.2.2. Loss of Circular Perception Consistency
For the general GAN, it is necessary to compare the reconstructed image and the original image in the training stage with a certain metric as content loss. The common choice of content loss is pixel-space loss, and the simplest is L1 or L2 loss. Because this kind of loss often produces excessively smooth pixel-space output, this leads to blurring artifacts on the generated image. This brings negative factors to the deblurring task, so the circular perception consistency loss suggested in [58] is adopted. The purpose of circular perception consistency loss is to preserve the original image structure by looking at the combination of high-level and low-level features extracted from the second and fifth pooling layers of the VGG-16 system [59]. Under the constraints of generator GB2S: IB⟶IS and generator GS2B: IS⟶IB, the following formula of circular perception consistency loss is given:
Among them,
3.2.3. Prior Loss Based on the Dark Channel and Bright Channel
Using the bright channel and dark channel presented in formulas (2) and (3), the following two different energies are defined:
[figures omitted; refer to PDF]
Based on this conclusion, it is considered that clear images and blurred images can be distinguished by dark energy and bright energy defined in (9) and (10). In order to improve the GAN from the perspective of domain knowledge, the prior judgment of the traditional blind image deblurring method is taken as the training loss function:
Combining formulas (4)–(12), the final losses adopted in this article are as follows:
In formula (13),
4. Experiment
All models are implemented by the PyTorch deep learning framework. FLIR_ADAS_1_3 dataset and LTIR dataset are used to train on a desktop with 2.20 GHz × 40 Intel Xeon (r) Silver4114 CPU, GeForce GTX 1080Ti, and 64GiB memory. In this section, the experimental results are introduced and compared with the results of mainstream methods. In addition, qualitative results are provided on real images.
4.1. Synthetic Blurring Dataset
There are two types of blurred images: the overall image is blurred due to the movement of the imaging device, and the partial image is blurred due to the movement of the imaging object. In order to verify that our deblurring method is effective for both types of blur, we simulate the two types of image blur through two different schemes.
For the overall image blur caused by the motion of the imaging device, we choose to use a linear blur kernel to create a synthetic blur image. Sun et al. [40] created a composite blurred image by convolving a clear natural image with one of 73 possible linear motion kernels. Xu et al. [60] also used the linear motion kernel to create synthetic blurred images. Chakrabarti [61] created a blurring kernel by sampling six random points and fitting splines to them. Levin et al. provided eight blurring kernels [62] that have been used for multiple datasets. However, the maximum blurring kernel size of these eight blurring kernels is 41 × 41, which is relatively small in practice. Therefore, we follow the algorithm in [63] to generate four uniform blur kernels from 51 × 51 to 101 × 101 by sampling random 6D camera trajectories. Then, a convolution model with 1% Gaussian noise is used to synthesize a blurred image.
For the local image blur caused by the motion of the imaging object, we choose to use the average frame of the video sequence to simulate. This is a typical method of simulating blurred image pairs [23, 37]. This method can create realistic blurred images but only limits the image space to scenes with video sequences; this makes the dataset limited. Figure 5 shows a comparison of two different blur types. The blurred image generated by averaging frames shows the blur caused by moving objects and static background. The car in Figure 5(b) is blurred, but the surrounding trees are clear. The blur kernel method simulates the motion blur of the whole image caused by the motion of the camera. In Figure 5(c), the car and the surrounding trees are blurred. In order to verify the universality of our algorithm, we use the blur kernel to synthesize blurred images for the LTIR dataset and use two synthetic methods of average frame and the blur kernel for the FLIR dataset to simulate motion blur. The blurred dataset synthesized by the blur kernel method is used as the FLIR-A dataset; the blurred dataset synthesized by the average frame method is used as the FLIR-B dataset.
[figures omitted; refer to PDF]
4.2. FLIR_ADAS_1_3 Dataset Results
FLIR_ADAS_1_3 datasets provide annotated thermal imaging datasets and corresponding unannotated RGB images for training and verifying neural networks. Data are acquired by using the RGB camera and thermal imaging camera installed on the vehicle. The dataset contains a total of 14,452 infrared images, of which 10,228 are from multiple short videos, and 4224 are from a video with a length of 144 s. All videos come from streets and highways. The sampling rate of most pictures is two frames per second. The frame rate of the video is 30 frames per second. When there are few targets in a few environments, the sampling rate is 1 frame per second. In the experiment, 8862 8-bit infrared images are divided into 7090 image training sets and 1772 image test sets. Figure 6 shows the test images on the FLIR-A blurred dataset, and the quantitative results are shown in Table 3.
[figure omitted; refer to PDF]Table 3
Comparison of quantitative deblurring performance on FLIR-A datasets.
DeepDeblur | DeblurGAN | CycleGAN | Cycle-Dehaze | Ours | |
SSIM | 0.8916 | 0.9899 | 0.9190 | 0.9788 | 0.9985 |
PSNR (dB) | 17.48 | 26.91 | 20.45 | 21.03 | 28.79 |
Time (s) | 40.03 | 1.05 | 4.59 | 7.01 | 0.14 |
In order to further compare the deblurring effects of various methods on different types of blurred images, we compare the deblurring results of FLIR-A and FLIR-B blurred datasets. Figure 7 shows the deblurred images of different methods on the two types of blurred datasets, and the evaluation indicators are shown in Table 4. It can be seen from the subjective and objective results that our method has better deblurring performance than several other methods. This result is particularly obvious on the FLIR-B blurred dataset. For partially blurred images caused by the motion of the imaging object, the deblurring effect of other methods is significantly reduced, the original clear background becomes more blurred, and the blurred area does not achieve the ideal deblurring effect. However, our method can restore the blurred area clearly while keeping the background clear. This has a lot to do with the idea of channel prior discrimination adopted in our method. The channel prior discrimination algorithm is based on local color patches. This makes our method have better deblurring performance in the local blurred image.
[figures omitted; refer to PDF]
Table 4
Comparison of the deblurring performance of various methods on different types of blurred images.
DeepDeblur | DeblurGAN | CycleGAN | Cycle-Dehaze | Ours | ||
FLIR-A | SSIM | 0.8916 | 0.9899 | 0.9190 | 0.9788 | 0.9985 |
PSNR | 17.48 | 26.91 | 20.45 | 21.03 | 28.79 | |
FLIR-B | SSIM | 0.7458 | 0.8161 | 0.7997 | 0.8364 | 0.9589 |
PSNR | 16.76 | 18.47 | 17.20 | 19.51 | 21.22 |
4.3. LTIR_v1_0 Dataset Results
LTIR dataset is a thermal infrared dataset used to evaluate the tracking of a single object (STSO) in a short time. Currently, only one version is available. Version 1.0 consists of 20 infrared thermal sequences with an average length of 563 frames. This dataset is a subchallenge of the 2015 Visual Object Recognition (VOT) Challenge. In the experiment, 11,262 8-bit images are divided into a training set of 9010 images and a test set of 2252 images. Figure 8 shows the test image on the LTIR dataset. The quantitative results are shown in Table 5.
[figure omitted; refer to PDF]Table 5
Comparison of quantitative deblurring performance on LTIR datasets.
DeepDeblur | DeblurGAN | CycleGAN | Cycle-Dehaze | Ours | |
SSIM | 0.7535 | 0.8576 | 0.6977 | 0.7110 | 0.9697 |
PSNR (dB) | 15.85 | 22.48 | 17.51 | 10.55 | 25.85 |
Time (s) | 37.62 | 0.82 | 3.56 | 5.74 | 0.06 |
4.4. Ablation Research and Analysis
We conduct ablation research on the effect of the loss function component in the deblurring method proposed in this paper. The results are summarized in Table 6. We can see that our proposed dark channel and bright channel a priori determination components are steadily improving PSNR and SSIM. In particular, the dark channel a priori determination module contributes the most. When we replace the perceptual loss function with L1 and L2 loss functions, the average SSIM and PSNR both decrease. It can be seen from Figure 9 that the deblurred image generated after replacing the perceptual loss function with the L1 and L2 loss function is too smooth. In summary, in the deblurring task, the perceptual loss function is more suitable than the L1 and L2 loss functions.
Table 6
Ablation study of the channel prior loss function.
SSIM | PSNR (dB) | |||
FLIR dataset | LTIR dataset | FLIR dataset | LTIR dataset | |
Remove the dark channel prior loss function | 0.9805 | 0.7463 | 21.66 | 13.95 |
Remove the bright channel prior loss function | 0.9823 | 0.8818 | 22.47 | 22.73 |
Replace perceptual loss with L1 loss | 0.9344 | 0.9191 | 19.43 | 20.20 |
Replace perceptual loss with L2 loss | 0.9421 | 0.9256 | 19.25 | 20.64 |
Ours | 0.9985 | 0.9697 | 28.79 | 25.85 |
[figures omitted; refer to PDF]
4.5. Use Advanced Vision Tasks to Compare Deblurring Results
Basic vision tasks, including image deblurring, serve for advanced vision tasks. In order to further verify the effectiveness of our method, we match the deblurred images generated by several methods with real clear images. Scale-Invariant Feature Transformation (SIFT) is a representation of Gaussian image gradient statistics in the field of feature points and is a commonly used image local feature extraction algorithm. In the matching result, the number of matching points can be used as a criterion for matching quality, and the corresponding matching points can also determine the similarity of the local features of the two images. Figure 10 shows the result of matching the deblurred image with the real clear image through the SIFT algorithm. It can be seen from the quantity that the deblurred image produced by our proposed method obtains more correct matching pairs than other methods.
[figures omitted; refer to PDF]
In this experiment, we use the classic YOLO [65] method for deblurring image target detection (Figure 11). As can be seen, the proposed method to generate a blurred image has better detection result, and more targets can be detected.
[figures omitted; refer to PDF]
5. Conclusion
Blind deblurring of a single infrared image is still a challenging computer vision problem. In this work, a method based on the GAN and channel prior discrimination is proposed for the problem of infrared image deblurring. Different from the previous deblurring work, we combine traditional blind deblurring and blind deblurring methods based on learning methods. Considering the different types of blur caused by the motion of the imaging device and the imaging object, extensive experiments were carried out on different public datasets. Experimental results show that the proposed method is more competitive than other popular image deblurring methods in terms of deblurring quality (subjective and objective) and efficiency.
[1] A. Agrawal, Motion Deblurring: Motion Deblurring Using Fluttered Shutter, 2014.
[2] N. Joshi, S. B. Kang, C. Lawrence Zitnick, R. Szeliski, "Image deblurring using inertial measurement sensors," ACM Transactions on Graphics, vol. 29 no. 4CD, pp. 30-39, DOI: 10.1145/1833349.1778767, 2010.
[3] B. Oswald-Tranta, M. Sorger, P. O’Leary, "Motion deblurring of infrared images from a microbolometer camera," Infrared Physics & Technology, vol. 53 no. 4, pp. 274-279, DOI: 10.1016/j.infrared.2010.04.003, 2010.
[4] B. Oswald-Tranta, "Temperature reconstruction of infrared images with motion deblurring," Journal of Sensors and Sensor Systems, vol. 7 no. 1, pp. 13-20, DOI: 10.5194/jsss-7-13-2018, 2018.
[5] N. Wang, W. Jing, Y. Zhang, X. Sun, "Restoration of the infrared image blurred by motion," Proceedings of the 2016 SPIE Society of Photo-optical Instrumentation Engineers,DOI: 10.1117/12.2268426, .
[6] Y. Luo, T. Xu, N. Wang, F. Liu, "Restoration of non-uniform exposure motion blurred image," Proceedings of the 2014 International Symposium on Optoelectronic Technology & Application,DOI: 10.1117/12.2082837, .
[7] L. I. Jing, M. Wang, J. Sha, B. Xujmet, "Research on wavelet transform based motion deblurring method of infrared target," 2016.
[8] X. Liua, Y. Chena, Z. Penga, J. Wu, "Total variation with overlapping group sparsity and lp quasinorm for infrared image deblurring under salt-and-pepper noise," Journal of Electronic Imaging, vol. 28 no. 4,DOI: 10.1117/1.JEI.28.4.043031, 2018.
[9] R. Fergus, B. Singh, A. Hertzmann, S. T. Roweis, W. T. Freeman, "Removing camera shake from a single photograph," ACM Transactions on Graphics, vol. 25 no. 3, pp. 787-794, DOI: 10.1145/1141911.1141956, 2006.
[10] D. Perrone, P. Favaro, "Total variation blind deconvolution: the devil is in the details," Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2014.372, .
[11] L. Xu, S. Zheng, J. Jia, "Unnatural L0 sparse representation for natural image deblurring," Proceedings of the 2013 IEEE Conference on Computer Vision & Pattern Recognition,DOI: 10.1109/CVPR.2013.147, .
[12] W. S. Lai, J. J. Ding, Y. Y. Lin, Y. Y. Chuang, "Blur kernel estimation using normalized color-line priors," Proceedings of the 2015 IEEE Computer Vision & Pattern Recognition,DOI: 10.1109/CVPR.2015.7298601, .
[13] W. S. Lai, J. B. Huang, Z. Hu, N. Ahuja, M. H. Yang, "A comparative study for single image blind deblurring," Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2016.188, .
[14] T. Michaeli, M. Irani, "Blind deblurring using internal patch recurrence," Proceedings of the 2014 European Conference on Computer Vision,DOI: 10.1007/978-3-319-10578-9_51, .
[15] J. Pan, D. Sun, H. Pfister, M. H. Yang, "Blind image deblurring using dark channel prior," Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/CVPR.2016.180, .
[16] J. Pan, H. Zhe, Z. Su, M. H. Yang, "Deblurring text images via L0-regularized intensity and gradient prior," Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2014.371, .
[17] D. Perrone, P. Favaro, "A logarithmic image prior for blind deconvolution," International Journal of Computer Vision, vol. 117 no. 2, pp. 159-172, DOI: 10.1007/s11263-015-0857-2, 2016.
[18] D. Perrone, P. Favaro, "A clearer picture of total variation blind deconvolution," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38 no. 6, pp. 1041-1055, DOI: 10.1109/TPAMI.2015.2477819, 2015.
[19] W.-Z. Shao, H.-S. Deng, Q. Ge, H.-B. Li, Z.-H. Wei, "Regularized motion blur-kernel estimation with adaptive sparse image prior learning," Pattern Recognition, vol. 51 no. C, pp. 402-424, DOI: 10.1016/j.patcog.2015.09.034, 2016.
[20] W. Zuo, D. Ren, D. Zhang, S. Gu, L. Zhang, "Learning iteration-wise generalized shrinkage-thresholding operators for blind deconvolution," IEEE Transactions on Image Processing, vol. 25 no. 4, pp. 1751-1761, DOI: 10.1109/tip.2016.2531905, 2016.
[21] Z. Hu, L. Xu, M. H. Yang, "Joint depth estimation and camera shake removal from single blurry image," Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2014.370, .
[22] T. H. Kim, K. M. Lee, "Segmentation-free dynamic scene deblurring," Proceedings of the 2014 Computer Vision & Pattern Recognition,DOI: 10.1109/CVPR.2014.348, .
[23] M. Noroozi, P. Chandramouli, P. Favaro, "Motion deblurring in the wild," Proceedings of the 2017 German Conference on Pattern Recognition, .
[24] J. Pan, H. Zhe, Z. Su, H. Y. Lee, M. H. Yang, "Soft-segmentation guided object motion deblurring," Proceedings of the 2016 Computer Vision & Pattern Recognition,DOI: 10.1109/CVPR.2016.56, .
[25] O. Whyte, "Non-uniform deblurring for shaken images: derivation of parameter update equations for blind de-blurring," 2010.
[26] S. Zheng, X. Li, J. Jia, "Forward motion deblurring," Proceedings of the 2013 IEEE International Conference on Computer Vision,DOI: 10.1109/ICCV.2013.185, .
[27] A. Gupta, N. Joshi, C. L. Zitnick, M. F. Cohen, B. Curless, "Single image deblurring using motion density functions," Proceedings of the 2010 European Conference on Computer Vision,DOI: 10.1007/978-3-642-15549-9_13, .
[28] Y. W. Tai, P. Tan, M. S. Brown, "Richardson-lucy deblurring for scenes under a projective motion path," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33 no. 8, pp. 1603-1618, DOI: 10.1109/TPAMI.2010.222, 2011.
[29] H. Zhang, D. Wipf, Y. Zhang, "Multi-image blind deblurring using a coupled adaptive sparse prior," Proceedings of the 2013 Computer Vision & Pattern Recognition,DOI: 10.1109/CVPR.2013.140, .
[30] M. Hirsch, C. J. Schuler, S. Harmeling, B. Schlkopf, "Fast removal of non-uniform camera shake," Proceedings of the 2011 International Conference on Computer Vision,DOI: 10.1109/ICCV.2011.6126276, .
[31] M. Hirsch, S. Sra, B. Scho¨Lkopf, G. Spemannstrae, "Efficient filter flow for space-variant multiframe blind deconvolution," Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2010.5540158, .
[32] H. Ji, K. Wang, "A two-stage approach to blind spatially-varying motion deblurring," Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition,DOI: 10.1109/CVPR.2012.6247660, .
[33] A. Levin, "Blind motion deblurring using image statistics," Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, .
[34] D. Gong, J. Yang, L. Liu, "From motion blur to motion flow: a deep learning solution for removing heterogeneous motion blur," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3806-3815, DOI: 10.1109/CVPR.2017.405, .
[35] M. Hradiš, "Convolutional neural networks for direct text deblurring," Proceedings of the 2015 British Machine Vision Conference,DOI: 10.5244/C.29.6, .
[36] O. Kupyn, V. Budzan, M. Mykhailych, D. Mishkin, J. Matas, "DeblurGAN: blind motion deblurring using conditional adversarial networks," Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2018.00854, .
[37] S. Nah, T. H. Kim, K. M. Lee, "Deep multi-scale convolutional neural network for dynamic scene deblurring," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/CVPR.2017.35, .
[38] S. Ramakrishnan, S. Pachori, A. Gangopadhyay, S. Raman, "Deep generative filter for motion deblurring," Proceedings of the 2017 IEEE International Conference on Computer Vision Workshop (ICCVW),DOI: 10.1109/ICCVW.2017.353, .
[39] C. J. Schuler, M. Hirsch, S. Harmeling, B. Scholkopf, M. Intelligence, "Learning to deblur," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38 no. 7, pp. 1439-1451, DOI: 10.1109/tpami.2015.2481418, 2016.
[40] J. Sun, W. Cao, Z. Xu, J. Ponce, "Learning a convolutional neural network for non-uniform motion blur removal," Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/CVPR.2015.7298677, .
[41] P. Svoboda, M. Hradis, L. Marsik, P. Zemcik, "CNN for license plate motion deblurring," Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP),DOI: 10.1109/ICIP.2016.7533077, .
[42] X. Tao, H. Gao, Y. Wang, X. Shen, J. Wang, J. Jia, "Scale-recurrent network for deep image deblurring," Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2018.00853, .
[43] A. Chakrabarti, "A neural approach to blind motion deblurring," Proceedings of the European Conference on Computer Vision, .
[44] S. Nah, T. Hyun Kim, K. Mu Lee, "Deep multi-scale convolutional neural network for dynamic scene deblurring," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, .
[45] P. Isola, J.-Y. Zhu, T. Zhou, A. A. Efros, "Image-to-image translation with conditional adversarial networks," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2017.632, .
[46] G. Huang, Z. Liu, L. Van Der Maaten, K. Q. Weinberger, "Densely connected convolutional networks," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2017.243, .
[47] M. Arjovsky, S. Chintala, L. Bottou, "Wasserstein GAN," 2017. http://arxiv.org/abs/1701.07875
[48] I. Goodfellow, J. Pouget-Abadie, M. Mirza, "Generative adversarial nets," 2014. http://arxiv.org/abs/1406.2661
[49] R. A. Yeh, C. Chen, T. Y. Lim, A. G. Schwing, M. Hasegawa-Johnson, M. N. Do, "Semantic image inpainting with deep generative models," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/CVPR.2017.728, .
[50] J. Johnson, A. Alahi, L. Fei-Fei, "Perceptual Losses for Real-Time Style Transfer and Super-resolution," Proceedings of the 2016 European Conference on Computer Vision, .
[51] J. Y. Zhu, T. Park, P. Isola, A. A. Efros, "Unpaired image-to-image translation using cycle-consistent adversarial networks," Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV),DOI: 10.1109/ICCV.2017.244, .
[52] B. Dai, S. Fidler, R. Urtasun, D. Lin, "Towards diverse and natural image descriptions via a conditional GAN," Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV),DOI: 10.1109/ICCV.2017.323, .
[53] C. Ledig, L. Theis, F. Huszar, "Photo-realistic single image super-resolution using a generative adversarial network," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/CVPR.2017.19, .
[54] R. Yeh, C. Chen, T. Y. Lim, M. Hasegawa-Johnson, M. Do, "Semantic image inpainting with perceptual and contextual losses," 2016. http://arxiv.org/abs/1607.07539
[55] K. He, J. Sun, X. Tang, "Single image haze removal using dark channel prior," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33 no. 12, pp. 2341-2353, DOI: 10.1109/TPAMI.2010.168, 2010.
[56] Y. Xu, X. Guo, H. Wang, F. Zhao, L. Peng, "Single image haze removal using light and dark channel prior," Proceedings of the 2016 IEEE/CIC International Conference on Communications in China (ICCC),DOI: 10.1109/ICCChina.2016.7636813, .
[57] Y. Yan, W. Ren, Y. Guo, R. Wang, X. Cao, "Image deblurring via extreme channels prior," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/CVPR.2017.738, .
[58] X. Wang, K. Yu, S. Wu, "ESRGAN: enhanced super-resolution generative adversarial networks," 2018. http://arxiv.org/abs/1809.00219
[59] K. Simonyan, A. Zisserman, "Very deep convolutional networks for large-scale image recognition," 2014. http://arxiv.org/abs/14091556
[60] L. Xu, J. S. J. Ren, C. Liu, J. Jia, "Deep convolutional neural network for image deconvolution," Proceedings of the 27th International Conference on Neural Information Processing Systems, pp. 1790-1798, .
[61] A. Chakrabarti, A Neural Approach to Blind Motion Deblurring, 2016.
[62] A. Levin, Y. Weiss, F. Durand, W. T. Freeman, "Understanding and evaluating blind deconvolution algorithms," Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2009.5206815, .
[63] U. Schmidt, C. Rother, S. Nowozin, J. Jancsary, S. Roth, "Discriminative non-blind deblurring," Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, pp. 604-611, DOI: 10.1109/CVPR.2013.84, .
[64] D. Engin, A. Genc, H. K. Ekenel, "Cycle-dehaze: enhanced CycleGAN for single image dehazing," Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW),DOI: 10.1109/CVPRW.2018.00127, .
[65] J. Redmon, S. Divvala, R. Girshick, A. Farhadi, "You only look once: unified, real-time object detection," Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/CVPR.2016.91, .
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2021 Yuqing Zhao et al. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. https://creativecommons.org/licenses/by/4.0/
Abstract
Blind deblurring of a single infrared image is a challenging computer vision problem. Because the blur is not only caused by the motion of different objects but also by the relative motion and jitter of cameras, there is a change of scene depth. In this work, a method based on the GAN and channel prior discrimination is proposed for infrared image deblurring. Different from the previous work, we combine the traditional blind deblurring method and the blind deblurring method based on the learning method, and uniform and nonuniform blurred images are considered, respectively. By training the proposed model on different datasets, it is proved that the proposed method achieves competitive performance in terms of deblurring quality (objective and subjective).
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer