3.1. The Test of Image Restoration in Simulation
In this section, we test the effectiveness and robustness of DeturNet using simulations. The training took approximately 3 h, with a total of 300 epochs. The initial learning rate was 1 × 10−4. The learning rate decay mode is cosine decay with a minimum learning rate of 1 × 10−6. The batch size is 16. The optimization algorithm was Adam, and the weight decay was set to 2 × 10−6.
The training platform was Pytorch 1.8.0, based on Python 3.7.11 in Sichuan, China. This is used to implement the network. An Intel Xeon(R) CPU (2.5 KHz) and NIVIDIA GeForce GTX 3090 GPU were used for the training and testing phases. The loss curve for the DeturNet training process is shown in
Figure 7.
The restoration results are shown in
Figure 8. The proposed results show that our method can effectively reduce the turbulence of an image. We also compared our method with other traditional and deep learning image recovery methods, such as the blind recovery method proposed by Jin et al. [
46], the deep learning algorithms U-Net [
42], and DeepRFT [
47]. All the methods were retrained to achieve optimal recovery. The effects of the image restoration are shown in
Figure 8. In order to evaluate the image restoration effectively, the results were analyzed using both subjective and objective methods. From a visual point of view, DeturNet has a better effect on turbulence image recovery, and the result is closer to the original image than those of the other methods. The other methods still have some degree of ambiguity in the edge information of the image target. Meanwhile, two evaluation functions, peak signal-to-noise ratio (PSNR) and structural similarity (SSIM), are proposed to perform an objective evaluation.
where
X denotes the pixel value of the reference image,
Y denotes the pixel value of the evaluated image, and
(i, j) denotes the pixel coordinates. MaxValue is the maximum value of the color grey scale in an image, which is usually 255 on a uniform statement.
l,
c, and
s represent the equations for brightness, contrast, and structure, respectively [
48]. These two evaluation functions show different degradations of the images. PSNR usually shows the ratio between the maximum energy of the signal and the energy of noise, which affects the fidelity of its representation. SSIM usually shows the similarity between two images. A higher PSNR and SSIM closer to 1 often mean a better imaging quality [
49,
50,
51].
In the outdoor experiments, the PSNR and SSIM evaluation criteria were not applicable due to the lack of corresponding labels. Therefore, we utilized three additional unreferenced evaluation criteria, namely variance, information entropy, and average gradient (
), to objectively assess the recovery quality. The calculations for these criteria are as follows:
where
represents the image size,
represents the gray value of the pixel
, and
P(k) represents the ratio of the number of pixels with a gray value of k to the total number of pixels.
represents the gradients in the
x and
y directions, respectively.
In
Figure 8, we can see that the evaluation index under the restored image of the DeturNet is the highest among all the methods in all four images. The average results of the two evaluation indices in the test sets are also shown in
Table 1, where PSNR_std represents the variance of the PSNR of the test set. For the test datasets, the recovery results show that DeturNet has an average improvement of 3.16 (16.8%) in the PSNR and 0.0899 (13.3%) in the SSIM, which is the highest among the four methods. Thus, our method performed better on all test sets and showed the superiority of our proposed method. We also compared our method with the other 4 methods in terms of time, as shown in
Table 2. The results show that our method has a better real-time capability compared to other methods. With the development of computer computing power, the proposed method has the potential to perform real-time imaging reconstruction.
3.3. The Robustness of DeturNet on Different Noises
To test the robustness of the network on different noises, four Gaussian-type noises with means of 0 and variances of 0.01, 0.03, 0.05, and 0.08 are added to the test set, and we feed it directly into our trained network. The test images and reconstructed images are shown in
Figure 9. The results for the entire test set are shown in
Table 4.
As we can see, the increase in noise affects both the blurred and restored images in
Figure 9. When the noise is small, the reconstructed image is less affected by noise. When the noise increases, the reconstructed image shows a ringing effect; however, it still preserves the most detailed information. The target could still be clearly recognized. From
Table 4, we can see that the PSNR and the SSIM are in a uniform decline, which indicates that the capability of our algorithm decreases slowly with an increase in noise, but the overall reconstruction capability is still good.
3.4. The Robustness of DeturNet on Different Turbulences
To test the robustness of our network under different turbulences, different turbulence intensities (D/r
0 = 5 and D/r
0 = 10) are proposed in this paper. So, we re-stimulated the datasets with their corresponding intensities and retrained them. The results are shown in
Figure 10. The degraded image shows that a stronger turbulence intensity causes a more severely blurred image, which makes the recovery of turbulent degraded images extremely challenging.
The test results in
Figure 10 show that, although the image is completely blurred by the turbulence of D/r
0 = 10, the recovery results of DeturNet can still reconstruct the target information, such as the aircraft outline and building edges. However, some high-frequency information is lost, such as the edge details in the recovered results, when the turbulence intensity increases further.
Table 5 shows that the reconstruction results are valid for the different turbulence intensity tests.
3.5. Laboratory Experiment Results and Discussions
In this section, we build a laboratory experimental imaging system to verify the performance and robustness of our network in practice. The laboratory connects the simulation and the outdoor experiments. Compared with the simulation dataset, it provides a more realistic atmospheric turbulence situation. Compared with outdoor experiments, it offers a dataset that is not affected by turbulence as the basis of the training dataset. As shown in
Figure 11, LED and Digital Micromirror Devices (DMD) are used to generate the dynamic targets loaded by aircraft images of the NWPU-RESISC45 dataset, where the LED provides a stable, broad-spectrum light source, and DMD projects the loaded targets rapidly. A turbulent screen was used to generate the stochastic atmospheric turbulence, which had the same intensity as the simulation situations.
Through the experimental process, a dataset of 700 turbulence-blurred images is obtained and divided into a training set, a validation set, and a test set in the ratio of 8:1:1. The subjectivity of the recovered images and the two evaluation criteria show that our proposed method has a better reconstruction ability than the other methods. The partial recovery result images are shown in
Figure 12. The average results for the test set are shown in
Table 6. Both the simulation and experimental results show that the recovery method based on DeturNet has better effects and robustness for turbulently degraded image restoration, which means that the proposed method has the potential to be used in image restoration for astronomical observation, remote sensing observation, and traffic detection.
3.6. Outdoor Experiment Results and Discussions
In order to further validate the effectiveness of our method, we collected real turbulence degradation pictures in the natural environment in the outside world. The experimental scene is shown in
Figure 13a. In this case, the telescope focal length was 1250 mm, the object distance of the target was about 200 m, the CCD camera pixel size was 5.5 µm, the CCD camera exposure time was 3 ms, the acquisition time was 4 pm, the maximum temperature of the day was 33 °C and the minimum temperature was 24 °C. One of the acquired images is shown in
Figure 13b.
In the experiment, we captured images of a toy car and calibration targets. We selected some regions of interest and tested the recoverability and generalization effect directly using DeturNet trained on a laboratory experimental dataset. The tested DeturNet-based single-frame recovery results are shown in
Figure 14, from top to bottom, for scenarios 1–4. We observed the reconstructed images using both the subjective and objective methods. In a subjective evaluation, DeturNet exhibited a good recovery effect. The target contour edge information in the image was better recovered, and the boundary was more visible. Meanwhile, the variances of DeturNet restoration results were much larger than those of the blurred images and the results of the comparison methods.
Two objective evaluation criteria, information entropy and average gradient (AG), were used in this paper, as shown in
Table 7. In the outdoor experiments, although the DeturNet was trained using the laboratory dataset, we can see that the network has good recovery results. Thus, from the perspective of visual effect, histogram distribution, and the size of the variance, entropy, and average gradient, the results show that we can recover the image scene from completely uninvolved training to a certain extent. The variance and entropy can reflect the amount of information in the image; the larger the value, the richer the information in the image hierarchy. The average gradient reflects the ability of an image to express the contrast in minute detail. In general, a larger average gradient often means the image is sharper. Since the original data are unknown, these evaluation functions have some limitations. For example, as we can see in the first row of images, even though the U-Net has the highest value of variance, the edges of the toy car in its recovered images are obviously distorted, while the edges of the toy car reconstructed using DeturNet are more realistic and sharper. Through subjective judgments supplemented by objective metrics, it can be seen that DeturNet outperforms other deep learning algorithms in terms of generalization ability. Meanwhile, the proposed results also show that the generalization ability can be improved with uncorrelated datasets when performing sufficiently fitted experiments, and we will continue our research on this issue.