Holographic imaging poses significant challenges when facing real-time disturbances introduced by dynamic environments. The existing deep-learning methods for holographic imaging often depend solely on the specific condition based on the given data distributions, thus hindering their generalization across multiple scenes. One critical problem is how to guarantee the alignment between any given downstream tasks and pretrained models. We analyze the physical mechanism of image degradation caused by turbulence and innovatively propose a swin transformer-based method, termed train-with-coherence-swin (TWC-Swin) transformer, which uses spatial coherence (SC) as an adaptable physical prior information to precisely align image restoration tasks in the arbitrary turbulent scene. The light-processing system (LPR) we designed enables manipulation of SC and simulation of any turbulence. Qualitative and quantitative evaluations demonstrate that the TWC-Swin method presents superiority over traditional convolution frameworks and realizes image restoration under various turbulences, which suggests its robustness, powerful generalization capabilities, and adaptability to unknown environments. Our research reveals the significance of physical prior information in the optical intersection and provides an effective solution for model-to-tasks alignment schemes, which will help to unlock the full potential of deep learning for all-weather optical imaging across terrestrial, marine, and aerial domains. |
1.IntroductionHolographic imaging is an interdisciplinary field that combines optics, computer science, and applied mathematics to generate holographic images using numerical algorithms. Although the concept of using computers to generate holograms can be traced back to the 1960s, it was not until the emergence of digital imaging and processing techniques in the 1990s that computational holography began to develop into a viable technology.1,2 In the 1990s, digital holography started to gain more attention due to advancements in computer technology and digital image processing.3 In recent years, holographic imaging has continued to advance, with new research and technology enabling even more sophisticated holographic imaging capabilities. Researchers have developed increasingly sophisticated numerical algorithms for holographic imaging, such as compressive sensing, sparse coding, and deep-learning techniques.4–10 Spatial coherence (SC) is a critical factor that determines the quantity and quality of high-frequency information carried by the light beam in holographic imaging. High-frequency information is crucial for achieving high resolution and capturing fine details in an image. When the SC of the light source is low, the phase relationship of the beam becomes chaotic, causing the interference pattern to be washed out and resulting in insufficient transmission of high-frequency information. As a result, the reconstructed image has a lower resolution and less fine-detail information, as the high-frequency information needed to capture these details has been lost. Therefore, high SC light is preferred for holographic imaging to ensure that sufficient high-frequency information is present in the interference pattern and the hologram, resulting in high-resolution and detailed reconstructed images. However, the SC of light sources is often very low in complex scenes, which leads to image degradation and loss of details. Therefore, how to restore images under low-SC light sources is a challenging issue.11–15 Oceanic and atmospheric turbulence may profoundly influence optical imaging, engendering distortions and deterioration in photographs acquired through cameras and alternative optical detection devices. The distortion and degradation of images caused by oceanic turbulence occur because the turbulent motions in the water column cause variations in the refractive index of the water, which in turn leads to variations in the path of light as it travels through the water. Atmospheric turbulence occurs because the Earth’s atmosphere is not uniform and contains regions of varying temperature and density, which can cause variations in the refractive index of the air. Whether it is oceanic turbulence or atmospheric turbulence, as the beam passes through these regions of varying refractive index, phase correlation changes, and the SC is distorted, causing the image to become blurred and distorted, or even completely lost. Massive efforts were devoted to finding a solution for imaging in various turbulences.16–23 There is no denying the fact that it is difficult to use the same methods to simultaneously resolve holographic imaging problems with low-SC scenes and multiple intensities of turbulence. Although low-SC and turbulence may not appear to be correlated at first glance, their influence on computational holography can both be described through the concept of SC. As a result, we can transform the aforementioned issues into the imaging problem of different SCs and leverage the advantages of deep learning to train a generalized model that can achieve image restoration for any turbulence intensity and low SC. Artificial intelligence for optics has unparalleled advantages, especially in the field of holography. For example, deep learning can address challenging inverse problems in holographic imaging, where the objective is to recover the original scene or object properties from observed images or measurements and enhance the resolution of optical imaging systems beyond their traditional diffraction limit,24–30 etc. Intersection research of optics and deep learning aims to solve massive tasks with one model, and one important problem is how to guarantee the alignment between the distribution of any given downstream data and tasks with pretrained models. This means that the same model and weights can only be applied to a specific environment. Our research uses SC as adaptable real-time physical prior information to precisely align any scenes with pretrained models. By combining the most advanced deep-learning algorithms, residual network,31 and swin transformer,32 we proposed our deep-learning-based methodology, termed as train-with-coherence-swin (TWC-Swin) method. It can achieve the restoration of computational holographic imaging under any low SC and turbulence. We summarize the innovations of this paper as follows.
2.Materials and Methods2.1.Scheme of the LPRFigure 1(a) shows the LPR. The high-coherence light source generated by the solid-state laser (CNI, MLL-FN, 532 nm) is polarized horizontally after passing through a half-wave plate and a polarization beam splitter, allowing it not only to match the modulation mode of the SLM but also to adjust the beam intensity. The RD (DHC, GCL-201) is used to reduce the SC of the light source, with the degree of reduction depending on the radius of the incident beam on the RD—the larger the radius is, the lower the SC of the output light source is (see Note 2 in the Supplementary Material). In the experiment, we control the incident beam radius by adjusting the distance between lens 1 (L1, 100 mm) and the RD. After being collimated by lens 2 (L2, 100 mm), the beam is incident on the SLM1 (HDSLM80R) loaded with turbulent phase, which is continuously refreshed at a rate of 20 Hz. After passing through the turbulence, the beam is split into two parts by a beam splitter. The first part employs Michelson interference to capture interference fringes and measure the SC of the light. The second part is used for holographic imaging, with the phase hologram of the image loaded onto the SLM2 (PLUTO). The high-pass filter is employed to filter out the unmodulated zero-order diffraction pattern, and the final imaging result is captured by the complementary metal–oxide-semiconductor (CMOS, Sony, E3ISPM). In summary, we control the SC of the light source by adjusting the distance between lens L1 and the RD. We simulate a turbulent environment using the SLM1, with the intensity of the turbulence depending on the loaded turbulent phase. If turbulence is not required, the SLM1 can be turned off, and it functions as a mirror equivalent. 2.2.Oceanic Turbulence and Atmospheric TurbulenceThe turbulence intensity in the experiment is determined by the spatial power spectrum of the turbulence. The function of the spatial power spectrum of the turbulent refractive-index fluctuations used in this paper is based on the assumption that turbulence is homogeneous and isotropic. We use the Nikishov power spectrum to describe oceanic turbulence:33 where is the spatial wavenumber of turbulent fluctuations, . is the dissipation rate of turbulent kinetic energy per unit mass. is the Kolmogorov microscale (inner scale). is the index of the relative strength of temperature and salinity fluctuations. , , and . stands for a variate that represents the rate of dissipation of mean-square temperature, which varies from in deep water to in surface water. We only changed the oceanic turbulence intensity by adjusting ; the greater the value of is, the stronger the oceanic turbulence is. Detailed parameter settings for the power spectrum of oceanic turbulence can be found in Table S2 in the Supplementary Material.For atmospheric turbulence, we use the non-Kolmogorov power spectrum,34 where is the refractive index power spectral density power law. and represent inner and outer scales, respectively. denotes the refractive index structure constant. We only changed the atmospheric turbulence intensity by adjusting ; the greater the value of is, the stronger the atmospheric turbulence is. Detailed parameter settings for the power spectrum of atmospheric turbulence can be found in Table S2 in the Supplementary Material. After setting reasonable parameters and returning to the space domain through the inverse Fourier transform, the turbulent phase can be obtained, which will be input into SLM1 to simulate the turbulent scene.2.3.Data AcquisitionLow SC and turbulence are different physical scenarios, but the influence of these scenarios on holographic imaging can be described through SC. Based on the above method, we only use the data obtained under different SCs for model training, and any other data are used for testing [Fig. 1(g)]. The process of data acquisition is as follows.
Our original images consist of public data sets, such as the Berkeley segmentation data set (BSD),36 Celebfaces attributes high-quality data set (CelebA),37 Flickr data set (Flickr),38 Webvision data set (WED),39 and DIV2k data set (DIV).40 The training set is only composed of images captured by CMOS1 in steps 2 and 3. In the training phase, we divide the training data into 11 groups based on SC and send them to the network for training in turn. Therefore we can obtain a model space containing swin models with different weights. In the testing phase, the swin adapter is a program that needs to receive the SC information of the light source and selects the optimal model in model space to achieve the image restoration task. Here we set to distance priority mode, and the swin adapter will select the weight parameter closest to the measured SC. The test set comes from the images generated in steps 4 and 5. Note that none of the test sets have been trained; they are blinded to the network. Our model was implemented using PyTorch; the detailed architecture can be found in Note 1 in the Supplementary Material. We use adaptive moment estimation with weight decay (AdamW) as optimizer,41 which is utilized to update the weights with initial learning rates of 0.0005 with a 50% drop every 10 epochs. The total epoch is 100. Mean-squared error (MSE) is the loss function of the network. All training and testing stages are placed on the NVIDIA GTX3080Ti graphics card, and a full training period takes about 12 h. To effectively verify the performance of our method, a series of credible image quality assessment measures were applied. The full-reference measures include peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and Pearson correlation coefficient (PCC), which are used to provide an assessment of a single image in relation to perceived visual quality. See Note 4 in the Supplementary Material for descriptions of evaluation indices. 3.Results and DiscussionThis section primarily showcases the performance of our method under various SCs and turbulent scenes. We simulated different strengths of oceanic and atmospheric turbulence, enhancing the diversity of turbulence intensities and types. Additionally, we conducted comparative analyses with traditional convolutional-residual networks and performed ablation studies to reinforce the validity and efficiency of our proposed method. It is important to emphasize that our training data exclusively consisted of holographic imaging results obtained under different SC conditions, with none of the test data used during the training phase. 3.1.Performance on Low SCFigures 2 and 1(e) show the original images captured by CMOS1 and restored images processed by the TWC-Swin method under different SCs. We present 11 groups of test results, each representing a different SC level and containing samples from five distinct data sets. As described in Sec. 2, the SC of the light source can be altered by adjusting the distance between RD and L1. It is evident that as the SC decreases, the quality of holographic imaging deteriorates significantly, exhibiting high levels of noise and blurriness. Simultaneously, the decrease in SC corresponds to a reduction in light efficiency, resulting in darker images that ultimately become indiscernible. After processing through the trained network, these degraded images become smoother, with improved sharpness, enhanced details, and reduced noise. Remarkably, even in low SC conditions where the original images captured by the CMOS1 sensor lack any discernible details, our network successfully reconstructs a significant portion of the elements. To accurately evaluate the effectiveness of image restoration, we present the evaluation indices (SSIM and PCC), comparing the original and reconstructed images with respect to the ground truth for different SCs [Fig. 1(f) and Table 1]. Other indices are provided in Table S3 in the Supplementary Material. The quantitative results further validate the significant improvement achieved in various indicators of the reconstructed images compared to the original ones, approaching the ground truth. Figure 3 illustrates the average evaluation indices for each test set. Here only partial results are shown; more detailed results are included in Fig. S2 in the Supplementary Material. It can be seen that each evaluation index of images has risen significantly compared to the original images after being processed by the TWC-Swin method, indicating a substantial improvement in the image quality. Moreover, the network demonstrates its robust generalization capability by performing image restoration on multiple test sets, which are beyond the scope of the training set. This implies that our method has effectively learned the underlying patterns in the data during training and can apply these patterns to unseen data, resulting in successful image restoration. Table 1Quantitative analysis of evaluation indices (SSIM and PCC) at different SCs and test samplesa. f1 is the focal length of L1. SC means spatial coherence of the light source.
3.2.Performance on Oceanic Turbulence and Atmospheric TurbulenceOwing to the stochastic variations of the refractive index within oceanic and atmospheric turbulence, the phase information of light beams becomes distorted, thereby reducing SC and degrading the quality of computational holography images. This issue can be effectively addressed using the TWC-Swin method. It should be mentioned that all images captured under turbulent scenes were never trained by the network. Figure 4 demonstrates the remarkable image restoration capability of TWC-Swin method under varying intensities of oceanic and atmospheric turbulence. As discussed in Sec. 2, the turbulence intensity depends on certain variates of the power spectrum function, where stronger turbulence presents more complex simulated turbulence phases, as shown in Figs. 4(A5) and 4(O5). We carried out experiments under five distinct intensities of both oceanic and atmospheric turbulence, and simultaneously measured the SC of the light source for selecting the optimal model. It should be noted that the turbulence phase loaded on the SLM is continuously refreshed (20 Hz). To provide stronger evidence, we present the evaluation indices (SSIM and PCC) for oceanic and atmospheric turbulence in Tables 2 and 3 and Fig. 1(h), whereas additional indices (MSE and PSNR) can be found in Tables S4 and S5 in the Supplementary Material. Our analysis concluded that as the turbulence intensity increases, the SC experiences a downturn, which subsequently degrades image quality. Nevertheless, our proposed method is capable of overcoming these adverse effects and effectively improving the image quality regardless of the turbulence intensity. Our model learns the universal features of image degradation and restoration that depend on SC. This further demonstrates the strong generalization capability of the network trained with SC as physical prior information and the ability to apply learned knowledge from the training set to new, unseen scenes. This versatility is a desirable trait in a neural network, as it suggests the method’s potential for broad application. Table 2Quantitative analysis of evaluation indices (SSIM and PCC) at different oceanic turbulence intensitiesa.
Table 3Quantitative analysis of evaluation indices (SSIM and PCC) at different atmospheric turbulence intensitiesa.
3.3.Comparison between Different Methods and Ablation StudyIn this section, we conduct a comprehensive comparative study of different methodologies, assessing their performance and efficacy in restoring images under challenging conditions of low SC and turbulent scenes. Traditional convolution-fusion framework methods, U-net,42 and U-RDN13 were compared to demonstrate the power of the proposed swin model. In our network architecture, the swin transformer serves as a robust backbone module, responsible for extracting high-level features from input. The special working mechanism gives it powerful hierarchical representation and global perception capabilities. However, direct output from the swin transformer often exhibits artifacts and high-noise levels in image restoration tasks. Therefore, it is necessary to add lightweight convolutional layers as postprocessing blocks. Convolution layers capture local features of the image through local receptive fields, aiding in a better understanding of image details and textures while facilitating mapping from high-dimensional to low-dimensional spaces, resulting in high-quality output. To validate the effectiveness of the postprocessing block in the swin model, we conduct an ablation study. In the ablation study, we created a control group named pure swin, which was obtained by removing the postprocessing block from the swin model. The training processes and data sets of all methods are consistent. Figure 5 shows detailed comparisons of images processed by various methods. Figure 6 illustrates the quantitative results between different methods on various data sets. More qualitative results are provided in Figs. S3 and S4 in the Supplementary Material. Comparing the visual output results of pure swin and the swin model, we found that the output results of the pure swin framework will produce black spots, resulting in blurred perception; the SSIM is 0.8396, a 7% reduction. This is because the swin transformer lacks the ability to sense local features and dimensional mapping. Convolutional layers can fill this gap by offering a mechanism to refine and enhance local features past the swin transformer blocks. The ablation study (compared with pure swin) validates that the postprocessing module is indispensable for the swin model. We tested the performance of other networks under the same conditions. Our proposed network outperforms other methods by presenting the lowest noise and best evaluation index. Tables S6 and S7 in the Supplementary Material provide a detailed quantitative comparison of the performance across different models and different SCs. In the task of image restoration under low SC, our proposed methodology exhibits superior performance across all evaluative indices when juxtaposed with alternative approaches. Figure 7 shows the comparative performance of various methods when faced with image degradation due to various turbulence types and intensities. We observed that all networks trained with SC exhibit the ability to significantly improve the image quality under turbulent scenes and not just the swin model. This is an exciting result, as it signifies the successful integration of physical prior information into network training, enabling the networks to be applied to multiple tasks and scenarios. 4.ConclusionsBy leveraging the SC as physical prior information and harnessing advanced deep-learning algorithms, we proposed a methodology, TWC-Swin, which demonstrates exceptional capabilities in simultaneously restoring images in low SC and random turbulent scenes. Our multicoherence and multiturbulence holographic imaging data sets, consisting of natural images, are created by the LPR, which can simulate different SCs and turbulence scenes (see Sec. 2). Though the swin model used in the tests was trained solely on the multicoherence data set, it can achieve promising results on both low SC, oceanic turbulence and atmospheric turbulence scenes. The key is that we capture the common physical property in these scenes, SC, and use it as physical prior information to generate a training set, so that the TWC-Swin method exhibits remarkable generalization capabilities, effectively restoring images from unseen scenes beyond the training set. Furthermore, through a series of rigorous experiments and comparisons, we have established the superiority of the swin model over traditional convolutional frameworks and alternative methods in terms of image restoration from qualitative and quantitative analysis (see Sec. 3). The integration of SC as a fundamental guiding principle in network training has proven to be a powerful strategy in aligning downstream tasks with pretrained models. Our research findings offer guidance not only for the domain of optical imaging but also for the integration with the segment anything model (SAM),43 extending its applicability to multiphysics scenarios. For instance, in turbulent scenes, our methodology can be implemented for preliminary image processing, enabling the utilization of unresolved images for precise image recognition and segmentation tasks facilitated by SAM. Moreover, our experimental scheme also provides a simple idea for turbulence detection. Our research contributes valuable insights into the use of deep-learning algorithms for addressing image degradation problems in multiple scenes and highlights the importance of incorporating physical principles into network training. It is foreseeable that our research can serve as a successful case for the combination of deep learning and holographic imaging in the future, which facilitates the synergistic advancement of the fields of optics and computer science. Code and Data AvailabilityThe codes of the TWC-Swin method, trained model, as well as some example images for testing, are publicly available at https://github.com/tongxinoptica/TWC-Swin. All relevant data that support the findings of this work are available from the corresponding author upon reasonable request. The parameter settings of TWC-Swin used in synthesizing the training and evaluation data sets will be publicly available along with this paper. AcknowledgmentsThis work was mainly supported by the National Natural Science Foundation of China (Grants Nos. 12174338 and 11874321) received by D.M.Z. All authors contributed to the discussions and preparation of the manuscript.44 The authors declare no competing interests. ReferencesL. B. Lesem, P. M. Hirsch and J. A. Jordan,
“Scientific applications: computer synthesis of holograms for 3D display,”
Commun. ACM, 11 661
–674 https://doi.org/10.1145/364096.364111 CACMA2 0001-0782
(1968).
Google Scholar
M. Lurie,
“Fourier-transform holograms with partially coherent light: holographic measurement of spatial coherence,”
J. Opt. Soc. Am., 58 614
–619 https://doi.org/10.1364/JOSA.58.000614 JOSAAH 0030-3941
(1968).
Google Scholar
U. Schnars and W. Jüptner,
“Direct recording of holograms by a CCD target and numerical reconstruction,”
Appl. Opt., 33 179
–181 https://doi.org/10.1364/AO.33.000179 APOPAI 0003-6935
(1994).
Google Scholar
R. Horisaki et al.,
“Compressive propagation with coherence,”
Opt. Lett., 47 613
–616 https://doi.org/10.1364/OL.444772 OPLEDP 0146-9592
(2022).
Google Scholar
D. Blinder et al.,
“Signal processing challenges for digital holographic video display systems,”
Signal Process. Image Commun., 70 114
–130 https://doi.org/10.1016/j.image.2018.09.014 SPICEF 0923-5965
(2019).
Google Scholar
H. Ko and H. Y. Kim,
“Deep learning-based compression for phase-only hologram,”
IEEE Access, 9 79735
–79751 https://doi.org/10.1109/ACCESS.2021.3084800
(2021).
Google Scholar
L. Shi et al.,
“Towards real-time photorealistic 3D holography with deep neural networks,”
Nature, 591 234
–239 https://doi.org/10.1038/s41586-020-03152-0
(2021).
Google Scholar
C. Lee et al.,
“Deep learning based on parameterized physical forward model for adaptive holographic imaging with unpaired data,”
Nat. Mach. Intell., 5 35
–45 https://doi.org/10.1038/s42256-022-00584-3
(2023).
Google Scholar
X. Guo et al.,
“Stokes meta-hologram toward optical cryptography,”
Nat. Commun., 13 6687 https://doi.org/10.1038/s41467-022-34542-9 NCAOBW 2041-1723
(2022).
Google Scholar
H. Yang et al.,
“Angular momentum holography via a minimalist metasurface for optical nested encryption,”
Light Sci. Appl., 12 79 https://doi.org/10.1038/s41377-023-01125-2
(2023).
Google Scholar
R. Fiolka, K. Si and M. Cui,
“Complex wavefront corrections for deep tissue focusing using low coherence backscattered light,”
Opt. Express, 20 16532
–16543 https://doi.org/10.1364/OE.20.016532 OPEXFF 1094-4087
(2012).
Google Scholar
S. Lim et al.,
“Optimal spatial coherence of a light-emitting diode in a digital holographic display,”
Appl. Sci., 12 4176 https://doi.org/10.3390/app12094176
(2022).
Google Scholar
Y. Deng and D. Chu,
“Coherence properties of different light sources and their effect on the image sharpness and speckle of holographic displays,”
Sci. Rep., 7 5893 https://doi.org/10.1038/s41598-017-06215-x SRCEC3 2045-2322
(2017).
Google Scholar
X. Tong et al.,
“A deep-learning approach for low-spatial-coherence imaging in computer-generated holography,”
Adv. Photonics Res., 4 2200264 https://doi.org/10.1002/adpr.202200264
(2023).
Google Scholar
Y. Peng et al.,
“Speckle-free holography with partially coherent light sources and camera-in-the-loop calibration,”
Sci. Adv., 7 5040 https://doi.org/10.1126/sciadv.abg5040 STAMCV 1468-6996
(2021).
Google Scholar
F. Wang et al.,
“Propagation of coherence-OAM matrix of an optical beam in vacuum and turbulence,”
Opt. Express, 31 20796
–20811 https://doi.org/10.1364/OE.489324 OPEXFF 1094-4087
(2023).
Google Scholar
D. Jin et al.,
“Neutralizing the impact of atmospheric turbulence on complex scene imaging via deep learning,”
Nat. Mach. Intell., 3 876
–884 https://doi.org/10.1038/s42256-021-00392-1
(2021).
Google Scholar
Q. Zhang et al.,
“Effect of oceanic turbulence on the visibility of underwater ghost imaging,”
J. Opt. Soc. Am. A, 36 397
–402 https://doi.org/10.1364/JOSAA.36.000397 JOAOD6 0740-3232
(2019).
Google Scholar
K. Wang et al.,
“Deep learning wavefront sensing and aberration correction in atmospheric turbulence,”
PhotoniX, 2 8 https://doi.org/10.1186/s43074-021-00030-4
(2021).
Google Scholar
Y. Chen et al.,
“A wavelet based deep learning method for underwater image super resolution reconstruction,”
IEEE Access, 8 117759
–117769 https://doi.org/10.1109/ACCESS.2020.3004141
(2020).
Google Scholar
L. Zhang et al.,
“Restoration of single pixel imaging in atmospheric turbulence by Fourier filter and CGAN,”
Appl. Phys. B, 127 45 https://doi.org/10.1007/s00340-021-07596-8
(2021).
Google Scholar
Y. Baykal, Y. Ata and M. C. Gökçe,
“Underwater turbulence, its effects on optical wireless communication and imaging: a review,”
Opt. Laser Technol., 156 108624 https://doi.org/10.1016/j.optlastec.2022.108624 OLTCAS 0030-3992
(2022).
Google Scholar
J. Bertolotti and O. Katz,
“Imaging in complex media,”
Nat. Phys., 18 1008
–1017 https://doi.org/10.1038/s41567-022-01723-8 NPAHAX 1745-2473
(2022).
Google Scholar
T. Zeng, Y. Zhu and E. Y. Lam,
“Deep learning for digital holography: a review,”
Opt. Express, 29 40572
–40593 https://doi.org/10.1364/OE.443367 OPEXFF 1094-4087
(2021).
Google Scholar
A. Khan et al.,
“GAN-Holo: generative adversarial networks-based generated holography using deep learning,”
Complexity, 2021 6662161 https://doi.org/10.1155/2021/6662161 COMPFS 1076-2787
(2021).
Google Scholar
M. Liao et al.,
“Scattering imaging as a noise removal in digital holography by using deep learning,”
New J. Phys., 24 083014 https://doi.org/10.1088/1367-2630/ac8308
(2022).
Google Scholar
T. Shimobaba et al.,
“Deep-learning computational holography: a review,”
Front. Photonics, 3 854391 https://doi.org/10.3389/fphot.2022.854391
(2022).
Google Scholar
Y. Rivenson, Y. Wu and A. Ozcan,
“Deep learning in holography and coherent imaging,”
Light Sci. Appl., 8 85 https://doi.org/10.1038/s41377-019-0196-0
(2019).
Google Scholar
Z. Chen et al.,
“Physics-driven deep learning enables temporal compressive coherent diffraction imaging,”
Optica, 9 677 https://doi.org/10.1364/OPTICA.454582
(2022).
Google Scholar
Y. Jo et al.,
“Holographic deep learning for rapid optical screening of anthrax spores,”
Sci. Adv., 3 e1700606 https://doi.org/10.1126/sciadv.1700606 STAMCV 1468-6996
(2023).
Google Scholar
K. He et al.,
“Deep residual learning for image recognition,”
in IEEE Conf. Comput. Vis. and Pattern Recognit. (CVPR),
770
–778
(2016). https://doi.org/10.1109/CVPR.2016.90 Google Scholar
Z. Liu et al.,
“Swin transformer: hierarchical vision transformer using shifted windows,”
in Proc. IEEE/CVF Int. Conf. Comput. Vis.,
10012
–10022
(2021). https://doi.org/10.1109/ICCV48922.2021.00986 Google Scholar
V. V. Nikishov,
“Spectrum of turbulent fluctuations of the sea-water refraction index,”
Int. J. Fluid Mech. Res., 27 82
–98 https://doi.org/10.1615/InterJFluidMechRes.v27.i1.70
(2000).
Google Scholar
B. E. Stribling, B. M. Welsh and M. C. Roggemann,
“Optical propagation in non-Kolmogorov atmospheric turbulence,”
Proc. SPIE, 2471 181
–195 https://doi.org/10.1117/12.211927 PSISDG 0277-786X
(1995).
Google Scholar
R. W. Gerchberg,
“A practical algorithm for the determination of phase from image and diffraction plane pictures,”
Optik, 35 237
–246 OTIKAJ 0030-4026
(1972).
Google Scholar
D. Martin et al.,
“A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,”
in Proc. Eighth IEEE Int. Conf. Comput. Vis.,
416
–423
(2001). https://doi.org/10.1109/ICCV.2001.937655 Google Scholar
Z. Liu et al.,
“Deep learning face attributes in the wild,”
in IEEE Int. Conf. Comput. Vis.,
3730
–3738
(2015). https://doi.org/10.1109/ICCV.2015.425 Google Scholar
P. Young et al.,
“From image descriptions to visual denotations: new similarity metrics for semantic inference over event descriptions,”
Trans. Assoc. Comput. Linguist., 2 67
–78 https://doi.org/10.1162/tacl_a_00166
(2014).
Google Scholar
W. Li et al.,
“WebVision database: visual learning and understanding from web data,”
(2017). Google Scholar
E. Agustsson and R. Timofte,
“NTIRE 2017 challenge on single image super-resolution: dataset and study,”
in IEEE Conf. Comput. Vis. and Pattern Recognit. Workshops (CVPRW),
1122
–1131
(2017). https://doi.org/10.1109/CVPRW.2017.150 Google Scholar
I. Loshchilov and F. Hutter,
“Decoupled weight decay regularization,”
(2017). Google Scholar
O. Ronneberger, P. Fischer and T. Brox,
“U-Net: convolutional networks for biomedical image segmentation,”
Lect. Notes Comput. Sci., 9351 234
–241 https://doi.org/10.1007/978-3-319-24574-4_28 LNCSD9 0302-9743
(2015).
Google Scholar
A. Kirillov et al.,
“Segment anything,”
(2023). Google Scholar
F. Gori and M. Santarsiero,
“Devising genuine spatial correlation functions,”
Opt. Lett., 32 3531
–3533 https://doi.org/10.1364/OL.32.003531 OPLEDP 0146-9592
(2007).
Google Scholar
BiographyXin Tong is a PhD student at the School of Physics, Zhejiang University, Hangzhou, China. He received his BS degree in physics from Zhejiang University of Science and Technology, Hangzhou, China. His current research interests include holographic imaging, deep learning, computational imaging, and partial coherence theory. Renjun Xu received his PhD from the University of California, Davis, California, United States. He is a ZJU100 Young Professor and a PhD supervisor at the Center for Data Science, Zhejiang University, Hangzhou, China. He was the senior director of data and artificial intelligence at VISA Inc. His research interests include machine learning, alignment techniques for large-scale pretrained models, transfer learning, space editing, transformation, generation, and the interdisciplinarity of physics and mathematics. Pengfei Xu is a PhD student at the School of Physics, Zhejiang University, Hangzhou, China. He received his BS degree in physics from Zhejiang University, Hangzhou, China, in 2017. His current research interests include computational holographic imaging, partially coherent structured light field, and vortex beam manipulation techniques. Zishuai Zeng is a PhD student at the School of Physics, Zhejiang University, Hangzhou, China. He received his BS degree in 2019 from the School of Information Optoelectronic Science and Engineering at South China Normal University. His current research interests include computer-generated holography, as well as beam propagation transformation and computational imaging. Shuxi Liu is a PhD student at the School of Physics, Zhejiang University, China. He received his BS degree in physics from Zhejiang University in 2022. His current research interests include catastrophe optics, optical vortex, and computational imaging. Daomu Zhao received his PhD from Zhejiang University, Hangzhou, China. Since 2003, he has been as a professor of the School of Physics at Zhejiang University. Now, he is the director of the Institute of Optoelectronic Physics at Zhejiang University. He has broad research interests in beam transmission, coherence and polarization theory, diffraction optics, holographic imaging, and deep learning. |