Abstract
Deep learning-based image reconstruction methods have achieved remarkable success in phase recovery and holographic imaging. However, the generalization of their image reconstruction performance to new types of samples never seen by the network remains a challenge. Here we introduce a deep learning framework, termed Fourier Imager Network (FIN), that can perform end-to-end phase recovery and image reconstruction from raw holograms of new types of samples, exhibiting unprecedented success in external generalization. FIN architecture is based on spatial Fourier transform modules that process the spatial frequencies of its inputs using learnable filters and a global receptive field. Compared with existing convolutional deep neural networks used for hologram reconstruction, FIN exhibits superior generalization to new types of samples, while also being much faster in its image inference speed, completing the hologram reconstruction task in ~0.04 s per 1 mm2 of the sample area. We experimentally validated the performance of FIN by training it using human lung tissue samples and blindly testing it on human prostate, salivary gland tissue and Pap smear samples, proving its superior external generalization and image reconstruction speed. Beyond holographic microscopy and quantitative phase imaging, FIN and the underlying neural network architecture might open up various new opportunities to design broadly generalizable deep learning models in computational imaging and machine vision fields.
Similar content being viewed by others
Introduction
Digital holography provides unique advantages in microscopic imaging, by reconstructing the complex optical fields of input samples1,2,3,4,5,6,7,8,9,10,11,12. Due to the missing phase information, various computational approaches have been developed to digitally reconstruct holograms13,14,15,16,17,18,19,20,21,22. Recent work has also utilized deep neural networks23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45 to reconstruct the complex sample field from a hologram in a single forward inference step, achieving an image reconstruction quality comparable to iterative hologram reconstruction algorithms that are based on physical wave propagation. Some of the earlier results have also reported simultaneous performance of phase retrieval and autofocusing in a single network architecture, demonstrating holographic imaging over an extended depth-of-field25,34,42. In these earlier demonstrations, various deep network architectures, such as e.g., U-net-based convolutional neural networks (CNNs)23,25,27,33,37, recurrent neural networks (RNNs)42,46, as well as generative adversarial networks (GANs)31,37,42,43,47 have been proven to be effective for phase retrieval and hologram reconstruction for new (unseen) objects that belong to the same sample type of interest used during the training process. Stated differently, this earlier body of work has successfully demonstrated the “internal generalization” of the hologram reconstruction and phase retrieval networks to new objects of the same sample type as used in training.
On the other hand, “external generalization” to new objects from entirely new types of samples, never seen by the network before, remains a major challenge for deep neural networks, which might lead to image reconstruction degradation or hallucinations. Some studies have explored using transfer learning to address this challenge, which requires fine-tuning the network using a subset of the new types of samples48,49,50. In addition to this external generalization issue, it is, in general, difficult for CNN-based image reconstruction networks to accurately reconstruct raw holograms of samples due to the limited receptive field of convolutional layers, which casts another challenge considering the relatively large scale of holographic diffraction patterns of samples. As a result, existing end-to-end hologram reconstruction deep neural networks24,26,33,35,42 could only achieve decent reconstruction performance on relatively sparse samples. Alternatively, a pre-processing step, such as zero phase-padded free space propagation (FSP), has also been utilized to better deal with this issue23,25,26,40,42, which requires a precise physical forward model with the correct estimate of the axial propagation distance.
Here we introduce an end-to-end deep neural network, termed Fourier Imager Network (FIN), to rapidly implement phase recovery and holographic image reconstruction from raw holograms of new types of samples, achieving unprecedented success in external generalization. This framework takes in two or more input raw holograms captured at different sample-to-sensor distances without any pre-processing steps involved. By comprehensively utilizing the global spatial-frequency information processed by its trained spatial Fourier transform (SPAF) modules, FIN accurately reconstructs the complex field of the specimen, successfully demonstrating external generalization to new types of samples never used during its training. To experimentally demonstrate the success of this approach, we trained FIN models using human lung tissue samples (i.e., thin histopathology sections of connected tissue) and blindly tested the resulting trained FIN models on prostate and salivary gland tissue sections as well as Pap smear samples without compromising the image reconstruction quality. Compared to iterative hologram reconstruction algorithms based on wave propagation between different measurement planes, FIN is >27-fold faster to reconstruct an image. Compared to existing CNN-based deep learning models, FIN exhibits an unprecedented generalization performance, and is also much faster in its inference speed. We expect FIN to be widely used in various image reconstruction and enhancement tasks commonly employed in the computational microscopy field. In addition to coherent imaging, FIN can be applied to other image reconstruction or enhancement tasks in different imaging modalities, including e.g., fluorescence and brightfield microscopy.
Results
FIN provides an end-to-end solution for phase recovery and holographic image reconstruction, and its architecture is schematically presented in Fig. 1 (also see the Methods section). To acquire raw holograms of specimens, we used a lens-free in-line holographic microscope, as detailed in the Methods section, to image transmissive samples, such as human tissue samples and Pap smears, using a set of sample-to-sensor distances, i.e., z2,i, i = 1,⋯, M. The input images to FIN consist of M intensity-only raw holograms captured at z2,1 to z2,M and the network outputs are the reconstructed real and imaginary images of the object, revealing the complex-valued sample field. The corresponding ground truth images for supervised learning of FIN are obtained using an iterative multi-height phase retrieval (MH-PR) algorithm18 with M = 8 holograms acquired at different sample-to-sensor distances.
To demonstrate the success of FIN, we trained it using raw holograms of human lung tissue sections and tested the trained model on four different types of samples: (1) lung tissue samples from different patients never used in the training set (testing internal generalization), (2) Pap smear samples, (3) prostate tissue samples, and (4) salivary gland tissue samples, where (2,3,4) test external generalization, referring to new types of samples. The z2,i distances that we used in these holographic imaging experiments were 300, 450, and 600 μm (M = 3). After its training, our blind testing results (Fig. 2) reveal that FIN can not only reconstruct new lung tissue sections from new patients (internal generalization) but also achieves a strong external generalization performance on new sample types never seen by the network before. Furthermore, compared to the output of the MH-PR algorithm using the same input (raw hologram) data (M = 3), FIN is ~27.3 times faster in its inference speed per image (see Table 1) and delivers better reconstruction quality on all the test samples, as highlighted by the yellow arrows and the zoomed-in regions in Fig. 2.
To further showcase the generalization ability of FIN, we separately trained four FIN models using lung tissue, prostate tissue, salivary gland tissue, and Pap smear hologram datasets (i.e., one type of sample for each network model), and blindly tested the trained models on unseen FOVs from four types of samples using M = 3 raw holograms for each FOV. Similar to our conclusions reported in Fig. 2, Fig. 3a also shows that the FIN model trained using each sample type can successfully generalize to other types of samples. Even when FIN was only trained using relatively sparse samples such as Pap smear slides, the resulting network model successfully generalized to reconstruct the raw holograms of connected tissue sections that significantly deviate from the structural distribution and sparsity observed in Pap smear samples. The success of the reconstruction performance of FIN was also quantified using the structural similarity index (SSIM)51 across all four FIN networks that were trained using different types of samples (see Fig. 3b), demonstrating the superior generalization of these FIN models regardless of the distribution shifts observed between the training and testing data.
Next, we evaluated the hologram reconstruction performance of FIN when only two input holograms were measured, i.e., M = 2. For this, we trained ten different FIN models from scratch using the same human lung tissue sections but with different sets of z2,1 and z2,2, such that the sample-to-sensor distances for different FIN models were different. These trained FIN models were then blindly tested on new lung tissue sections from new patients (internal generalization); Fig. 4 reports the amplitude and phase root mean square error (RMSE) of the reconstructed holographic images generated by FIN (M = 2) and MH-PR (M = 2) for different combinations of z2,1 and z2,2. Both the amplitude and phase RMSE values show that FIN achieves a significant reconstruction quality improvement compared to MH-PR (M = 2), and the RMSE values of FIN models are consistently better with different sample-to-sensor distances varying from 300 μm to 600 μm. The visualization of the reconstructed holograms shown in Fig. 4 further confirms the same conclusion that FIN achieves consistently better image reconstruction compared to MH-PR for various combinations of z2,1 and z2,2. We also confirmed that the same conclusions apply to the external generalization tests of FIN (M = 2).
In addition to MH-PR based comparisons, we also extended our performance analysis to other deep learning-based phase retrieval and hologram reconstruction methods. For this additional set of comparisons, we used a state-of-the-art deep learning model based on a recurrent convolutional neural network, termed RH-M, that was developed for multi-height holographic image reconstruction42. Using the same training hologram data, we trained FIN and RH-M models for different M values, the blind testing results of which are compared in Fig. 5. As for the internal generalization performance shown in Fig. 5, both FIN and RH-M can successfully generalize to new lung tissue samples from new patients. However, for the external generalization to new sample types (prostate and salivary gland tissue as well as Pap smear samples), FIN provides superior image reconstruction performance even though it uses a smaller number of trainable parameters compared to RH-M (see Table 1); in comparison, RH-M has reconstruction artifacts on external testing sets for both M = 3 and M = 4, also confirmed by the significantly lower SSIM values (for RH-M reconstructions) reported in Fig. 5.
In addition to its superior generalization performance, FIN also has faster inference speed compared to deep learning-based or iterative phase retrieval algorithms. In Table 1, we compared the inference time of FIN, RH-M, and MH-PR algorithms. Noticeably, FIN has the shortest inference time among these methods using any number of raw input holograms. For the case of M = 3, FIN is ~9.3-fold faster than RH-M and ~27.3-fold faster than MH-PR, which highlights the computational efficiency of our network. We can further accelerate the inference speed of FIN by using parallelization, which reduces the computation time to 0.04 s/mm2 under an image batch size of 20 (see Table 1). We should also note that the number (M) of input holograms has a negligible impact on the inference time of FIN, since it uses a fixed channel size for most parts of the network model, and M only affects the first 1 × 1 convolutional layer. That is why the inference times of FIN (M = 3) and FIN (M = 4) are approximately the same as shown in Table 1. Refer to the Methods section for further details.
To further demonstrate the advantages of FIN, we performed a transfer learning-based comparison; for this, we transferred FIN and RH-M models to unseen sample types to compare their generalization performance as shown in Table 2 (see the Methods). In this comparison, the original models of FIN and RH-M were trained on lung tissue sections and then separately transferred to Pap smear, human prostate, and salivary gland tissue sections, creating 3 models for each method. The transferred FIN and RH-M models were then tested on blind test sets of the corresponding sample types. Table 2 reveals that (1) FIN outperforms RH-M in all cases, and (2) RH-M gains significant performance improvement after transfer learning on all three types of samples that were never seen before. On the contrary, FIN only got differential improvement after the corresponding transfer learning step, due to its originally strong generalization to new types of samples (see Table 2).
Discussion
We demonstrated an end-to-end phase retrieval and hologram reconstruction network that is highly generalizable to new sample types. FIN outperforms other phase retrieval algorithms in terms of both the reconstruction quality and speed. This method presents superior generalization capability to new types of samples without any prior knowledge about these samples or any fine-tuning of its trained model. This strong external generalization of our model mainly stems from the regularization effect of the SPAF modules in its architecture. In a lensfree holographic imaging system, the Fourier transforms of the fields at the sample plane and the measurement plane are related by a frequency-dependent phasor, which can be effectively learned through the element-wise multiplication module in SPAF. The strong external generalization of FIN allows the model to be robust and effective on testing data distributions that are different from the training data distribution, which can result from e.g., imaging set-up misalignments, poor signal-to-noise (SNR) ratio, and sample related deformations or changes. The application of FIN for different imaging methods to potentially build cross-modality generalization is left as future work.
Besides, SPAF modules provide a global receptive field to FIN, in contrast to the limited, local receptive fields of common CNNs. The global receptive field helps the FIN model more effectively process the holographic diffraction patterns for various samples, regardless of the morphologies and dimensions of the objects. In fact, previous research has already shown that end-to-end hologram reconstruction requires a larger network receptive field, which can be partially addressed by using e.g., dilated convolution42. In our method, the Fourier transform intrinsically captures the global spatial information of the sample and thus provides a maximized receptive field for FIN, contributing to its performance gain over CNN-based hologram reconstruction models reported in Fig. 5. Like FIN, other deep neural networks52 have also utilized learnable spatial Fourier transform modules for inference, for example, to successfully map the initial and/or boundary conditions of partial differential equations (PDEs) and infer numerical solutions.
Unlike fully convolutional networks, in FIN architecture, the size of the input raw hologram FOV is fixed at the beginning, i.e., we cannot use a larger FOV in the testing phase because of the element-wise multiplication in our SPAF module. A larger FOV raw hologram can be reconstructed using FIN by dividing the hologram into smaller FOVs and running them through FIN in parallel. This parallelization of a large FOV hologram reconstruction is feasible since FIN has a significant speed advantage in its inference, and can reconstruct ~1 mm2 sample area within 0.04 sec using a standard GPU (see Table 1). The total inference time for an image FOV with ~4000 × 3000 pixels can be reduced to ~0.07 sec after parallelization, allowing real-time hologram reconstruction and imaging using FIN. This fast inference speed of FIN also opens up new opportunities to establish generalizable deep learning models for a broad spectrum of computational imaging tasks.
Materials and methods
Holographic Imaging
A lens-free in-line holographic microscope was utilized to capture the raw holograms of the specimens. A broadband light source (WhiteLase Micro, NKT Photonics) and an acousto-optic tunable filter (AOTF) were used as the illumination source emitting 530 nm light. The image sensor was a complementary metal-oxide-semiconductor (CMOS) RGB image sensor (IMX081, Sony). The light source, sample, and the CMOS image sensor were aligned vertically. The sample was directly placed between the light source and the sensor such that the sample-to-source distance (z1) was about 10 cm, and the sample-to-sensor distance (z2) ranged from 300 to 600 μm. The CMOS sensor was placed on and controlled by a 6-axis stage (MAX606, Thorlabs) to perform lateral and axial shifts. All hardware was connected to a computer and controlled by a customized LabVIEW program to capture holograms automatically.
All the human samples involved in this work were deidentified and prepared from existing specimens that were captured before this research. Human prostate, salivary gland, and lung tissue slides were provided by the UCLA Translational Pathology Core Laboratory (TPCL). Pap smear slides were prepared by the UCLA Department of Pathology.
Pre-processing
The captured raw holograms were firstly processed by a pixel super-resolution algorithm18,53,54. The 6-axis stage was programmed to automatically capture in-line holograms at 6 × 6 lateral positions with sub-pixel shifts. The super-resolution algorithm estimated the relative shifts for each hologram and merged these holograms using a shift-and-add algorithm18. The effective pixel size of the generated super-resolved holograms decreases to 0.37 μm from the original CMOS pixel size of 2.24 μm. The resulting super-resolved holograms were cropped into unique patches of 512 × 512 pixels, without any overlap. Hologram datasets of each sample type were partitioned into training and testing sets, at a ratio of 6:1, comprising ~600 unique FOVs in each training set and ~100 FOVs for the testing set. The testing FOVs were strictly obtained from different whole slides (new patients) excluded in the training sets.
The ground truth sample fields were retrieved by an iterative multi-height phase retrieval algorithm18. At each sample FOV, M = 8 in-line holograms were captured at different sample-to-sensor distances, which were later estimated by an autofocusing algorithm using the edge sparsity criterion55. In each iteration, the estimated sample field is digitally propagated to each hologram plane using the angular spectrum propagation56. The propagated complex field is updated according to the measurement at each hologram plane, by averaging the amplitude of the propagated field with the measured amplitude and retaining the new estimated phase. One iteration is completed after all the hologram planes are used, and this MH-PR algorithm converges within 100 iterations.
Network structure
The FIN network architecture has a Residual in Residual architecture shown in Fig. 1, inspired by RCAN57 to have a deeper network structure and better information flow. Our network, FIN, consists of several SPAF modules with a long skip connection to form the large-scale residual connection, in conjunction with two 1 × 1 convolutional layers at the head and tail of the network. Each SPAF group contains two recursive SPAF modules, which share the same parameters to improve the network capacity without significantly enlarging the size of the network. A short skip connection is introduced for every SPAF group to form the middle-scale residual connection, and a small-scale residual connection is used to connect the inputs and outputs of each SPAF module. SPAF module, as shown in Fig. 1b, has a linear transformation applied to the tensor after it was transformed into the frequency domain using the 2D Discrete Fourier transform, following a similar architecture as in Ref. 52; a half window size of k is applied to truncate the higher frequency signals, i.e.,
where \(F \in {\mathbb{C}}^{c,2k + 1,2k + 1}\) is the truncated frequency domain representation of the input to the SFAP module after performing the 2D Discrete Fourier Transform, \(W \in {\mathbb{R}}^{c,c,2k + 1,2k + 1}\) represents the trainable weights, c is the channel number, and k is the half window size. After this linear transformation, the inverse 2D Discrete Fourier transform is used to obtain the processed data back in the spatial domain, followed by a PReLU activation function.
where a is a learnable parameter.
To adapt the SPAF module to high-resolution image processing in a deeper network, we shrank the matrix W allowing a significant model size reduction. The optimized linear transformation is defined as
where \(F \in {\mathbb{C}}^{c,2k + 1,2k + 1}\) is the truncated frequency components, and \(W^{\prime} \in {\mathbb{R}}^{c,2k + 1,2k + 1}\) represents the trainable weights.
To further optimize the network structure for high-resolution holographic image reconstruction, a set of decreasing half window sizes (k) was chosen for the SPAF modules. Specifically, both of the SPAF modules in each SPAF group have shared hyperparameters, and we set a decreasing half window size k for the SPAF groups in the sequence of the network structure, which forms a pyramid-like structure. This pyramid-like structure provides a mapping of the high-frequency information of the holographic diffraction patterns to low-frequency regions in the first few layers and passes this low-frequency information to the subsequent layers with a smaller window size, which better utilizes the features at multiple scales and at the same time considerably reduces the model size, avoiding potential overfitting and generalization issues.
Network implementation
The networks are implemented using PyTorch58 with GPU acceleration and are trained and tested on the same computer with an Intel Xeon W-2195 CPU, 256 GB memory, and NVidia RTX 2080 Ti GPUs. During the training phase, the input FOVs of 512 × 512 pixels were randomly selected from the training hologram dataset, and data augmentation was applied to each FOV, which includes random image rotations of 0, 90, 180, or 270 degrees.
The training loss is the weighted sum of three different loss terms:
where α, β, and γ are set as 0.5, 1, and 0.5, respectively. The MAE loss and complex domain loss can be expressed as:
where y is the ground truth, \(\hat y\) is the network’s output, n is the total number of pixels, and \({{{\mathcal{F}}}}\) stands for the 2D Discrete Fourier Transform operation. For the perceptual loss term59, we used a pre-trained VGG16 network as the feature extractor to minimize the Euclidean distance between the low-level features of the reconstructed images and the ground truth images.
The trainable parameters of the deep neural network models were learned iteratively using the Adam optimizer60 and the cosine annealing scheduler with warm restarts61 was used to dynamically adjust the learning rate during the training phase. Every model went through 1,000 epochs and we selected the best one with the lowest validation loss. When performing transfer learning to a new sample type (e.g., Table 2), we transferred FIN and RH-M models using a smaller dataset containing 100 unique FOVs of the target sample type.
In the testing phase, a batch of test holograms with the same resolution (512 × 512 pixels) is fed to the network, and the inference time for one FOV at a time (batch size is set to 1) is 0.52 s/mm2. Additionally, using the same Nvidia RTX 2080 Ti GPU, the inference can be parallelized with a batch size of 20, resulting in 0.04 s/mm2 inference time (Table 1).
References
Kou, S. S. et al. Transport-of-intensity approach to differential interference contrast (TI-DIC) microscopy for quantitative phase imaging. Opt. Lett. 35, 447–449 (2010).
Popescu, G. Quantitative phase imaging of cells and tissues. (New York: McGraw-Hill, 2011).
Chhaniwal, V. et al. Quantitative phase-contrast imaging with compact digital holographic microscope employing Lloyd’s mirror. Opt. Lett. 37, 5127–5129 (2012).
Jericho, M. H. et al. Quantitative phase and refractive index measurements with point-source digital in-line holographic microscopy. Appl. Opt. 51, 1503–1515 (2012).
Su, T. W., Xue, L. & Ozcan, A. High-throughput lensfree 3D tracking of human sperms reveals rare statistics of helical trajectories. Proc. Natl Acad. Sci. USA 109, 16018–16022 (2012).
Greenbaum, A. et al. Wide-field computational color imaging using pixel super-resolved on-chip microscopy. Opt. Express 21, 12469–12483 (2013).
Greenbaum, A. et al. Wide-field computational imaging of pathology slides using lens-free on-chip microscopy. Sci. Transl. Med. 6, 267ra175 (2014).
Tian, L. & Waller, L. Quantitative differential phase contrast imaging in an LED array microscope. Opt. Express 23, 11394–11403 (2015).
Merola, F. et al. Tomographic flow cytometry by digital holography. Light Sci. Appl. 6, e16241 (2017).
Park, Y., Depeursinge, C. & Popescu, G. Quantitative phase imaging in biomedicine. Nat. Photonics 12, 578–589 (2018).
Barbastathis, G., Ozcan, A. & Situ, G. H. On the use of deep learning for computational imaging. Optica 6, 921–943 (2019).
Javidi, B. et al. Roadmap on digital holography [Invited]. Opt. Express 29, 35078–35118 (2021).
Fienup, J. R. Phase retrieval algorithms: a comparison. Appl. Opt. 21, 2758–2769 (1982).
Teague, M. R. Deterministic phase retrieval: a Green’s function solution. J. Optical Soc. Am. 73, 1434–1441 (1983).
Yang, G. Z. et al. Gerchberg–Saxton and Yang–Gu algorithms for phase retrieval in a nonunitary transform system: a comparison. Appl. Opt. 33, 209–218 (1994).
Allen, L. J. & Oxley, M. P. Phase retrieval from series of images obtained by defocus variation. Opt. Commun. 199, 65–75 (2001).
Marchesini, S. Invited article: a unified evaluation of iterative projection algorithms for phase retrieval. Rev. Sci. Instrum. 78, 011301 (2007).
Greenbaum, A. & Ozcan, A. Maskless imaging of dense samples using pixel super-resolution based multi-height lensfree on-chip microscopy. Opt. Express 20, 3129–3143 (2012).
Greenbaum, A., Sikora, U. & Ozcan, A. Field-portable wide-field microscopy of dense samples using multi-height pixel super-resolution based lensfree imaging. Lab a Chip 12, 1242–1245 (2012).
Luo, W. et al. Synthetic aperture-based on-chip microscopy. Light Sci. Appl. 4, e261 (2015).
Rivenson, Y. et al. Sparsity-based multi-height phase recovery in holographic microscopy. Sci. Rep. 6, 37862 (2016).
Luo, W. et al. Propagation phasor approach for holographic image reconstruction. Sci. Rep. 6, 22738 (2016).
Rivenson, Y. et al. Phase recovery and holographic image reconstruction using deep learning in neural networks. Light Sci. Appl. 7, 17141 (2018).
Wang, H., Lyu, M. & Situ, G. H. eHoloNet: a learning-based end-to-end approach for in-line digital holographic reconstruction. Opt. Express 26, 22603–22614 (2018).
Wu, Y. C. et al. Extended depth-of-field in holographic imaging using deep-learning-based autofocusing and phase recovery. Optica 5, 704–710 (2018).
Goy, A. et al. Low photon count phase retrieval using deep learning. Phys. Rev. Lett. 121, 243902 (2018).
Zhang, G. et al. Fast phase retrieval in off-axis digital holographic microscopy through deep learning. Opt. Express 26, 19388–19405 (2018).
Ren, Z. B., Xu, Z. M. & Lam, E. Y. Learning-based nonparametric autofocusing for digital holography. Optica 5, 337–344 (2018).
Rivenson, Y., Wu, Y. C. & Ozcan, A. Deep learning in holography and coherent imaging. Light Sci. Appl. 8, 85 (2019).
Liu, T. R. et al. Deep learning-based color holographic microscopy. J. Biophotonics 12, e201900107 (2019).
Wu, Y. C. et al. Bright-field holography: cross-modality deep learning enables snapshot 3D imaging with bright-field contrast using a single hologram. Light Sci. Appl. 8, 25 (2019).
Jo, Y. et al. Quantitative phase imaging and artificial intelligence: a review. IEEE J. Sel. Top. Quantum Electron. 25, 6800914 (2019).
Wang, K. Q. et al. Y-Net: a one-to-two deep learning framework for digital holographic reconstruction. Opt. Lett. 44, 4765–4768 (2019).
Byeon, H., Go, T. & Lee, S. J. Deep learning-based digital in-line holographic microscopy for high resolution with extended field of view. Opt. Laser Technol. 113, 77–86 (2019).
Ren, Z. B., Xu, Z. M. & Lam, E. Y. M. End-to-end deep learning framework for digital holographic reconstruction. Adv. Photonics 1, 016004 (2019).
Li, H. Y. et al. Deep DIH: single-shot digital in-line holography reconstruction by deep learning. IEEE Access 8, 202648 (2020).
Moon, I. et al. Noise-free quantitative phase imaging in Gabor holography with conditional generative adversarial network. Opt. Express 28, 26284–26301 (2020).
Zeng, T., So, H. K. H. & Lam, E. Y. RedCap: residual encoder-decoder capsule network for holographic image reconstruction. Opt. Express 28, 4876–4887 (2020).
Liu, T. R. et al. Deep learning-based holographic polarization microscopy. ACS Photonics 7, 3023–3034 (2020).
Deng, M. et al. Learning to synthesize: robust phase retrieval at low photon counts. Light Sci. Appl. 9, 36 (2020).
Yin, D. et al. Digital holographic reconstruction based on deep learning framework with unpaired data. IEEE Photonics J. 12, 3900312 (2020).
Huang, L. Z. et al. Holographic image reconstruction with phase recovery and autofocusing using recurrent neural networks. ACS Photonics 8, 1763–1774 (2021).
Yang, X. et al. High imaging quality of Fourier single pixel imaging based on generative adversarial networks at low sampling rate. Opt. Lasers Eng. 140, 106533 (2021).
Shimobaba, T. et al. Deep-learning computational holography: a review (invited). Front. Photonics 3, 854391 (2022).
Pirone, D. et al. Speeding up reconstruction of 3D tomograms in holographic flow cytometry via deep learning. Lab a Chip 22, 793–804 (2022).
Huang, L. Z. et al. Recurrent neural network-based volumetric fluorescence microscopy. Light Sci. Appl. 10, 62 (2021).
Goodfellow, I. et al. Generative adversarial networks. Commun. ACM 63, 139–144 (2020).
Kim, S. J. et al. Deep transfer learning-based hologram classification for molecular diagnostics. Sci. Rep. 8, 17003 (2018).
Zhu, Y. M., Yeung, C. H. & Lam, E. Y. Digital holographic imaging and classification of microplastics using deep transfer learning. Appl. Opt. 60, A38–A47 (2021).
Huang, L. Z. et al. Few-shot transfer learning for holographic image reconstruction using a recurrent neural network. APL Photonics 7, 070801 (2022).
Wang, Z., Simoncelli, E. P. & Bovik, A. C. Multiscale structural similarity for image quality assessment. The Thirty-Seventh Asilomar Conference on Signals, Systems & Computers, 2003. Pacific Grove: IEEE, 2003, 1398-1402.
Li, Z. Y. et al. Fourier neural operator for parametric partial differential equations. The 9th International Conference on Learning Representations (ICLR-21). https://iclr.cc/virtual/2021/poster/3281 (2021).
Bishara, W. et al. Lensfree on-chip microscopy over a wide field-of-view using pixel super-resolution. Opt. Express 18, 11181–11191 (2010).
Greenbaum, A. et al. Imaging without lenses: achievements and remaining challenges of wide-field on-chip microscopy. Nat. Methods 9, 889–895 (2012).
Zhang, Y. B. et al. Edge sparsity criterion for robust holographic autofocusing. Opt. Lett. 42, 3824–3827 (2017).
Goodman, J. W. Introduction to Fourier Optics. 3rd edn. (Greenwoood Village: Roberts & Company Publishers, 2005).
Zhang, Y. L. et al. Image super-resolution using very deep residual channel attention networks. The 15th European Conference on Computer Vision. Munich, Germany: Springer, 2018, 294-310.
Paszke, A. et al. PyTorch: an imperative style, high-performance deep learning library. Proceedings of the 33rd International Conference on Neural Information Processing Systems. Vancouver: ACM, 2019, 8026-8037.
Johnson, J., Alahi, A. & Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. The 14th European Conference on Computer Vision. Amsterdam: Springer, 2016, 694-711.
Kingma, D. P. & Ba, J. Adam: a method for stochastic optimization. Print at https://doi.org/10.48550/arXiv.1412.6980 (2014).
Loshchilov, I. & Hutter, F. SGDR: stochastic gradient descent with warm restarts. Print at https://doi.org/10.48550/arXiv.1608.03983 (2016).
Acknowledgements
The Ozcan Research Group at UCLA acknowledges the support of NSF.
Author information
Authors and Affiliations
Contributions
A.O. conceived the research, H.C. and L.H. conducted the numerical experiments and analyzed the results. T.L. captured the experiment data, L.H. and T.L. processed and analyzed the data. All the authors contributed to the preparation of the manuscript. A.O. supervised the research.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Chen, H., Huang, L., Liu, T. et al. Fourier Imager Network (FIN): A deep neural network for hologram reconstruction with superior external generalization. Light Sci Appl 11, 254 (2022). https://doi.org/10.1038/s41377-022-00949-8
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41377-022-00949-8
This article is cited by
-
Digital in-line holographic microscopy for label-free identification and tracking of biological cells
Military Medical Research (2024)
-
All-optical complex field imaging using diffractive processors
Light: Science & Applications (2024)
-
On the use of deep learning for phase recovery
Light: Science & Applications (2024)
-
Liquid lens based holographic camera for real 3D scene hologram acquisition using end-to-end physical model-driven network
Light: Science & Applications (2024)
-
Non-orthogonal optical multiplexing empowered by deep learning
Nature Communications (2024)