Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

DeepLSR: a deep learning approach for laser speckle reduction

Open Access Open Access

Abstract

Speckle artifacts degrade image quality in virtually all modalities that utilize coherent energy, including optical coherence tomography, reflectance confocal microscopy, ultrasound, and widefield imaging with laser illumination. We present an adversarial deep learning framework for laser speckle reduction, called DeepLSR (https://durr.jhu.edu/DeepLSR), that transforms images from a source domain of coherent illumination to a target domain of speckle-free, incoherent illumination. We apply this method to widefield images of objects and tissues illuminated with a multi-wavelength laser, using light emitting diode-illuminated images as ground truth. In images of gastrointestinal tissues, DeepLSR reduces laser speckle noise by 6.4 dB, compared to a 2.9 dB reduction from optimized non-local means processing, a 3.0 dB reduction from BM3D, and a 3.7 dB reduction from an optical speckle reducer utilizing an oscillating diffuser. Further, DeepLSR can be combined with optical speckle reduction to reduce speckle noise by 9.4 dB. This dramatic reduction in speckle noise may enable the use of coherent light sources in applications that require small illumination sources and high-quality imaging, including medical endoscopy.

© 2019 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Laser illumination offers many advantages over incoherent light for imaging, including high power densities, efficient light generation, narrow spectral bandwidths, robust stability, long lifetimes, and fast triggering capabilities. Unfortunately, coherent illumination also introduces speckle artifacts that are caused by constructive and destructive interference between emitted wavefronts [1]. The poor image quality resulting from speckle noise prohibits lasers from being used in many widefield imaging applications. For example, commercial endoscopes utilize arc lamps or light-emitting diodes (LEDs) as illumination sources, and consequently require large-diameter light guides to transmit sufficient illumination power. Speckle noise also corrupts image quality in optical coherence tomography (OCT) [2], reflectance confocal microscopy [3], and ultrasound imaging [4]. To mitigate laser speckle noise, several optical methods have been explored [5,6]. In general, optical approaches add cost and complexity, reduce power throughput, and place fundamental limitations on imaging speed.

While image processing methods for laser speckle reduction have been most prominently developed for OCT [7–10], conventional model-based algorithms may be applied to speckle reduction in widefield imaging. In general, these denoising approaches can be grouped into three categories: total variation-based, non-local methods, and sparse filtering. Rudin et al. first introduced the concept of Total Variation (TV) denoising [11], which has been shown to work extremely well on piecewise-like images, but often suffers from a characteristic “staircase effect” in non-constant regions of images. Abergel et al. expanded TV to include iterative conditional expectations [12], and more recent work incorporated a priori models for noise in TV [13]. Multiplicative Image Denoising by Augmented Lagrangian (MIDAL) [14] utilizes a multiplicative noise model (with knowledge of the noise standard deviation) and optimization through a Lagrangian framework, and generally mitigates noise with only some loss in image texture. Buades et al. were the first to propose a non-local algorithm for noise reduction, called Non-Local Means (NLM) [15]. NLM relies on the assumption that for each noisy feature in an image, there exists similar non-local features that may be combined to separate noise from the common, underlying image features. NLM has been adapted for a probabilistic Poisson noise model [16], and also paired with Principal Component Analysis (PCA) [17] as well as the Wiener filter [18]. Sparse filtering methods seek to maximize the variety of image features in a learned dictionary that is later averaged to reduce image noise. Aharon et al. introduced a joint K-mean clustering and Singular Value Decomposition (K-SVD) approach [19], which was further optimized by Rubenstein et al [20]. Block-Matching 3D Filtering (BM3D) and its variants are the de-facto state-of-the-art for image denoising using sparse filtering [21–23]. Similar to K-SVD, BM3D relies on the aggregation of noisy blocks with comparable features for collaborative filtering and weighted averaging to reduce noise. This method was later expanded to RGB images by way of a transformation to the YUV image space [24]. An in-depth review of prior work in relevant image denoising techniques is provided by Meiniel et al [23].

In general, conventional model-based image processing techniques are computationally complex and require parameter tuning. Machine learning approaches, on the other hand, can generate complex transformation functions by training on datasets that contain example input and desired output images. Moreover, deep learning has emerged as a powerful technique to learn complex representations of imaging data using multi-layer neural networks. Here, we present a deep convolutional neural network for laser speckle reduction (‘DeepLSR’) on widefield images formed from multi-wavelength, red-green-blue laser illumination. We describe a method for effectively learning the distribution of speckle artifacts to target and reduce noise in images not previously seen by the network. This technique relies on a training set of coherent- and incoherent-illuminated image pairs of a variety of objects to learn a transformation from speckled to speckle-free images. Previous efforts in OCT have explored shallow neural networks for estimating filter parameters in a speckle reduction model [25], and deep networks for speckle reduction using a set of registered and averaged volumes of retinal tissue as ground truth [26]. CNNs have previously been shown to be effective for recovering image information degraded by scattering media [27,28]. In widefield imaging, deep learning networks have been applied for general image denoising [29], but not specifically for speckle reduction. DeepLSR is novel in its use of a true incoherent source as a target ground truth, the use of a diverse set of objects for training a generalizable model, and in its application of deep learning to widefield laser-illumination images. We benchmark this approach against conventional speckle reduction methods on images of laser-illuminated objects previously unseen by the network. We further provide step-by-step instructions for adapting DeepLSR to new data sets contaminated with speckle noise (see Appendix).

Standard deep learning models employ handcrafted loss functions that utilize repeated pixel-wise comparisons between the model’s prediction and the ground truth for model refinement. However, such pixel-wise loss functions do not capture higher order statistics that exist in the training data, such as non-local dependencies [30]. To address relationships beyond the second order and to capture spatial relationships amongst distant pixels, recent focus has shifted to generative models to improve translations between higher dimensional data. Such methods have been used for image-to-image translation tasks with applications that include artistic style transfer [31], super-resolution imaging [32], and synthetic data refinement [33].

DeepLSR utilizes a conditional Generative Adversarial Network (cGAN) to reduce laser speckle by posing the problem as an image-to-image translation task [34]. The overall architecture involves simultaneously training a speckle-free image generator and a real-versus-fake image discriminator, given a conditional input (Fig. 1). While the generator learns to generate a realistic mapping from an input speckled image to an output speckle-free image, the discriminator learns to classify pairs of input and generated output images as either real or fake. During this adversarial training, the discriminator provides feedback to the generator. The trained generator is then capable of reducing speckle noise in images it has never seen. Previous research in adversarial image-to-image translation has been challenged by instability caused by the complexity of back-and-forth training between the generator and the discriminator. Furthermore, this training paradigm can struggle with learning multiple sub-distributions of data that may exist within a certain distribution. This problem is often referred to as mode collapse. To overcome this challenge, we utilize spectral normalization instead of the more commonly used batch normalization for training DeepLSR [35]. With this adversarial framework, we trained networks to reduce speckle in images of a wide assortment of objects to evaluate robustness, and in images of tissue for evaluation in an endoscopic setting.

 figure: Fig. 1

Fig. 1 DeepLSR Architecture. (a) Training architecture for image-to-image translation-based laser speckle reduction using a conditional Generative Adversarial Network. A generator learns to transform between pairs of images acquired with coherent and incoherent illumination while a discriminator learns to classify input images as real or fake. (b) Once training is complete, the discriminator is discarded and the trained generator (DeepLSR) reduces laser speckle noise in images not previously seen by the generator.

Download Full Size | PDF

2. Methods

We trained and tested DeepLSR using images acquired of assorted household and laboratory objects picked to represent a wide range of textures, shapes, and bidirectional reflectance distribution functions and images of ex-vivo porcine esophagus, intestine, and stomach from three animals. These samples were illuminated using a red-green-blue laser for coherent illumination, the same laser with a commercial optical laser speckle reducer (oLSR) utilizing an oscillating diffuser, and an LED for incoherent illumination. We trained two networks to learn two transformations: (1) from laser illuminated images to LED illuminated images (DeepLSR), and (2) from optically laser speckle reduced images to LED illuminated images (DeepLSR+oLSR).

2.1. Imaging setup

Paired images for adversarial training were captured using laser illumination, laser illumination with optical speckle reduction, and light-emitting diode (LED) illumination (Fig. 2). A laser unit with 445nm, 520nm, and 638nm diodes (Optlasers microRGB) and a white-light LED (Luxeon Star Tri-Star) with Rebel 448nm, 530nm, and 655nm diodes were used to illuminate samples. Diodes were selected so that wavelengths emitted by the laser and LED were similar. The laser unit’s beam was positioned normal to the aperture of an Optical Laser Speckle Reducer [36] (oLSR, Optotune LSR-3005-24D), and the oLSR was toggled on and off for imaging with and without speckle reduction. A frosted poly-carbonate triple lens (Luxeon Star #10508) and a 600 grit diffusion lens were used to match the full-width at half maximum of the LED’s illumination intensity profile to that of the laser. For profile-matching, the laser and LED were both aimed at a Teflon imaging target placed below a color, 8-bit CMOS detector (ThorLabs #DCC3240C, 12mm/F1.8) with an integration time set to 50 ms. A linear polarizing sheet (Thorlabs, #LPVISE2X3) was placed in front of the light sources and a linear polarizer (Edmund Optics #47316) was mounted to the detector and adjusted to minimize specular reflection by cross polarization. The illumination intensities of the laser and LED sources were matched by imaging the Teflon sheet and modulating the power of each individual laser diode color channel to achieve the same average pixel value as the corresponding LED channel. Periodic adjustments were made as the diodes’ power drifted over time. The light source and detector positions were kept constant, and all image scenes were kept static while collecting each triad of images illuminated with laser light, laser light with optical speckle reduction, and LED illumination. Because the imaging setup was fixed when illumination was toggled, image registration and further calibration were not needed.

 figure: Fig. 2

Fig. 2 Imaging setup for acquiring images with: (1) laser illumination (oLSR turned off), (2) laser illumination with optical laser speckle reduction (oLSR turned on), and (3) LED illumination.

Download Full Size | PDF

2.2. Data acquisition and preprocessing

Data was acquired from: (1) 113 assorted household and laboratory objects imaged at up to 9 different positions, with each illumination source, resulting in 1533 images, and (2) 449 images of ex-vivo porcine esophagus, intestine, and stomach from three animals, with each illumination source, resulting in 1347 images. The position of objects and tissues were varied using translation and rotation to collect multiple images of the same sample with different speckle interference patterns for both training and testing. Before training, the histograms of each laser and oLSR image were adjusted to match the corresponding LED image using uniform histogram matching to correct for any white-balancing discrepancies. The 2880 images were resized from 1280×1024 to 1024×1024 pixels using bicubic interpolation. Images were then divided into sets of laser, oLSR, and LED images to be paired with their corresponding ground truth images for network training. 90 images (39 images from 6 objects, and 51 images acquired from porcine tissue) were removed from the dataset for final network testing. The objects imaged for testing were not represented in the training set. Similarly, the porcine tissue images for testing came from a different animal than the images used for training.

2.3. Network training

An adversarial deep learning paradigm was used to train the networks. To this effect, two deep networks, a generator and a discriminator, were iteratively trained. The generator was tasked with generating target images and the discriminator was tasked with classifying the generator output as real or fake and giving feedback to the loss function of the generator. This paradigm enables the network to use non-local information when making determinations.

The network proposed here has two loss terms: a GAN loss which is updated every iteration based on the feedback from the discriminator and an 1 loss. The 1 loss compliments the GAN loss to minimize the difference between the output and ground truth. By inserting this term, the objective of the generator is not only to fool the discriminator but also to be close to the ground truth output. Recent work by Kuarch et al. has demonstrated that in GANs, hand-crafting additional loss functions is not significantly beneficial to the overall adversarial training process [37]. Because 2 is more sensitive to outliers in the prediction, there is a possibility that it will try to adjust the model according to these outlier values, even at the expense of other well predicted samples. For this reason, we use an 1 loss function, as do many GAN-based image-to-image translation methods [34]. To prevent mode collapse, images were pooled and fed to the discriminator in batches rather than individual images in each iteration. Spectral normalization was used to stabilize GAN training when learning simultaneously from assorted objects and tissue. The problem was solved using Adam for stochastic optimization [38]. Further details about the generator and discriminator architectures can be found in [39].

The network was trained for 400 epochs. The learning rate was set to 0.0002 for the first 200 epochs and linearly decayed to a learning rate of zero over the remaining 200 epochs. The size of the image buffer that stores generated images was set to 64. The networks were implemented using PyTorch 0.4 and the training was run on Nvidia P100 GPUs using Google Cloud. The average training time for each epoch was 303 seconds and the entire network was trained in approximately 33.66 hours. Once the training process is complete, the trained network computes speckle-reduced images at 6 frames per second on a virtual workstation with 4 CPUs on a 2.6 GHz Intel Xeon E5 processor and at 27 frames per second when using a P100 GPU.

2.4. Evaluation metrics

To quantify the performance of DeepLSR, we measured the peak signal-to-noise (PSNR) ratio and Structural Similarity Index (SSIM) [40] between computed images and the incoherent, speckle-free images. PSNR assesses the relative noise of an image and was computed using,

PSNR=10log10(R2MSE)
where the maximum possible intensity R is 255 and MSE is the mean squared error between images. SSIM is a valuable metric for image comparison using quantities that are important for human perception (image contrast, luminance, and structure). SSIM was computed using,
SSIM(x,y)=(2μxμy+C1)(2σxy+C2)(μx2+μy2+C1)(σx2+σy2+C2)
where x and y are images for comparison, the image mean is μ, variance is σ, covariance between x and y is σxy, and the constant C is added for avoiding instability when the denominator is close to 0. Average image SSIM was calculated using windows of 11×11 pixels. The resultant SSIM index is a value between −1 and 1, where an index of 1 indicates equivalent image inputs. Images of a slanted edge and of a 1951 United States Air force Resolution Target were used to assess DeepLSR’s effect on image resolution. The Slanted Edge MTF plugin available for ImageJ [41] was used to compute modulation transfer functions.

2.5. Performance

Both trained networks were evaluated using the reserved test images of assorted objects and porcine tissues. For performance benchmarks, we report PSNR and SSIM comparisons for laser vs. LED and oLSR vs. LED. We also compared DeepLSR to other common image processing denoising techniques: median filtering [42], non-local means [15], K-SVD [19], and CBM3D [24]. The input parameters for these algorithms were determined by multi-objective optimization, where PSNR and SSIM were the objective functions, using the same training data set as was used for DeepLSR (Table 1).

Tables Icon

Table 1. Parameters for noise reduction methods

3. Results

Validation tests on assorted objects imaged with laser illumination (Fig. 3a) show that DeepLSR reduces speckle noise by 5.3 dB, compared to a 2.7 dB reduction by non-local means filtering, a 3.6 dB reduction from CBM3D, and a 4.4 dB reduction by oLSR (Fig. 4). We also found that a network trained to transform from input images with optical laser speckle reduction to LED illumination (DeepLSR+oLSR) benefits from greater speckle reduction compared to DeepLSR or oLSR alone. SSIM comparisons between coherent- and incoherent-illuminated images indicate a 25% improvement in structural similarity for assorted object images and a 26% improvement for porcine tissue images after applying DeepLSR. While this improvement in recovering object structure is similar to the result of NLM and CBM3D processing, DeepLSR more-effectively reduced speckle noise while maintaining high-frequency image features, demonstrated by the larger PSNR improvement. This result can be seen in NLM- and CBM3D-processed images, which reduce both laser speckle and high-frequency object features, resulting in a blurry appearance of the lines and text in Fig. 3.

 figure: Fig. 3

Fig. 3 DeepLSR compared to conventional speckle reduction methods. DeepLSR was trained on an assortment of images that represent a variety of textures, shapes, and bidirectional reflectance distribution functions. (a) Images of two test objects illuminated with laser illumination, laser illumination with optical speckle reduction (oLSR), median filtering, non-local means, K-SVD, CBM3D, DeepLSR applied to the laser illuminated image, DeepLSR+oLSR applied to the optically speckle reduced image (DeepLSR+oLSR), the target speckle-free image illuminated with a light-emitting diode (LED), and the speckle artifacts removed from the laser illuminated image by DeepLSR. (b) Modulation transfer functions for LED illumination and laser illumination with DeepLSR found using a slanted edge. (c) Images of a 1951 United States Air Force Target with each illumination strategy and laser illumination with DeepLSR.

Download Full Size | PDF

 figure: Fig. 4

Fig. 4 Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) results from reserved test images of assorted objects and porcine tissues.

Download Full Size | PDF

The DeepLSR method has a small effect on resolution measured by a slanted edge test, as demonstrated through the modulation transfer functions of LED illumination compared to laser illumination with DeepLSR. DeepLSR resulted in a reduction in spatial resolution of 17% when comparing the spatial frequency at which each MTF reached half-modulus (Fig. 3c). The pixel-level effects of image processing (NLM, CBM3D, DeepLSR) and optical speckle reduction (oLSR, DeepLSR+oLSR) methods are compared in the line profiles shown in Fig. 5. While non-local means and CBM3D processing suppress high spatial-frequency details of both speckle and original object features, DeepLSR appears to reduce speckle while maintaining edge features of the original object. Lastly, Fig. 5 shows these same trends are observed in object regions that contain primarily red light intensity, indicating that DeepLSR does not rely on information from other color channels for despeckling.

 figure: Fig. 5

Fig. 5 (a) Images of a test object for each speckle reduction technique. (b) The red channels from the color images were studied to assess speckle reduction in the absence of information from other channels. Line profiles from a reserved test image patch are reported, comparing image processing methods (NLM, CBM3D, DeepLSR) and optical methods (oLSR, DeepLSR+oLSR) to the input (Laser) and ground truth (LED) images.

Download Full Size | PDF

In applications involving tissue imaging, the object of interest is often a turbid medium that naturally blurs speckle artifacts. To assess the applicability of DeepLSR in this scenario, we applied our model to images of gastrointestinal tissue illuminated with laser light. In these tissue validation tests, DeepLSR reduced speckle noise by 6.4 dB, compared to 2.9 dB reduction by non-local means filtering, a 3.0 dB reduction by CBM3D, and a 3.7 dB reduction by oLSR. Fig. 6 shows representative images of test samples with laser illumination, conventional noise reduction methods, DeepLSR, DeepLSR combined with optical speckle reduction, and LED illumination. While CBM3D resulted in only a marginal PSNR improvement in tissue imaging, it did outperform other methods in recovering structural features assessed by SSIM. K-SVD saw minimal improvements in image quality in both imaging experiments. Images from comparative methods in Fig. 6 have remnant color artifacts that could be removed with further processing [43, 44]. As in the assorted object tests, DeepLSR removed a greater number of speckle artifacts than conventional imaging processing approaches while retaining structural features (Fig. 4). Compared to the assorted object results, the performance metrics from tissue imaging were elevated due to inherent speckle reduction caused by the blurring effect of turbid media.

 figure: Fig. 6

Fig. 6 DeepLSR applied to images of laser-illuminated ex-vivo porcine gastrointestinal tissues not previously seen by the network.

Download Full Size | PDF

4. Discussion

Compared to many model-based image processing methods, DeepLSR considers large portions of the image’s distribution, allowing it to predict the effective spatial relationships across the image instead of considering only individual pixels. The robustness and generalizability of DeepLSR would improve with increased quantity and variety of training data. Additionally, training data with a larger variety of laser sources, working distances, and imaging lens parameters will improve the generalizability of DeepLSR. To minimize the required quantity of manually captured data, a forward model may be utilized to generate synthetic data with varying speckle size, wavelengths, and small perturbations to prevent the network from learning shape and object texture [45]. Moreover, future work with unsupervised networks may extend the DeepLSR approach to applications where images with and without speckle noise can be obtained but not registered.

DeepLSR may be applied to virtually any modality in which paired coherent-illuminated and speckle reduced samples can be obtained, such as OCT or ultrasound. Another application that may benefit from DeepLSR is automated quality inspection of products in a factory setting, where remote lighting and bright sources with long lifetimes are required. DeepLSR may be particularly useful in endoscopy applications that require bright illumination or small-diameter endoscopes. Incoherent light sources for endoscopy, such as arc lamps and LEDs, require large-diameter light guides to deliver sufficient optical power through an endoscope. Laser illumination enables the delivery of greater illumination power through fiber optics, and can generate incoherent-like images after DeepLSR or DeepLSR+oLSR is applied. Moreover, in widefield applications that require coherent light, such as laser speckle contrast imaging for mapping flow [46], DeepLSR allows both a computational image and a conventional image to be acquired simultaneously.

5. Conclusion

We have presented the application of deep learning to laser speckle reduction by utilizing pairs of images illuminated with coherent and incoherent light for training. DeepLSR removes more speckle artifacts than conventional imaging processing approaches, while retaining structural features. As a data-driven approach, DeepLSR should be trained on images that span the target domain. We make available the nearly 3,000 images collected for training and the models described here so that other researchers may use them to augment their collected training images. We have made the DeepLSR model and source code for widefield laser illumination available here: https://durr.jhu.edu/DeepLSR, and provide step-by-step instructions for installing and applying this framework to new data sets in the Appendix.

6. Appendix

6.1. Instructions for using DeepLSR

Setting up cloud computing resources to run DeepLSR

DeepLSR requires performance-capable graphics processing units (GPUs) as training is computationally intensive. We utilized Google Cloud computing to train the models reported in this publication.

  1. In order to run DeepLSR, set up a Google Cloud instance with the necessary dependencies (Ubuntu, PyTorch, and CUDA) by following the instructions provided here: https://cloud.google.com/deep-learning-vm/docs/pytorch_start_instance
    • * The code released with this publication was tested using Ubuntu 16.04. We recommend utilizing the same operating system to avoid complications.
    • * For best performance, we suggest selecting multiple GPUs.
  2. With the Google Cloud instance setup, a few dependencies must be installed in order to run the models on a cloud instance. Install torchvision, dominate (v2.3.1+), visdom (v0.1.8.3+) and scipy using the following commands:
    • For users utilizing Anaconda:
      • conda install torchvision -c soumith
      • conda install -c conda-forge dominate
      • conda install -c conda-forge visdom
      • conda install -c anaconda scipy
    • For users utilizing pip:
      • pip install torchvision
      • pip install dominate
      • pip install visdom
      • pip install scipy

Training DeepLSR for laser speckle reduction

  1. Begin by cloning the DeepLSR GitHub repository found at (https://durr.jhu.edu/DeepLSR).
  2. The directory structure for the dataset should be organized as follows:
    • boe-10-6-2869-i001.gif
  3. To utilize the training data and models referenced in our publication, visit our GitHub repository for download and use instructions (https://durr.jhu.edu/DeepLSR).
  4. If you intend to train models using your own dataset:
    • All test and train data should be in either .jpeg, .jpg or .png formats. .tiff and other raw formats can result in extremely slow training.
    • All data must be paired side-by-side i.e. the input and output should be concatenated end-to-end in a single image.
    • The size of each individual image should be 2n×2n.
    • Once the dataset is setup use the following command for training:
      • python train.py -dataroot < datapath> -name DeepLSR -gpu_ids 0 -display_id 0 -lambda_L1 70 -niter 200 -niter_decay 200 -pool_size 64 -loadSize < image_size> -fineSize < image_size>

    Training Parameters:

    • -niter is the number of epochs trained with a constant learning rate
    • -niter_decay is the number of epochs trained with a linearly decaying learning rate.
    • -lr adjusts the learning rate (default = 0.0002).
    • -gpu_ids is the number of GPUs used (0 is one GPU, 1 is two GPUs and −1 is no GPU).
    • -lambda_L1 is the 1 regularization parameter used for training. The default is 70 because on our data, the tuned range was [67,74].
    • -load_Size is the size of the input/output image.
    • -fine_size is the size of the random crop from within the image to introduce jitter.
  5. To view training losses and results, run python -m visdom.server and click the URL http://localhost:8097. For cloud servers replace localhost with your IP.
  6. To view epoch-wise intermediate training results, visit /checkpoints/DeepLSR/web/index.html

Testing the trained network

  1. To test with our pre-trained models, visit our GitHub repository for download instructions.
  2. Once the model has been uploaded to Google Cloud and configured, run the following code: python test.py -dataroot < datapath> -name DeepLSR -gpu_ids 0 -display_id 0 -loadSize < image_size> -fineSize < image_size>
    • * To test with our data <image_size> should be set to 1024.
    • * The test results will be saved to a html file here: /results/DeepLSR/test_latest/index.html

Funding

This work was supported in part with funding from the NIH Trailblazer Award (R21 EB024700).

Disclosures

TLB (P), FM (P), NJD (P)

References

1. J. W. Goodman, “Some fundamental properties of speckle,” J. Opt. Soc. Am. 66, 1145–1150 (1976). [CrossRef]  

2. M. Bashkansky and J. Reintjes, “Statistics and reduction of speckle in optical coherence tomography,” Opt. Lett. 25, 545–547 (2000). [CrossRef]  

3. M. R. Christopher Glazowski, “Optimal detection pinhole for lowering speckle noise while maintaining adequate optical sectioning in confocal reflectance microscopes,” J. Biomed. Opt. 17, 085001 (2012). [CrossRef]   [PubMed]  

4. J. G. Abbott and F. Thurstone, “Acoustic speckle: Theory and experimental analysis,” Ultrason. Imaging 1, 303–324 (1979). [CrossRef]   [PubMed]  

5. J. Goodman, Speckle phenomena in optics: Theory and applications (W. H. Freeman, 2007).

6. O. Liba, M. D. Lew, E. D. SoRelle, R. Dutta, D. Sen, D. M. Moshfeghi, S. Chu, and A. de la Zerda, “Speckle-modulating optical coherence tomography in living mice and humans,” Nat. Commun. 8, 15845 (2017). [CrossRef]   [PubMed]  

7. D. C. Adler, T. H. Ko, and J. G. Fujimoto, “Speckle reduction in optical coherence tomography images by use of a spatially adaptive wavelet filter,” Opt. Lett. 29, 2878 (2004). [CrossRef]  

8. H. M. Salinas and D. C. Fernandez, “Comparison of pde-based nonlinear diffusion approaches for image enhancement and denoising in optical coherence tomography,” IEEE Transactions on Med. Imaging 26, 761–771 (2007). [CrossRef]  

9. Z. Jian, L. Yu, B. Rao, B. J. Tromberg, and Z. Chen, “Three-dimensional speckle suppression in optical coherence tomography based on the curvelet transform,” Opt. Express 18, 1024–1032 (2010). [CrossRef]   [PubMed]  

10. A. Wong, A. Mishra, K. Bizheva, and D. A. Clausi, “General bayesian estimation for speckle noise reduction in optical coherence tomography retinal imagery,” Opt. Express 18, 8338–8352 (2010). [CrossRef]   [PubMed]  

11. L. I. Rudin, S. Osher, and E. Fatemi, “Nonlinear total variation based noise removal algorithms,” Phys. D: Nonlinear Phenom. 60, 259–268 (1992). [CrossRef]  

12. C. Louchet and L. Moisan, “Total variation denoising using iterated conditional expectation,” in 2014 22nd European Signal Processing Conference (EUSIPCO), (2014), pp. 1592–1596.

13. R. Abergel, C. Louchet, L. Moisan, and T. Zeng, “Total variation restoration of images corrupted by poisson noise with iterated conditional expectations,” in Scale Space and Variational Methods in Computer Vision, J.-F. Aujol, M. Nikolova, and N. Papadakis, eds. (Springer International Publishing, Cham, 2015), pp. 178–190.

14. J. M. Bioucas-Dias and M. A. T. Figueiredo, “Multiplicative noise removal using variable splitting and constrained optimization,” IEEE Transactions on Image Process. 19, 1720–1730 (2010). [CrossRef]  

15. A. Buades, B. Coll, and J.-M. Morel, “A Non-Local Algorithm for Image Denoising,” 2005 IEEE Comput. Soc. Conf. on Comput. Vis. Pattern Recognit. (CVPR’05)2, 60–65.

16. C. Deledalle, F. Tupin, and L. Denis, “Poisson nl means: Unsupervised non local means for poisson noise,” in 2010 IEEE International Conference on Image Processing, (2010), pp. 801–804. [CrossRef]  

17. J. Salmon, Z. Harmany, C. A. Deledalle, and R. Willett, “Poisson noise reduction with non-local pca,” J. Math. Imaging Vis. 48, 279–294 (2014). [CrossRef]  

18. A. A. Bindilatti, M. A. Vieira, and N. D. Mascarenhas, “Poisson wiener filtering with non-local weighted parameter estimation using stochastic distances,” Signal Process. 144, 68–76 (2018). [CrossRef]  

19. M. Aharon, M. Elad, and A. Bruckstein, “K-svd: An algorithm for designing overcomplete dictionaries for sparse representation,” IEEE Transactions on Signal Process. 54, 4311–4322 (2006). [CrossRef]  

20. R. Rubinstein, T. Peleg, and M. Elad, “Analysis k-svd: A dictionary-learning algorithm for the analysis sparse model,” IEEE Transactions on Signal Process. 61, 661–677 (2013). [CrossRef]  

21. K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Image denoising by sparse 3-d transform-domain collaborative filtering,” IEEE Transactions on Image Process. 16, 2080–2095 (2007). [CrossRef]  

22. B. Ahn and N. I. Cho, “Block-Matching Convolutional Neural Network for Image Denoising,” https://arxiv.org/abs/1704.00524 (2017).

23. W. Meiniel, J. Olivo-Marin, and E. D. Angelini, “Denoising of microscopy images: A review of the state-of-the-art, and a new sparsity-based method,” IEEE Transactions on Image Process. 27, 3842–3856 (2018). [CrossRef]  

24. K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Color image denoising via sparse 3d collaborative filtering with grouping constraint in luminance-chrominance space,” in 2007 IEEE International Conference on Image Processing, vol. 1 (2007), pp. I – 313–I – 316.

25. M. R. N. Avanaki, P. P. Laissue, T. J. Eom, A. G. Podoleanu, and A. Hojjatoleslami, “Speckle reduction using an artificial neural network algorithm,” Appl. Opt. 52, 5050 (2013). [CrossRef]   [PubMed]  

26. Y. Ma, X. Chen, W. Zhu, X. Cheng, D. Xiang, and F. Shi, “Speckle noise reduction in optical coherence tomography images based on edge-sensitive cGAN,” Biomed. Opt. Express 9, 5129 (2018). [CrossRef]   [PubMed]  

27. S. Li, M. Deng, J. Lee, A. Sinha, and G. Barbastathis, “Imaging through glass diffusers using densely connected convolutional networks,” Optica 5, 803–813 (2018). [CrossRef]  

28. Y. Li, Y. Xue, and L. Tian, “Deep speckle correlation: a deep learning approach toward scalable imaging through scattering media,” Optica 5, 1181–1190 (2018). [CrossRef]  

29. F. Agostinelli, M. R. Anderson, and H. Lee, “Adaptive multi-column deep neural networks with application to robust image denoising,” Proc. 26th Int. Conf. on Neural Inf. Process. Syst. - Vol. 1 pp. 1493–1501 (2013).

30. R. Chen, F. Mahmood, A. L. Yuille, and N. J. Durr, “Rethinking monocular depth estimation with adversarial training,” http://arxiv.org/abs/1808.07528 (2018).

31. L. A. Gatys, M. B. Alexander, and S. Ecker, “Image style transfer using convolutional neural networks,” The IEEE Conf. on Comput. Vis. Pattern Recognit. (CVPR) (2016).

32. C. Dong, C. C. Loy, K. He, and X. Tang, “Image super-resolution using deep convolutional networks,” IEEE Transactions on Pattern Analysis Mach. Intell. 38, 295–307 (2016). [CrossRef]  

33. F. Mahmood, R. Chen, and N. J. Durr, “Unsupervised reverse domain adaptation for synthetic medical images via adversarial training,” IEEE Transactions on Med. Imaging 37, 2572–2581 (2018). [CrossRef]  

34. P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-Image Translation with Conditional Adversarial Networks,” https://arxiv.org/abs/1611.07004 (2016).

35. T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida, “Spectral Normalization for Generative Adversarial Networks,” https://arxiv.org/abs/1802.05957 (2018).

36. C. Graetzel, M. Suter, and M. Aschwanden, “Reducing laser speckle with electroactive polymer actuators,” Int. Soc. for Opt. Photonics p. 943004 (2015).

37. K. Kurach, M. Lucic, X. Zhai, M. Michalski, and S. Gelly, “The GAN Landscape: Losses, Architectures, Regularization, and Normalization,” https://arxiv.org/abs/1807.04720 (2018).

38. D. P. Kingma and J. Ba, “Adam: A Method for Stochastic Optimization,” https://arxiv.org/abs/1412.6980 (2014).

39. A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” https://arxiv.org/abs/1511.06434 (2015).

40. Z. Wang, A. Bovik, H. Sheikh, and E. Simoncelli, “Image Quality Assessment: From Error Visibility to Structural Similarity,” IEEE Transactions on Image Process. 13, 600–612 (2004). [CrossRef]  

41. J. Schindelin, I. Arganda-Carreras, E. Frise, V. Kaynig, M. Longair, T. Pietzsch, S. Preibisch, C. Rueden, S. Saalfeld, B. Schmid, J. Y. Tinevez, D. J. White, V. Hartenstein, K. Eliceiri, P. Tomancak, and A. Cardona,“Fiji: an open-source platform for biological-image analysis,” Nat. Methods 9, 676–682 (2012). [CrossRef]   [PubMed]  

42. T. Huang, G. Yang, and G. Tang, “A fast two-dimensional median filtering algorithm,” IEEE Transactions on Acoust. Speech, Signal Process. 27, 13–18 (1979). [CrossRef]  

43. S. Ono and I. Yamada, “Color-line regularization for color artifact removal,” IEEE Transactions on Comput. Imaging 2, 204–217 (2016). [CrossRef]  

44. J. Mairal, M. Elad, and G. Sapiro, “Sparse representation for color image restoration,” IEEE Transactions on Image Process. 17, 53–69 (2008). [CrossRef]  

45. F. Mahmood, R. Chen, S. Sudarsky, D. Yu, and N. J. Durr, “Deep learning with cinematic rendering: Fine-tuning deep neural networks using photorealistic medical images,” Phys. Medicine Biol. 63, 18 (2018). [CrossRef]  

46. A. K. Dunn, H. Bolay, M. A. Moskowitz, and D. A. Boas, “Dynamic Imaging of Cerebral Blood Flow Using Laser Speckle,” J. Cereb. Blood Flow & Metab. 21, 195–201 (2001). [CrossRef]  

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (6)

Fig. 1
Fig. 1 DeepLSR Architecture. (a) Training architecture for image-to-image translation-based laser speckle reduction using a conditional Generative Adversarial Network. A generator learns to transform between pairs of images acquired with coherent and incoherent illumination while a discriminator learns to classify input images as real or fake. (b) Once training is complete, the discriminator is discarded and the trained generator (DeepLSR) reduces laser speckle noise in images not previously seen by the generator.
Fig. 2
Fig. 2 Imaging setup for acquiring images with: (1) laser illumination (oLSR turned off), (2) laser illumination with optical laser speckle reduction (oLSR turned on), and (3) LED illumination.
Fig. 3
Fig. 3 DeepLSR compared to conventional speckle reduction methods. DeepLSR was trained on an assortment of images that represent a variety of textures, shapes, and bidirectional reflectance distribution functions. (a) Images of two test objects illuminated with laser illumination, laser illumination with optical speckle reduction (oLSR), median filtering, non-local means, K-SVD, CBM3D, DeepLSR applied to the laser illuminated image, DeepLSR+oLSR applied to the optically speckle reduced image (DeepLSR+oLSR), the target speckle-free image illuminated with a light-emitting diode (LED), and the speckle artifacts removed from the laser illuminated image by DeepLSR. (b) Modulation transfer functions for LED illumination and laser illumination with DeepLSR found using a slanted edge. (c) Images of a 1951 United States Air Force Target with each illumination strategy and laser illumination with DeepLSR.
Fig. 4
Fig. 4 Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) results from reserved test images of assorted objects and porcine tissues.
Fig. 5
Fig. 5 (a) Images of a test object for each speckle reduction technique. (b) The red channels from the color images were studied to assess speckle reduction in the absence of information from other channels. Line profiles from a reserved test image patch are reported, comparing image processing methods (NLM, CBM3D, DeepLSR) and optical methods (oLSR, DeepLSR+oLSR) to the input (Laser) and ground truth (LED) images.
Fig. 6
Fig. 6 DeepLSR applied to images of laser-illuminated ex-vivo porcine gastrointestinal tissues not previously seen by the network.

Tables (1)

Tables Icon

Table 1 Parameters for noise reduction methods

Equations (2)

Equations on this page are rendered with MathJax. Learn more.

PSNR = 10 log 10 ( R 2 MSE )
SSIM ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.