Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Spatial resolution improved fluorescence lifetime imaging via deep learning

Open Access Open Access

Abstract

We present a deep learning approach to obtain high-resolution (HR) fluorescence lifetime images from low-resolution (LR) images acquired from fluorescence lifetime imaging (FLIM) systems. We first proposed a theoretical method for training neural networks to generate massive semi-synthetic FLIM data with various cellular morphologies, a sizeable dynamic lifetime range, and complex decay components. We then developed a degrading model to obtain LR-HR pairs and created a hybrid neural network, the spatial resolution improved FLIM net (SRI-FLIMnet) to simultaneously estimate fluorescence lifetimes and realize the nonlinear transformation from LR to HR images. The evaluative results demonstrate SRI-FLIMnet’s superior performance in reconstructing spatial information from limited pixel resolution. We also verified SRI-FLIMnet using experimental images of bacterial infected mouse raw macrophage cells. Results show that the proposed data generation method and SRI-FLIMnet efficiently achieve superior spatial resolution for FLIM applications. Our study provides a solution for fast obtaining HR FLIM images.

Published by Optica Publishing Group under the terms of the Creative Commons Attribution 4.0 License. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

1. Introduction

Fluorescence lifetime imaging (FLIM) has become a powerful tool for biomedical discoveries. The fluorescence lifetime is a robust indicator for better observing living cells [1]. FLIM provides abundant cellular microenvironmental information that is unattainable in fluorescence intensity imaging. It can probe biophysical parameters such as temperature, O2, viscosity, metabolism, pH, and ion concentrations [2,3]. In combination with Förster resonance energy transfer (FRET) principles, FLIM-FRET has been widely used to assess protein-protein interactions and conformational changes [24]. Meanwhile, FLIM is more robust than intensity imaging because it is less susceptible to experimental parameters such as the fluorophore concentration, excitation intensity, and optical path. FLIM can distinguish fluorophores with identical emission spectra. It can also identify tagged fluorochromes on specified cells from background signals or unbounded fluorochromes [2]. It has been widely used for cellular study, assessing drug delivery, cancer diagnosis, and ophthalmological diagnosis [5,6].

The time-correlated single photon counting (TCSPC) technique is the gold standard for FLIM imaging due to its superior timing resolution, photon efficiency, and signal-to-noise ratio performances [7]. However, traditional TCSPC-based FLIM systems are slow, especially for acquiring high-resolution (HR) images. TCSPC requires a low photon-counting rate (typically < 5% of the laser repetition rate) to avoid pile-up effects [8]. There are two different instrumental configurations for acquiring FLIM images: laser scanning microscopy (LSM) and wide-field illumination microscopy (WFIM) [9]. LSM FLIM systems acquire images through scanning the focal point across samples. Obtaining HR images requires a minimized focal point and a smaller scanning step, leading to a long acquisition, photobleaching fluorophores, and causing phototoxicity. On the other side, WFIM systems adopt a parallel light field to excite samples, offering a higher frame rate and causing less damage to samples. However, WFLIM’s spatial resolution is poor, and images are easily contaminated by out-of-focus surrounding light. Besides, there is a trade-off between the spatial and temporal resolutions due to the crosstalk [10]. Hence, it is a great challenge to acquire HR images rapidly.

Meanwhile, since FLIM is an indirect imaging technique, traditional algorithms such as the least-squares fitting method (LSF), maximum likelihood estimation methods (MLE), Bayesian analysis, or Laguerre deconvolution [3,11,12] are applied to analyze the fluorescence lifetime. They are model-based iterative optimization methods (generally slow) and rely on prior knowledge to set proper models and initial conditions. Other fast and model-free analysis algorithms were also developed, including the center-of-mass method (CMM) and phasor method with graphic representation [13,14]. However, the image acquisition and data processing are too slow for HR or 3D images, hampering widespread applications such as cellular dynamics imaging or high-throughput screening.

New sensor and microscopy technologies have been introduced to improve FLIM systems’ performances, including short-deadtime high-throughput single-photon sensors [15] and multifocal scanning systems [16]. This paper proposed an alternative computational strategy using deep learning techniques to rapidly obtain HR FLIM images from raw image data without modifying the system configuration. Our method combines DL-based fluorescence lifetime determination and single image resolution enhancement (DL-SIRE) techniques. DL techniques provide a new approach to delivering fast analysis with different models. They have achieved unprecedented accuracy and fast speed [1719]. ‘SIRE’ refers to reconstructing a single HR image directly from its LR counterpart. It is a notoriously ill-posed problem because an infinite number of HR images correspond to a given LR image. DL’s powerful hierarchical representation learning ability allows DL-SIRE to reconstruct state-of-the-art HR images through learning a single image’s intrinsic and high-level statistical features [20,21]. DL-SIRE has been widely used in face recognition, remote sensing, and video surveillance [22,23]. It has become a research trend in improving the spatial resolution of optical and fluorescence microscopes [2428]. However, DL-SIRE has not yet been used in FLIM systems, and there are two main challenges. First, there is a lack of massive FLIM data for training neural networks. Unlike natural images, FLIM images are three-dimensional (spatial and temporal) data. Acquiring massive HR FLIM images is labor-intensive and time-consuming. Second, FLIM images simultaneously possess complex spatial and temporal features.

To address the above challenges, 1) we proposed a method to generate a large-scale semi-synthetic FLIM dataset for training neural networks. The samples in the FLIM dataset include distinct morphological characteristics of subcellular structures and complex decays with a wide range of fluorescence lifetimes. 2) a hybrid neural network architecture, SRI-FLIMnet, was proposed to map LR FLIM data to corresponding HR images. SRI-FLIMnet serves two purposes: A) to directly map a decay histogram to its corresponding lifetime without iterative optimization processes and B) to nonlinearly transform LR FLIM to HR FLIM images. LR images can be fast obtained through both LSM and WFIM FLIM systems. SRI-FLIMnet was evaluated using semi-synthetic FLIM images and validated with experimental FLIM data. Results confirm SRI-FLIMnet’s superiority in reconstructing HR images. SRI-FLIMnet is expected to fast capture HR FLIM images and find potential applications in high-throughput and real-time FLIM systems.

2. Dataset preparation

Deep learning techniques are data-driven approaches. A large amount of suitable training data is necessary for developing a robust DL algorithm. However, for DL tasks in microscopy, there are limited datasets. Unlike open-source natural image datasets used for DL computer vision applications [29], most previously reported studies only considered specific tissues or cells, consisting of several dozens to hundreds of samples for network training [2428]. Data argumentation techniques should be applied to increase the training data. In addition, it is challenging to generalize previously reported DL algorithms for general applications. The lack of FLIM datasets severely handicaps the development of DL FLIM techniques. Considering that FLIM data possess many spatial and temporal features, traditional data argumentation techniques cannot achieve desirable performance. One effective method to circumvent this difficulty is to generate model-based synthetic data to train networks. Data synthesis is a mature and widely used method to approximate realistic data and serve network training purposes. It has been successfully applied to many DL tasks, including light detection and ranging (LiDAR), fluorescence lifetime estimation, and super-resolution microscopy [17,18,27]. In this section, a theoretical framework was proposed to obtain massive semi-synthetic FLIM data. The forward TCSPC model was first developed to describe the time-stamped photon detection process. Then semi-synthetic FLIM images were generated by converting immunofluorescence intensity images from the Human Protein Atlas (HPA) dataset to 3D FLIM data using the TCSPC model. Meanwhile, the degrading model between HR and LR images was also developed to obtain LR-HR training pairs. We demonstrate that semi-synthetic data can be used for neural network training and successfully applied to different experimental data.

2.1 TCSPC model

In TCSPC, a train of laser pulses excites the sample periodically. In each cycle, at most one photon is detected by a single-photon detector to avoid the pile-up effect. According to the photon counting theory, such a process is governed by a non-homogeneous Poisson process [30]. For a pixel (i, j) in an M×M×L FLIM image, the rate function ${\lambda _{i,j}}$ can be expressed as:

$${\mathrm{\lambda }_{i,j}} = {A_{i,j}} \cdot I\textrm{RF}(t )\mathrm{\ast }\mathop \sum \limits_{p = 1}^D {\mathrm{\alpha }_p}{e^{ - t/{\mathrm{\tau }_p}}} + \mathrm{\epsilon }, \textrm{i},\textrm{j} \in [{1,M} ],\textrm{t} \in [{1,L} ], $$
where M is the length of spatial dimension and L is the number of recorded time bins. ${A_{i,j}}$ is the detected fluorescence amplitude at pixel (i, j), which relates to the emitted fluorescence intensity and the detector’s quantum efficiency; IRF(t) is the instrument response function; ${\alpha _p}$ is the fraction ratio of the lifetime component ${\tau _p}$, and D is the number of lifetime components. The asterisk * denotes the convolution operator. An additional term $\mathrm{\epsilon }$ represents added noise from surrounding light and dark count noise in the sensor. We assume fluorescence signals are collected at the focal plane without considering the spatial blur function. After S cycles, the detected photon number at pixel (i, j) follows a Poisson distribution:
$${N_{i,j}} \sim Poisson\left( {S\mathop \smallint \limits_0^T {\mathrm{\lambda }_{i,j}}(t )dt} \right). $$
The N photon arriving times denoted by $t_{i,j}^n({n \in [{1,{N_{i,j}}} ]} )$ at pixel (i, j) are independent and identically distributed random variables with the probability-density function described as:
$${f_{i,j}}(t )= \frac{{{\mathrm{\lambda }_{i,j}}(t )}}{{\mathop \smallint \nolimits_0^T {\mathrm{\lambda }_{i,j}}(t )dt}}. $$
Equations (1)–(3) define the TCSPC forward model. Given the decay model and lifetime parameters, we can simulate the recorded timestamps of first N arrival photons [30]. Poisson noise is dominant in modern TCSPC measurements as low noise sensors have been used. We have compared synthetic and measured decays (analysis not included), showing that the TCSPC model can precisely describe fluorescence decays with arbitrary signal-to-noise ratio (SNR) levels.

2.2 Semi-synthetic FLIM images

The HPA dataset provides a comprehensive subcellular map of human proteomes in cells, tissues, and organs [31]. It contains more than 80,000 cellular images with a large morphological diversity obtained by confocal immunofluorescence microscopy systems. The images in HPA dataset cover a wide range of cellular morphologies. We used the HPA image classification dataset [32], including more than 30,000 samples with 27 different cell types. Each sample contains four color channels, and each color channel contains one cellular organelle. The blue, red, yellow, and green channels contain nuclei, microtubules, endoplasmic reticulums, and proteins (such as endosomes, lysosomes, or mitochondria). Figure 1 shows some original image samples in the HPA dataset and generated 3D FLIM data. The 1st column shows RGB images of three different samples, and their corresponding color-channel images are shown in the 2nd column. The cellular images contain more statistical features, including more refined details, complex spatial variations, and more sophisticated correlations than open-access natural images (mostly used by computer vision communities for DL training). To generate FLIM data with complex decay features and various cellular morphologies, The samples in HPA datasets are converted to semi-synthetic FLIM data. the pixel values of intensity images are converted to time-stamped photon counts using the forward TCSPC model in Eqs. (1)–(3). The pixel values from 0 to 255 were linearly scaled to 0 to 5100. The photon count range corresponds to the typical intensity range in practical experiments and offers a sufficient SNR for accurate lifetime estimations. Instead of restricting to specified fluorophore lifetimes, each color channel was randomly assigned a single lifetime in the range from 0.5 to 5.5 ns to include more general cases. The semi-synthetic FLIM data were obtained by summing up all the considered color channels. Since different color channels have diverse intensities and spatial overlaps, each pixel naturally shows complex decays with different lifetime components. Meanwhile, they also have a variety of shapes. The total photon count in a pixel of a FLIM image remains the same as the intensity image, while the temporal photon distribution follows an exponential decay with given lifetime parameters. In simulations, M = L = 256, D = 1, and the total recorded time is 10 ns. The IRF is modeled by a Gaussian distribution with FWHM = 167.3 ps. This matches the two-photon FLIM system we used for experiments (see details in Section 5). The added noise can be reasonably neglected because Poisson noise dominates in TCSPC measurements.

 figure: Fig. 1.

Fig. 1. Original intensity images in the HPA dataset and the semi-synthetic FLIM data. The 1st column is the RGB intensity images of three samples. Their corresponding four color-channel images are shown in the 2nd column; each organelle is labeled by one color. The 3rd and 4th columns show the synthetic 3D FLIM data cloud and lifetime images. In 3D point plots (3rd column), the dark blue to bright yellow point color indicates a high to low photon density.

Download Full Size | PDF

The generated data cloud and the corresponding fluorescence lifetime images calculated by the least-square fitting method are shown in the 3rd and 4th columns of Fig. 1, respectively. Each point corresponds to a detected photon’s spatial and temporal positions in the data cloud. FLIM images have lower contrast than RGB intensity images in revealing cell structures. Many delicate structures are blurred or even disappear, and this is because fluorescence lifetime is independent of the intensity variation. To ensure diversity, FLIM data with only one, two, or three decay components are also generated. The whole dataset contains 30,000 FLIM samples for training purposes, providing sufficient data for network training.

2.3 Degrading model

An accurate degrading model for developing LR-HR image pairs correctly is crucial for DL-SIRE’s performance [21]. For general DL-SIRE frameworks, the degrading model is:

$${I^{LR}} = {D_k}({{I^{HR}}\ast K} )+ \mathrm{\varepsilon }, $$
where ILR and IHR denote the LR and HR images, respectively, K denotes the blurring kernel, and Dk is the down-sampling operator with a scaling ratio k. $\varepsilon $ is added noise. In typical computer vision tasks, the bicubic interpolation is usually applied as the degrading model [19]. The Gaussian blurring kernel is widely used in super-resolution microscopy tasks [24, 26]. However, the situation is different in FLIM systems. In LSM FLIM systems, LR images are measured using coarser scanning steps. Therefore, the LR image, ILR with the scaling ratio k, is obtained by summing a patch of k×k pixels in the corresponding HR image, IHR. This is also applicable to WFIM-FLIM systems, in which a low-resolution wide-field camera like single-photon avalanche diode (SPAD) arrays is used [11]. LR images are equivalent to spatially pixel-binned HR images. Thus, the degrading model is expressed as:
$${I^{LR}}({m,m,L} )= {1^{k \times k \times 1}}{ \otimes _{s = ({k,k,1} )}}{I^{HR}}({M,M,L} ), $$
where ${1^{k \times k \times 1}}$ denotes k×k×1 matrix with all 1 elements, ${ \otimes _{s = ({k,k,1} )}}$ is a three-dimensional cross-correlation operator with a stride s = $({k,k,1} )$ along the spatial and temporal dimensions. The size of ILR is m×m×L with m = mod (M, k). For simplification, the intensity changes in LR pixels are not considered because the fluorescence lifetime is independent of the intensity, given the SNR is large enough. Equation (5) reveals that in LR images, low- and high-frequency information is lost through spatial pixel binning. Meanwhile, the temporal information among neighboring pixels fuses, resulting in information loss and a shift in the lifetime distribution. Achieving HR FLIM is more challenging compared to other situations. In semi-synthetic FLIM data processing, when applying Eq. (5) to HR images, a threshold of 100 photon counts was set for each color channel image to exclude low-intensity pixels.

3. Network design and training

3.1 Network architecture

It has been a great challenge to process the spatial and temporal features of FLIM data simultaneously. One intuitionistic approach is to design a 3D neural network architecture to map 3D FLIM data to HR images. However, the spatial and temporal information could interfere, leading to difficulty in network training. It would also cause an enormous computation burden and a prolonged inference speed. Our proposed SRI-FLIMnet adopts a temporal and spatial decoupling approach to address this problem. The SRI-FLIMnet, as shown in Fig. 2, uses a hybrid network architecture composed of two sub-nets: τ-Net and SRI-Net. The former aims to estimate fluorescence lifetimes, whereas the latter is to produce spatial-resolution improved lifetime images so that the crosstalk between spatial and temporal dimensions can be eliminated. τ-Net uses a one-dimensional convolutional (1DCNN) neural network to calculate pixel-wise fluorescence lifetimes [19]. 1DCNN is efficient and hardware-friendly, suitable for fast analysis, and can accurately analyze single- and multi-exponential decay models. The design and characterization of 1DCNN have been reported elsewhere [18, 19]. τ-Net contains five 1D convolutional blocks. Each block has a convolution layer, a batch normalization layer, and a nonlinear activation layer (Rectified Linear Unit, ReLU). The hyperparameters of all layers in τ-Net are annotated in Fig. 2. We adopt the residual dense network (RDN) with hybrid dilated convolution as the backbone for SRI-Net. RDN can fully exploit local and global hierarchical features from LR images [33]. It consists of four parts, the shallow feature extraction (SFE) layers, residual dense blocks (RDB), global feature fusion (GFF) layers, and the sub-pixel convolution up-sampling (SCUS) block [23]. SFE layers extract high-level features of input LR images for later network blocks. RDBs are for learning and fusing local features with different levels. In an RDB, 8 2D convolutional layers are densely connected. The input feature of the first layer is 64, and the growth rate of the feature number in the following layers is 64, too. A concatenation layer is put at the end of the RDB, followed by a residual connection across the whole block. To achieve a better performance, RDN requires a deeper architecture with many RDBs to obtain a large receptive field and high nonlinearity. The receptive field is a critical factor in the network’s performance because it decides how much contextual information is used to reconstruct high-frequency components. However, a deeper architecture consumes more computation resources and is, therefore, slower in training and inference. To overcome this problem, we further improved the network by introducing the hybrid dilated convolution in each RDBs [34]. By inserting zeros in the convolution kernel, dilated convolutions can exponentially expand the receptive field without extra parameters. In each RDB, the dilation rate for 8 convolutional layers is [1, 2, 5, 9, 1, 2, 5, 9]. We particularly used mixed saw-toothed dilation rates for different layers to expand the convolution kernels to different scales while avoiding the gridding effect and information loss brought by dilated convolutions [35]. The GFF block is for global feature fusion and global residual learning, and the final SCUS block up-samples images to the size of corresponding HR images. Figure 2 only shows the network architecture for a k = 4. Note that for k = 2 and 3, there are only one 2D convolutional layer and one Pixel-Shuffle layer in the Up-Scaling part for obtaining the desirable data size of output images. The corresponding feature number of the convolutional layer are 64×k2, and the kernel size and stride remain 3 and 1, respectively.

 figure: Fig. 2.

Fig. 2. Network architecture of the proposed SRI-FLIMnet with a scaling ratio k = 4. SRI-FLIMnet contains two subnets, τ-Net, and SRI-Net, responsible for lifetime evaluation and nonlinear transformation for improving spatial resolution, respectively. The parameters of convolutional layers are labeled in the parentheses as (feature number, kernel size, stride). SRI-Net contains 8 residual dense blocks (RDB) with hybrid dilated convolution. In each RDB, 8 2D convolutional layers with a kernel size of 3 and a stride of 1 are densely connected. Both the input feature number and growth rate are 64. In the Up-Scaling part of SRI-Net, the parameters are (1024, 3, 1) for both convolutional layers. The input data of SRI-FLIMnet is 64×64×256, and the final output data size is 256×256.

Download Full Size | PDF

3.2 Loss function

For τ-Net, the Mean-Square Error (MSE) loss or L2 norm loss is used for network training:

$${L_2}({\hat{\tau },\tau } )= \frac{1}{N}\mathop \sum \limits_{i = 1}^N {({{{\hat{\tau }}_i} - {\tau_i}} )^2}, $$
where $\hat{\tau }$ is the predicted lifetime by τ-Net and N is the batch size. As for SRI-Net, although the L2 loss tends to obtain a high peak signal-to-noise ratio (PSNR), the reconstructed images are over-smoothed and unappealing to human visual perception. L2 also inherently assumes that noise follows a Gaussian distribution. It is hard to conform to this condition in our study. Therefore, we adopted the perceptual loss to restore high-frequency contents as many as possible. The perceptual loss, hereafter named Lpercep, is composed of two terms, the multiscale structural similarity index measure (MS-SSIM) loss [36] and L1 loss, which is expressed as:
$${L_{percep}}({{{\hat{I}}^{HR}},{I^{HR}}} )= {L_{MS - SSIM}}({{{\hat{I}}^{HR}},{I^{HR}}} )+ \gamma {L_1}({{{\hat{I}}^{HR}},{I^{HR}}} ), $$
where ${\hat{I}^{HR}}$ denotes the reconstructed HR image by the network. LMS-SSIM is defined as:
$${L_{\textrm{MS} - \textrm{SSIM}}} = 1 - {[{{l_{\textrm{Ms}}}({{{\hat{I}}^{HR}},{I^{HR}}} )} ]^{{\mathrm{\alpha }_{Ms}}}}\mathop \prod \limits_{j = 1}^{Ms} {[{c{s_j}({{{\hat{I}}^{HR}},{I^{HR}}} )} ]^{{\mathrm{\beta }_j}}}, $$
where $l({{{\hat{I}}^{HR}},{I^{HR}}} )= ({2{\mu_{{{\hat{I}}^{HR}}}}{\mu_{{I^{HR}}}} + {C_1}} )/({\mu_{{{\hat{I}}^{HR}}}^2 + \mu_{{I^{HR}}}^2 + {C_1}} )$ is the luminance comparison function with μ the mean of ${\hat{I}^{HR}}$ or ${I^{HR}}$. $cs({{{\hat{I}}^{HR}},{I^{HR}}} )= ({2{\sigma_{{{\hat{I}}^{HR}}}}{\sigma_{{I^{HR}}}} + {C_2}} )/({\sigma_{{{\hat{I}}^{HR}}}^2 + \sigma_{{I^{HR}}}^2 + {C_2}} )$ is the contrast-structure comparison function and $\sigma $ is the standard deviation of ${\hat{I}^{HR}}$ and ${I^{HR}}$. C1 and C2 are constants to ensure stability when the denominator is close to 0. Ms denotes the resolution level of SSIM. To estimate MS-SSIM, ${C_1} = {0.01^2}$ and ${C_2} = {0.03^2}$ Ms = 5 and $\alpha = \beta = [{0.0448,0.2856,0.3001,0.2363,0.1333} ].\; $An 11×11 Gaussian sliding window with σ = 1.5 is used to calculate l and cs for different scales in a pixel-by-pixel manner. The overall LMS-SSIM is the average value of the whole image [36]. The L1 loss serves as a content loss to enforce the reconstructed HR images close to the ground truth HR images. $\gamma \; $= 0.2 is a parameter to adjust the weight of L1 losses.

3.3 Network training

We implemented SRI-FLIMnet using the Pytorch library in Python 3.8 and performed network training and inference on a workstation containing an Intel Core i9-10900X CPU and an Nvidia Quadro RTX5000 GPU card. The two subnets were trained independently. For τ-Net, the training dataset contains 50,000 decay histograms generated from Eqs. (1)–(3). For each histogram, the total photon counts vary from 100 to 10,000, and the number of lifetime components ranges from 1 to 4, with a lifetime ranging from 0.5 to 5.5 ns. All random variables follow uniform distributions. The optimizer is Adam with a learning rate of 1e-4, and the batch size is 256. An early stop with 20 patient epochs was added during the training process to prevent over-fitting. The total training time is about 5 mins for τ-Net. For SRI-net, the semi-synthetic dataset mentioned above was used for training. The sizes of LR FLIM data are 128×128×256, 85×85×256, and 64×64×256 for k = 2, 3, and 4, respectively, which are obtained using Eq. (5). The optimizer is Adam. The initial learning rate is 1e-4, and it dropped 0.9 times every 20 epochs. The batch sizes for the k = 2, 3, and 4 are 6, 12, and 24, respectively, to fully use GPU memory. The total training epoch for different scaling ratios is 200, and the training time is about 400 h, 160 h, and 50 h for k = 2, 3, and 4. Once the networks were well trained, the inference time for different scaling ratios typically took several seconds to map the 3D LR FLIM data to 256×256 HR images.

4. Characterization of the neural network

After the training procedure, SRI-FLIMnet was evaluated with a new testing dataset. The performance of SRI-FLIMnet on lifetime determination was first quantitatively assessed. Figure 3 (a) shows three normalized decay histograms used for testing τ-Net. The samples denoted as Decays 1–3 have different SNR, lifetimes, and decay components. As mentioned above, the decay components may vary in different pixels in a FLIM image, and we do not know the exact number of decay components. Consequently, applying multi-exponential decay model analysis can cause model-mismatch problems and wrong interpretation. As suggested in Ref. [37], the amplitude-weighted average lifetime τA was used as a lifetime indicator, suitable for FLIM-FRET measurements. τA is defined as ${\tau _A} = \mathop \sum \limits_{i = 1}^D {\alpha _i}{\tau _i}$, where D is the number of lifetime components. τ-Net directly predicts τA for input histograms. The accuracy of lifetime determination in terms of SNR is shown in Fig. 3(b). The model-free algorithm, CMM and model-based LSF were also compared to SRI-FLIMnet. For LSF, the mono-exponential model is used for evaluating τA of multi-exponential decays. The SNR of a decay histogram equals to $\sqrt N $ with N the total photon count. The SNR(dB) is calculated as $SNR({dB} ) = 20log\sqrt N $. The mean absolute error (MAE) determines the analysis accuracy. Figure 3(b) indicates that τ-Net shows the best accuracy, much better than traditional CMM and LSF. The median MAE of τ-Net is 0.169 for SNR 20 ∼ 30 dB (100 ∼ 1000 photon counts). As a comparison, the median MAE of LSF is 0.46 for SNR 38 ∼ 40 dB (6400 ∼ 10, 000 photon counts). The accuracy of τ-Net for low SNR is even better than that of LSF for high SNR. The lifetime determination performance over the whole dynamic lifetime range was investigated in Fig. 3 (c). Generally, given a fixed SNR, decays with larger lifetimes are nosier because photons are distributed in more time bins, resulting in fewer photons in a single time bin. So, the accuracy is poor for decays with larger lifetimes. Nevertheless, τ-Net keeps a low MAE and small changes compared with the other two traditional methods over the whole lifetime ranges. The results in Fig. 3 indicate τ-Net is robust and accurate to calculate lifetimes. In addition, the calculation speed for τ-Net is more than 300-fold faster than LSF.

 figure: Fig. 3.

Fig. 3. Fluorescence lifetime determination performance of SRI-FLIMnet. (a) Decay samples used for testing τ-Net. The lifetime components for Decays 1–3 are 2, 3, and 4; the lifetimes for Decays 1–3 are [0.6, 1.5], [0.8, 2, 2.6], and [1.2, 1.8, 3.1, 4.5]; the total photon counts are 400, 2000, and 5000. The red dashed line represents the IRF. (b) Lifetime determination performance under different SNR conditions. In each group, the lifetime ranges from 0.5 to 5.5 ns and the lifetime components randomly vary from 1 to 4. The SNR ranges from 20 ∼ 40dB, corresponding to 100 ∼ 10,000 photon counts. (c) Lifetime determination performance for different average lifetimes. In each group, the lifetime components randomly vary from 1 to 4, and SNR ranges from 20 ∼ 40dB. In (b) and (c), 2000 samples were taken into consideration in each group.

Download Full Size | PDF

The performance of spatial resolution improvement for SRI-FLIMnet was further investigated. The testing FLIM dataset contains 1,000 samples. As an example, we evaluated reconstructed HR images with k = 4. In this case, suppose that the dwell time in each scanning step is fixed, the image acquisition speed improves 16 times for LSM-FLIM systems. At the same time, LR images undergo severe degradation as much spatial and temporal information is lost. Figure 4 shows the input LR, interpolated HR, SRI-FLIMnet-reconstructed HR, and ground truth (GT) FLIM images of four samples chosen from the testing dataset to illustrate the results better. The four samples labeled S1 to S4 include different morphological features such as a sizeable cellular population and complex cell patterns. They also have distinct lifetime distributions. As shown in Fig. 4, the 64×64 LR images predicted by τ-Net are shown in the 1st row, with a poor spatial resolution and a strong mosaic pattern. Compared with the GT images in the 4th row, the boundaries are unclear, and many delicate structures are missing. The lifetime distributions also have significant deviations. The 2nd row shows referenced interpolated 256×256 HR images using the bicubic interpolating method, indicating that the direct interpolation fails to improve the spatial resolution and introduces more artifacts. For example, the boundaries in all samples are severely distorted. In contrast, SRI-FLIMnet-reconstructed HR images in the 3rd row have sharper and smoother edges. They also recover more refined textures compared to LR images.

 figure: Fig. 4.

Fig. 4. Spatial resolution improved FL images of four samples with distinct morphology and lifetime distributions. The samples are denoted by S1 – S4. The low-resolution (LR), bicubic interpolation, reconstructed HR by SRI-FLIMnet, and ground-truth HR (GT) FL images are shown in the 1st row to 4th rows, respectively. The white dashed boxes in each subfigure depict the regions of interest (ROIs) for further analysis.

Download Full Size | PDF

Figure 5 shows the zoomed-in regions of interest (ROIs) marked by white dashed boxes in Fig. 4 to illustrate the performance of SRI-FLIMnet better. The results indicate that SRI-FLIMnet can generate shapes, boundaries, and general lifetime distributions closer to GT images from minimal information in LR images. SRI-FLIMnet-reconstructed images significantly outperform those using traditional interpolation methods. However, SRI-FLIMnet fails to reconstruct pixel-level refined structures such as gaps between cells in S1 and irregular boundaries in S3. In addition, it cannot recover random lifetime variations in cells. For example, the spotty lifetime patterns in S4 are blurred and smoothed. Therefore, SRI-FLIMnet has a solid ability to generate visually clear shapes and boundaries, whereas it cannot precisely reconstruct pixel-wise details and variations.

 figure: Fig. 5.

Fig. 5. Zoomed ROIs corresponding to the white dashed boxes in Fig. 3. The LR images lose most of the high- and low-frequency information and show strong mosaic patterns. The reconstructed HR images are more visually pleasing and close to the GT images than the interpolating method.

Download Full Size | PDF

The precision of SRI-FLIMnet-reconstructed HR images was also evaluated. Figure 6 shows absolute error maps (Fig. 6(a) – (d)) and lifetime distribution histograms (Fig. 6(e) – (h)) of S1 – S4 shown in Fig. 4. The absolute error maps are the differences between reconstructed images and GT images. Since different samples have various lifetime ranges, absolute error maps are normalized to 0 ∼ 1 for a better comparison. Interestingly, the largest errors always occur at outer boundaries for all four samples, where the spatial differentiation reaches its maximum value. The spatial differentiation has a significant effect on reconstruction precision. An area with sizeable spatial differentiation tends to have an apparent absolute error and vice versa. Figures 6 (e)- (h) show lifetime distribution histograms of S1 – S4, respectively. The two histograms of SRI-FLIMnet-reconstructed and GT images have primarily overlapped distributions for all four samples, revealing that SRI-FLIMnet can correctly restore the lifetime distribution in a wide dynamic range. However, there is a slight difference in lifetime distribution between reconstructed and GT images. The reconstructed images have more pixels with lifetime centering in the peak values. In addition, the lifetime distributions of REC images are relatively narrower compared to corresponding GT images.

 figure: Fig. 6.

Fig. 6. Normalized absolute error maps (a) - (d) and lifetime distributions of the GT and SRI-FLIMnet reconstructed HR images (e) - (h) for S1 – S4. The normalized absolute error maps have an arbitrary unit.

Download Full Size | PDF

We further evaluated SRI-FLIMnet on the whole testing dataset. Table 1 summarizes the results of k = 2, 3, and 4. Two image quality metrics, SSIM and image MAE (IMMAE), were used for quantitative evaluation. The SSIM quantifies how SRI-FLIMnet can reconstruct close LR images to corresponding GT images. In contrast, the IMMAE measures the overall deviation of the lifetime distribution of reconstructed images from that of GT images. The IMMAE for an image is calculated as$|{{I^{Rec}} - {I^{GT}}} |/{M^2}$, where ${\textrm{I}^{\textrm{Rec}}}$and ${\textrm{I}^{\textrm{GT}}}$denote the reconstructed and GT images, respectively. M is the length of the spatial dimension. The dynamic lifetime range for calculating SSIM is 0 ∼ 5.5 ns. In Table 1, the results are the average values of 10,000 samples. Both metrics of SRI-FLIMnet significantly outperform those of bicubic interpolation algorithms for different scaling ratios. In comparison, The SSIM of SRI-FLIMnet is approximately twice that of the bicubic interpolation method, whereas the IMMAE of SRI-FLIMnet is also much better. The results reveal that SRI-FLIMnet has superior performance in reconstructing morphologies and lifetime distributions. It is worth noting that the testing samples we considered have complex structures and an extensive dynamic lifetime range. Both metrics for SRI-FLIMnet can be further improved if we only consider specified cell types and narrow the lifetime range.

Tables Icon

Table 1. Quantitative Evaluation of Different Scaling Ratios on Testing Dataset

5. Experimental validations

5.1 Sample preparation

We validated the trained SRI-FLIMnet using FLIM data of bacteria-infected mouse raw macrophage cells cultured in DMEM (Dulbecco’s modified Eagle medium) containing 10% FCS (fetal calf serum), maintained with 5% CO2 in a humidified chamber at 37°C. Cells were seeded onto glass coverslips within 24-well cell culture plates. Lactobacillus and Shigella Sonnei bacteria were used in FLIM experiments. They were engineered to express a green fluorescent protein, which was harvested at an early exponential phase. Cells cultures were infected at a multiplicity of infection (MOI) of 100. About one hour of infection, the wells were washed 3 times with phosphate buffer saline (PBS) to remove extracellular bacteria. Then, cells were further incubated with DMEM containing 50 μs/mL gentamicin. The DMEM with gentamicin was removed at the chosen time, and cells were washed 3 times with antibiotic-free PBS. Cells were fixed with 3.7% paraformaldehyde for 15 min and permeabilized with 0.1% Triton X-100 for 5 min after being washed 3 times with PBS. Cells were stained with phalloidin Alexa Flour 546 after being washed 3 times with PBS and then were covered by coverslips with a ProLong antifade solution for microscopic examination.

5.2 Experimental data acquisition and analysis

A two-photon FLIM system was used to obtain FLIM data. The system contains a confocal microscope (LSM 510, Carl Zeiss), a femtosecond Ti: sapphire laser (Chameleon, Coherent) with an 850 nm wavelength as the excitation source, and a TCSPC acquisition system (SPC-830, Becker & Hickl GmbH) for time-tagging detected photons. The laser repetition rate is 80MHz, and the laser pulse duration is less than 200 fs. The fluorescence signals were collected using a 63× water-immersion objective lens (N.A. = 1.0) after passing a 500-550 nm bandpass filter. The TCSPC resolution is 0.039 ns, and each measured histogram contains 256 time-bins. FLIM data (with a dimension of 256×256×256) were taken as ground truth HR images. LR FLIM data is obtained through pixel binning as described in Eq. (5). We used a k = 4 as an example. Figure 7 shows 4 different samples denoted by S1 to S4. The peak intensities of the samples range from 4,000 to 16,000 photon counts, providing a sufficient SNR for determining lifetimes. The bacteria have a smaller lifetime of around 2 ns, contrasting to macrophage cells. FLIM images can better identify cells infected by bacteria and localize bacteria. The 1st column shows LR lifetime images of the samples. The 2nd and 3rd columns are reconstructed HR images by bicubic interpolation and SRI-FLIMnet, respectively. The last column is GT images for reference. SRI-FLIMnet can generate clear boundaries, whereas bicubic interpolated HR images show distinct jagged artifacts. This can be observed in all samples. For S2, small rod-shaped bacteria structures were well recovered in reconstructed HR images from LR images. The boundaries between cells are difficult to distinguish in the LR and interpolated HR images of S4. Reconstructed HR images, however, show clearer boundaries close to GT images. SRI-FLIMnet also has good performance in reconstructing correct lifetime distributions. The lifetime distributions of LR images shift towards a smaller lifetime due to the pixel binning of decay histograms among neighboring pixels. This problem is resolved in reconstructed HR images, which generate similar lifetimes with GT images. More apparently, the outer boundaries with a higher lifetime are also recovered. These results demonstrate that SRI-FLIMnet can reconstruct more useful HR images from their LR counterparts without prior knowledge.

 figure: Fig. 7.

Fig. 7. Spatial resolution improved images of experimental FLIM data of bacteria-infected mouse raw macrophage cells. Four different samples denoted by S1 to S4 are shown in the 1st to 4th columns. The 1st row shows the normalized intensity images of the four samples. Their FL images are shown in the 2nd row. The reconstructed HR images by SRI-FLIMnet are shown in the 3rd row. The last row is the GT HR images for reference.

Download Full Size | PDF

Figure 8 shows the normalized absolute error maps and lifetime distribution histograms in Fig. 7. Like Fig. 6, large discrepancy only occurs at boundaries. The lifetime distributions are also close to GT images. Figures 7 and 8 indicate that SRI-FLIMnet trained by semi-synthetic datasets can be well applied to new experimental datasets. The semi-synthetic datasets and the proposed SRI-FLIMnet are expected to find more DL-based FLIM applications. However, it should be noted that SRI-FLIMnet is limited to recovering very fine structures (high-frequency contents) and producing smooth images. This can be seen from the reconstructed images of S1 to S4 in the 3rd column, in which irregular and rough outer edges of cells are over-smoothed.

 figure: Fig. 8.

Fig. 8. Normalized absolute error maps (a) - (d) and lifetime distributions of the GT and SRI-FLIMnet reconstructed HR images (e) - (h) for experimental data S1 – S4. The normalized absolute error maps have an arbitrary unit.

Download Full Size | PDF

6. Discussion and conclusion

We first developed a theoretical method to generate massive semi-synthetic FLIM data in this work. It is challenging to use traditional FLIM experiments to image every cellular morphology and various fluorophores with an extensive range of lifetimes and complex decays. Therefore, there are no available large-scale FLIM datasets in scientific communities. Our work offers a feasible way to fill the gap and develop FLIM-related deep learning techniques. A degrading model was also proposed to obtain LR-HR image pairs with different scaling ratios, essential for spatial resolution enhancement tasks. The semi-synthetic FLIM dataset can be further optimized according to specific tasks by considering cell types, fluorophores, SNR, and other noise sources to perform better. It is expected that our method will significantly boost deep learning techniques for FLIM applications.

In addition to the semi-synthetic FLIM data preparation, a high-efficient neural network architecture, SRI-FLIMnet, was designed to map LR FLIM data to HR images. Achieving HR for FLIM is more challenging and complicated. The lifetime should be evaluated from measured decay histograms. Meanwhile, LR FLIM data has significant information loss in spatial and temporal dimensions. To address these issues, SRI-FLIMnet adopts a combined strategy to estimate lifetimes and improve spatial resolution for LR images sequentially. Two subnets, τ-Net and SR-Net are responsible for the two steps, respectively. This configuration is flexible to tailor the network for different situations. For example, it has been demonstrated that a narrower input lifetime range will provide higher estimation accuracy in τ-Net [18]. If samples’ lifetime range is known in advance, τ-Net can be fast re-trained within a minute to get better performance. Both semi-synthetic and experimental FLIM datasets were applied to show that SRI-FLIMnet can show more apparent boundaries and closer lifetime distributions to GT images. However, it should be noted that SRI-FLIMnet cannot precisely reconstruct images at the pixel level. Theoretically, mapping LR to HR images is an ill-posed problem, as LR images do not contain detailed information. Deep learning approaches can learn statistical features but cannot fully recover all information. Even when we use generative models like generative adversarial networks (GANs), the network would only generate plausible and artificial details, which may misrepresent original images.

In summary, we propose a deep learning-based spatial resolution improved FLIM algorithm. The proposed SRI-FLIMnet trained by a semi-synthetic dataset shows superior performance in reconstructing HR images from LR 3D FLIM data. The performance is also verified by experimental data for bacterial infected mouse raw macrophage cells. Our approach offers a practical computational method to fast obtain HR images without conducting extra experiments and modifying the system setting.

Funding

Medical Research Scotland (1179-2017); Biotechnology and Biological Sciences Research Council (BB/V019643/1).

Acknowledgements

The authors would like to acknowledge Photon Force, Ltd. for supporting this project, and we also acknowledge support from NVIDIA for donating Quadro P5000 GPUs.

Disclosures

The authors declare no conflict of interest.

Data availability

Data underlying the results presented in this paper are not publicly available but may be obtained from the authors upon reasonable request.

References

1. J. R. Lakowicz, Principles of fluorescence Spectroscopy. (Springer, 2006).

2. K. Suhling, L. M. Hirvonen, J. A. Levitt, P. Chung, C. Tregidgo, A. L. Marois, D. A. Rusakov, K. Zheng, S. Ameer-Beg, S. Poland, S. Coelho, R. Henderson, and N. Krstajic, “Fluorescence lifetime imaging (FLIM): Basic concepts and some recent developments,” Med Photon 27, 3–40 (2015). [CrossRef]  

3. R. Datta, T. M. Heaster, J. T. Sharick, A. A. Gillette, and M. C. Skala, “Fluorescence lifetime imaging microscopy: fundamentals and advances in instrumentation, analysis, and applications,” J. Biomed. Opt. 25(07), 1 (2020). [CrossRef]  

4. Y. Long, Y. Stahl, S. Weidtkamp-Peters, M. Postma, W. Zhou, J. Goedhart, M. Sánchez-Pérez, T. Gadella Jr, R. Simon, B. Scheres, and I. Blilou, “In vivo FRET–FLIM reveals cell-type-specific protein interactions in arabidopsis roots,” Nature 548(7665), 97–102 (2017). [CrossRef]  

5. L. Marcu, “Fluorescence lifetime techniques in medical applications,” Ann Biomed Eng. 40(2), 304–331 (2012). [CrossRef]  

6. L. Sauer, A. S. Vitale, N. K. Modersitzki, and P. S. Bernstein, “Fluorescence lifetime imaging ophthalmoscopy: autofluorescence imaging and beyond,” Eye 35(1), 93–109 (2021). [CrossRef]  

7. W. Becker, Advanced Time-Correlated Single Photon Counting Applications. (Springer, 2015).

8. M. Wahl, “Time-Correlated Single Photon Counting,” Technical Note, PicoQuant GmbH (2014) https://www.picoquant.com/images/uploads/page/files/7253/technote_tcspc.pdf

9. J. B. Pawley, Handbook of Biological Confocal Microscopy, (Springer, 2006).

10. M. Castello, G. Tortarolo, M. Buttafava, T. Deguchi, F. Villa, S. Koho, L. Pesce, M. Oneto, S. Pelicci, L. Lanzanó, P. Bianchini, C. J. R. Sheppard, A. Diaspro, A. Tosi, and G. Vicidomini, “A robust and versatile platform for image scanning microscopy enabling super-resolution FLIM,” Nat. Methods 16(2), 175–178 (2019). [CrossRef]  

11. B. Kaye, P. J. Foster, T. Y. Yoo, and D. J. Needleman, “Developing and Testing a Bayesian Analysis of Fluorescence Lifetime Measurements,” PLoS ONE 12(1), e0169337 (2017). [CrossRef]  

12. S. Wang, J. V. Chacko, A. K. Sagar, K. W. Eliceiri, and M. Yuan, “Nonparametric empirical Bayesian framework for fluorescence-lifetime imaging microscopy,” Biomed. Opt. Express 10(11), 5497–5517 (2019). [CrossRef]  

13. D. D. Li, B. Rae, R. Andrews, J. Arlt, and R. K. Henderson, “Hardware implementation algorithm and error analysis of high-speed fluorescence lifetime sensing systems using center-of-mass method,”,” J Biomed Opt 15(1), 017006 (2010). [CrossRef]  

14. S. Ranjit, L. Malacrida, D. M. Jameson, and E. Gratton, “Fit-free analysis of fluorescence lifetime imaging data using the phasor approach,” Nat Protoc 13(9), 1979–2004 (2018). [CrossRef]  

15. R. K. Henderson, N. Johnston, F. M. D. Rocca, H. Chen, D. D. Li, G. Hungerford, R. Hirsch, D. McLoskey, P. Yip, and D. J. S. Birch, “A 192 ( 128 Time Correlated SPAD Image Sensor in 40-nm CMOS Technology,” IEEE J. Solid-State Circuits 54(7), 1907–1916 (2019). [CrossRef]  

16. S. P. Poland, N. Krstajić, J. Monypenny, S. Coelho, D. Tyndall, R. J. Walker, V. Devauges, J. Richardson, N. Dutton, P. Barber, D. D. Li, K. Suhling, T. Ng, R. K. Henderson, and S. M. Ameer-Beg, “A high speed multifocal multiphoton fluorescence lifetime imaging microscope for live-cell FRET imaging,” Biomed Opt Express 6(2), 277–296 (2015). [CrossRef]  

17. J. T. Smith, R. Yao, N. Sinsuebphon, A. Rudkouskaya, N. Un, J. Mazurkiewicz, M. Barroso, P. Yan, and X. Intes, “Fast fit-free analysis of fluorescence lifetime imaging via deep learning,” Proc Natl Acad Sci U S A 116(48), 24019–24030 (2019). [CrossRef]  

18. D. Xiao, Y. Chen, and D. D. Li, “One-Dimensional Deep Learning Architecture for Fast Fluorescence Lifetime Imaging,” IEEE J Sel Top Quantum Electron 27(4), 1–10 (2021). [CrossRef]  

19. D. Xiao, Z. Zang, N. Sapermsap, Q. Wang, W. Xie, Y. Chen, and D. D. U. Li, “Dynamic fluorescence lifetime sensing with CMOS single-photon avalanche diode arrays and deep learning processors,” Biomed. Opt. Express 12(6), 3450–3462 (2021). [CrossRef]  

20. C. Dong, C. C. Loy, K. He, and X. Tang, “Image super-resolution using deep convolutional networks,” IEEE Trans. Pattern Anal. Mach. Intell. 38(2), 295–307 (2016). [CrossRef]  

21. W. Yang, X. Zhang, Y. Tian, W. Wang, J. Xue, and Q. Liao, “Deep learning for single image super-resolution: a brief review,” IEEE Trans. Multimed. 21(12), 3106–3121 (2019). [CrossRef]  

22. K. Jiang, Z. Wang, P. Yi, and J. Jiang, “A progressively enhanced network for video satellite imagery super resolution,” IEEE Signal Process. Lett. 25(11), 1630–1634 (2018). [CrossRef]  

23. W. Shi, J. Caballero, F. Huszar, J. Totz, A.P. Aitken, R. Bishop, D. Rueckert, and Z. Wang, “Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 1874–1883 (2016).

24. Y. Rivenson, Z. Göröcs, H. Günaydin, Y. Zhang, H. Wang, and A. Ozcan, “Deep learning microscopy,” Optica 4(11), 1437–1443 (2017). [CrossRef]  

25. H. Wang, Y. Rivenson, Y. Jin, Z. Wei, R. Gao, H. Günaydın, L. A. Bentolila, C. Kural, and A. Ozcan, “Deep learning enables cross-modality super-resolution in fluorescence microscopy,” Nat Methods 16(1), 103–110 (2019). [CrossRef]  

26. H. Zhang, C. Fang, X. Xie, Y. Yang, W. Mei, D. Jin, and P. Fei, “High-throughput, high-resolution deep learning microscopy based on registration-free generative adversarial network,” Biomed. Opt. Express 10(3), 1044–1063 (2019). [CrossRef]  

27. C. Qiao, D. Li, Y. Guo, C. Liu, T. Jiang, Q. Dai, and D. Li, “Evaluation and development of deep neural networks for image super-resolution in optical microscopy,” Nat Methods 18(2), 194–202 (2021). [CrossRef]  

28. L. Fang, F. Monroe, S. W. Novak, L. Kirk, C. R. Schiavon, S. B. Yu, T. Zhang, M. Wu, K. Kastner, A. A. Latif, Z. Lin, A. Shaw, Y. Kubota, J. Mendenhall, Z. Zhang, G. Pekkurnaz, K. Harris, J. Howard, and U. Manor, “Deep learning-based point-scanning super-resolution imaging,” Nat Methods 18(4), 406–416 (2021). [CrossRef]  

29. Dataset collections in Pytorch, https://pytorch.org/vision/stable/datasets.html

30. S. Ross, Stochastic Processes, 2nd Edition, (Wiley, 1995).

31. The Human Protein Atlas, https://www.proteinatlas.org/

32. Human Protein Atlas - Single Cell Classification, https://www.kaggle.com/c/hpa-single-cell-image-classification

33. Y. Zhang, Y. Tian, Y. Kong, B. Zhong, and Y. Fu, “Residual Dense Network for Image Super-Resolution,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2472–2481 (2018).

34. F. Yu and V. Koltun, “Multi-scale context aggregation by dilated convolutions,” in International Conference on Learning Representations (ICLR), (2016).

35. P. Wang, P. Chen, Y. Yuan, D. Liu, Z. Huang, X. Hou, and G. Cottrell, “Understanding Convolution for Semantic Segmentation,” in 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), 1451–1460, (2018).

36. Z. Wang, E. P. Simoncelli, and A. C. Bovik, “Multiscale structural similarity for image quality assessment,” The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers 2, 1398–1402 (2003). [CrossRef]  

37. Y. Li, S. Natakorn, Y. Chen, M. Safar, M. Cunningham, J. Tian, and D. D.-U. Li, “Investigations on Average Fluorescence Lifetimes for Visualizing Multi-Exponential Decays,” Front. Phys. 8, 576862 (2020). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (8)

Fig. 1.
Fig. 1. Original intensity images in the HPA dataset and the semi-synthetic FLIM data. The 1st column is the RGB intensity images of three samples. Their corresponding four color-channel images are shown in the 2nd column; each organelle is labeled by one color. The 3rd and 4th columns show the synthetic 3D FLIM data cloud and lifetime images. In 3D point plots (3rd column), the dark blue to bright yellow point color indicates a high to low photon density.
Fig. 2.
Fig. 2. Network architecture of the proposed SRI-FLIMnet with a scaling ratio k = 4. SRI-FLIMnet contains two subnets, τ-Net, and SRI-Net, responsible for lifetime evaluation and nonlinear transformation for improving spatial resolution, respectively. The parameters of convolutional layers are labeled in the parentheses as (feature number, kernel size, stride). SRI-Net contains 8 residual dense blocks (RDB) with hybrid dilated convolution. In each RDB, 8 2D convolutional layers with a kernel size of 3 and a stride of 1 are densely connected. Both the input feature number and growth rate are 64. In the Up-Scaling part of SRI-Net, the parameters are (1024, 3, 1) for both convolutional layers. The input data of SRI-FLIMnet is 64×64×256, and the final output data size is 256×256.
Fig. 3.
Fig. 3. Fluorescence lifetime determination performance of SRI-FLIMnet. (a) Decay samples used for testing τ-Net. The lifetime components for Decays 1–3 are 2, 3, and 4; the lifetimes for Decays 1–3 are [0.6, 1.5], [0.8, 2, 2.6], and [1.2, 1.8, 3.1, 4.5]; the total photon counts are 400, 2000, and 5000. The red dashed line represents the IRF. (b) Lifetime determination performance under different SNR conditions. In each group, the lifetime ranges from 0.5 to 5.5 ns and the lifetime components randomly vary from 1 to 4. The SNR ranges from 20 ∼ 40dB, corresponding to 100 ∼ 10,000 photon counts. (c) Lifetime determination performance for different average lifetimes. In each group, the lifetime components randomly vary from 1 to 4, and SNR ranges from 20 ∼ 40dB. In (b) and (c), 2000 samples were taken into consideration in each group.
Fig. 4.
Fig. 4. Spatial resolution improved FL images of four samples with distinct morphology and lifetime distributions. The samples are denoted by S1 – S4. The low-resolution (LR), bicubic interpolation, reconstructed HR by SRI-FLIMnet, and ground-truth HR (GT) FL images are shown in the 1st row to 4th rows, respectively. The white dashed boxes in each subfigure depict the regions of interest (ROIs) for further analysis.
Fig. 5.
Fig. 5. Zoomed ROIs corresponding to the white dashed boxes in Fig. 3. The LR images lose most of the high- and low-frequency information and show strong mosaic patterns. The reconstructed HR images are more visually pleasing and close to the GT images than the interpolating method.
Fig. 6.
Fig. 6. Normalized absolute error maps (a) - (d) and lifetime distributions of the GT and SRI-FLIMnet reconstructed HR images (e) - (h) for S1 – S4. The normalized absolute error maps have an arbitrary unit.
Fig. 7.
Fig. 7. Spatial resolution improved images of experimental FLIM data of bacteria-infected mouse raw macrophage cells. Four different samples denoted by S1 to S4 are shown in the 1st to 4th columns. The 1st row shows the normalized intensity images of the four samples. Their FL images are shown in the 2nd row. The reconstructed HR images by SRI-FLIMnet are shown in the 3rd row. The last row is the GT HR images for reference.
Fig. 8.
Fig. 8. Normalized absolute error maps (a) - (d) and lifetime distributions of the GT and SRI-FLIMnet reconstructed HR images (e) - (h) for experimental data S1 – S4. The normalized absolute error maps have an arbitrary unit.

Tables (1)

Tables Icon

Table 1. Quantitative Evaluation of Different Scaling Ratios on Testing Dataset

Equations (8)

Equations on this page are rendered with MathJax. Learn more.

λ i , j = A i , j I RF ( t ) p = 1 D α p e t / τ p + ϵ , i , j [ 1 , M ] , t [ 1 , L ] ,
N i , j P o i s s o n ( S 0 T λ i , j ( t ) d t ) .
f i , j ( t ) = λ i , j ( t ) 0 T λ i , j ( t ) d t .
I L R = D k ( I H R K ) + ε ,
I L R ( m , m , L ) = 1 k × k × 1 s = ( k , k , 1 ) I H R ( M , M , L ) ,
L 2 ( τ ^ , τ ) = 1 N i = 1 N ( τ ^ i τ i ) 2 ,
L p e r c e p ( I ^ H R , I H R ) = L M S S S I M ( I ^ H R , I H R ) + γ L 1 ( I ^ H R , I H R ) ,
L MS SSIM = 1 [ l Ms ( I ^ H R , I H R ) ] α M s j = 1 M s [ c s j ( I ^ H R , I H R ) ] β j ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.