Conference PaperPDF Available

Wide-field Diffuse Optical Tomography Using Deep Learning

Authors:
Wide-field Diffuse Optical Tomography Using Deep Learning
Navid Ibtehaj Nizam†, *, Marien Ochoa, Jason T. Smith, and Xavier Intes
Center for Modeling, Simulation and Imaging in Medicine, Rensselaer Polytechnic Institute, Troy, NY 12180, USA
These authors contributed equally.
*nizamn@rpi.edu
Abstract: A modified AUTOMAP architecture, with micro-CT validation, is used for 3D
reconstructions in Diffuse Optical Tomography, employing wide-field illumination and detection.
The performance is compared with a regularized Least Squares technique.
1. Introduction
Diffuse Optical Tomography (DOT) enables monitoring of the physiological state of deep tissues with high
sensitivity. DOT performances depend on the spatial, spectral, and temporal density of the data acquired. Recently, a
hyperspectral wide-field time-domain single-pixel instrumental strategy has been proposed to collect dense data
along all these dimensions efficiently [1]. Still, the image formation component is challenging. Traditionally,
techniques like the least-squares (LSQ), conjugate gradient (CGS), and total-variation minimization (TVAL) have
been mainly deployed to solve the optical inverse problem. However, the major drawback of these techniques is that
there is no “one-size-fits-all” approach. Additionally, optimizing the parameters associated with these traditional
solvers is often an expert and cumbersome process. Because of the availability of high-powered GPUs and the
associated rise in computational power, there has been an increased interest in investigating the potential of Deep-
Learning (DL)-based approaches for 3D image formation [2]. Herein, we report on modified AUTOMAP (ModAM)
[3], a Convolutional Neural Network (CNN)-based architecture, that directly reconstructs the δμa contrast based on
single-pixel tomographic data sets [4].
2. Methods
For simplicity, in this report, we consider a Continuous-wave (CW) approach throughout. An in silico workflow is
used to train the network, which remains suitable for an in vitro experiment. Additionally, widefield illumination
and detection strategies are utilized leveraging sparse and low-frequency patterns (total of 36) as in [2] (shown in
Fig. 1(a)). We use the open-source Monte Carlo (MC) based software, MCX [5], to generate large volumes of in
silico optical phantoms. The homogenous embeddings in the phantoms are generated from binary characters
obtained from the EMNIST dataset (Fig. 1(b), introduced in [6] and shown to contain enough spatial heterogeneity
to cover the complex 3D biodistributions associated with in vivo imaging). These in silico phantoms have a range of
values for δµa and depth. MCX is deployed to generate perturbed (φ) and unperturbed (φo) measurement vectors for
each in silico phantom (Fig. 1(c)). The ModAM network is trained using 5,000 measurement vectors under the
Rytov approximation (log φo/ φ) with an 80/20 training/validation split and an Adam optimizer. The overall
structure of the network is illustrated in Fig. 1(d).
For experimental validation, an agar-based in vitro phantom is prepared. A mixture of ink and intralipid solution is
used to generate the absorption contrast and scattering, respectively. In a homogenous background, two thin
capillaries are filled with the same absorption contrast and embedded at a high depth (for diffuse optics) of 8.5 mm
from the illumination plane (δµa=0.176 mm-1 and reduced scattering, μs’=1 mm-1). A schematic of the in vitro
phantom is shown in Fig. 2(a). Our single-pixel hyperspectral system (equipped with Digital Micro-Mirror Devices)
is used for projecting the 36 illumination and detection patterns (same as the ones shown in Fig. 1(a)). The perturbed
and unperturbed measurement vectors are recorded with a 16-channel PMT. The details of the experimental
Fig. 1. (a) Illumination and detection
patterns (36 each). (b) The EMNIST
dataset used for training. (c) Perturbed
(red) and Unperturbed (blue)
measurements simulated using MCX.
(d) The ModAM architecture. (e) The
reconstructed iso-volumes with their
GT. (f) The 2D cross-sections at a
depth of 2 mm. (g) Table showing the
quantitative results in terms of the VE
and
the maximum reconstructe
d
δμ
a
.
OW4D.7 Biophotonics Congress: Biomedical Optics (Translational,
Microscopy, OCT, OTS, BRAIN) © Optica Publishing Group 2022
© 2022 The Author(s)
apparatus and protocol can be found in [4]. Additionally, to obtain the exact position, depth, and separation of the
two tubes, a micro-CT scan is carried out (Fig. 2(b)). The volume obtained from the micro-CT is treated as the
Ground-Truth (GT) for the experiment (to calculate the Volume Error (VE)).
3. Results
Representative reconstruction results for an in silico phantom, not part of the training dataset, are presented in Figs.
1(e)-1(g) for both the ModAM network and the regularized LSQ-based technique. The in silico phantom has a
dimension of 30x40x20 mm3 (to match the in vitro experiment) and a homogenous embedding in a homogenous
background (δµa=0.2 mm-1). The embedding (thickness=3 mm) is placed at a shallow depth of 2 mm from the
illumination plane. We present the results in terms of the iso-volume (Fig. 1(e)), the 2D cross-sections at the 2 mm
depth (Fig. 1(f)), and the quantitative evaluation of the reconstructions in terms of the maximum value of the
reconstructed δµa and the VE (Fig. 1(g)). The results obtained from the ModAM network are superior to the
regularized LSQ, both in terms of the VE and the maximum value of δµa reconstructed. Although the ModAM
network takes approximately 5.25 hours to train (NVIDIA RTX 2080 Ti), the reconstruction time for the ModAM
network is a few milliseconds, while that for the regularized LSQ is approximately 20 minutes. The reconstruction
results of the in vitro experiment are presented in Figs. 2(c) and 2(d) (the iso-volumes and the 2D cross-section at a
depth of 10 mm, respectively). Here, the ModAM results are significantly better than those obtained from the
regularized LSQ (as shown quantitatively in Fig. 2(e)). The time advantage in reconstruction is similar to the in
silico case, with the added bonus that the ModAM network need not be re-trained for the in vitro experiment.
Fig. 2. (a) Schematic of the in vitro phantom (b)
Volume generated from micro-CT scan. (c)
Reconstructed iso-volumes. (d) 2D cross-sections at a
depth of 10 mm. (e) Table showing the quantitative
results in terms of the VE and the maximum
reconstructed δμa.
4. Discussion/Conclusion
Our presented in silico and in vitro results reveal that the ModAM network can lead to faster and more accurate δμa
reconstructions than the traditional techniques even at high depths (hence, high scattering). It has also been
demonstrated that the ModAM network, enhanced by the spatial heterogeneity in the EMNIST dataset, can
reconstruct a wide array of structures, both in silico and in vitro. However, re-training the network will be necessary
to change the source-detector configuration and/or phantom dimensions. A detailed investigation, with more in vitro
data and pre-clinical in vivo imaging, will be carried out in future works.
Acknowledgements
The authors acknowledge the funding support from National Institutes of Health (NIH) under grants R01CA237267,
R01CA207725 and R01CA250636. We would like to thank Mr. Mengzhou Li and Mr. Xiaodong Guo for providing
the raw micro-CT data.
References
[1] Q. Pian, et al., “Compressive Hyperspectral Time-resolved Wide-Field Fluorescence Lifetime Imaging,” Nature Photonics 11, 411-417
(2017).
[2] L. Tian, et al., “Deep learning in biomedical optics,” Laser in Surgery and Medicine 53(6), 748-775 (2021).
[3] B. Zhu, et al., “Image reconstruction by domain-transform manifold learning,” Nature 555, 487-492 (2018).
[4] Q. Pian, et al., “Hyperspectral wide-field time domain single-pixel diffuse optical tomography platform,” BOE 9(12), 6258-6272 (2018).
[5] R.Yao et al. "Direct approach to compute Jacobians for diffuse optical tomography using perturbation Monte Carlo-based photon “replay”."
Biomedical optics express 9(10), 4588-4603 (2018).
[6] R. Yao et al. "Net-FLICS: fast quantitative wide-field fluorescence lifetime imaging with compressed sensing–a deep learning approach."
Light: Science & Applications 8(1), 1-7 (2019).
OW4D.7 Biophotonics Congress: Biomedical Optics (Translational,
Microscopy, OCT, OTS, BRAIN) © Optica Publishing Group 2022
... D: DESIGN APPROACH; FF: FEED-FORWARD, I: ITERATIVE UNROLLED BASED MODEL; M: MULTI-MODAL/FREQUENCY; S/P: IN SILICO/PHANTOM DATA; AND C: CLINICAL PATIENT DATA. D M S/P C Approach to mitigate ill-posedness [29]- [31] FF × ✓ × CNN learns the nonlinear end-to-end mapping [25] FF × ✓ × Promote appearance similarity [27], [32] I × ✓ × Augment Gauss-Newton algorithm with deep learning [33] FF ✓ ✓ × Micro-CT structural prior [26] FF × ✓ × Model based on Lippmann-Schwinger equation [34] FF × ✓ × Reflection model as sum of features from different depths [35] I × ✓ × Data-driven unrolled network promoting appearance similarity [36]- [38] FF ✓ ✓ ✓ Multi-modal representation learning (US+DOT) [16] FF × ✓ ✓ Deep spatial-wise attention network Ours FF ✓ ✓ ✓ Orthogonal multi-frequency representation learning condition like a rest stage in brain DOT, absolute imaging approaches use a single set of measurements to reconstruct optical coefficients. In this manuscript, we focus on absolute imaging. ...
Preprint
Full-text available
p>Identifying breast cancer lesions with a portable diffuse optical tomography (DOT) device can improve early detection while avoiding otherwise unnecessarily invasive, ionizing, and more expensive modalities such as CT, as well as enabling pre-screening efficiency. Critical to this capability is not just the identification of lesions but rather the complex problem of discriminating between malignant and benign lesions. To accurately capture the highly heterogeneous tissue of a cancer lesion embedded in healthy breast tissue with non-invasive DOT, multiple frequencies can be combined to optimize signal penetration and reduce sensitivity to noise. However, these frequency responses can overlap, capture common information, and correlate, potentially confounding reconstruction and downstream end tasks. We show that an orthogonal fusion loss of multi-frequency DOT can improve reconstruction. More importantly, the orthogonal fusion leads to more accurate end-to-end identification of malignant versus benign lesions, illustrating its regularization properties in the multi-frequency input space. While the deployment of portable DOT probes requires a severely constrained computational budget, we show that our raw-to-task model, for direct prediction of the end task from signal, significantly reduces computational complexity without sacrificing accuracy, enabling a high real-time throughput, desiredin medical settings. Furthermore, our results indicate that image reconstruction is not necessary for unbiased classiication of lesions with a balanced accuracy of 77% and 66% on the synthetic dataset and clinical dataset, espectively, using the raw-to-task model. Code is available at https: //github.com/sfu-mial/FuseNet </p
... Nizam et al. and Deng et al. have demonstrated progress in improving CW diffuse optical tomography reconstruction quality using a variation of the automated transform by manifold approximation (AUTOMAP) architecture. [43][44][45] Li et al. 46 has applied deep learning and DOT to clinical applications, improving the accuracy of breast tumor imaging. Deep learning has also been applied for generating data in DOT applications. ...
Article
Full-text available
Significance: Imaging through scattering media is critical in many biomedical imaging applications, such as breast tumor detection and functional neuroimaging. Time-of-flight diffuse optical tomography (ToF-DOT) is one of the most promising methods for high-resolution imaging through scattering media. ToF-DOT and many traditional DOT methods require an image reconstruction algorithm. Unfortunately, this algorithm often requires long computational runtimes and may produce lower quality reconstructions in the presence of model mismatch or improper hyperparameter tuning. Aim: We used a data-driven unrolled network as our ToF-DOT inverse solver. The unrolled network is faster than traditional inverse solvers and achieves higher reconstruction quality by accounting for model mismatch. Approach: Our model "Unrolled-DOT" uses the learned iterative shrinkage thresholding algorithm. In addition, we incorporate a refinement U-Net and Visual Geometry Group (VGG) perceptual loss to further increase the reconstruction quality. We trained and tested our model on simulated and real-world data and benchmarked against physics-based and learning-based inverse solvers. Results: In experiments on real-world data, Unrolled-DOT outperformed learning-based algorithms and achieved over 10× reduction in runtime and mean-squared error, compared to traditional physics-based solvers. Conclusion: We demonstrated a learning-based ToF-DOT inverse solver that achieves state-of-the-art performance in speed and reconstruction quality, which can aid in future applications for noninvasive biomedical imaging.
Article
Full-text available
This article reviews deep learning applications in biomedical optics with a particular emphasis on image formation. The review is organized by imaging domains within biomedical optics and includes microscopy, fluorescence lifetime imaging, in vivo microscopy, widefield endoscopy, optical coherence tomography, photoacoustic imaging, diffuse tomography, and functional optical brain imaging. For each of these domains, we summarize how deep learning has been applied and highlight methods by which deep learning can enable new capabilities for optics in medicine. Challenges and opportunities to improve translation and adoption of deep learning in biomedical optics are also summarized.
Article
Full-text available
Macroscopic fluorescence lifetime imaging (MFLI) via compressed sensed (CS) measurements enables efficient and accurate quantification of molecular interactions in vivo over a large field of view (FOV). However, the current data-processing workflow is slow, complex and performs poorly under photon-starved conditions. In this paper, we propose Net-FLICS, a novel image reconstruction method based on a convolutional neural network (CNN), to directly reconstruct the intensity and lifetime images from raw time-resolved CS data. By carefully designing a large simulated dataset, Net-FLICS is successfully trained and achieves outstanding reconstruction performance on both in vitro and in vivo experimental data and even superior results at low photon count levels for lifetime quantification.
Article
Full-text available
We present the design and comprehensive instrumental characterization of a time domain diffuse optical tomography (TD-DOT) platform based on wide-field illumination and wide-field hyperspectral time-resolved single-pixel detection for functional and molecular imaging in turbid media. The proposed platform combines two digital micro-mirror devices (DMDs) to generate structured light and a spectrally resolved multi-anode photomultiplier tube (PMT) detector in time domain for hyperspectral data acquisition over 16 wavelength channels based on the time-correlated single-photon counting (TCSPC) technique. The design of the proposed platform is described in detail and its characteristics in spatial, temporal and spectral dimensions are calibrated and presented. The performance of the system is further validated through a phantom study where two absorbers in glass tubes with spectral contrast are mapped in a turbid medium of ~20 mm thickness. The method presented here offers the potential of accelerating the imaging process and improving reconstruction results in TD-DOT and thus facilitates its wide spread use in preclinical and clinical in vivo imaging scenarios.
Article
Full-text available
Perturbation Monte Carlo (pMC) has been previously proposed to rapidly recompute optical measurements when small perturbations of optical properties are considered, but it was largely restricted to changes associated with prior tissue segments or regions-of-interest. In this work, we expand pMC to compute spatially and temporally resolved sensitivity profiles, i.e. the Jacobians, for diffuse optical tomography (DOT) applications. By recording the pseudo random number generator (PRNG) seeds of each detected photon, we are able to “replay” all detected photons to directly create the 3D sensitivity profiles for both absorption and scattering coefficients. We validate the replay-based Jacobians against the traditional adjoint Monte Carlo (aMC) method, and demonstrate the feasibility of using this approach for efficient 3D image reconstructions using in vitro hyperspectral wide-field DOT measurements. The strengths and limitations of the replay approach regarding its computational efficiency and accuracy are discussed, in comparison with aMC, for point-detector systems as well as wide-field pattern-based and hyperspectral imaging systems. The replay approach has been implemented in both of our open-source MC simulators - MCX and MMC (http://mcx.space)
Article
Full-text available
Spectrally resolved fluorescence lifetime imaging and spatial multiplexing have offered information content and collection-efficiency boosts in microscopy, but efficient implementations for macroscopic applications are still lacking. An imaging platform based on time-resolved structured light and hyperspectral single-pixel detection has been developed to perform quantitative macroscopic fluorescence lifetime imaging (MFLI) over a large field of view (FOV) and multiple spectral bands simultaneously. The system makes use of three digital micromirror device (DMD)-based spatial light modulators (SLMs) to generate spatial optical bases and reconstruct N by N images over 16 spectral channels with a time-resolved capability (∼40 ps temporal resolution) using fewer than N2 optical measurements. We demonstrate the potential of this new imaging platform by quantitatively imaging near-infrared (NIR) Förster resonance energy transfer (FRET) both in vitro and in vivo. The technique is well suited for quantitative hyperspectral lifetime imaging with a high sensitivity and paves the way for many important biomedical applications.
Article
Full-text available
Image reconstruction plays a critical role in the implementation of all contemporary imaging modalities across the physical and life sciences including optical, MRI, CT, PET, and radio astronomy. During an image acquisition, the sensor encodes an intermediate representation of an object in the sensor domain, which is subsequently reconstructed into an image by an inversion of the encoding function. Image reconstruction is challenging because analytic knowledge of the inverse transform may not exist a priori, especially in the presence of sensor non-idealities and noise. Thus, the standard reconstruction approach involves approximating the inverse function with multiple ad hoc stages in a signal processing chain whose composition depends on the details of each acquisition strategy, and often requires expert parameter tuning to optimize reconstruction performance. We present here a unified framework for image reconstruction, AUtomated TransfOrm by Manifold APproximation (AUTOMAP), which recasts image reconstruction as a data-driven, supervised learning task that allows a mapping between sensor and image domain to emerge from an appropriate corpus of training data. We implement AUTOMAP with a deep neural network and exhibit its flexibility in learning reconstruction transforms for a variety of MRI acquisition strategies, using the same network architecture and hyperparameters. We further demonstrate its efficiency in sparsely representing transforms along low-dimensional manifolds, resulting in superior immunity to noise and reconstruction artifacts compared with conventional handcrafted reconstruction methods. In addition to improving the reconstruction performance of existing acquisition methodologies, we anticipate accelerating the discovery of new acquisition strategies across modalities as the burden of reconstruction becomes lifted by AUTOMAP and learned-reconstruction approaches.
  • Q Pian
Q. Pian, et al., "Compressive Hyperspectral Time-resolved Wide-Field Fluorescence Lifetime Imaging," Nature Photonics 11, 411-417 (2017).