|
1.IntroductionDiffuse optical tomography (DOT) has shown a great potential for breast imaging1–9 and functional brain imaging,10–12 which use near-infrared light in the spectral range of 600 to 950 nm to quantify tissue optical (absorption and scattering) coefficients. There is a critical need to develop an efficient image reconstruction algorithm for DOT. Recovering the internal distribution of optical properties is a severely ill-posed and under-determined inverse problem, due to light propagation in highly scattering biological tissues and limited number of measurements,13,14 which makes image reconstruction challenging. Although both linear and nonlinear reconstruction algorithms for DOT are available,14 considerable efforts have been made to develop various reconstruction algorithms to improve quantitative accuracy and image quality.14–22 To date, the ill-posedness of the inverse problem in DOT can be alleviated by employing a regularization technique, which utilizes a data fitting term together with a regularizer ( or norm, etc.) to suppress the effect of measurement noise and modeling errors.23 When the regularizer is a norm, the reconstruction algorithm becomes the well-known Tikhonov regularization method, which imposes restrictions on the norm of the optical properties23 and is optimal when the distribution of optical properties subjects to Gaussian distribution.24 The merits of Tikhonov regularization are simple, and easy to be implemented. However, the norm will oversmooth reconstructed images and yield low-quality images by penalizing large values.23 An alternative regularizer is total variation (TV) norm, which is the ideal choice when the distribution of optical properties is known to be piecewise constant.24 Another possible regularizer is norm (), which poses a sparsity constraint on the optical properties. The quality of reconstructed images can be improved with the use of sparsity regularization.23,25 The abovementioned algorithms are generally iterative reconstruction algorithms.14 These algorithms require heavy computation and large storage memory because the forward problem must be solved repeatedly, and an updated distribution of optical properties must be found at each iterative step.14 However, iterative reconstruction algorithms have limited capability in terms of reconstruction accuracy and image quality,26 which are important for accurate diagnosis of breast cancers. In addition, how to accurately choose parameters, particularly regularization parameter, for iterative reconstruction algorithms needs to be further considered.27 Recently, artificial neural networks with various network architectures, including deep convolutional neural network,28,29 generative adversarial networks,30 and multilayer perceptron,31 have achieved significant improvements over existing iterative reconstruction methods in the quality of reconstructed images. It is likely that image recovery in DOT benefits from these important developments. In this work, we investigated the feasibility and effectiveness of a back-propagation (BP) neural network (BPNN) to recover the distribution of optical properties in DOT. BPNN is a widely used neural network because it has many advantages. For example, BPNN is simple, efficient at computing the gradient descent, and straightforward to implement. The basic procedure of BPNN includes the forward propagation of input data and the reverse transmission of output error.32 More detailed introduction about BPNN can be found in Refs. 3334.35.36.37.38.–39. The forward propagation of input data is to transmit input data from the input layer through a series of hidden layers toward the output layer, which builds the relationship between input data and output data. The reverse transmission of output error between the calculated and the ground true output is backward propagated from the output layer through the hidden layers to the first layer to adjust the connection weights and bias variables of neurons. By repeatedly applying this procedure, the output error is adjusted to an expected range. We validate the proposed method using simulation experiments and compare BPNN with the popular Tikhonov regularization. Our results demonstrate that our method provides higher accuracy and superior image quality than Tikhonov regularization in recovering a single inclusion or two closely spaced inclusions. The remainder of the paper is organized as follows. Section 2 describes the light propagation model, BPNN, and evaluation metrics. Experimental results and comparisons are presented in Sec. 3. Finally, we present a discussion of results with our conclusions and future work in Sec. 4. 2.Methods2.1.Forward ModelThe light propagation in biological tissues can be modeled by the steady diffusion equation,13,14 which can be described as follows: where is the imaged object, is the photon fluence rate at position , is the diffusion coefficient (), is the absorption coefficient (), is the reduced scattering coefficients (), and is the source term.Here, the boundary condition used for Eq. (1) is Robin-type condition, which can be expressed as follows:13,14 where is the surface boundary of imaged object , is the boundary mismatch parameter, and is the outer normal on .When the distributions of and are known, light measurements at the detectors can be calculated by solving Eqs. (1) and (2) based on the finite-element method,40 which can be modeled with the following equation: where and represent the optical properties ( and ) and the measurements at the detectors, respectively; and are the number of finite-element nodes and the number of boundary measurements, respectively; and is the forward operator that relates the unknown distribution of optical properties to the boundary measurements.2.2.Back-Propagation Neural Network-Based ReconstructionTo improve the performances of iterative reconstruction algorithms in DOT, here we develop a reconstruction algorithm based on a BPNN. BPNN is divided into three types of layers: the input layer (), the fully connected hidden layer (), and the predictable output layer (). The architecture of the three-layer BPNN used for DOT image reconstruction is shown in Fig. 1. We train the neural network from the boundary measurements to learn a DOT reconstruction. In this work, the reduced scattering coefficient () is assumed to be spatially constant and known and we recover only the absorption coefficient (). Therefore, for the network training, boundary measurements (i.e., amplitude) are regarded as the input vector, which are generated by solving the forward model using open source software Nirfast,40 and the ground true distribution of absorption coefficient is served as the expected output. During the training process, the error between the predicted output and the true output is backward propagated from the output layer to the hidden layer to adjust the weights and biases in the opposite direction to the signal flow with respect to each individual weight.32 By repeatedly applying this procedure for each sample in the training set, the learning process can converge. The BPNN-based reconstruction algorithm can be described as follows:36
In our algorithm, the Tansig function is adopted as the activation function. Its formula is given in Eq. (9): and its derivative is shown in Eq. (10):2.3.Evaluation MetricsThe performance of the proposed algorithm is accessed with four evaluation metrics, including the absolute bias error (ABE), mean square error (MSE),41,42 peak signal-to-noise ratio (PSNR),43 and structural similarity index (SSIM).44 These parameters are defined as follows: where and are the true and reconstructed absorption coefficients at the finite node , respectively; and are the mean and standard derivation for the ground true () or reconstructed () absorption coefficients, respectively; is the covariance between the ground true and the reconstructed absorption coefficients, and are stabilization constants used to prevent division by a small denominator;44 is the above mentioned number of finite-element nodes. The ABE and MSE are used to compare the accuracy of reconstructed images. The PSNR (unit: dB) is used to compare the restoration of the images, without depending strongly on the image intensity scaling.15 SSIM is used to measure the similarity between the true and the reconstructed images, and an SSIM value of 1.0 refers to identical images. We expect lower ABE and MSE, while higher PSNR and SSIM, which show better performance.3.Results3.1.Data PreparationA 2-D circular phantom with a diameter of 80 mm was used to generate dataset. It was discretized into 2001 finite-element nodes and 3867 triangular elements. The absorption coefficient () and the reduced scattering coefficient () of the phantom were and , respectively. A total of 16 sources and 16 detectors were uniformly arranged along the circumference of the phantom. For each source illumination, data were collected at the remaining 15 detector locations, thus leading to a total of 240 () measurements. To generate simulation datasets, the phantom includes circular inclusions associated with varied sizes, locations, and absorption coefficients. Initially, an inclusion with the diameter of 6, 8, or 10 mm was randomly placed at different locations over the phantom. In this case, the absorption coefficients of the inclusions were varied from 0.015 to , leading to 17075 geometries. Next, two inclusions, which have the same radius of 8 mm, were placed at different edge-to-edge distances over the phantom. In this case, the phantoms were assigned different absorption coefficients (0.015, 0.02, 0.04, 0.06, or ) to each of its inclusions, leading to 5015 geometries. In all cases, the reduced scattering coefficients of inclusions were assumed to be , which were the same as those of the background. Examples of geometries of generating data are shown in Fig. 2. Software Nirfast was used to generate the simulation data,40 and 2% random Gaussian noise was added to the measurement data. A total of 22,090 samples which are data pairs contained input data and desired output data were obtained, and were separated into training, validation, and testing datasets. About 20,000 samples were used for training, 1045 for validation, and 1045 for testing. In the following experiments, the neural network has three layers: an input layer (240 neurons), a fully connected hidden layer, and an output layer (2001 neurons). To determine the number of neurons in the hidden layer, an empirical formula that has been introduced in Ref. 45 was used. The formula is given by where is the number of the hidden layer units, and are the numbers of input neurons and output neurons. In our experiments, the values of and are 240 and 2001, respectively. According to Eq. (16), the calculated number of neurons in hidden layer is 694.3. In our experiments, the number of neurons in the hidden layer was set to 695. The learning rate , the stopping threshold , the maximum number of iteration and the threshold were set to be 10, , 50,000 and , respectively. BPNN was trained for 16,000 epochs to minimize the MSE between the true and the recovered images. It took about 26 h to train the BP neural network. The experiments ran on a personal computer with Intel Core i7 CPU at 2.8 GHz and 8 GB RAM.For the purpose of comparison, we also performed Tikhonov regularization-based DOT reconstruction based on the Nirfast software.40 Tikhonov regularization-based DOT reconstruction is achieved using a least squares (LS) minimization technique, which is solved in the Levenberg–Marquardt procedure.40 The objective function in the Tikhonov regularization-based reconstruction algorithm typically consists of a data fidelity term of weighted LS and a regularization term of norm, balanced by a regularization parameter . For at the ’th iteration, it was setting as , where is the Jacobian matrix at the ’th iteration. The stopping criterion for Tikhonov regularization is defined such that the algorithm stops when the change in the difference between the forward data and the reconstructed data of two successive iterations is less than 2% or the maximum number of iteration (50) is reached. The initial regularization parameter is set to be 10. More detailed information about Tikhonov regularization-based DOT reconstruction can be found in Ref. 40. 3.2.Experimental ResultsIn this subsection, we provide numerical simulations to illustrate recovered results using BPNN and compare it with widely used Tikhonov regularization-based reconstruction method. Figure 3 shows some examples of recovered absorption coefficient using the two algorithms in the case of single inclusion. In Fig. 3, the size, the location, and the absorption coefficient of inclusions are varied. The ground true images are shown in the top row of Fig. 3, the reconstructed images using Tikhonov regularization and BPNN are given in the second and the third rows of Fig. 3, respectively. The corresponding cross-section profiles through their centers of the inclusions and along the axis are plotted in the last row of Fig. 3. From the last row of Fig. 3, we can see that the maximum values of recovered using Tikhonov regularization are much higher compared to their ground truths, the recovered using BPNN matched with their true values. The quantitative comparisons for the five cases in Fig. 3 are listed in Table 1. Compared with Tikhonov regularization, the values of ABE and MSE obtained using BPNN are significantly reduced, and the values of PSNR and SSIM are greatly improved. As an example, the values of ABE and MSE in Fig. 3(a) are reduced by 80% and 61%, respectively, which show that BPNN can yield higher reconstruction accuracy. It is evident that BPNN provides high-quality images with less artifacts in the background than those of Tikhonov regularization. Therefore, BPNN can have a PSNR gain of about 4.4 dB over Tikhonov regularization while the higher value of SSIM is obtained. The value of SSIM is improved by 424%, which indicates that the recovered image is nearly the same with the ground truth image. The similar results can also be observed in other images of Fig. 3. These results show that BPNN outperforms Tikhonov regularization in terms of higher accuracy and better image quality. In addition, the results also show that the image quality obtained by Tikhonov regularization is improved with increment of size and absorption coefficient of inclusions. By contrast, BPNN can always obtain robust reconstruction results. Table 1The quantitative comparisons presented in Fig. 3.
Figure 4 shows the capability of BPNN to recover images with two inclusions. And the corresponding quantitative results are compiled in Table 2. The two inclusions are observable and reconstructed with their centers at the correct positions for the two algorithms. But it can be seen that the shapes and the edges of the inclusions can be clearly observed by BPNN even when the edge-to-edge distance is 1.3 mm. However, the absorption coefficients of the inclusions far away from the boundary were underestimated and the images were somehow distorted for the Tikhonov regularization-based reconstruction method. For each evaluation metric, we draw a bar plot of test images in Fig. 5. As the higher accuracy of BPNN can be clearly observed in Fig. 5 and Table 2, we do not compare the cross-sections of different images here. Similar to the previous simulation results of one inclusion, BPNN reduces the artifacts and has offered more than 55.0%, 46.6%, and 16.7% improvement in ABE, MSE, and SSIM, respectively, compared with Tikhonov regularization-based reconstruction algorithm. The PSNRs in Figs. 4(b) and 4(c) are slightly higher than those of the Tikhonov regularization-based reconstruction algorithm because the absorption coefficients reconstructed by Tikhonov regularization have been overestimated with peak values of 0.03 and , respectively. By contrast, the peak values obtained by BPNN are 0.02 and , which are the same as the true targeted values. For Fig. 4, BPNN can have an average PSNR gain of about 2.56 dB over Tikhonov regularization. Overall, BPNN can obtain better performance. Table 2The quantitative comparisons in the case of two inclusions presented in Fig. 4.
Here, we reported the results of 1045 samples, which was randomly selected from the dataset to further evaluate the performance of our algorithm. We use the mean and standard deviation (SD) of ABE, MSE, PSNR, and SSIM to evaluate the performances of the two algorithms. The boxplots for the statistical results are presented in Fig. 6. And the corresponding quantitative comparisons are shown in Table 3. Table 3Mean ± SD of ABE/MSE/PSNR/SSIM for N=1045.
*Significant values are marked. For BPNN, the ABE had a mean value (SD) of (); for Tikhonov regularization, the ABE had a mean value (SD) of (). The accuracy improvement is 77.3% compared with Tikhonov regularization. The MSE of BPNN has an average of , while the average for Tikhonov regularization is . The improvement in the average MSE of BPNN with respect to Tikhonov regularization is 74.0%. Furthermore, the average PSNR of BPNN is improved from 24.34 to 27.79 dB compared with Tikhonov regularization. The average value of SSIM (0.91) obtained by BPNN is significant as compared to the average value of SSIM (0.46) produced by Tikhonov regularization. For every one of these metrics, BPNN performs better than Tikhonov regularization being compared. A student’s -test is used to determine whether there is statistical significance of the improvement between the evaluation metrics of Tikhonov regularization and BPNN. Significance is achieved at the 95% confidence interval using a two-tailed distribution. The corresponding -values are also listed in Table 3. Statistically significant differences ( values ) were found in the four evaluation metrics for BPNN versus Tikhonov regularization results, which further confirms that BPNN achieves more stable and effective performance than Tikhonov regularization. 4.Discussion and ConclusionIterative reconstruction algorithms with regularization have become the dominant approach for solving DOT inverse problem over the past few decades. However, it remains difficult to provide high-quality images. In this study, we explored using a BPNN to recover optical properties to improve the reconstruction accuracy and image quality of DOT. This was motivated by the fact that popular Tikhonov regularization-based reconstruction algorithms tend to produce oversmoothed images, which leads to poor reconstruction accuracy and bad image quality. The superior performance of the proposed algorithm was presented with numerical simulation experiments. Our results indicate that significant improvements of accuracy and image quality can be achieved by the proposed algorithm when compared with the Tikhonov regularization-based algorithm. Qualitative analysis demonstrated that our method can outperform Tikhonov regularization up to 77.3%, 74.0%, 14.2%, and 97.8% in terms of ABE, MSE, PSNR, and SSIM, respectively. Furthermore, we compared the reconstructed results of the proposed method to those of the and TV regularized reconstruction algorithms. The examples of reconstructed images are shown in Fig. 7. For the regularized reconstruction algorithm, it was solved with the GPSR algorithm.46 For the TV regularized reconstruction algorithm, it was solved by the Split Bregman algorithm.47 As for the regularization parameters used in the two reconstruction algorithms, their values were the same and were set to 0.01, which was tuned manually to get the best performance. In the case of single inclusion, the - or TV-based reconstruction algorithm can obtain better images than the Tikhonov regularization-based reconstruction algorithm in terms of less artifacts. In the case of two inclusions, decreasing the edge-to-edge distance between inclusions leads to degraded image quality for the - and TV-based reconstruction algorithms. When the edge-to-edge distance of inclusions decreased to 4.5 mm, the two inclusions could not be discriminated for the TV-based reconstruction algorithm. Compared to regularization-based reconstruction algorithms (, and TV), our algorithm performs the best in reconstructing DOT images. It is possible to obtain high-quality DOT images by training a neural network, even when iterative reconstruction algorithms underperform. Compared to iterative reconstruction algorithms, BPNN-based reconstruction algorithm is capable of: (1) improving the reconstruction accuracy with relatively stable performance; (2) enhancing the image quality with fewer image artifacts in the background; and (3) improving the spatial resolution. Currently, computational speed is still an open active research area in DOT. To the best of our knowledge, the widely used iterative reconstruction algorithms in DOT require to solve the forward model and to calculate the Jacobian matrix at each iteration.40 Therefore, the computational speed is relatively slow. For example, the computational time for each reconstruction in this study is about 1 to 2 min for Tikhonov regularization. For this reason, an algorithm that can fast reconstruct the distribution of optical properties is preferred. Although it takes a rather longer time to train the neuron network for the proposed algorithm, the training is implemented off-line. Once the training is finished, the time for the reconstruction is in a few seconds, which is practically useful for in vivo data. The effects of different activation functions, including ReLU, Sigmoid, and Tansig, were also examined. The representative results are shown in Fig. 8. Figure 8 shows that the distribution of absorption properties cannot be accurately reconstructed when using either ReLU or sigmoid as the activation function. The mean values of SSIM were 0.02 and 0.03 when using ReLU and sigmoid as the activation function, respectively. There are significant differences between the reconstructed and true images. By contrast, better images can be achieved with the activation function of Tansig and the mean SSIM is 0.91. However, it is still unclear why the activation function of Tansig works for DOT reconstruction. How to determine the number of neurons in the hidden layer needs to be investigated. Except for Eq. (16), one of the typical equations which can be used to determine the number of neurons in the hidden layer is ,48 where is the number of input neurons. In our experiments, is 240. Therefore, is about 8. We trained our network with eight hidden neurons. Using the trained network, DOT image reconstruction was performed. The examples of reconstructed images are shown in Fig. 9. Figure 9 shows that the distribution of absorption coefficient cannot be accurately reconstructed. The reason is that using too few neurons in the hidden layer will result in underfitting. The effects of the learning rate on the performance of BPNN were also analyzed. The learning rate was set to 0.01, 0.1, 1, 10, or 50. The optimal learning rate was the one with the lowest validation loss. In the experiments, the activation function was fixed as Tansig. When the learning rate was set to 50, the validation loss fluctuated over epochs. Therefore, the learning rate of 50 was discarded. The validation losses were 0.15, 0.09, 0.04, and 0.02, corresponding to the learning rates, 0.01, 0.1, 1, and 10, respectively. When learning rate was 10, the lowest validation loss was obtained. We also note that a higher learning rate effectively speeds up the convergence for the training procedure. Therefore, the learning rate was set to 10 in our experiments. Note that several recent publications have applied various deep learning architectures for solving inverse problems.49–51 For example, a recent work by Sun et al. has shown that a convolutional neural network can be applied to perform image reconstruction under multiple scattering problem in diffraction tomography.49 However, our focus is on diffuse optical tomography. Recently, an artificial neural network-based approach is developed to estimate the inclusion location, then the estimated inclusion location is used as a-priori knowledge in DOT reconstruction.50 Therefore, the approach is not to reconstruct DOT images directly by learning an artificial neural network. Deep learning technique has been recently applied to reconstruct DOT images, and its superiority has been indicated by comparing with an analytic technique.51 The average value of SSIM they reported is 0.46, which is relatively low. By contrast, a higher value of SSIM is obtained by our algorithm, and its value is 0.91. A future study would be to compare the performances of the two algorithms in the same datasets. Although the proposed BPNN method has achieved promising results, our work could still be considered as a preliminary attempt of applying neural network in DOT; its application in DOT is still very challenging. The performance of BPNN depends on the training data; however, it would be difficult to acquire a sufficient number of real data for training in patient studies. An available strategy is to create training data pairs from breast geometries with known optical properties and the real data is for evaluation. Breast geometries can be obtained from breast MRI images. Since breast geometries have different sizes and shapes, the finite-element meshes generated from breast geometries are different, which leads to different BPNN architectures. To deal with the problem, pixel basis provides a solution.40 Each finite-element mesh is mapped to the same pixel basis for network training. Therefore, the trained BPNN will be a universal network. Certain challenges remain which are the subject of further study, including the effect of heterogeneity in the background region as well as performance evaluation using clinical patient data. AcknowledgmentsThis work has been sponsored by the National Science Foundation of China under Grant No. 81871394, the Beijing Municipal Education Committee Science Foundation under Grant No. KM201810005030, and Beijing Laboratory of Advanced Information Networks under Grant No. 040000546618017. ReferencesJ. Feng et al.,
“Addition of T2-guided optical tomography improves noncontrast breast magnetic resonance imaging diagnosis,”
Breast Cancer Res., 19 117
(2017). https://doi.org/10.1186/s13058-017-0902-x BCTRD6 Google Scholar
Q. Zhu et al.,
“Assessment of functional differences inmalignant and benign breast lesions and improvement of diagnostic accuracy by using US-guided diffuse optical tomography in conjunction with conventional US,”
Radiology, 280
(2), 387
–397
(2016). https://doi.org/10.1148/radiol.2016151097 RADLAX 0033-8419 Google Scholar
C. Xu et al.,
“Ultrasound-guided diffuse optical tomography for predicting and monitoring neoadjuvant chemotherapy of breast cancers: recent progress,”
Ultrason. Imaging, 38
(1), 5
–18
(2016). https://doi.org/10.1177/0161734615580280 ULIMD4 0161-7346 Google Scholar
V. Ntziachristos et al.,
“Concurrent MRI and diffuse optical tomography of breast after indocyanine green enhancement,”
Proc. Natl. Acad. Sci. U. S. A., 97
(6), 2767
–2772
(2000). https://doi.org/10.1073/pnas.040570597 Google Scholar
B. Brooksby et al.,
“Imaging breast adipose and fibroglandular tissue molecular signatures by using hybrid MRI-guided near-infrared spectral tomography,”
Proc. Natl. Acad. Sci. U. S. A., 103
(23), 8828
–8833
(2005). https://doi.org/10.1073/pnas.0509636103 Google Scholar
M. A. Mastanduno et al.,
“Sensitivity of MRI-guided near-infrared spectroscopy clinical breast exam data and its impact on diagnostic performance,”
Biomed. Opt. Express, 5 3103
–3115
(2014). https://doi.org/10.1364/BOE.5.003103 BOEICL 2156-7085 Google Scholar
S. Jiang et al.,
“Predicting breast tumor response to neoadjuvant chemotherapy with diffuse optical spectroscopic tomography prior to treatment,”
Clin. Cancer Res., 20
(23), 6006
–6015
(2014). https://doi.org/10.1158/1078-0432.CCR-14-1415 Google Scholar
M. Mastanduno et al.,
“MR-guided near-infrared spectral tomography increases diagnostic performance of breast MRI,”
Clin. Cancer Res., 21
(17), 3906
–3912
(2015). https://doi.org/10.1158/1078-0432.CCR-14-2546 Google Scholar
V. Ntziachristos et al.,
“MRI-guided diffuse optical spectroscopy of malignant and benign breast lesions,”
Neoplasia, 4 347
–354
(2002). https://doi.org/10.1038/sj.neo.7900244 Google Scholar
C. Chen et al.,
“Diffuse optical tomography enhanced by clustered sparsity for functional brain imaging,”
IEEE Trans. Med. Imaging, 33
(12), 2323
–2331
(2014). https://doi.org/10.1109/TMI.2014.2338214 ITMID4 0278-0062 Google Scholar
M. A. Franceschini and D. A. Boas,
“Noninvasive measurement of neuronal activity with near-infrared optical imaging,”
Neuroimage, 21
(1), 372
–386
(2004). https://doi.org/10.1016/j.neuroimage.2003.09.040 NEIMEF 1053-8119 Google Scholar
C. W. Lee, R. J. Cooper and T. Austin,
“Diffuse optical tomography to investigate the newborn brain,”
Pediatr. Res., 82 376
–386
(2017). https://doi.org/10.1038/pr.2017.107 PEREBL 0031-3998 Google Scholar
S. R. Arridge,
“Optical tomography in medical imaging,”
Inverse Prob., 15
(2), R41
–R93
(1999). https://doi.org/10.1088/0266-5611/15/2/022 INPEEY 0266-5611 Google Scholar
S. R. Arridge and J. C. Schotland,
“Optical tomography: forward and inverse problems,”
Inverse Probl., 25
(12), 123010
(2009). https://doi.org/10.1088/0266-5611/25/12/123010 INPEEY 0266-5611 Google Scholar
J. Feng et al.,
“Impact of weighting function on reconstructed breast images using direct regularization imaging method for near-infrared spectral tomography,”
Biomed. Opt. Express, 9
(7), 3266
–3283
(2018). https://doi.org/10.1364/BOE.9.003266 BOEICL 2156-7085 Google Scholar
R. Baikejiang, W. Zhang and C. Li,
“Diffuse optical tomography for breast cancer imaging guided by computed tomography: a feasibility study,”
J. X-Ray Sci. Technol., 25 341
–355
(2017). https://doi.org/10.3233/XST-16183 JXSTE5 0895-3996 Google Scholar
J. Feng et al.,
“Multiobjective guided priors improve the accuracy of near-infrared spectral tomography for breast imaging,”
J. Biomed. Opt., 21
(9), 090506
(2016). https://doi.org/10.1117/1.JBO.21.9.090506 JBOPFO 1083-3668 Google Scholar
L. Zhang et al.,
“Direct regularization from co-registered anatomical images for MRI-guided near-infrared spectral tomographic image reconstruction,”
Biomed. Opt. Express, 6 3618
–3630
(2015). https://doi.org/10.1364/BOE.6.003618 BOEICL 2156-7085 Google Scholar
Q. Fang et al.,
“Combined optical and X-ray tomosynthesis breast imaging,”
Radiology, 258
(1), 89
–97
(2011). https://doi.org/10.1148/radiol.10082176 RADLAX 0033-8419 Google Scholar
P. K. Yalavarthy et al.,
“Structural information within regularization matrices improves near infrared diffuse optical tomography,”
Opt. Express, 15 8043
–8058
(2007). https://doi.org/10.1364/OE.15.008043 OPEXFF 1094-4087 Google Scholar
Z. Yuan and H. Jiang,
“Image reconstruction scheme that combines modified Newton method and efficient initial guess estimation for optical tomography of finger joints,”
Appl. Opt., 46
(14), 2757
–2768
(2007). https://doi.org/10.1364/AO.46.002757 APOPAI 0003-6935 Google Scholar
P. K. Yalavarthy et al.,
“Weight-matrix structured regularization provides optimal generalized least-squares estimate in diffuse optical tomography,”
Med. Phys., 34
(6), 2085
–2098
(2007). https://doi.org/10.1118/1.2733803 MPHYA6 0094-2405 Google Scholar
N. Cao, A. Nehorai and M. Jacobs,
“Image reconstruction for diffuse optical tomography using sparsity regularization and expectation-maximization algorithm,”
Opt. Express, 15
(21), 13695
–13708
(2007). https://doi.org/10.1364/OE.15.013695 OPEXFF 1094-4087 Google Scholar
A. P. Gibson, J. C. Hebden and S. R. Arridge,
“Recent advances in diffuse optical imaging,”
Phys. Med. Biol., 50 R1
–R43
(2005). https://doi.org/10.1088/0031-9155/50/4/R01 PHMBA7 0031-9155 Google Scholar
J. Prakash et al.,
“Sparse recovery methods hold promise for diffuse optical tomographic image reconstruction,”
IEEE J. Sel. Top. Quantum Electron., 20
(2), 74
–82
(2014). https://doi.org/10.1109/JSTQE.2013.2278218 Google Scholar
J. Prakash et al.,
“Model-resolution based basis pursuit deconvolution improves diffuse optical tomographic imaging,”
IEEE Trans. Med. Imaging, 33
(4), 891
–901
(2014). https://doi.org/10.1109/TMI.2013.2297691 ITMID4 0278-0062 Google Scholar
Y. Zhao et al.,
“Optimization of image reconstruction for magnetic resonance imaging-guided near-infrared diffuse optical spectroscopy in breast,”
J. Biomed. Opt., 20
(5), 056009
(2015). https://doi.org/10.1117/1.JBO.20.5.056009 JBOPFO 1083-3668 Google Scholar
X. Yang et al.,
“Low-dose x-ray tomography through a deep convolutional neural network,”
Sci. Rep., 8 2575
(2018). https://doi.org/10.1038/s41598-018-19426-7 SRCEC3 2045-2322 Google Scholar
D. M. Pelt and K. J. Batenburg,
“Fast tomographic reconstruction from limited data using artificial neural networks,”
IEEE Trans. Image Process., 22
(12), 5238
–5251
(2013). https://doi.org/10.1109/TIP.2013.2283142 IIPRE4 1057-7149 Google Scholar
G. Yang et al.,
“DAGAN: deep de-aliasing generative adversarial networks for fast compressed sensing MRI reconstruction,”
IEEE Trans. Image Process., 37
(6), 1310
–1321
(2018). https://doi.org/10.1109/TMI.2017.2785879 IIPRE4 1057-7149 Google Scholar
B. Yang, L. Ying and J. Tang,
“Artificial neural network enhanced Bayesian PET image reconstruction,”
IEEE Trans. Image Process., 37
(6), 1297
–1309
(2018). https://doi.org/10.1109/TMI.2018.2803681 IIPRE4 1057-7149 Google Scholar
C. Lv et al.,
“Levenberg-Marquardt backpropagation training of multilayer neural networks for state estimation of a safety critical cyber-physical system,”
IEEE Trans. Ind. Inf., 14
(8), 3436
–3446
(2018). https://doi.org/10.1109/TII.2017.2777460 Google Scholar
D. E. Rumelhart, G. E. Hinton and R. J. Williams,
“Learning representations by back-propagating errors,”
Nature, 323
(6088), 533
–536
(1986). https://doi.org/10.1038/323533a0 Google Scholar
P. D. Wasserman, Neural Computing: Theory and Practice, Van Nostrand Reinhold, New York
(1989). Google Scholar
R. Hecht-Nielsen,
“Theory of the backpropagation neural network,”
in Int. Joint Conf. on Neural Networks,
593
–605
(1989). https://doi.org/10.1109/IJCNN.1989.118638 Google Scholar
S. Haykin, Neural Networks, Macmillan College Publishing Company, Inc., New York
(1994). Google Scholar
I. A Basheer and M Hajmeer,
“Artificial neural networks: fundamentals, computing, design, and application,”
J. Microbiol. Methods, 43
(1), 3
–31
(2000). https://doi.org/10.1016/S0167-7012(00)00201-3 JMIMDQ 0167-7012 Google Scholar
F. Amato et al.,
“Artificial neural networks in medical diagnosis,”
J. Appl. Biomed., 11 47
–58
(2013). https://doi.org/10.2478/v10136-012-0031-x Google Scholar
E. Kuriscak et al.,
“Biological context of Hebb learning in artificial neural networks, a review,”
Neurocomputing, 152 27
–35
(2015). https://doi.org/10.1016/j.neucom.2014.11.022 NRCGEO 0925-2312 Google Scholar
H. Dehghani et al.,
“Near infrared optical tomography using NIRFAST: algorithm for numerical model and image reconstruction,”
Commun. Num. Methods Eng., 25
(6), 711
–732
(2009). https://doi.org/10.1002/cnm.v25:6 Google Scholar
B. W. Pogue et al.,
“Statistical analysis of nonlinearly reconstructed near-infrared tomographic images: part I-theory and simulations,”
IEEE Trans. Med. Imaging, 21
(7), 755
–763
(2002). https://doi.org/10.1109/TMI.2002.801155 ITMID4 0278-0062 Google Scholar
X. Song et al.,
“Statistical analysis of nonlinearly reconstructed near-infrared tomographic images: part II-experimental interpretation,”
IEEE Trans. Med. Imaging, 21
(7), 764
–772
(2002). https://doi.org/10.1109/TMI.2002.801158 ITMID4 0278-0062 Google Scholar
A. P. Cuadros and G. R. Arce,
“Coded aperture optimization in compressive X-ray tomography: a gradient descent approach,”
Opt. Express, 25
(20), 23833
–23849
(2017). https://doi.org/10.1364/OE.25.023833 OPEXFF 1094-4087 Google Scholar
Z. Wang et al.,
“Image quality assessment: from error visibility to structural similarity,”
IEEE Trans. Image Process., 13
(4), 600
–612
(2004). https://doi.org/10.1109/TIP.2003.819861 IIPRE4 1057-7149 Google Scholar
D. Gao and S. Wu,
“An optimization method for the topological structures of feed-forward multi-layer neural networks,”
Pattern Recognit., 31
(9), 1337
–1342
(1998). https://doi.org/10.1016/S0031-3203(97)00160-X Google Scholar
M. Figueiredo, R. Nowak and S. Wright,
“Gradient projection for sparse reconstruction: application to compressed sensing and other inverse problems,”
IEEE J. Sel. Top. Signal Process., 1
(4), 586
–597
(2007). https://doi.org/10.1109/JSTSP.2007.910281 Google Scholar
J. Feng et al.,
“Total variation regularization for bioluminescence tomography with the split Bregman method,”
Appl. Opt., 51
(19), 4501
–4512
(2012). https://doi.org/10.1364/AO.51.004501 APOPAI 0003-6935 Google Scholar
R. P. Gorman and T. J. Sejnowski,
“Analysis of hidden units in a layered network trained to classify sonar target,”
Neural Networks, 1 75
–89
(1988). https://doi.org/10.1016/0893-6080(88)90023-8 NNETEB 0893-6080 Google Scholar
Y. Sun, Z. Xia and U. S. Kamilov,
“Efficient and accurate inversion of multiple scattering with deep learning,”
Opt. Express, 26
(11), 14678
–14688
(2018). https://doi.org/10.1364/OE.26.014678 OPEXFF 1094-4087 Google Scholar
R. Patra and P. K. Dutta,
“Improved dot reconstruction by estimating the inclusion location using artificial neural network,”
Proc. SPIE, 8668 86684C
(2013). https://doi.org/10.1117/12.2007905 PSISDG 0277-786X Google Scholar
H. B. Yedder et al.,
“Deep learning based image reconstruction for diffuse optical tomography,”
in Int. Workshop on Machine Learning for Medical Image Reconstruction,
112
–119
(2018). Google Scholar
|