14 December 2015 Reconstruction-classification method for quantitative photoacoustic tomography
Author Affiliations +
J. of Biomedical Optics, 20(12), 126004 (2015). doi:10.1117/1.JBO.20.12.126004
Abstract
We propose a combined reconstruction-classification method for simultaneously recovering absorption and scattering in turbid media from images of absorbed optical energy. This method exploits knowledge that optical parameters are determined by a limited number of classes to iteratively improve their estimate. Numerical experiments show that the proposed approach allows for accurate recovery of absorption and scattering in two and three dimensions, and delivers superior image quality with respect to traditional reconstruction-only approaches.
Malone, Powell, Cox, and Arridge: Reconstruction-classification method for quantitative photoacoustic tomography

1.

Introduction

Photoacoustic tomography (PAT) is an emerging technique for in vivo imaging of soft biological tissue.1 This hybrid modality uses ultrasound to detect optical contrast, combining the high resolution of acoustic methods with the spectroscopic capability of optical imaging. To generate a PA image, a short laser pulse is shined into the object, the ultrasonic waves emitted following the heating of the tissue are measured, and an image of the absorbed optical energy field is recovered. Whereas purely optical methods suffer from poor spatial resolution, acoustic waves propagate with minimal scattering, and PAT can achieve 100μm resolution at depths of several centimeters. However, PA images provide only qualitative information about the tissue, and are not directly related to tissue morphology and functionality. The principal difficulty is that the PA image is the product of both the optical absorption coefficient (which is directly related to underlying tissue composition) and the light distribution (which is not). This severely restricts the range of applications for which PAT is suitable.

Quantitative photoacoustic tomography (QPAT) aims to provide clinically valuable images of the optical absorption and scattering coefficients, or chromophore (light-absorbing molecules) concentrations from conventional PA images via an image reconstruction method.2,3 A model of light propagation is required to relate the absorbed optical energy to the light fluence and tissue parameters. The primary challenge of QPAT is solving the nonlinear imaging problem. In particular, recovering the scattering coefficient is especially difficult due to the weak dependence of the absorbed energy density on scattering.

In this paper, we develop a method for solving the image reconstruction problem for QPAT by alternating reconstruction and segmentation steps in an automated iterative process. We introduce a probabilistic model that describes optical properties in terms of a limited number of optically distinct classes, which may correspond to tissues or chromophores. These are identified and characterized by a classification, or segmentation, algorithm. This approach allows for the use of information retrieved by the classification in the reconstruction stage and vice versa. The aim of the reconstruction is to choose solutions for which the image parameters take values close to a finite set of discrete points. The aim of the classification algorithm is to progressively improve the parametric optical model and correct for errors in the initial assumptions. Multinomial models have been employed previously in the related fields diffuse optical tomography4 and electrical impedance tomography.5 For QPAT, the main advantage is that this approach enables accurate recovery of both the absorption and scattering coefficients simultaneously.

2.

Numerical Methods

2.1.

Quantitative Photoacoustic Imaging

A conventional PAT image is proportional to the absorbed optical energy

(1)

H(r)=Γ^(r)μa(r)φ[μa(r),μs(r)]rΩ,
where r is a position vector within the domain Ω, μa and μs are the optical absorption and reduced scattering coefficients, φ is the optical fluence, and Γ^ is the Grüneisen parameter. The Grüneisen parameter represents the efficiency with which the tissue converts heat into acoustic pressure, and is often taken to be the constant Γ^(r)=1 rΩ. The fluence is dependent on the optical parameters and illumination pattern in the whole domain. The problem of recovering the optical parameters (μa,μs) from a conventional PAT image is known as the “quantitative” problem. The optical absorption μa is of particular interest because it is fundamentally related to underlying tissue physiology and functionality, and encodes clinically useful information such as tissue oxygenation levels and chromophore concentrations. Conversely, the absorbed energy density H depends nontrivially on optical absorption and thus is not directly related to tissue morphology because it is distorted structurally and spectrally by the nonuniform light fluence.

2.2.

Diffusion Model of Light Transport

In order to recover the optical parameters (μa,μs), a model of light propagation within the tissue is required. For highly scattering media and those far from boundaries and sources, a low-order spherical harmonic approximation to the “radiative transfer equation” is suitable. The “diffusion approximation” is given by6

(2)

[μa·κ(r)]φ(r)=q(r),
where q(r) is an isotropic source term and κ=1/3μs is the diffusion coefficient.

We set Robin boundary conditions

(3)

φ(r)+12Aκ(r)n^·φ(r)=0rδΩ,
where A accounts for the refractive index mismatch at the boundary.

2.3.

Minimization-Based Quantitative Photoacoustic Tomography Imaging

In this paper, we adopt a gradient-based minimization approach to image reconstruction. Typically, both μa and μs are unknown and need to be recovered simultaneously from the absorbed energy density. An objective function is defined, which measures the distance between the conventional PAT image Hm and the data predicted by the model for the current estimates H(μa,μs)

(4)

E=12Ω[HmH(μa,μs)]2dΩ.
In order to treat the problem for a generic geometry, the finite element method is employed, whereby a weak formulation of the diffusion approximation [Eq. (2)] is considered. A discretization of the domain is defined, and the fluence and optical parameters are expressed in terms of the same piecewise-linear basis functions ui(r): χiχiui(r) for χ{μa,μs,φ}, where χi are nodal coefficients and i=1,,N.

We assume that the data dm is the absorbed energy density Hm, projected onto a particular basis {Ψj},

(5)

dm={djm,j=1,,N},djm=ΩHm(r)Ψj(r)dΩ=Ψj,Hm.
Choices for {Ψj} include:

  • 1. Point sampling Ψj(r)=δ(rrj),

  • 2. Piecewise-linear sampling Ψj=uj,

  • 3. Sinc sampling Ψj=sinc(|rrj|).

Substituting into the objective function [Eq. (4)] leads to the discrete form of the objective function

(6)

E=12j[djmΨj,H(μa,μs)]2=12j[djmΨj,μaφ]2.

If a single illumination source is used and both absorption and scattering are undetermined, the problem is ill posed.2 In this study, the nonuniqueness of the solution was removed by using multiple illumination patterns,78.9 thus the objective function must be summed over the number of sources. In the following, we have omitted this sum for ease of notation. Prior information regarding the solution can be included by adding a regularization term

(7)

E=12j(djmΨj,μaφ)2+R(μa,μs).
In the Bayesian framework, an image is obtained by maximizing the posterior probability of the parameters, given the data

(8)

p(μa,μs|dm)p(dm|μa,μs)p(μa,μs).
Under this interpretation, the regularization term R is given by the negative log of the prior probability distribution

(9)

R(μa,μs)=logp(μa,μs).

2.4.

Gradient Calculations

Cox et al.10 have shown that, for the continuous case, the gradient of Eq. (4) with respect to μa at position r0 is given by

(10)

Eμa|r0=φ(HmH)|r0+φ·φ*|r0,
where the “adjoint” light field φ* is the solution to the equation

(11)

[μa·κ(r)]φ*(r)=μa(HmH).

In the following, we derive the expression for the gradient in the discrete case. The sampled forward model can be expressed as a vector H={Hj,j=1,,N}

(12)

Hj=ΩH(r)Ψj(r)dΩ=Ψj,H.=ikμaiφkΩΨj(r)ui(r)uk(r)dΩ=φTCjμa,
where Cj is a sparse matrix indexed by i, k where the support of the basis functions Ψj(r), ui(r), uk(r) overlap. Taking the derivative of Eq. (6) with respect to μai, we have

(13)

Eμai=j(Hjμai)(djmHj).
Using the expression for the absorbed energy density [Eq. (12)],

(14)

Hjμai=eiTCjφ+μaTCjφμai,
where ei is a vector of zeros with a single 1 in position i. Substituting into Eq. (13) gives

(15)

Eμai=j(eiTCjφ+μaTCjφμai)(djmHj).
The first term in Eq. (15) is

(16)

jeiTCjφ(djmHj)=j,i,keiCikjφk(djmHj)=j,kφk(djmHj)ΩΨj(r)ui(r)uk(r)dΩ=φTEi(dmH),
where Ei is given by a reordering of Cikj

(17)

Ekji=ΩΨj(r)ui(r)uk(r)dΩ.
Note that while Cj is symmetric, in general, Ei is not.

It remains to determine φ/μai. The discrete form of the DA model [Eq. (2)] assumes the form11

(18)

(M+K+F)φ=Q,
where

(19)

Mjk=iμaiΩuiujukdΩ,

(20)

Kjk=iκiΩuiuj·ukdΩ,

(21)

Fjk=i12AΩujukdS,

(22)

Qj=iqiΩuiujdΩ.
Taking the derivative of Eq. (18) with respect to the i’th coefficient of μa,

(23)

(M+K+F)φμai=Vμaiφ,
where

(24)

Vμa,jki=ΩuiujukdΩ
is given by the derivative of the system matrix. We define the adjoint field φ* as the solution to the equation

(25)

(M+K+F)φ*=Q*,
where

(26)

Q*=jμaTCj(djmHj)
is the adjoint source. Taking φ* Equation (23) (φ/μai). Eq. (25), we obtain

(27)

jμaTCjφμai(djmHj)=φTVμaiφ*.
Substituting into Eq. (15) gives the expression for the derivative with respect to μai

(28)

Eμai=φT[Vμaiφ*Ei(dmH)].
The derivative with respect to μsi can be derived analogously

(29)

Eμsi=κiμsiφTVμsiφ*,
where

(30)

Vμs,jki=Ωuiuj·ukdΩ
and (κi/μsi)=1/3μsi2. Note that calculation of the gradient requires only two runs of the forward model. The forward problem was solved using the Toast++ software package.11

Choosing point-sampling Ψj(r)=δ(rrj) gives simply Cj=Ei=I. In this study, we chose piecewise-linear sampling Ψj=uj, so we had Cj=Ei=Vμai and

(31)

Eμai=φTVμai(φ*dm+H).

3.

Reconstruction-Classification Method for Quantitative Photoacoustic Tomography

A reconstruction-classification scheme is devised, which enables the recovery μa and μs by approaching the image reconstruction and segmentation problems simultaneously. At each reconstruction step, we minimize a regularized objective function, where the regularization term is given by a mixture model. At each classification step, the result of the previous reconstruction step is employed to update the class parameters for the multinomial model. We alternate between reconstruction and classification steps for a fixed number of iterations (Fig. 1).

3.1.

Mixture Model for μa and μs

In this section, we introduce a probability model for μa and μs, which encodes prior knowledge about the optical parameters and allows us to bias the solution of the imaging problem accordingly. We assume that an array of labels ζi can be determined for each node, such that

(32)

ζij={1iftheithnode is assigned to thejthclass,0otherwise.
The labels constitute “hidden variables” on which the image parameters are dependent. For each class j=1,,J, a mean vector mj=(μ¯aj,μ¯sj)R2 is defined, and the covariance of each class is described by matrix ΣjR2×2.

We assume that if ζij=1, the probability distribution for xi=(μai,μsi) is given by a multivariate Gaussian distribution

(33)

p(xi|θj)=N(mj,Σj),
where θj indicates the set of class parameters (mj,Σj).

The prior probability distribution of the class properties θj is given by the conjugate prior to the Gaussian distribution. Prior information about the distribution of the class means or covariances can be encoded by choosing the parameters of the conjugate prior accordingly. Using a noninformative prior for the class means we have p(mj)1. The conjugate prior distribution for the covariance of a normal distribution is given by the normal inverse Wishart distribution (NIW):

(34)

NIW(νj,Γj)=|Σj|(ν+d+1)/2exp[12Tr(ΓjΣj1)],
where d is the dimension of the domain, νj indicates the number of degrees of freedom, and Γj is a scaling matrix. If the prior is noninformative, then νj=0 and Γj=0, and the probability distribution of the class parameters becomes

(35)

p(θj)|Σj|(d+1)/2,
which is known as Jeffreys prior.

The probability that the set of labels ζi={ζi1,,ζij,,ζiJ} is assigned to the i’th node is given by a multinomial distribution

(36)

p(ζi|λ)=jλjζij,
where λj is the overall probability that a node is assigned to the j’th class. Therefore, the joint probability for (xi,ζi) is given by the product

(37)

p(xi,ζi|θ,λ)=p(xi|ζi,θ)p(ζi|λ)=j[λjp(xi|θj)]ζij.
By marginalizing over all possible values of the indicator variables ζij, a “mixture of Gaussians” model for the optical parameters is obtained

(38)

p(xi|θ,λ)=ζip(xi,ζi|θ,λ)dζi=jλjp(xi|θj).
Finally, for independent nodes, the prior of the image is given by

(39)

p(x|θ,λ)=ijλjp(xi|θj).

3.1.1.

Reconstruction step

The objective function takes the form of Eq. (7), where at iteration t of the reconstruction-classification algorithm, the regularization is given by Eqs. (9) and (39)

(40)

Rt(μa,μs)=logp(x|θt,λt)=logN(x¯,Σx¯)=τ2Lx¯(xx¯)2,
where τ is a regularization parameter and

(41)

x¯i=jζij·mj|MAP(ζ)=mjR2
is obtained by fixing the labels to the “maximum a posteriori” estimate, given the results of the previous iteration

(42)

MAP(ζ)=argmaxζp(ζ|xt1,θt1,λt1),
which is calculated in the classification step (see Sec. 3.1.2). The weighting matrix Lx¯ is the Cholesky decomposition of Σx¯1, where Σx¯R2N×2N is a sparse matrix of which the i’th 2×2 block along the diagonal is Σj if the i’th element belongs to the j’th class.

In order to sphere the solution space, that is, to render the space dimensionless, we performed a change of variables μaμa/μa0 and μsμa/μs0, where (μa0,μs0) is the initial guess for the optical parameters (in this study, we initialized to the homogeneous background). Given the size of the problem, we chose a gradient-based optimization method in order to reduce memory use and computational expense.12 The minimization was performed using the limited-memory Broyden–Fletcher–Goldfarb–Shanno (L-BFGS) method,13 with a storage memory of six iterations.

3.1.2.

Classification

The purpose of the classification step is to update the multinomial model using the result of the previous reconstruction step. First, the expected values of the labels ζt+1 are computed for the current class parameters (θt,λt) and image xt=(μat,μst) (E-step). Then the model parameters are updated by maximizing the posterior probability (M-step)

(43)

p(θ,λ|xt)p(xt|θ,λ)p(θ,λ).

  • E-step: The “responsibility” rijt is a measure of the probability that the i’th node is assigned to the j’th class. Using Bayes’ theorem and the Gaussian mixture model [Eq. (38)], we have

    (44)

    p(ζij=1|xit,θt,λt)=p(xi|ζij=1,θt)p(ζij=1)p(xi|θ,λ)=λjtp(xit|θjt)jλjtp(xit|θjt)=rnjt.
    The expectation for the indicator values is

    (45)

    E(ζij|xit,θt,λt)=ζijp(ζij=1|xit,θt,λt)dζij=0×p(ζij=0|xit,θt,λt)+1×p(ζij=1|xit,θt,λt)=rijt.
    Therefore, the MAP estimate for the labels is

    (46)

    ζijt+1={1ifrijtismaximumj,0otherwise,
    which can be used in Eq. (42).

  • M-step: The parameters (θ,λ) are chosen in order to maximize the log posterior

    (47)

    (θt+1,λt+1)=argmax(θ,λ)logp(xt|θ,λ)+logp(θ,λ).
    Averaging over all possible values of ζ gives

    (48)

    logp(xt|θ,λ)+logp(θ,λ)=ζlogp(xt,ζ|θ,λ)dζ+logp(θ,λ).
    Using “Jensen’s inequality”14 and ignoring terms that do not depend on (θ,λ), we obtain a lower bound for the log prior

    (49)

    B(θ,λ)=ijrijtlog[λjp(σn|θj)]+logp(λ)+logp(θ)=ijrijt[log(λj)+log(|Σj|)12(xi(n)mj)Σj1(xi(n)mj)]+j[(αj1)log(λj)νj+d+12log|Σj|].
    Maximizing B(θ,λ) for jλj=1 and using noninformative priors, we obtain the update rules for the model parameters

    (50)

    λjt+1=irijtN,

    (51)

    mjt+1=irijtxiirijt,

    (52)

    Σjt+1=irijt(ximj)(ximj)T+Γjirijt+νj+d+1.

3.2.

Class Means Initialization

The number of classes J and the class means mj were initialized by automatically segmenting the result of the first reconstruction step and averaging over the segmented areas. To segment the image [e.g., see Fig. 2(a)], we looked at a binned histogram of the image of μa and chose the value μah for which the number of occurrences was highest [Fig. 2(c), column 1]. We found the first node index h for which the value μah occurs, and identified the corresponding scattering value μsh. Having chosen a covariance matrix Σh, we computed a map of the multivariate normal probability of the (μa,μs) images, with mean (μah,μsh) [Fig. 2(c), column 2]. A suitable choice for Σh is the initial covariance of the classes. Then we selected a tolerance level tolh at which to truncate the probability map, and selected all nodes with probability higher than the tolerance as belonging to the same class as node h [Fig. 2(c), column 3]. We repeated this process on the remaining nodes until all nodes were classified. Thus, the number of classes was set to the number of iterations, and the average of the optical parameters over each class was used to initialize the class means [Fig. 2(b)].

Fig. 1

Reconstruction-classification algorithm outline.

JBO_20_12_126004_f001.png

Fig. 2

Class initialization example: (a) original image of μa to which we apply the segmentation; (b) result of taking average image values over the segmented areas; (c) first column, histogram of occurrences of values of μa in the portion of the image requiring segmentation—value with highest number of occurrences is μah (indicated by a red cross); second column, probability density function with mean (μah,μsh) and covariance Σh; third column, labels identifying nodes with probability density higher than tolerance value tolh; each row corresponds to an iteration and a distinct class, so in this case, J=3.

JBO_20_12_126004_f002.png

3.3.

Visualization of the Results

Results obtained using the reconstruction-classification method are displayed alongside scatter plots of the nodal values recovered in the two-dimensional (2-D) feature space (μa,μs) [e.g., see Fig. 2(c), final column in 4]. The positions of the class means mj=(μ¯aj,μ¯sj) are identified by a cross, and the class covariances Σj are represented by ellipses. These are color-coded by class, and are indicative of the clustering of image nodal values around the class means.

4.

Results

4.1.

Two-Dimensional Validation and Reconstruction

We chose a numerical phantom defined on a 2-D circular mesh with 1331 nodes and radius 25 mm. Four illumination sources were placed on the boundary at angles 0, π/2, π, and 3π/2rad. In all cases, the illumination profile was a normalized Gaussian with radius (distance from the center at which the profile drops to 1/e) 6 mm. The background optical parameters were set to μa=0.01mm1 and μs=1mm1. Two circular perturbations of radius 6 mm were added in positions (6 mm, 10 mm) and (6mm,10mm) [Fig. 3(a)]. The values of the perturbations were μa=0.02mm1, μs=1.5mm1 and μa=0.03mm1, μs=1.25mm1, respectively. The absorbed energy field was simulated for each illumination, and 1% white Gaussian noise was added [Fig. 3(b)]. The class covariances were initialized to

(53)

Σj=(10600101)j=1,,3,
where the first variable was the absorption and the second was the reduced scattering. The parameters of the Jeffreys prior were set to Γj=Σj j, ν(1)=1 for the background class and ν(2,3)=10 for the perturbation classes. The number of classes and optical parameters was initialized using the class means initialization method (Sec. 3.2) with tolh=105 and Σh=Σj [Eq. (53)], and the labels were initialized to 1 for the background class and zero for all other classes. The tolerance of the L-BFGS algorithm was set to tol=1011, and the total number of reconstruction-classification iterations was set to MaxIt=10 (Fig. 4). The regularization parameter τ=1010 was chosen by inspection. For comparison, images were reconstructed without introducing a prior (Fig. 5); the images were reconstructed by minimizing Eq. (6) using the L-BFGS method with tol=1012.

Fig. 3

Two-dimensional (2-D) model: (a) circular mesh and (b) absorbed energy for each illumination pattern.

JBO_20_12_126004_f003.png

Fig. 4

2-D reconstruction-classification results at iteration 1 (first row), 5 (second row), and 10 (third row). Reconstructed values of μa and μs (first and second columns), labels recovered for perturbation classes (third and fourth columns), and scatter plot (fifth column).

JBO_20_12_126004_f004.png

Fig. 5

2-D model and reconstruction: first column, model of μs and μs; second column, reconstructed values of μa and μs without multinomial prior; third column, reconstructed values of μa and μs with multinomial prior.

JBO_20_12_126004_f005.png

4.2.

Three-Dimensional Validation and Reconstruction

We chose a three-dimensional (3-D) phantom analogous to the 2-D case, defined on a cylinder with 27,084 nodes, radius 25 mm, and height 25 mm. Two spherical inclusions of radius 6 mm were placed in (6, 10, and 0 mm) and (6, 10, and 0 mm) [Fig. 6(a)]. Illumination sources were Gaussian in the xy-plane constant in the z-axis, with radius 6 mm and length 25 mm [Figs. 6(b) and 6(c)]. PAT images were simulated for four illuminations at the cardinal points, and 1% noise was added to the absorbed energy [Fig. 6(d)]. The optical, covariance, and reconstruction parameters were set to the same values used in the 2-D case. The class initialization parameters were set to tolh=107 and Σh=Σj. Images were reconstructed by performing 10 iterations of the reconstruction-classification method (Fig. 7).

Fig. 6

Three-dimensional (3-D) model: (a) numerical phantom and perturbation locations, (b) all illumination sources, (c) cross section of optical parameters used to simulate the data for z=0, (d) cross section of absorbed energy for each illumination pattern.

JBO_20_12_126004_f006.png

Fig. 7

3-D reconstruction-classification results at iteration 1 (first row), 5 (second row), and 10 (third row). Reconstructed values of μa and μs (first and second columns), labels recovered for perturbation classes (third column), and scatter plot (fourth column).

JBO_20_12_126004_f007.png

5.

Discussion

5.1.

Summary of Findings

We applied the proposed reconstruction-classification algorithm to a 2-D numerical phantom with three tissues, a background, and two perturbations (Fig. 3). The optical absorption was recovered reliably within a small number of iterations, and the scattering was recovered with sufficient accuracy after approximately 10 iterations (Fig. 4). We compared the optical model with images obtained by the reconstruction-classification method and by a traditional reconstruction-only (no regularization) method (Fig. 5). We found that the reconstruction-classification method delivered superior image quality, particularly with regards to the scattering parameter. We applied the reconstruction-classification algorithm to a much larger 3-D problem (Fig. 6) and observed similar results (Fig. 7) as in the 2-D case.

5.2.

Choice of Parameters

The parametric optical model and classification algorithm introduce a number of parameters that require tuning by the user. In addition to the regularization parameter, the parameters of the Jeffreys prior Γ and ν and the initial guess of the class variances Σj must be set before performing the classification. However, their significance is fairly intuitive, and with experience of a certain type of problem, the choice of parameters becomes natural. Visualizing the class covariance matrix Σj as an ellipse, changing the value of Γ varies its eccentricity, and changing ν varies the length of its axes. Further, given that in the first iteration the optical absorption is recovered with superior accuracy to the scattering, it is preferable to initialize the variance of the former to a smaller value than the latter, indicating greater confidence in the imaging solution.

5.3.

Initialization of the Class Means

The purpose of the means initialization scheme is to increase automation of the method so that minimum user intervention and no prior knowledge of the number of tissues or their optical properties is required. The algorithm simply performs a segmentation of the image, then takes averages over the segmented areas to initialize the class properties (Fig. 1). Alternative segmentation techniques could have been employed; however, the advantage of the proposed approach is that it directly exploits the mixture of Gaussians model to identify the tissues. Our choice to investigate a node h with μa belonging to the bin with a maximum number of occurrences leads to the background tissue being identified first, followed by the perturbation tissues. The choice of the node index h could have been randomized so that tissues would be identified in random order. This approach is equally valid; however, we found that in cases where tissue values were close together (such as after a single reconstruction-classification iteration), it was preferable to identify the largest classes first because the mean was estimated with greater accuracy for the classes with a larger number of samples. Further, for a given image and tolerance level, our choice renders the result of the segmentation process unique and reproducible.

5.4.

Recovery of the Scattering

From the comparison with the reconstruction-only case with no regularization (Fig. 5), it is evident that the introduction of the parametric prior enables better recovery of the scattering. The inconsistency between the quality of the recovered absorption and scattering parameters in the nonregularized case is due to the weaker dependence of the latter on the absorbed energy density with respect to the former. This results in the scattering gradient being approximately an order of magnitude smaller than the absorption gradient. Although the problem can be mitigated by sphering the solution space, variations in the data due to the scattering often fall below the noise floor. In the reconstruction-classification case, typically the absorption is recovered with good accuracy within a small number of iterations. Thus, the absorption takes values very close to the class means (resulting in small clusters), and the variance along the μa direction converges to a small value. Given that the regularization term is weighted by the inverse of the covariance matrix, the dependence of the absorption gradient on the data becomes weaker at each iteration, until its magnitude is comparable or smaller to that of the scattering. In the iterations that follow, the descent of the data term of the objective function is primarily due to updates to the scattering, which converges to the correct values.

5.5.

Computational Demands

Computational performance was found to be strongly dependent on the problem size. In the 2-D case with 1331 nodes (Fig. 4), the total reconstruction time (10 outer reconstruction-classification iterations) using MATLAB on a 16-processor PC with 128 GB RAM was only 77 s. In the 3-D case with 27,084 nodes (Fig. 7), the total reconstruction time increased linearly with the number of nodes and was approximately 3.7 h on the same workstation. The increase in computation time was mostly due to much longer processing times for the L-BFGS algorithm in the reconstruction step.

5.6.

Experimental Application

In experimental situations, prior information on tissue properties may be held, such as knowledge of the characteristic optical absorption and scattering spectra of chromophores of interest. These may be obtained from the literature15 or gained through tissue sample measurements. This information could be used in one of two ways. First, a library of typical chromophores could be used to initialize the class parameters instead of the proposed class means initialization method. The classification process could then perform the function of correcting for uncertainty, errors, or local variations in the real optical properties with respect to the prior information. Alternatively, it could be used to label the chromophores found by the segmentation process and identify these as certain tissues such as, e.g., “oxygenated blood” or “fat,” on the basis of the closeness of the recovered means to the characteristic properties.

5.7.

Additional Priors

In this study, we assumed independence between nodal values; however, the mixture of Gaussian models could be used in conjunction with a spatial prior. Knowledge of smoothness or sparsity properties of the solution could be employed to introduce a homogeneous spatial regularizer such as first-order Tikhonov16 or total variation.7,17 Knowledge of structural information, such as that provided by an alternative imaging method or anatomical library, could be exploited by introducing a spatially varying probability map for the optical properties.

6.

Conclusions

In this paper, we proposed a method for performing image reconstruction in QPAT. We introduced a parametric class model for the optical parameters and implemented a minimization-based reconstruction algorithm. We suggested an automated method by which to initialize the parameters of the class model and proposed a classification algorithm by which to progressively update and improve those parameters after each reconstruction step. We demonstrated though 2-D and 3-D numerical examples that the reconstruction-classification method allows for the simultaneous recovery of optical absorption and scattering. In particular, we found that this approach delivered superior accuracy in the recovery of the scattering with respect to traditional gradient-based reconstruction.

Acknowledgments

This work was funded by the EPSRC Doctoral Prize Fellowship EP/M506448/1.

References

1. 

P. Beard, “Biomedical photoacoustic imaging,” Interface Focus 1, 602–631 (2011).http://dx.doi.org/10.1098/rsfs.2011.0028Google Scholar

2. 

B. Cox et al., “Quantitative spectroscopic photoacoustic imaging: a review,” J. Biomed. Opt. 17, 061202 (2012).JBOPFO1083-3668http://dx.doi.org/10.1117/1.JBO.17.6.061202Google Scholar

3. 

H. Gao, S. Osher and H. Zhao, “Quantitative photoacoustic tomography,” Lect. Notes Math. 2035, 131–158 (2012).http://dx.doi.org/10.1007/978-3-642-22990-9_5Google Scholar

4. 

P. Hiltunen, S. J. D. Prince and S. Arridge, “A combined reconstruction-classification method for diffuse optical tomography,” Phys. Med. Biol. 54, 6457–6476 (2009).PHMBA70031-9155http://dx.doi.org/10.1088/0031-9155/54/21/002Google Scholar

5. 

E. Malone et al., “A reconstruction-classification method for multifrequency electrical impedance tomography,” IEEE Trans. Med. Imaging 34(7), 1486–1497 (2015).ITMID40278-0062http://dx.doi.org/10.1109/TMI.2015.2402661Google Scholar

6. 

S. Arridge, “Optical tomography in medical imaging,” Inverse Prob. 15, R41 (1999).INPEEY0266-5611http://dx.doi.org/10.1088/0266-5611/15/2/022Google Scholar

7. 

G. Bal and K. Ren, “Multiple-source quantitative photoacoustic tomography in a diffusive regime,” Inverse Prob. 27(7), (2011).INPEEY0266-5611http://dx.doi.org/10.1088/0266-5611/27/7/075003Google Scholar

8. 

P. Shao, B. Cox and R. J. Zemp, “Estimating optical absorption, scattering, and Grueneisen distributions with multiple-illumination photoacoustic tomography,” Appl. Opt. 50(19), 3145–3154 (2011).APOPAI0003-6935http://dx.doi.org/10.1364/AO.50.003145Google Scholar

9. 

H. Gao, J. Feng and L. Song, “Limited-view multi-source quantitative photoacoustic tomography,” Inverse Prob. 31(6), 065004 (2015).INPEEY0266-5611http://dx.doi.org/10.1088/0266-5611/31/6/065004Google Scholar

10. 

B. T. Cox, S. R. Arridge and P. C. Beard, “Gradient-based quantitative photoacoustic image reconstruction for molecular imaging,” Proc. SPIE 6437, 64371T (2007).PSISDG0277-786Xhttp://dx.doi.org/10.1117/12.700031Google Scholar

11. 

M. Schweiger and S. Arridge, “The Toast++ software suite for forward and inverse modeling in optical tomography,” J. Biomed. Opt. 19, 040801 (2014).JBOPFO1083-3668http://dx.doi.org/10.1117/1.JBO.19.4.040801Google Scholar

12. 

T. Saratoon et al., “3D quantitative photoacoustic tomography using the δ-Eddington approximation,” Proc. SPIE 8581 85810V (2013).PSISDG0277-786Xhttp://dx.doi.org/10.1117/12.2004105Google Scholar

13. 

J. Nocedal and S. Wright, “Numerical optimization,” in Springer Series in Operations Research and Financial Engineering, Springer-Verlag, New York (1999).Google Scholar

14. 

S. Prince, Computer Vision: Models, Learning, and Inference, Cambridge University Press, Cambridge, United Kingdom (2012).Google Scholar

15. 

S. L. Jacques, “Optical properties of biological tissues: a review,” Phys. Med. Biol. 58, 5007–5008 (2013).PHMBA70031-9155http://dx.doi.org/10.1088/0031-9155/58/14/5007Google Scholar

16. 

T. Saratoon et al., “A gradient-based method for quantitative photoacoustic tomography using the radiative transfer equation,” Inverse Prob. 29, 075006 (2013).INPEEY0266-5611http://dx.doi.org/10.1088/0266-5611/29/7/075006Google Scholar

17. 

T. Tarvainen et al., “Reconstructing absorption and scattering distributions in quantitative photoacoustic tomography,” Inverse Prob. 28, 084009 (2012).INPEEY0266-5611http://dx.doi.org/10.1088/0266-5611/28/8/084009Google Scholar

Biographies for the authors are not available.

Emma Malone, Samuel Powell, Ben T. Cox, Simon Arridge, "Reconstruction-classification method for quantitative photoacoustic tomography," Journal of Biomedical Optics 20(12), 126004 (14 December 2015). http://dx.doi.org/10.1117/1.JBO.20.12.126004
Submission: Received ; Accepted
JOURNAL ARTICLE
10 PAGES


SHARE
KEYWORDS
Scattering

Tissue optics

Absorption

Photoacoustic tomography

Image segmentation

Geometrical optics

Reconstruction algorithms

Back to Top