Open Access
1 September 2007 Standardized platform for coregistration of nonconcurrent diffuse optical and magnetic resonance breast images obtained in different geometries
Author Affiliations +
Abstract
We present a novel methodology for combining breast image data obtained at different times, in different geometries, and by different techniques. We combine data based on diffuse optical tomography (DOT) and magnetic resonance imaging (MRI). The software platform integrates advanced multimodal registration and segmentation algorithms, requires minimal user experience, and employs computationally efficient techniques. The resulting superposed 3-D tomographs facilitate tissue analyses based on structural and functional data derived from both modalities, and readily permit enhancement of DOT data reconstruction using MRI-derived a-priori structural information. We demonstrate the multimodal registration method using a simulated phantom, and we present initial patient studies that confirm that tumorous regions in a patient breast found by both imaging modalities exhibit significantly higher total hemoglobin concentration (THC) than surrounding normal tissues. The average THC in the tumorous regions is one to three standard deviations larger than the overall breast average THC for all patients.

1.

Introduction

Near-infrared (NIR) diffuse optical tomography (DOT) relies on functional processes, and provides unique measurable parameters with potential to enhance breast tumor detection sensitivity and specificity. For example, several groups have demonstrated the feasibility of breast tumor characterization based on total hemoglobin concentration, blood oxygen saturation, water and lipid concentration and scattering. 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17

The functional information derived with DOT is complementary to structural and functional information available to conventional imaging modalities such as magnetic resonance imaging (MRI), x-ray mammography, and ultrasound. Thus the combination of functional data from DOT with structural/anatomical data from other imaging modalities holds potential for enhancing tumor detection sensitivity and specificity. To achieve this goal of data fusion, two general approaches can be employed. The first, concurrent imaging, physically integrates the DOT system into the conventional imaging instrument. This approach derives images in the same geometry and at the same time. The second approach, nonconcurrent imaging, employs optimized stand-alone DOT devices to produce 3-D images that must then be combined with those of the conventional imaging modalities via software techniques. In this case, the images are obtained at different times and often in different geometries.

Thus far a few DOT systems have been physically integrated into conventional imaging modalities such as MRI,18, 19, 20, 21, 22 x-ray mammography,5 and ultrasound9 for concurrent measurements. By doing so, however, these DOT systems are often limited by the requirements of the “other” imaging modality, for example, restrictions on metallic instrumentation for MRI, hard breast compression for x-ray mammography, limited optode combinations for ultrasound (and MRI, x-ray), and time constraints. On the other hand, among the stand-alone DOT systems available today, only a few attempts have been made to quantitatively compare DOT images of the same breast cancer patient to those of other imaging modalities9, 22 obtained at different times, because the nonconcurrent coregistration problem presents many challenges. It is therefore desirable to develop quantitative and systematic methods for data fusion that utilize the high-quality data and versatility of the stand-alone imaging systems.

In this work, we introduce and demonstrate a novel software platform for combining nonconcurrent MRI and DOT: the optical and multimodal imaging platform for research assessment and diagnosis (OMIRAD). To our knowledge, OMIRAD is the first software prototype platform capable of fusing and jointly analyzing multimodal optical imaging data with x-ray tomosynthesis and MR images of the breast.23, 24, 25, 26, 27, 28 The OMIRAD platform enables multimodal 3-D image visualization and manipulation of datasets based on a variety of 3-D rendering techniques. Through its ability to simultaneously control multiple fields of view, OMIRAD can streamline quantitative analyses of structural and functional data. OMIRAD is the result of four years of work to develop and test a prototype platform specifically designed for multimodal optical data visualization, fusion, and analysis, including the ability to share data and analysis results across several institutions. It is possible that other postprocessing software exists, but to our knowledge, none of it has been integrated into a full-fledged software prototype, nor tested within several leading institutions in optical imaging research such as within the Network for Translational Research in Optical Imaging (NTROI). Our preliminary study takes an important step toward improved diagnosis and treatment of breast cancer patients with DOT and MRI. Coregistration, facilitated by the software platform, combines structural and functional data from multiple modalities. Segmentation and fusion will also enable a-priori structural information derived from MRI to be incorporated into the DOT reconstruction algorithms. The combined MRI/DOT dataset provides information in a more useful format than the sum of the individual datasets, and we expect the platform to have substantial impact on the standardization of diffuse optical imaging systems, and therefore on the translation of optical imaging research prototypes into viable clinical systems.

The work is organized as follows. We first summarize the structure of the DOT system. We then present the multimodal software platform developed to enable nonconcurrent multimodal data analysis. We describe procedures used for image coregistration and segmentation, and we demonstrate the approach using a simulated phantom breast model. Last, we apply the platform to coregister DOT and MRI images of three breast cancer patients.

2.

Diffuse Optical Tomography System Overview

A schematic of the DOT instrument is shown in Fig. 1 .29 This hybrid continuous-wave (cw) and frequency-domain (FD) parallel-plane DOT system has been extensively characterized for breast cancer imaging using tissue phantoms and normal breast.30 The breast is softly compressed between the source plate and a viewing window, to a thickness of 5.5to7.5cm . The breast box is filled with a matching fluid (i.e., Intralipid and indian ink) that has optical properties similar to human tissue. Four laser diodes (690-, 750-, 786-, and 830-nm wavelength), amplitude modulated at 70MHz , were used as light sources. We use a grid of 9×5=45 source positions with a spacing of 1.6cm . For (cw) transmission detection, we sample a 24×41=984 grid of pixels from the charge-coupled device (CCD), which corresponds to a detector separation of about 3mm on the detection window. For remission detection (FD), we use a 3×3=9 grid ( 1.6-cm spacing) of detector fibers located on the source plate. Remission detection is used to determine the average optical properties of the breast. These values are used as an initial guess for the nonlinear image reconstruction. The CCD data are used for the image reconstruction. For each source position and wavelength, FD measurements were obtained via nine detector fibers on the source plate, and cw measurements were obtained simultaneously via CCD camera in transmission. The amplitude and phase information obtained from the FD measurements are used to quantify bulk optical properties, and the cw transmission data are used to reconstruct a 3-D tomography of optical properties within the breast.

Fig. 1

Schematic of the parallel plate DOT instrument with patient subject lying in the prone position (refer to Ref. 29 for the technical details).

051902_1_023705jbo1.jpg

To reconstruct the absorption and scattering image, an inverse problem associated with the photon diffusion equation is solved by iterative gradient-based optimization.31 The algorithm reconstructs chromophore concentrations (CHb,CHbO2) and scattering coefficients directly using data from all wavelengths simultaneously. We use a variation of the open-source software package Time-resolved optical absorption and Scattering tomography (TOAST) for these reconstructions. TOAST determines the optical properties inside a medium by adjusting these parameters, such that the difference between the modeled and experimental light measurements at the sample surface is minimized.32 Images of physiologically relevant variables, such as total hemoglobin concentration (THC), blood oxygenation saturation (StO2) , and scattering are thus obtained. Figure 2 shows an example of a 3-D distribution of THC (μM) in a patient breast with an invasive ductal carcinoma.

Fig. 2

3-D distribution of THC ( μM concentration) in a patient breast containing an invasive ductal carcinoma. Consecutive 2-D patient slices are adjacent.

051902_1_023705jbo2.jpg

The resulting DOT dataset is a finite element (FE) model containing on average 50,000 nodes and 200,000 tetrahedral elements. Each node is associated with the reconstructed physiological values such as THC and StO2 . To facilitate registration of DOT and MR images, the FE model is automatically resampled into a 3-D voxelized volume. The smallest bounding box surrounding the FE model is identified; this volume is divided into voxels ( 1283 by default). Every voxel is associated to the tetrahedral element to which it belongs, and finally, using the element’s shape functions, the correct physiological value is interpolated at the location of the voxel.

3.

Software Platform Description

The OMIRAD platform enables multimodal integration and visualization of data from DOT and MRI. Figure 3 describes a typical workflow that a user can follow:

  • 1. Input: the software platform accepts two types of data formats. 1. For MRI datasets: Digital Imaging and Communications in Medicine (DICOM), the widely accepted format.25 2. For DOT datasets: TOAST, developed at University College London,30 and Near Infrared Frequency Domain Absorption and Scatter Tomography (NIRFAST), developed at Dartmouth College (Hanover, New Hampshire),33 two popular formats used in the DOT image reconstruction community. Datasets are converted into a common binary format through a user-friendly interface. Then a patient browser (in the import module shown in Fig. 3) allows the user to select any two 3-D datasets for visualization and/or registration.

  • 2. Visualization: the visualization stage permits the user to inspect each dataset, both through volume rendering and multiplanar reformatting (MPR) visualization, and to define the volume of interest (VOI) through morphological operations such as punching. Punching involves determining a 3-D region of an object from the 2-D region specified on the orthographic projection of the same object. This 3-D region can then be removed or retained. This type of operation enables an easy editing of 3-D structures. This is a particularly important stage, as the user removes parts of the data that should not be used in the registration process.

  • 3. Segmentation: the breast MR image segmentation technique enables a-priori structural information derived from MRI to be incorporated into the reconstruction of DOT data (details are given in the next section).

  • 4. Registration: the user may decide to roughly align one volume to the other, before starting the automatic registration procedure (details are given in the next section).

  • 5. Analysis: once the registration is completed, several tools are available to the user for assessment of the results, including fused synchronized MPR and volume manipulation.

The visualization platform showing same patient MRI and DOT (blood volume) datasets before registration is exhibited in Fig. 4 . After the appropriate color transfer functions are applied, one can clearly observe the location of the invasive ductal carcinoma diagnosed in this patient breast.

Fig. 4

Visualization platform showing MRI and DOT (blood volume) datasets from the same patient before registration. After the appropriate transfer functions are applied, one can clearly observe the location of the invasive ductal carcinoma diagnosed in this patient breast. The following components are shown from left to right: 1. orientation cube, 2. transfer function editors, 3. data attribute windows, 4. volume rendering window, 5. MPR windows, and 6. command tabs.

051902_1_023705jbo4.jpg

4.

Three-Dimensional/Three Dimensional Diffuse Optical Tomography to Magnetic Resonance Imaging Image Registration Algorithm

3-D DOT/3-D MRI image registration presents several new challenges. Because registration of DOT to MR acquired nonconcurrently has not been extensively studied, to our knowledge no standard approach has been established for this problem. In this work, we present a new approach, specifically devised for addressing the challenges of DOT/MR image registration. DOT images have much lower anatomical resolution and contrast than MRI, and the optical reconstruction process typically uses a geometric model of the breast. In our case, the constraining geometric model of the breast is a semiellipsoid. At the University of Pennsylvania, the patient breast is compressed axially in the DOT imaging device and sagitally in the MRI machine, and, of course, the breast is a highly deformable organ.

For this task, we require that registration be automatic with little prior user interaction and be robust enough to handle the majority of patient cases. In addition, the process should be computationally efficient for applicability in practice, and yield results useful for combined MRI/DOT analysis.

At this time, automatic image registration is an essential component in medical imaging systems. The basic goal of intensity-based image registration techniques is to align anatomical structures in different modalities. This is done through an optimization process, which assesses image similarity and iteratively changes the transformation of one image with respect to the other, until an optimal alignment is found.34 Computation speed is a critical issue and dictates applicability of the technology in practice. Although feature-based methods are computationally more efficient, they are notoriously dependent on the quality of the extracted features from the images.35

In intensity-based registration, volumes are directly aligned by iteratively computing a volumetric similarity measure based on the voxel intensities. Since the amount of computations per iteration is high, the overall registration process is very slow. In the cases where mutual information (MI) is used, sparse sampling of volume intensity could reduce the computational complexity while compromising the accuracy.36, 37 In Ref. 38, a projection-based method for 2-D/2-D image registration is proposed. In this method, the projections along the two axes of the image are computed. Horizontal and vertical components of the shift are then computed using a 1-D cross-correlation-based estimator. They show the method is robust in the presence of temporal and spatial noise, and computationally efficient compared to the 2-D correlation-based shift estimator. In Ref. 39, the authors formulate 3-D/3-D registration cost function as the summation of three 2-D/3-D optimization cost functions. The optimization is then done concurrently on the sum of the cost functions, which are identically parameterized. Furthermore, images are preprocessed to extract a binary segmentation. Projection images from the binary segmentation are used for computing similarity measures. The key is to choose a well-behaved similarity measure that can robustly characterize a metric for the volumes.40 To make such an algorithm practical, the computational time must also be reduced. In Ref. 40, researchers suggest random sampling of the volume datasets and computation performance based only on these random samples to decrease the computational load. In Ref. 41, authors propose a hybrid technique, which selects a set of high-interest points (i.e., landmarks) within the volume and tries to do registration based on those points only.

Let us consider two datasets to be registered to each other. One dataset is considered the reference and is commonly referred to as the “fixed” dataset. The other dataset is the one onto which the registration transformation is applied. This dataset is commonly referred to as the “moving” dataset. Registration of volumetric datasets (i.e., fixed and moving) involves three steps: first, computation of the similarity measure quantifying a metric for comparing volumes; second, an optimization scheme that searches through the parameter space (e.g., 6-D rigid body motion) to maximize the similarity measure; and third, a volume warping method that applies the latest computed set of parameters to the original moving volume to bring it a step closer to the fixed volume.

Our proposed approach is a novel combination of the methods described in Refs. 42, 43: we compute 2-D projection images from the two volumes for various projection geometries, and set up a similarity measure with an optimization scheme that searches through the parameter space. These images are registered within a 2-D space, which is a subset of the 3-D space of the original registration transformations. Finally, we perform these registrations successively and iteratively to estimate all the registration parameters of the original problem.

We further optimize the performance of projection and 2-D/2-D registration similarity computation through the use of graphics processing units (GPU). Details and general validation of this novel approach have been recently presented.44 Multiple 2-D signatures (or projections) can represent the volume robustly depending on the way the signatures are generated. An easy way to understand the idea is to derive the motion of an object by looking at three perpendicular shadows of the object (see Fig. 5 ).

Fig. 5

Generation of 2-D signatures from 3-D volumes, in the three mutually orthogonal directions.

051902_1_023705jbo5.jpg

Figure 6 provides an illustration of different transformation models used in medical image registration: rigid, affine, and free-form transformations. Nonrigid registration, depending on complexity, can be classified in two ways: 1. affine transformations [see Fig. 6c], which include nonhomogeneous scaling and/or shearing; and 2. free-form transformations [see Fig. 6d], which include arbitrary deformations at the voxel level. These transformations can be based on intensity, shape, or material properties. The dominant transformation observed across the MR and DOT datasets is due to the difference in compression axis (lateral compression for MR versus axial compression for DOT); this transformation can be modeled using affine parameters. DOT images do not possess enough local structure information for computation of a free-form deformation mapping to register a DOT to an MR dataset.

Fig. 6

Illustration of different transformation models applied to a 2-D image: (a) original image, (b) rigid transformation, (c) affine transformation, and (d) free-form transformation.

051902_1_023705jbo6.jpg

Given the previous challenges, we used the following parameters in the nonrigid registration algorithm.

  • Projection images: we use maximum intensity projection (MIP) techniques. MIP is a computer visualization method for 3-D data that projects in the visualization plane the voxels with maximum intensity that fall in the way of parallel rays traced from the viewpoint to the plane of projection.

  • Projection geometries: we use three mutually orthogonal 2-D MIPs to achieve greater robustness in the registration algorithm.

  • Similarity measure: we use normalized mutual information.40 Mutual information measures the information that two random variables A and B share. It measures how knowledge of one variable reduces the uncertainty in the other. For example, if A and B are independent, then knowing A does not give any information about B and vice versa, so their normalized mutual information is zero. On the other hand, if A and B are identical, then all information given by A is shared with B; therefore, knowing A determines the value of B and vice versa, and the normalized mutual information is equal to its maximum possible value of 1. Mutual information quantifies the distance between the joint distribution of A and B from what it would be if A and B were independent. In our case, the moving dataset is deformed until the normalized mutual information between it and the fixed dataset is maximized.

  • Parameter space: we use rigid body motion (translation and rotation), and independent linear scaling in all three dimensions. This results in a 9-D parameter space enabling nonrigid registration: three parameters for translation in x , y , and z , three parameters for rotations about three axes, and three parameters for linear scaling in each of the x , y , and z directions.

Mathematically, the estimate of the nine degrees-of-freedom (DOF) homogeneous transformation matrix T9 is initially given by

Eq. 1

T9=argmaxT9S3[If,ΓT93(Im)],
where ΓT93 is the six DOF mapping operator, S3 estimates the similarity metric between two volumes, and If and Im are the fixed and moving volumetric data, respectively. Both ΓT93 and S3 have a superscript of 3 to indicate that the operations are applied in three dimensions. We can reformulate the registration optimization process so it can be applied to each of the 2-D signatures, or projections, using the five DOF homogeneous transformation matrix defined in the plane of projection TP5 . The five degrees of freedom in the plane of projection correspond to horizontal and vertical translation, horizontal and vertical scaling, and in-plane rotation. The estimate of the transformation matrix is given by:

Eq. 2

TP5=argmaxTP5S2{ΦP(If),ΓTP52[ΦP(Im)]},
where ΦP is an orthographic projection operator, which projects the volume points onto an image plane, P is a 4×4 homogeneous transformation matrix, which encodes the principal axis of the orthographic projection, ΓTP52 is a three DOF mapping operator, and S2 computes the similarity metric between 2-D projections. Here, ΓTP52 and S2 have a superscript of 2 to indicate that the operations are applied in two dimensions.

Since here the similarity metric is mutual information, i.e., S2h(A)+h(B)h(A,B) , Eq. 2 can be rewritten as:

Eq. 3

TP5=argmaxTP5[h(A)+h(B)h(A,B)],
where A=ΦP(If) , B=ΓTP52[ΦP(Im)] , h(x) is the entropy of a random variable x , and h(x,y) is the joint entropy of two random variables x and y .

Entropy is a measure of variability and is defined as: h(x)p(x)lnp(x)dx , and h(x,y)p(x,y)ln p(x,y)dxdy ,40 where p(x) is the probability density function (PDF) of variable x , and p(x,y) is the joint PDF of variables x and y . The entropy h is discretely computed as:

Eq. 4

H(II) =I=LHpII(I)logpII(I)andH(II,IJ) =I=LHJ=LHpII,IJ(I,J)logpII,IJ(I,J),
where II and IJ are two given images; and I and J are the intensities ranging from lower limit L (e.g., 0) to higher limit H (e.g., 255) for II and IJ , respectively. pII(I) is the PDF of image II , and pII,IJ(I,J) is the joint PDF of images II and IJ Here, a PDF is represented by a normalized image histogram.

The algorithm flowchart is shown in Fig. 7 . Figure 7a shows the global registration flowchart. For a number of iterations n (typically n=3 ), the three mutually orthogonal 2-D signatures are generated (sagittal, coronal, and axial) for both the fixed and moving volumes. After each 2-D signature generation, the moving 2-D signature is registered to the fixed 2-D signature. This process is shown schematically in Fig. 7b, and explained in detail next.

Fig. 7

Flowcharts for the registration process: (a) global registration flowchart, and (b) 2-D signatures registration flowchart.

051902_1_023705jbo7.jpg

First, the Δ variables are initialized:

Δscale=Δscale̱initial;Δtrans=Δtrans̱initial;Δrot=Δroṯinitial.

Then, step k=1 to m is as follows.

  • 1. Compute the deformation steps:

    Δscale=Δscaledivider;Δtrans=Δtransdivider;Δrot=Δrotdivider.

  • 2. Compute the initial similarity measure Sinitial2 between the two 2-D signatures.

  • 3. Scale moving volume vertically by ±Δscale , then estimate Sscale-vert2 . If an improvement has been made, i.e., Sscale-vert2> Sinitial2 , then go to next step, otherwise do not apply this scaling operation.

  • 4. Scale moving volume horizontally by ±Δscale , then estimate Sscale-horiz2 . If an improvement has been made, i.e., Sscale-horiz2> Sscale-vert2 , then go to next step, otherwise do not apply this scaling operation.

  • 5. Translate moving volume vertically by ±Δtrans , then estimate Strans-vert2 . If an improvement has been made, i.e., Strans-vert2> Sscale-vert2 , then go to next step, otherwise do not apply this translation operation.

  • 6. Translate moving volume horizontally by ±Δtrans , then estimate Strans-horiz2 . If an improvement has been made, i.e., Strans-horiz2> Strans-vert2 , then go to next step, otherwise do not apply this translation operation.

  • 7. Rotate moving volume in-plane by ±Δrot , then estimate Srot2 . If an improvement has been made, i.e., Srot2> Strans-horiz2 , then go to next step, otherwise do not apply this rotation operation.

  • 8. Convergence criteria: if 0<Srot2Sinitial2ΔS2 or divider> divideṟthreshold , then end k-loop.

  • 9. If no improvements have been made, i.e., Srot2=Sinitial2 , then decrease the deformation steps (i.e., divider=divider×2 ).

The variables are initialized at the beginning of the registration process and are typically set to the following: n=3 , m=40 , Δscale̱initial=4mm , Δtrans̱initial=4mm , Δroṯinitial=4deg , divideṟthreshold=40 .

5.

Breast Magnetic Resonance Imaging Image Segmentation

Our proposed segmentation approach is based on the random walker algorithm. In this case, the segmentation technique requires little user interaction and is computationally efficient for practical applications.

This algorithm, originally developed in Ref. 45 and extended Ref. 46 to incorporate intensity priors, can perform multilabel, semiautomated image segmentation. Given a small number of pixels with user-defined labels, one can analytically (and quickly) determine the probability that a random walker starting at each unlabeled pixel will first reach one of the prelabeled pixels. By assigning each pixel to the label for which the greatest probability is calculated, high-quality image segmentation can be obtained (Fig. 8 ). This algorithm is formulated in discrete space (i.e., on a graph) using combinatorial analogs of standard operators and principles from continuous potential theory, allowing it to be applied in arbitrary dimensions.

Fig. 8

Breast MRI 3-D image segmentation based on “random walkers:” (a) segmenting fatty from nonfatty tissue, and (b) segmenting tumor from nontumor tissue.

051902_1_023705jbo8.jpg

Usually T1 -weighted MR imaging is performed. Images show lipid as bright and parenchyma as dark. Tumor also tends to be dark. Minimal user initialization is required. We have developed a workflow customized for this type of breast MRI segmentation.

  • 1. Using a custom-made interactive visual interface, the user scrolls through axial, sagittal, and coronal views of the MRI dataset. In each view, the user selects one or two slices that best incorporate all tissue types.

  • 2. The user draws three types of seed points using a virtual “brush” on each of the selected slices to indicate different tissue types: fatty tissue, nonfatty tissue (parenchyma and/or tumor), and outside the breast.

  • 3. The algorithm generates a mask file representing the result of the segmentation. Each voxel in the generated mask is assigned a value (fatty, nonfatty, or outside), indicating the type of tissue.

  • 4. The segmented mask file can finally be incorporated in a more accurate reconstruction of physiological quantities (such as THC) to generate the DOT dataset.

The algorithm takes two minutes on average for a MRI volume of size 256×256×50 . This algorithm can be used to distinguish fatty from nonfatty tissue and tumor from nontumor tissue, as shown in Fig. 8. Since the objective of this initial study is to show how nonconcurrent data may be registered and jointly analyzed, we should note that we did not use the results of MRI segmentation to improve the DOT reconstructions. Rather, we used the MRI segmentation to isolate the tumor tissue in the image.

One significant advantage of spatially registering DOT to MRI data is the ability to treat anatomical information from MRI data as prior information in the DOT chromophore concentration and scattering variables reconstruction process. By segmenting fatty from nonfatty tissue in a MR dataset for example, we can provide a-priori data about the tissue, which interacts with light in a DOT imaging device. This information can further be incorporated in solving the inverse problem associated with the photon diffusion equation, and lead to a more precise reconstruction of physiological quantities (such as hemoglobin concentration). Note that in this work, the MR segmentation is not incorporated in the DOT reconstruction process. This approach is the subject of an ongoing study to be published in the near future.

6.

Initial Validation of Registration Using a Simulated Phantom Model

6.1.

Building a Simulated Phantom Model

To obtain reference results, we tested our methodology using a virtual model of the breast. This model consists of a hemispherical form representing the breast and containing a second sphere of twice the background intensity, representing the tumor. The diameter of the tumor is 25.6mm (20% of the spherical form diameter), and the diameter of the spherical form is 128mm (see Fig. 9 ).

Fig. 9

Compressed models: (a) 3-D sagittal perspective view of superimposed MRI (dark volume) and DOT (light volume) models. (b) Sagittal cross section of MRI model going through the center of the tumor. (c) Spatially corresponding sagittal cross section of DOT model.

051902_1_023705jbo9.jpg

The semispherical model is first compressed in the axial direction to simulate the DOT image. The initial model is again compressed in the sagittal direction to simulate the MR image. The amount of compression used is 25% for both optical and MR images, respectively, in the axial direction (along the z axis) and the sagittal direction (along the x axis).

For the axial compression, the z component of the voxel size was decreased by 25% and the x and y components are proportionally increased to keep the same volume size as the uncompressed model. The sagittal compression is simulated in a similar way by decreasing the x component of the voxel size by 25%, and the z and y components are proportionally increased to keep the same volume size. The new tumor center position after compression is determined by multiplying the tumor center position in pixels, by the new voxel size (see Fig. 9).

6.2.

Simulation Description

The experiments conducted test the registration algorithm’s sensitivity to changes in breast compression, translation, and rotation between the MRI and DOT datasets, mainly due to patient positioning differences.

6.2.1.

First set of simulations: incremental translations along the x , y , and z axis

Initial translations along the x axis are applied incrementally to the DOT image, i.e., the moving volume. The registration algorithm was tested after each translation. These translations simulate the difference in patient placement between the two image acquisition processes (translations of +50mm ). The simulation is repeated with translations applied along the y and z axes, and the registration is tested for each translation. Figure 10 shows the visual results of translations along the z and x axes. The MR model (top row) is the fixed volume in the simulation, and therefore remains unchanged. The DOT model is the moving volume shown in the center row in Fig. 10. This center row shows different initial starting points for the DOT model. In Figs. 10c and 10d, the tumor appears very small because the cross sections shown are spatially corresponding to those of the MR model, and show the edge of the tumor. Note that the bottom row (the DOT model after registration) should look as much as possible like the top row (the MR model).

Fig. 10

Simulation of compression and translation. Spatially corresponding cross sections of MRI model (top row), DOT model before registration (center row), and DOT model after registration (bottom row). (a) and (b) coronal cross sections: DOT model is translated ±50mm along z direction. (c) and (d) Axial cross sections: DOT model translated ±50mm along x direction.

051902_1_023705jbo10.jpg

6.2.2.

Second set of simulations: incremental rotations about the x axis

Several incremental rotations about the x axis (clockwise and counterclockwise directions) are applied to the DOT volume, and the registration algorithm is tested after each rotation. This is repeated for rotations around the y and z axes, and the algorithm is tested for each rotation step. Figure 11 shows examples of rotations applied and the resulting alignments (rotations of +18deg are shown). Here again, the MR model (top row) is the fixed volume in the simulation, and therefore remains unchanged. The DOT model is the moving volume shown in the center row in Fig. 11. This center row shows different initial rotations for the DOT model. In Fig. 11c, the tumor appears very small because the cross sections shown are spatially corresponding to those of the MR model, and show the edge of the tumor. Note that the bottom row (the DOT model after registration) should look as much as possible like the top row (the MR model).

Fig. 11

Simulation of compression and rotation. Spatially corresponding cross sections of MRI model (top row), DOT model before registration (center row), and DOT model after registration (bottom row). (a) Sagittal cross sections: DOT model is rotated±18deg about the x axis. (b) Coronal cross sections: DOT model is rotated±18deg about the y axis. (c) Axial cross sections: DOT model is rotated±18deg about the z axis.

051902_1_023705jbo11.jpg

6.2.3.

Third set of simulations: incremental axial compression of the simulated diffuse optical tomography dataset

Different incremental amounts of compression are applied to the DOT images in the axial direction (along the z axis). To simulate the axial compression, the z component of the voxel size was decreased by 10% for each test, and the x and y components are proportionally increased to keep the same volume size as the uncompressed model. The range of compression used is from 0% compression to 40% compression with a step of 10% for each simulation. Note, no figure is shown in this section.

7.

Application to Nonconcurrent Magnetic Resonance Imaging and Diffuse Optical Tomography Data of Human Subjects

A study involving three patients was performed. This study provides an initial answer to a vital question regarding MRI/DOT data analysis: how can functional information on a tumor obtained from DOT data be combined with the anatomical information about the tumor derived from MRI data?

Three MRI and three DOT (displaying THC) datasets are used in this experiment.

  • 1. Patient 1: MRI ( 256×256×22 with 0.63×0.63×4.0mmpixel size) and mastectomy show an invasive ductal carcinoma of the left breast. The size of the tumor was 2.1cm , as measured from pathology,

  • 2. Patient 2: MRI ( 256×256×60 with 0.7×0.7×1.5mmpixel size) and biopsy show an invasive ductal carcinoma of the left breast. The size of the tumor was 5.3cm , as measured from the MRI (patient 2 was a neoadjuvant chemo patient and did not have surgery until later),

  • 3. Patient 3: MRI ( 512×512×56 with 0.35×0.35×3.93mmpixel size) and mastectomy show an invasive in-situ carcinoma of the right breast. The size of the tumor was 2.0cm , as measured from pathology.

Fig. 3

Typical user workflow for the OMIRAD platform.

051902_1_023705jbo3.jpg

All DOT image acquisitions are similar and show the patient total hemoglobin concentration (THC). The procedure described in the typical workflow (Fig. 3) was used for visualizing, editing, and registering the MRI and DOT datasets. However, since the objective of this initial study is to show how nonconcurrent data may be registered and jointly analyzed, we did not use the results of MRI segmentation to improve the DOT reconstructions.

A quantitative analysis of the resulting data is not trivial. We propose a simple analysis method that provides valuable functional information about the carcinoma. Using the MRI/DOT registered data, we calculate the differences in total hemoglobin concentration (THC) between the volumes inside and outside the segmented tumor, as follows.

  • 1. Segment tumor from nontumor tissue in the breast MRI dataset, using our segmentation approach.

  • 2. Calculate the following statistical quantities from the DOT dataset, within the resulting segmented tumor and nontumor volumes (this is a trivial task, since the DOT and MRI datasets are now registered): average THC value over the entire breast α ; average THC value within the tumor volume defined by the MRI segmentation β ; standard deviation of THC for the entire breast σ .

  • 3. Calculate a new difference measure, defined as the distance from α to β in terms of σ : μ=(βα)σ .

The computed quantities are described in Fig. 12 . The results are shown in next section.

Fig. 12

THC distribution in a DOT dataset showing the resulting computed quantities after DOT-MRI image registration.

051902_1_023705jbo12.jpg

8.

Results and Discussion

8.1.

Phantom Model Validation

For most registration tasks, the most significant error measure is the target registration error (TRE), i.e., the distance after registration between corresponding points not used in calculating the registration transform. The term “target” is used to suggest that the points are typically points within, or on, the boundary of lesions. The registration algorithm gives us the absolute transformation Tresult that should be applied to the DOT volume to be aligned to the MRI volume. This transformation is applied to the tumor center and 26 neighboring points. The points are arranged on a cube in which the tumor is inscribed. The cube shares the same center as the tumor (as shown in Fig. 13 ). It has a side length of 25.6mm (equal to the diameter of the tumor). The point positions resulting from the application of the absolute transformation are then compared to the corresponding point positions resulting from the application of the ground truth transformation TGT , which gives us the expected point positions. This allows us to determine the average TRE for each simulation. The TRE is computed as the average Euclidian distance between the 27 pairs of points (PGTi,Presulti) :

Eq. 5

TRE=127i=127d(PGTi,Presulti).

Fig. 13

26 points arranged on the cube and used to compute the TRE. The tumor is inscribed in the cube, and shares the same center as that of the cube, noted with an ×.

051902_1_023705jbo13.jpg

The volume of the tumor after registration is also compared to the initial one, and the percentage error is computed. It is important to note that the range of translations chosen during simulations is 40mm (from 20to20mm ) to keep a realistic aspect of the simulations. Indeed, the translations represent the patient displacements during the image acquisition, so a range of 40mm is reasonable. Also, the range of rotations chosen is 36deg (from 18to18deg ) for the same reasons as before.

Tables 1, 2, 3 show the percent volume errors and the resulting target registration errors. As can be observed, the algorithm is more sensitive to rotations than translations, as the error exceeds 5% in some instances. This is explained by the fact that the algorithm uses 2-D signatures of the 3-D volume. By applying a rotation to the volume, the shape of the 2-D signature changes, whereas by applying a translation, the signature is moved compared to the volume of reference while keeping the same form. The change in form due to rotation makes the convergence of the algorithm more difficult. However, the higher level of rotations (more than ±10deg ) will seldom be encountered in reality, where patients usually lie prone in a reproducible manner, and will not cause such high levels of initial rotation. We conducted tests at these higher rotations to explore limitations of the registration technique. For certain points, the error rate increases considerably. This is also explained by the use of the 2-D signatures. Indeed, when the displacement of the image exceeds the limit of the projector that captures the signature, part of the information on volume is lost, leading to a potential divergence of the algorithm. Even though the algorithm is not strictly volume preserving, because of the scaling transformation, the volume percent error shows that within the practical range of deformations, the tumor volume is preserved within an average of about 3% of its original size, which is a reasonable error. Finally, the error due to compression is always under 5%.

Table 1

Percent volume errors with respect to the original moving volume, and resulting target registration error due to incremental translations applied along the x , y , and z axes.

Translation amount(mm)Translation along x axisTranslation along y axisTranslation along z axis
Volume%errorAverageTRE(mm)Volume% errorAverageTRE(mm)Volume% errorAverageTRE(mm)
20 2.973.772.590.601.050.89
10 2.271.761.470.87 1.51 2.34
01.792.621.792.621.792.62
104.513.022.800.711.241.00
203.953.03 1.43 3.035.554.21

Table 2

Percent volume errors with respect to the original moving volume, and resulting target registration errors due to incremental rotations applied about the x , y , and z axes.

Rotation amount(Degrees)Rotation about x axisRotation about y axisRotation about z axis
Volume% errorAverageTRE(mm)Volume%errorAverageTRE(mm)Volume% errorAverageTRE(mm)
18 7.527.452.422.663.2910.59
9 10.0811.310.710.907.004.29
01.792.621.792.621.792.62
92.882.584.772.985.671.77
180.700.70 0.34 0.34 4.244.24

8.2.

Patient Study

Figure 14 shows the computed statistical quantities as well as the difference measures. As expected, all DOT datasets show average tumor THC values, one to three standard deviations higher than the average breast THC values. The results also show large variability in average breast THC values from one patient to another (varying from 21to31μM ). This justifies the use of our difference measure μ , which defines a normalized quantity allowing interpatient comparisons.

Fig. 14

Statistical values computed in the registered DOT datasets: the segment middle points are the average THC values ( α inside the breast, β inside the tumor). The segment endpoints represent one standard deviation spread σ and μ is the difference measure (distance from α to β in terms of σ) .

051902_1_023705jbo14.jpg

These results confirm that the tumor areas in the patient breasts exhibit significantly higher THC than their surroundings. Figures 15, 16, 17 show the visual results of the registration algorithm when applied to real patient datasets. We show superimposed MRI and DOT images (3-D renderings and 2-D fused images) before and after registration. As can be qualitatively ascertained from the figures, registration has greatly improved the alignment of the DOT and MRI datasets. The images also show a significant overlap between the location of the tumors in the MRI and DOT datasets. Patient 3 shows particularly good correlation between the two modalities.

Fig. 15

Superimposed MRI and DOT images (3-D renderings and 2-D fused images) of patient 1 before and after registration. The 2-D fused images show the cross sections going through the center of the tumor.

051902_1_023705jbo15.jpg

Fig. 16

Superimposed MRI and DOT images (3-D renderings and 2-D fused images) of patient 2 before and after registration. The 2-D fused images show the cross sections going through the center of the tumor.

051902_1_023705jbo16.jpg

Fig. 17

Superimposed MRI and DOT images (3-D renderings and 2-D fused images) of patient 3 before and after registration. The 2-D fused images show the cross sections going through the center of the tumor.

051902_1_023705jbo17.jpg

The combination of DOT and MR image resolution, the registration technique, and the segmentation accuracy in MR all affect the final outcome. Can registration errors significantly affect the quantification of the results? Certainly this is possible, but a larger-scale study is required to better characterize the effect of the registration error. Indeed, this issue is significant when automatic segmentation is employed and when quantitative values are derived, especially in the case of small pathologies. Variations in the target registration error (TRE) cause variations in the overlap of the MR segmentation to the THC in the DOT dataset, which in turn cause variations in the quantification of the computed difference measure μ . However, because the THC is a slowly varying quantity in the DOT dataset, we expect small variations in μ .

To test this hypothesis, we simulated variations in the TRE by translating incrementally the MR segmentation area in the direction of maximum THC gradient in the DOT dataset. This enabled us to assess the upper bound of the quantification error due to TRE variations. The MR segmentation area was translated 1, 2, 3, 4, and 5mm . Then the different statistics were computed again, and variations in the difference measure μ are shown in Fig. 18 .

Fig. 18

Variations in μ for each patient, due to translations of the MR segmentation area inside the THC DOT dataset.

051902_1_023705jbo18.jpg

Table 3

Percent volume errors with respect to the original moving volume, and resulting target registration errors due to incremental axial compression of the moving volume.

% Amount ofaxialcompressionVolume % errorAverage TRE (mm)
04.911.21
103.660.87
201.801.06
30 0.36 2.37
400.112.76

As Fig. 18 shows, in all cases the difference measure decreases in amplitude as the translation distance is increased. This shows that the MR segmentation area is translated away from the THC “hotspot” in the DOT datasets. The variations of μ from the baseline (translation=0mm) in all cases are less than 15%, and μ remains equal to or larger than 1, i.e., the average THC inside the segmentation area remains more than one standard deviation away from the overall dataset average THC. Even though these results are limited to only three patients, they exhibit a relative robustness of the registration-segmentation-quantification approach to errors in automatic registration and segmentation. It is also worth noting that these results may apply more generally to patients with breast cancer tumors of sizes within the size range tested, between 2 and 5cm , which is typical.

The most efficient way to improve the coregistration technique in the near future will be to provide additional structural information on the DOT dataset. One way to achieve this goal is to provide a more accurate surface map of the patient’s breast as it is scanned in the DOT device, using stereo cameras for example. A more precise surface map may enable us in the near future to use more complex voxel-based and/or physics-based nonrigid registration techniques, and achieve better MR-DOT dataset alignment, especially for patients whose cancer tumors are smaller than 1cm in size.

9.

Conclusion

A software platform is developed and tested for multimodal integration and visualization of DOT and MRI datasets. The platform, OMIRAD, enables qualitative and quantitative analysis of DOT and MRI breast datasets, and combines the functional and structural data extracted from both types of images. To our knowledge, no other software systems combine nonconcurrent DOT and MRI data, as this developed platform does. The present work introduces this platform and its applications. Multiple simulation results are obtained and demonstrate the robustness of the algorithm.

The method integrates advanced multimodal registration and segmentation algorithms along with a straightforward and well-defined workflow. The method requires little prior user interaction, and is robust enough to handle a majority of patient cases, computationally efficient for practical applications, and yields results useful for combined MRI/DOT analysis. This method presents additional advantages:

  • more flexibility than integrated MRI/DOT imaging systems in the system design and patient positioning

  • the ability to independently develop a stand-alone DOT system without the stringent limitations imposed by the MRI device environment

  • combined analysis of structural and functional data derived simultaneously from both modalities

  • enhancement of DOT data reconstruction through the use of MRI-derived a-priori structural information.

We also present an initial patient study that explores an important question: how can functional information on a tumor obtained from DOT data be combined with the anatomy of that tumor derived from MRI data? The study confirms that the tumor areas in the patient breasts exhibit significantly higher THC than their surroundings. The results show significance in intrapatient THC variations, and justify the use of our normalized difference measure.

The completion of this study and the use of the OMIRAD platform contribute to the important long-term goal of enabling a standardized direct comparison of the two modalities (MRI and DOT). We anticipate it will have a positive impact on standardization of optical imaging technology through establishing common data formats and processes for sharing data and software, which in turn will allow direct comparison of different modalities, validation of new versus established methods in clinical studies, development of commonly accepted standards in postprocessing methods, creation of a standardized MR-DOT technology platform, and eventually, translation of research prototypes into clinical imaging systems.

Acknowledgments

We would like to thank all the interns who have worked on the design of OMIRAD (Jan Hegner, Antoine Azar, Pierre-Antoine Manzagol, Nassim Hajjioui, Mazen el-Bawab, Jia Kang), as well as Tania Monteiro and Yoo Kyung Choe for helping in the design of figures and diagrams. We acknowledge support from the NIH/NCI program U54 CA105480, “A Network for Translational Research in Optical Imaging: Multi-Dimensional Diffuse Optical Imaging in Breast Cancer,” and also from the NIH CA breast grant R01-EB-002109.

References

1. 

S. B. Colak, M. B. van der Mark, G. W. Hooft, J. H. Hoogenraad, E. S. van der Linden, and F. A. Kuijpers, “Clinical optical tomography and NIR spectroscopy for breast cancer detection,” IEEE J. Quantum Electron., 5 1143 –1158 (1999). 0018-9197 Google Scholar

2. 

B. W. Pogue, S. P. Poplack, T. O. McBride, W. A. Wells, K. S. Osterman, U. L. Osterberg, and K. D. Paulsen, “Quantitative hemoglobin tomography with diffuse near-infrared spectroscopy: Pilot results in the breast,” Radiology, 218 261 –266 (2001). 0033-8419 Google Scholar

3. 

H. B. Jiang, N. V. Iftimia, Y. Xu, J. A. Eggert, L. L. Fajardo, and K. L. Klove, “Near-infrared optical imaging of the breast with model-based reconstruction,” Acad. Radiol., 9 186 –194 (2002). 1076-6332 Google Scholar

4. 

A. Li, “Tomographic optical breast imaging guided by three-dimensional mammography,” Appl. Opt., 42 5181 –5190 (2003). https://doi.org/10.1364/AO.42.005181 0003-6935 Google Scholar

5. 

A. Torricelli, L. Spinelli, A. Pifferi, P. Taroni, R. Cubeddu, and G. M. Danesini, “Use of a nonlinear perturbation approach for in vivo breast lesion characterization by multi-wavelength time-resolved optical mammography,” Opt. Express, 11 853 –867 (2003). 1094-4087 Google Scholar

6. 

V. E. Pera, E. L. Heffer, H. Siebold, O. Schutz, S. Heywang-Kobrunner, L. Gotz, A. Heinig, and S. Fantini, “Spatial second-derivative image processing: an application to optical mammography to enhance the detection of breast tumors,” J. Biomed. Opt., 8 517 –524 (2003). https://doi.org/10.1117/1.1578496 1083-3668 Google Scholar

7. 

D. B. Jakubowski, A. E. Cerussi, F. Bevilacqua, N. Shah, D. Hsiang, J. Butler, and B. J. Tromberg, “Monitoring neoadjuvant chemotherapy in breast cancer using quantitative diffuse optical spectroscopy: a case study,” J. Biomed. Opt., 9 230 –238 (2004). https://doi.org/10.1117/1.1629681 1083-3668 Google Scholar

8. 

Q. Zhu, S. H. Kurtzma, P. Hegde, S. Tannenbaum, M. Kane, M. Huang, N. G. Chen, B. Jagjivan, and K. Zarfos, “Utilizing optical tomography with ultrasound localization to image heterogeneous hemoglobin distribution in large breast cancers,” Neoplasia, 7 263 –270 (2005). https://doi.org/10.1593/neo.04526 1522-8002 Google Scholar

9. 

R. Choe, “Diffuse optical tomography of breast cancer during neoadjuvant chemotherapy: a case study with comparison to MRI,” Med. Phys., 32 1128 –1139 (2005). https://doi.org/10.1118/1.1869612 0094-2405 Google Scholar

10. 

B. Chance, S. Nioka, J. Zhang, E. F. Conant, E. Hwang, S. Briest, S. G. Orel, M. D. Schnall, and B. J. Czerniecki, “Breast cancer detection based on incremental biochemical and physiological properties of breast cancers: A six-year, two-site study,” Acad. Radiol., 12 925 –933 (2005). 1076-6332 Google Scholar

11. 

D. Grosenick, K. T. Moesta, M. Möller, J. Mucke, H. Wabnitz, B. Gebauer, C. Stroszczynski, B. Wassermann, P. M. Schlag, and H. Rinneberg, “Time-domain scanning optical mammography: I. Recording and assessment of mammograms of 154 patients,” Phys. Med. Biol., 50 2429 –2450 (2005). https://doi.org/10.1088/0031-9155/50/11/001 0031-9155 Google Scholar

12. 

D. Grosenick, H. Wabnitz, K. T. Moesta, J. Mucke, P. M. Schlag, and H. Rinneberg, “Time-domain scanning optical mammography: II. Optical properties and tissue parameters of 87 carcinomas,” Phys. Med. Biol., 50 2451 –2468 (2005). https://doi.org/10.1088/0031-9155/50/11/002 0031-9155 Google Scholar

13. 

P. Taroni, A. Torricelli, L. Spinelli, A. Pifferi, F. Arpaia, G. Danesini, and R. Cubeddu, “Time-resolved optical mammography between 637 and 985nm: clinical study on the detection and identification of breast lesions,” Phys. Med. Biol., 50 2469 –2488 (2005). https://doi.org/10.1088/0031-9155/50/11/003 0031-9155 Google Scholar

14. 

C. H. Schmitz, “Design and implementation of dynamic near-infrared optical tomographic imaging instrumentation for simultaneous dual-breast measurements,” Appl. Opt., 44 2140 –2153 (2005). https://doi.org/10.1364/AO.44.002140 0003-6935 Google Scholar

15. 

T. Yates, J. C. Hebden, A. Gibson, N. Everdell, S. R. Arridge, and M. Douek, “Optical tomography of the breast using a multi-channel time-resolved imager,” Phys. Med. Biol., 50 2503 –2518 (2005). https://doi.org/10.1088/0031-9155/50/11/005 0031-9155 Google Scholar

16. 

X. Intes, “Time-Domain optical mammography SoftScan: Initial results,” Acad. Radiol., 12 934 –947 (2005). 1076-6332 Google Scholar

17. 

A. Cerussi, N. Shah, D. Hsiang, A. Durkin, J. Butler, and B. J. Tromberg, “In vivo absorption, scattering, and physiologic properties of 58 malignant breast tumors determined by broadband diffuse optical spectroscopy,” J. Biomed. Opt., 11 044005 (2006). https://doi.org/10.1117/1.2337546 1083-3668 Google Scholar

18. 

V. Ntziachristos, A. G. Yodh, M. Schnall, and B. Chance, “Concurrent MRI and diffuse optical tomography of breast after indocyanine green enhancement,” Proc. Natl. Acad. Sci. U.S.A., 97 2767 –2772 (2000). https://doi.org/10.1073/pnas.040570597 0027-8424 Google Scholar

19. 

B. Brooksby, B. W. Pogue, S. Jiang, H. Dehghani, S. Srinivasan, C. Kogel, T. D. Tosteson, J. Weaver, S. P. Poplack, and K. D. Paulsen, “Imaging breast adipose and fibroglandular tissue molecular signatures by using hybrid MRI-guided near-infrared spectral tomography,” Proc. Natl. Acad. Sci. U.S.A., 103 8828 –8833 (2006). https://doi.org/10.1073/pnas.0509636103 0027-8424 Google Scholar

20. 

D. Li, P. M. Meaney, T. D. Tosteson, S. Jiang, T. Kerner, T. O. McBrice, B. W. Pogue, A. Hartov, and K. D. Paulsen, “Comparisons of three alternative breast modalities in a common phantom imaging experiment,” Med. Phys., 30 (8), 2194 –2205 (2003). https://doi.org/10.1118/1.1586266 0094-2405 Google Scholar

21. 

B. Brooksby, H. Dehghani, B. W. Pogue, and K. D. Paulsen, “Near infrared (NIR) tomography breast image reconstruction with a priori structural information from MRI: algorithm development for reconstructing heterogeneities,” IEEE J. Sel. Top. Quantum Electron., 9 (2), 199 –209 (2003). 1077-260X Google Scholar

22. 

N. Shah, J. Gibbs, D. Wolverton, A. Cerussi, N. Hylton, and B. J. Tromberg, “Combined diffuse optical spectroscopy and contrast-enhanced magnetic resonance imaging for monitoring breast cancer neoadjuvant chemotherapy: a case study,” J. Biomed. Opt., 10 (5), 051503 (2005). https://doi.org/10.1117/1.2070147 1083-3668 Google Scholar

23. 

F. S. Azar, K. Lee, R. Choe, A. Corlu, J. Pearson, F. Sauer, and A. G. Yodh, “A software platform for multi-modal information integration & visualization: diffuse optical tomography and MRI of breast cancer,” Intl. Soc. Molecular Imag., 96 (2004) Google Scholar

24. 

F. S. Azar, M. ElBawab, A. Khamene, K. Lee, R. Choe, A. Corlu, S. D. Konecky, A. G. Yodh, and F. Sauer, “Multimodal 3D registration of non-concurrent diffuse optical tomography with MRI of breast cancer,” Intl. Soc. Molecular Imagi., 4 (3), 264 (2005). Google Scholar

25. 

C. Florin, M. ElBawab, and F. S. Azar, “A new method for increasing the flexibility of DICOM tags management in application-specific integration,” 16 –17 (2005). Google Scholar

26. 

F. S. Azar, K. Lee, R. Choe, A. Corlu, S. D. Konecky, A. G. Yodh, and F. Sauer, “A novel approach for joint analysis of non-concurrent magnetic resonance imaging and diffuse optical tomography of breast cancer,” Intl. Soc. Molecular Imag., 5 (3), 275 (2006). Google Scholar

27. 

S. D. Konecky, R. Wiener, N. Hajjioui, R. Choe, A. Corlu, K. Lee, S. M. Srinivas, J. R. Saffer, R. Freifelder, F. S. Azar, J. S. Karp, and A. G. Yodh, “Characterization of breast lesions using diffuse optical tomography and positron emission tomography,” (2007). Google Scholar

28. 

F. S. Azar, K. Lee, R. Choe, A. Corlu, S. D. Konecky, and A. G. Yodh, “Joint analysis of non-concurrent magnetic resonance imaging and diffuse optical tomography of breast cancer,” Proc. SPIE, 6434 643419 (2007). 0277-786X Google Scholar

29. 

R. Choe, A. Corlu, K. Lee, T. Durduran, S. D. Konecky, M. Grosicka-Koptyra, S. R. Arridge, G. J. Czerniecki, D. L. Fraker, A. DeMichele, B. Chance, M. A. Rosen, and A. G. Yodh, “Diffuse optical tomography of breast cancer during neoadjuvant chemotherapy: A case study with comparison to MRI,” Med. Phys., 32 (4), 1128 –1139 (2005). https://doi.org/10.1118/1.1869612 0094-2405 Google Scholar

30. 

J. P. Culver, R. Choe, M. J. Holboke, L. Zubkov, T. Durduran, A. Slemp, V. Ntziachristos, B. Chance, and A. G. Yodh, “Three-dimensional diffuse optical tomography in the parallel plane transmission geometry: Evaluation of a hybrid frequency domain/continuous wave clinical system for breast imaging,” Med. Phys., 30 235 –247 (2003). https://doi.org/10.1118/1.1534109 0094-2405 Google Scholar

31. 

A. Corlu, R. Choe, T. Durduran, K. Lee, M. Schweiger, E. M. C. Hillman, S. R. Arridge, and A. G. Yodh, “Diffuse optical tomography with spectral constraints and wavelength optimization,” Appl. Opt., 44 2082 –2093 (2005). https://doi.org/10.1364/AO.44.002082 0003-6935 Google Scholar

32. 

M. Schweiger, “Application of the finite element method in infrared image reconstruction of scattering media,” Univ. of London, (1994). Google Scholar

34. 

J. Maintz and M. Viergever, “A survey of medical image registration,” Med. Image Anal, 2 1 –36 (1998). 1361-8415 Google Scholar

35. 

D. L. G. Hill, D. J. Hawkes, J. E. Crossman, M. J. Gleeson, T. C. S. Cox, E. C. M. L. Bracey, A. J. Strong, and P. Graves, “Registration of MR and CT images for skull base surgery using pointlike anatomical features,” Br. J. Radiol., 64:1030–1035 (1991). 0007-1285 Google Scholar

36. 

A. Colignon, F. Maes, D. Delaere, D. Vandermeulen, P. Suetens, and G. Marchal, “Automated multi-modality image registration based on information theory,” IPMI, 263 –274 (1995). Google Scholar

37. 

W. Wells III, P. Viola, H. Atsumi, S. Nakajima, and R. Kikinis, “Multi-modal volume registration by maximization of mutual information,” Med. Image Anal, 1 35 –52 (1996). 1361-8415 Google Scholar

38. 

S. C. Cain, M. M. Hayat, and E. E. Armstrong, “Projection-based image registration in the presence of fixed-pattern noise,” IEEE Trans. Img. Process., 10 (2), 1860 –1872 (2001). Google Scholar

39. 

H. Chan and A. C. S. Chung, “Efficient 3D-3D vascular registration based on multiple orthogonal 2D projections,” 2nd Intl. Workshop Biomedi. Image Reg. (WBIR), 301 –310 (2003). Google Scholar

40. 

M. Wells III, P. Viola, H. Atsumi, S. Nakajima, and R. Kikinis, “Multi-modal volume registration by maximization of mutual information,” Med. Image Anal, 1 (1), 35 –51 (1996). 1361-8415 Google Scholar

41. 

R. H. Huesman, G. J. Klein, J. A. Kimdon, C. Kuo, and S. Majumdar, “Deformable registration of multi-modal data including rigid structures,” IEEE Trans. Nucl. Sci., 50 (3), 1879 –1882 (2003). 0018-9499 Google Scholar

42. 

S. C. Cain, M. M. Hayat, and E. E. Armstrong, “Projection-based image registration in the presence of fixed-pattern noise,” IEEE Trans. Image Process., 10 1860 –1872 (2001). https://doi.org/10.1109/83.974571 1057-7149 Google Scholar

43. 

H. Chan and A. C. S. Chung, “Efficient 3D-3D vascular registration based on multiple orthogonal 2D projections,” 2nd Intl. Workshop Biomedi. Image Reg. (WBIR), 301 –310 (2003). Google Scholar

44. 

A. Khamene, R. Chisu, W. Wein, N. Navab, and F. Sauer, “A novel projection based approach for medical image registration,” (2006) Google Scholar

45. 

L. Grady and G. Funka-Lea, “Multi-label image segmentation for medical applications based on graph-theoretic electrical potentials,” (2004). Google Scholar

46. 

L. Grady, “Multilabel random walker image segmentation using prior models,” Proc. CVPR05, 1 763 –770 (2005). Google Scholar
©(2007) Society of Photo-Optical Instrumentation Engineers (SPIE)
Fred S. Azar, Kijoon Lee, Ali Khamene, Regine Choe, Alper Corlu, Soren David Konecky, Frank Sauer, and Arjun G. Yodh "Standardized platform for coregistration of nonconcurrent diffuse optical and magnetic resonance breast images obtained in different geometries," Journal of Biomedical Optics 12(5), 051902 (1 September 2007). https://doi.org/10.1117/1.2798630
Published: 1 September 2007
Lens.org Logo
CITATIONS
Cited by 35 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Magnetic resonance imaging

Image registration

Breast

Tumors

Image segmentation

Imaging systems

Visualization

Back to Top