Image segmentation is a fundamental problem in computer vision and image analysis. In the image segmentation community, level set-based approaches are important tools, because they are able to handle nuclei shapes and contours with complex variations. Chan and Vese initiated a region-based active contour model with a level set formulation based on Mumford-Shah’s functional.1,2 This model does not depend on gradient information and thus can detect nuclei contours with weak edges. However, Chan–Vese model does not include prior shape knowledge to restrain shapes in appearance and may, therefore, detect meaningless shapes. This drawback becomes more serious when nuclei are partially occluded, corrupted, or represented in low contrast imaging data. In the literature, some research work emerged to mitigate this problem by incorporating shape prior information into the level set formulation so that the detected shape can be regulated by a selected reference shape.3–5 For example, Chan and Zhu6 proposed a variational model introducing a shape difference term with a shape prior as the reference shape.1 With prior information in place, shapes similar to the shape prior can be successfully segmented. Meanwhile, nuclei presenting meaningless shapes are restrained. Based on the shape prior segmentation model, Yan et al.7 introduced a geodesic length term to drive contours to nuclei edges. Ali and Madabhushi8 proposed an active contour model to accommodate complex shapes by learning shape priors through principal component analysis.
Nevertheless, shape prior-based level set segmentation methods are still challenged by multiple problems. (1) It is difficult to segment nuclei from raw images without knowing the number and position of nuclei. Thus, one important step in this class of segmentation approaches is to appropriately identify number and positions of nuclei of interest for initialization purpose. (2) Thus far, shape prior-based segmentation approaches exploit libraries consisting of a small number of shape priors as reference to nuclei in similar shapes. However, in most real-world scenarios, such as nuclei segmentation, it could be very complex to model shape variations explicitly. Therefore, it is practically difficult to find a limited number of shape priors that could represent all shapes reasonably well. (3) Simultaneous segmentation of mutually occluded nuclei remains challenging. Recent development in level set methods introduced a repulsion term to prevent two nuclei from becoming identical through evolution. However, it is common to have more than two nuclei involved in mutual occlusion. As a result, we conceive that larger penalties should be given to regions shared by more nuclei.
Nuclei initialization is an important step that guides the following modules in segmentation. The geometric centers of nuclei, so-called seeds, are natural nuclei indicators in practice, as nuclei contours can be appropriately initialized with them. Parvin et al.9 proposed an iterative voting method for detecting seeds of overlapping nuclei. Al-Kofahi et al.10 proposed a multiscale Laplacian of Gaussian (LoG) filtering method for automatic detection and segmentation of nuclei. By distance-map-based adaptive scale selection, nuclear seed points were detected to perform an initial segmentation. However, the multiscale LoG method is sensitive to minor peaks on the distance map, resulting in oversegmentation and detection. Qi et al.11 improved the method by applying a shifted Gaussian kernel and mean shift to a single-pass voting algorithm. The basic idea is to search for the voting direction over a cone-shaped voting area. The two-dimensional (2-D) Gaussian kernel was designed to assign larger weights for area centers in the voting process. Finally, mean shift was applied to determine the seed points.
To accommodate the variation in nuclei shapes, a large set of shape priors can be used as a reference library. However, it is challenging to model the relationship between the nuclei of interest to be segmented and the shape priors. Representation learning methods, such as graph learning and sparse coding, were successfully applied in many areas of computer vision research, such as face recognition,12 image restoration,13 graph learning,14 and medical image analysis.15 Wright et al.12 proposed a robust face recognition algorithm based on sparse representation. The method can effectively handle occlusion and corruption errors by enforcing sparse distribution with respect to the pixel/sample basis. Zhang et al.15 proposed a sparse shape composition model to incorporate shape prior information for detection and segmentation tasks in radiologic images. Cheng et al.14 introduced a process to build graph and multiple algorithms for data clustering, subspace learning, and semisupervised learning based on graph. Inspired by theories in graph learning, spectral clustering, and sparse representation, an graph-based shape proximity was learned to cluster the shape priors with which a compact dictionary was created for sparse coding. Given a shape of interest, it can be approximated by a sparse representation based on the compact dictionary within the level set framework. The resulting reconstruction error can be included as a shape penalty term in the variational level set model.
To date, numerous investigators have devised methods to segment overlapping nuclei. These methods extended the shape prior segmentation approaches to segment multiple similar nuclei under mutual occlusion. In Refs. 1617.–18, various repulsive force terms were proposed and included in the cost functional to penalize two overlapping nuclei contours and prevent them from becoming identical. In our approach, we extend this idea by introducing an adaptive occlusion penalty term that penalizes regions occupied by multiple nuclei based on the scope of occlusion. Following this principle, we assign a larger penalty to a region overlapped by more nuclei.
Recently, deep learning-based approaches greatly improved the state-of-the-art in research fields, such as computer vision, speech recognition, and natural language processing. Deep learning methods require large-scale annotated data to preserve generative power and prevent overfitting. However, collecting both data and annotation for medical imaging tasks is time-consuming and requires much domain expertise. Specifically, a large number of bounding boxes or shape masks of nuclei are required to train the network for deep learning-based nuclei detection and segmentation methods. Compared to the formidable data scale required by deep learning methods, the only prior information used in our method is a set of representative shapes in vector representations with only nuclei contour co-ordinates.
In this paper, we propose a level set-based variational model that allows simultaneous segmentation of multiple nuclei with mutual occlusion. The main contributions of our work are summarized as follows:
As shown in Fig. 1, our method framework consists of a seed detection algorithm for nuclei contour initialization, and an integrated contour deformable model that incorporates region, shape, and boundary information. The final nuclei contours are converged through an iterative contour evolution process.
This journal paper extends our earlier work19,20 through substantial method improvement on shape prior library generation, more comprehensive experiments, more in-depth parameter sensitivity analysis. In detail, these important extensions include:
Let us consider an image that contains nuclei of interest , where each nuclei has a closed and bounded shape contour in image domain . The basic idea of the level set framework is to implicitly represent as a zero level set of a Lipschitz function . has positive and negative value when is inside and outside , respectively. Note that due to memory limit, we use image patches instead of whole slide images in our experiments.
Distance Map Function
A distance map function represents the shortest distance from a current pixel in the image domain to a given nuclei contour :
Shape Prior Modelling
Instead of a single shape prior by previous shape prior segmentation methods, we use a large set of shape priors to deal with the complex shape variation observed in most real-world applications. Training shape priors are manually extracted from raw images and aligned with generalized Procrustes analysis,21 as illustrated in Fig. 2. The resulting shapes are represented by a set of vectors of uniformly sampled local landmarks: , where is the size of the shape prior set.
For better computational efficiency, the shape priors are partitioned into clusters, where . We cluster the shape priors with a learned graph learned by solving the minimization problem for each :1422,23 is used to partition the shape priors into clusters. The resulting average shape of each cluster is computed and used as a representative shape for the cluster. For simplicity, we denote the corresponding distance maps of the average shapes as . These distance maps are normalized to have unit Frobenius norm so that the impact of each shape prior is balanced.
To represent a nuclei with shape priors in the training set , we first align the associated distance map to shape priors by rotation and translation. The transformation from a pixel to the corresponding point in shape prior after alignment is formulated as follows:
Sparse Shape Representation
Given the set of distance maps derived by mapping the ’th nuclei to shape prior set , we assume that the distance map of can be approximately represented as a linear composition of shape priors in . The distance maps and shape priors are vectorized and denoted as and , respectively. By the linearity assumption, can be represented as follows:
To avoid the curse of dimensionality problem, we reduce dimensionality, leading to a less computational cost. The process of dimension reduction can be modeled as left multiplication by a nonzero projection matrix , , where is the total number of pixels in an image. As proved in 12, the choice of matrix does not critically affect the ability to recover the sparse solution. For computation simplicity, we compose a matrix with entries randomly generated from standard Gaussian distribution and subject to unit length for all rows. The corresponding low dimension representations are denoted as , , and . Thus, we have12 we observe that given a large enough training set, a test shape should be sufficiently represented using a small number of similar shape priors. In addition, the number of corrupted pixels in the original image and the derived distance map is assumed to be small. This suggests that the coefficient vectors and can be sparse as most entries are zero. In this way, most dissimilar prior shapes are suppressed due to large penalty measured by the reconstruction error, which enables us to differentiate from dissimilar shape priors. To ensure the sparsity, we use norm regularization in the formulation. The sparse representation problem is formulated as follows:
However, the problem of finding the sparsest solution in an underdetermined linear system such as Eq. (7) is proved to be nondeterministic polynomial-time hard.24 Due to recent developments in sparse theory, it has been revealed that such a problem can be solved via relaxation:25–27
In this section, we present an algorithm to recognize nuclei spatial locations by nuclei seed detection. Nuclei seed detection is essential for the follow up nuclei segmentation, as it decides the number and initial contour locations. The proposed seed detection algorithm utilizes joint information of spatial connectivity, distance constraint, image edge map, and a shape-based voting map derived from eigenvalue analysis of Hessian matrix across multiple scales.
Initialization and Preprocessing
The pathology images of tumor specimens for routine diagnostics usually contain two primary chemical stains: hematoxylin and eosin (H&E). Hematoxylin presents a dark blue or violet color, which positively binds to nuclei. We decompose signal intensity components of H&E stains in original images and use only signals of hematoxylin stain for analysis. Based on Lambert–Beer’s law, the relationship between the intensity of light entering a specimen and that through a specimen can be characterized as , where and are the amount of stain and the absorption factor, respectively. By this law, we can compute the stain-specific strength values with a predefined stain decomposition matrix and the observed optical density (OD): .28 The stain composition matrix consists of three normalized column vectors representing OD values of red, green, and blue channels from hematoxylin, eosin, and null stain. Therefore, stain specific optical signals can be computed by . We retain the first entry of the resulting stain specific signal vector at each pixel and denote the decomposed hematoxylin channel signal as .
Note that is often coupled with noise produced during the slide preparation process, including uneven tissue slide cut, heterogeneous histology process, staining artifacts, and digital scanning noise. We can normalize the background noise in by morphological reconstruction with two morphological components, namely mask image and marker image . Initially, we set as complementary image of and as , where denotes the morphological dilation operator and is a circular structural element with radius . With marker and mask image, an image can be reconstructed by iterative dilation and intersection until convergence.29 The difference image consists of a near zero-level background and a group of enhanced foreground peaks, each representing an object of interest. In Fig. 4(a), we present a typical background normalization result with morphological reconstruction.
Voting Map Construction
Assuming a typical nuclei shape is close to either a circle or an ellipse, we proceed with a shape-based voting process by analyzing nuclei structures within local image regions. Before the voting process, we first enhance nuclei regions by eigenvalue analysis with the Hessian matrix convolved with Gaussian filters at multiple scales.30 With this approach, we search for circular structures based on geometric structures characterized by the neighboring pixel intensity profiles. Specifically, for a pixel at , its local image intensity change can be represented by Taylor expansion:Fig. 3, we present typical examples of overlapped nuclei with distinct overall clump shapes after different analysis steps. By comparison, voting maps produced from eigenvalue sign analysis encoding pixel intensity profile in nuclei neighborhoods have superior contrast to complementary image of decomposed hematoxylin signal and the difference image . It demonstrates that the voting map derived from eigenvalue sign analysis is an effective way to characterize overlapped nuclei local intensity profile features regardless of overall clump shapes. Next, we begin the voting process with a zero-valued voting map for all pixels in the image domain. We only increment the voting map value when . If , we consider pixel is not within a nuclei structure and hence set zero vote for such locations. This process is repeated for all scales. A voting map overlaid with a typical histological image region, where nuclei regions in dark exhibit high votes, is presented in Fig. 4(b).
Dynamic Seeds Detection and Merging
Given the derived voting map, we dynamically adjust a seed list based on candidate spatial connectivity, distance constraint, and image edge map. The proposed method can produce robust and accurate seed detection result, especially for overlapped nuclei.
As any pixel on the voting map is no larger than the number of scales , we consider the voting map as a surface in a three-dimensional space, as shown in Fig. 4(b). Those strong voting peaks, representing consistent negative eigenvalue pairs from Hessian matrix at different smoothing scales due to radially outward decreasing pixel intensity profiles in local regions, suggest the presence of nuclei. The strong peaks on the voting surface can be detected as we gradually slide down an imaginary horizontal plane (e.g., from the blue to green plane) intersecting with the voting surface. We can generate a binary image from the original voting map with threshold for each intersection plane. We begin with an empty seed list and append to it the resulting centroids of voting peaks satisfying all the following conditions: (a) such voting peak centers do not exist in the peak list yet. (b) They come from strong peaks that suggest consistent local intensity change property, a key to a reliable nuclei detection. Such strong peaks can be identified with their sizes no less than area threshold , where and are predefined scalars that determine the lower area threshold for detected peaks in the voting map at each step and is a function normalizing the voting value ; specifically, is the lowest peak cross-section size expectation when the imaginary horizontal plane is the highest. determines how peak areas are expected to grow as the imaginary horizontal plane slides down. Their choices depend on the nuclei texture homogeneity and nuclei size. We choose to be small enough not to miss any strong peaks, and big enough to represent the rapid growth of strong voting peak cross-sections as the imaginary horizontal plane slides down. (c) Such points are within the foreground region in the binary mask detected by the adaptive thresholding method.
With a list of seed candidates, we perform seed pruning to eliminate false positive seeds. We compute the pairwise distances for all peaks in and iteratively merge adjacent centroids within distance threshold . This is followed by a second round of distance-based merging with a relaxed threshold (). In the second round, two peaks are merged only if the following conditions are true: (a) distance between these two points is less than and (b) the path connecting a pair of points does not intersect with any canny edge derived from the original image. The edge map is used to prevent centroids of closely clumped nuclei from being merged. With this seed merging process, we can retain seeds from true clumped nuclei in histological images without tedious parameter tuning process.
Occluded Nuclei Segmentation
With nuclei detected in Sec. 3, we further develop an improved level set based segmentation method to drive initialized nuclei contours to true nuclei edges. Our development is driven by the use case of automatic analysis of occluded nuclei in whole-slide histopathologic images. In our method, each nuclei is described by a level set function and we aim to simultaneously obtain all level set functions by optimizing a variational model.
Prior Information Integration
As described in Sec. 2.4, given a large enough set of shape priors , a shape of interest can be encoded as sparse linear superposition of shape priors. Thus, we define a shape term as follows:
Adaptive Penalty on Mutual Occlusion
It is common to have mutually occluded nuclei in 2-D histopathologic images due to the fact that these images represent 2-D projected signals of tissue nuclei in 3-D space. It is a challenging task to segment mutually occluded nuclei and identify hidden boundaries across each other. This problem becomes exponentially complicated when there are more than two nuclei involved in occlusion. In the level set framework, intersecting nuclei may all have positive function values after contour evolutions, making it difficult to differentiate them from each other. To address this problem, we introduce an adaptive occlusion penalty term to dynamically suppress nuclei intersection events. The occlusion penalty term is determined by the number of nuclei that are overlapped. Meanwhile, this term prevents deformable contours from becoming identical after iterations of evolution. Specifically, we define the adaptive occlusion penalty term as follows:
Edge Guided Contour and Evolution Stability
To further encourage contour convergence and retain contour smoothness, we define an edge-guided contour regularity term as follows:
Additionally, we include an evolutionary stability term31 to regulate the property of level set function as follows:
Improved Variational Level Set Formulation
By combining terms in Eqs. (10)–(13), we formulate our method with a variational level set framework.1 To evolve nuclei contours to desired nuclei boundaries, we minimize the following functional:
In this section, we present numerical computations to minimize the functional in Eq. (15). We optimize the functional iteratively by updating functions and variables alternatively. We begin with updating functions first. Parameterizing iteration as an artificial time variable , we minimize the functional by solving Euler–Lagrange equation based on theory of calculus of variations:
Next, we fix and update transformation parameters. We derive updating equations for transformation parameters by computing gradient descent of functional :
Finally, we fix , and , and update and by setting the partial derivative of with respective to and and solving these equations, respectively. The optimal values turn out to be the average image intensities in corresponding area:1
In this way, this minimization problem is solved by iterative computation of Euler–Lagrange equation and gradient descent approach until the functional is converged. The zero level sets of the converged functions indicate the final nuclei contours.
Dataset and Parameter Setting
We present experimental results of our algorithm for analysis of nuclei within histopathologic images of glioblastoma brain tumor specimens. The effectiveness of our algorithm is verified with two datasets of H&E stained GBM specimens captured at magnification, namely GBM40 dataset and TCGA FFPE dataset: http://cancergenome.nih.gov/. These images are obtained after glioblastoma brain tissues are processed with a tissue preparation protocol. Image patches with size are used for experiments due to memory limit. For GBM40 dataset, there are 5396 and 1849 manually annotated nuclei for seed detection and boundary delineation, respectively. For TCGA dataset, the total number of annotated nuclei for boundary segmentation is 4961. Shape profiles of 27,000 glioblastoma brain nuclei are manually extracted from the dataset to form the set of shape priors. All shape priors are aligned with generalized Procrustes analysis.21 As discussed in Secs. 2.1 and 2.2, we represent shapes as distance maps, cluster shape priors with an graph into groups and use one representative shape from each group to form a training shape dictionary for sparse representation. We apply the proposed method to datasets with the following parameter setup: , , , , , , , , , , , . Note that our approach is an image data driven process. Therefore, the scanner setting such as magnification factor does not have significant impact on the parameter setting. As seeds are detected in the voting map that is produced by eigenvalue analysis of image local Hessian matrix associated with a set of Gaussian filter scales, transforming from the original image to voting map can partially help take care of nuclei scale variations. In our study, this set of Gaussian filter scales is chosen to cover varying nuclei sizes, with approximately equal to the radius of a typical large nucleus.
For the weighting parameters for nuclei segmentation, we assign parameter values so that all terms in Eq. (15) are appropriately balanced in numerical values. In our experiment, following three parameters have similar value and scale: the coefficient for edge-gradient weighted contour length , the coefficient for the occlusion penalty term , and the coefficient for squared fitting error of shape-derived distance map . The weight is numerically set by referring to typical value of the double-well potential function. We use the l1-ls MATLAB toolbox32 to solve the -minimization problem.
For quantitative analysis, we assess performance of seed detection method with reference to annotations by a pathologist. Note that we evaluate our approach with nontouching and occluded nuclei in each image separately. Five metrics are computed from each image to show seed detection performance: nuclei number error (), miss detection (), false recognition (), over segmentation (), under segmentation (), and count error (CE)%. Nuclei number error is used to demonstrate the absolute difference between the number of nuclei detected by machine and that reported by human expert. Miss and false detection represent the number of missing and false recognition events when machine-based seed detection method is used to detect individual nuclei with no occluded neighbors. Meanwhile, we use over- and undersegmentation to record events where the number of machined-identified nuclei from a nuclei clump is more and less than the true number marked by human expert, respectively. Finally, we use the count error% to represent the nuclei number error in reference to the true nuclei number. The resulting outputs are shown in Tables 1 and 2. Additionally, we compare our method with a single-pass voting method based on mean shift11 and an oriented kernel-based iterative voting method.9 By contrast to our method, both the single-pass voting method and the iterative voting method tend to miss detecting nuclei. Note that GBM40 dataset is produced by our local laboratory. Compared to the public TCGA dataset, GBM40 dataset is more carefully prepared with better staining and contrast, leading to higher SNR. This results in better performances of all methods for comparison consistently. We also illustrate seed detection results on several typical image patches in Fig. 5.
Image-wise seed detection performance on GBM40 dataset.
Note: Bold numbers indicate the best performance.
Image-wise seed detection performance on TCGA dataset.
Note: Bold numbers indicate the best performance.
We model the nuclei segmentation problem by the variational level set framework mathematically described in Eq. (15) and solve it by the numerical computations in Sec. 4. The level set functions do not stop evolving until they either reach convergence or exceed iteration number limit. We present the evolution results of zero level sets at iteration 10, 20, and 30 in Fig. 6. The detected nuclei shapes are well defined, as shown in Fig. 6. Notably, those overlapping nuclei can also be correctly segmented, with occluded boundaries reasonably recovered. It is also observed that most zero level sets can rapidly converge to true nuclei boundaries within 10 iterations. After that, zero level sets fine tune themselves to better fit to nuclei contour details, especially over those overlapped nuclei. With our experiment data, we only observe minor improvements in nuclei segmentation from 10 to 30 iterations. In general, the optimal number of iterations depends on data properties. End users can determine the accuracy-speed tradeoff and select the best number of iterations based on individual experiment scenarios.
In addition, we present experimental results of three images presenting the best, median, and worst overall segmentation performance in Fig. 7. The bar charts in Fig. 7 present our method efficacy measured by the Jaccard coefficient, precision, recall, and score. The small variation in these metrics over the best, median, and worst images suggests a good consistency and generalization ability of our method.
To quantitatively assess our method’s performance, we use human annotations from pathologist as ground truth. By comparing the experimental result with ground truth annotation , we evaluate the performance of our method with multiple metrics:36,37 Jaccard coefficient , precision rate , recall rate , F1 score , and Hausdorff distance , where . The testing results evaluated with these metrics are presented in Table 3. Our proposed method achieves better performance in most metrics than the other methods. A qualitative performance comparison across several compared methods and our proposed method is shown in Fig. 8. We also demonstrate the results of the proposed nuclei segmentation method with different seed detection methods in Fig. 10. It is notable that our proposed method is able to better capture nuclei boundaries than other methods for comparison. In particular, only our method can recover boundaries of overlapped nuclei by comparisons. This property makes our method superior to the other methods when analytics of occluded nuclei is crucial in investigations.
The performance comparison of different nuclei segmentation methods measured by region- and distance-based metrics on GBM40 dataset and TCGA dataset.
Note: Bold numbers indicate the best performance.
Parameter Sensitivity Analysis for Segmentation
We further investigate the segmentation contribution from individual terms in our variational model by testing with different parameters and comparing their associated segmentation results. The segmentation result with our default parameter setting is shown in Fig. 9(a). When , the shape prior fitting term does not take effect in the contour deformation process. The resulting segmentation outcome is presented in Fig. 9(b), where finalized nuclei contours are less regulated. Similarly, we can remove the occlusion penalty term from the variational model by setting . The associated result is illustrated in Fig. 9(c). Under this setting, the detected nuclei present a strong inclination to overlap with each other. When the shape prior fitting term , the dynamic occlusion penalty term , and the evolutionary stability term are all removed (), the resulting nuclei contours become significantly degraded, as shown in Fig. 9(d). Note that shapes appear to be less regulated in Figs. 9(b)–9(d) without one or more terms in the variational model. In addition to the final results with only a subset of terms in Figs. 9(b)–9(d), we also investigate the sensitivity of parameters to final results. Our investigations with parameter deviation from our proposed value set suggest that results remain similar even when we change by 22% (), 10% (), and by 67% (), as presented in Figs. 9(e)–9(g). Figure 9(h) presents results when and are simultaneously changed by 22% and 67%. Overall, a larger leads to a better contour convergence to true nuclei boundary by energy term ; a larger forces contours to look more from the reconstructed sparse shape priors by ; a larger tends to prevent nuclei more from overlapping with each other by .
To illustrate the effect of seed detection, we demonstrate the segmented nuclei contours in Fig. 10 by replacing the seed detection algorithm in our pipeline with other seed detection methods. Since the number of detected nuclei depends on the seed-based initialization, miss detection of seeds may result in undersegmentation of nuclei, as shown in Figs. 10(b) and 10(c). We also test the impact of the shape prior library size on segmentation by changing the number of shape prior clusters . The segmented nuclei contours in Fig. 11 show that the proposed method does not degrade significantly with less shape priors. In practice, we carefully choose the number of clusters so that is large enough to cover shape variations, but small enough to avoid high computation burden.
Limitation and Future Work
Although the proposed method can achieve good quantitative and visual results, it is limited by the following factors: (a) Nuclei contour evolution depends on accurate detection of seeds, therefore, the segmentation performance may degrade when seeds are not correctly detected. In cases where seeds are missing, our proposed method, like other level set based methods in literature, does not produce correct segmentation result as no level set function is correctly initialized for deformation. However, we have shown in our experimental tests that our proposed seed detection has a good performance with a very low missing rate, as suggested in Tables 1 and 2. When nuclei seeds are slightly shifted within nuclei regions, our approach is able to produce similar and correct segmented nuclei contours for overlapping nuclei. (b) Application of our method to whole slide images can be time-consuming compared to other state-of-the-art methods. To address the problem, we will further develop a MapReduce based high performance image analysis framework to make this process scalable and cost effective.43,44
In this paper, we propose a nuclei segmentation method based on the level set framework, aiming to simultaneously identify contours of multiple nuclei with mutual occlusion. We present our method with its application to nuclei segmentation using histopathologic images of glioblastoma brain tumors. First, a seed detection method is developed to automatically initialize nuclei contours. For better nuclei contour deformation, we incorporate into the model a set of typical nuclei shapes as prior information through spectral clustering on graph. Meanwhile, an adaptive nuclei occlusion penalty term is designed to dynamically penalize nuclei contour occlusion based on the number of overlapped nuclei. It also prevents recognized contours of overlapped nuclei from being identical. A numerical optimization algorithm is used to iteratively search for the desired level set functions. Experiments on glioblastoma brain histopathologic images produce better results as compared with other state-of-the-art experiments, suggesting the effectiveness of our detection and segmentation approach for nuclei analysis with glioblastoma histopathologic images.
This research is supported in part by grants from National Institute of Health under Grant No. K25CA181503, National Science Foundation under Grant Nos. ACI 1443054 and IIS 1350885, and CNPq.
Pengyue Zhang is a PhD candidate at the Department of Computer Science at Stony Brook University. His research interests include medical image analysis, computer vision, and machine learning.
Fusheng Wang is an associate professor at the Department of Biomedical Informatics and Department of Computer Science at Stony Brook University. He received his PhD in computer science from the University of California, Los Angeles, in 2004. Prior to joining Stony Brook University, he was an assistant professor at Emory University. His research interest crosscuts data management and biomedical informatics.
George Teodoro received his PhD in computer science from the Universidade Federal de Minas Gerais (UFMG), Brazil, in 2010. Currently, he is an assistant professor in the Computer Science Department at the University of Brasilia (UnB), Brazil. His primary areas of expertise include high performance runtime systems for efficient execution of biomedical and data-mining applications on distributed heterogeneous environments.
Yanhui Liang is works at Google Brain as a software engineer. She received her PhD in biomedical informatics from Stony Brook University. Her research interests include 2-D/3-D medical imaging, computer vision, machine learning, and large-scale spatial data analytics.
Mousumi Roy is a PhD student at Stony Brook University with research focus on computer vision, data analysis, and machine learning modeling for biomedical research.
Daniel Brat received his MD and PhD degrees from Mayo Medical and Graduate Schools. He completed residency and a fellowship at Johns Hopkins Hospital. He is Magerstadt professor and chair of pathology at the Northwestern University Feinberg School of Medicine and pathologist-in-chief of Northwestern Memorial Hospital. He directs a basic and translational research lab that investigates mechanisms of glioma progression, including the contributions of hypoxia, genetics, tumor microenvironment, and stem cells.
Jun Kong received a PhD in electrical and computer engineering from the Ohio State University. Currently, he is an associate professor in the Department of Mathematics and Statistics at Georgia State University. He is an affiliated faculty and a member of the Winship Cancer Institute in Emory university. His primary areas of expertise include biomedical image analysis algorithms, computer-aided diagnosis systems, and large-scale integrative analysis for quantitative biomedical research.