Open Access
1 March 2004 Skin image segmentation based on energy transformation
Liangen Zhu, Shiyin Qin, Fugen Zhou
Author Affiliations +

1.

Introduction

Human skin gradually changes as time flies by, it is very fresh and elastic in the young days. The tension of skin is lost as one grows older. Use of cosmetics that is suitable for the human skin surface condition of a respective human is effective to prevent skin aging. Recently, there are many choices of cosmetic products. It is very desirable to develop an automatic evaluation system for the human skin surface condition to show the effect of cosmetics.

The human skin surface has the pattern called grid texture. This pattern is composed of the valleys that spread vertically, horizontally, and obliquely and the hills are separated by valleys. Changes of the grid are closely linked to the condition of the human skin surface. They can serve as a good indicator for the skin condition. By measuring the skin grid using digital image processing technologies, we can evaluate the human skin surface about its aging, its health, and its alimentary status.

Besides the grid, there are also a lot of pits and bulges textured among the lines and some impurities like the sebum scattered on the skin surface. This complexity of dermal structure makes it difficult to extract the centerline. What is more, some external disadvantages caused in the acquisition of the skin image by clinical photographing, such as illuminating conditions, noise level, spherical effect, etc., make it more difficult. To our knowledge, no published work has done well in the extraction.

In this paper, the centerlines is detected first and a subsequent postprocessing is performed to form connected, closed regions. Through measuring and analyzing some parameters of the grid. the condition of the skin can be monitored. The paper is arranged as follows. In Sec. 2, we start with some characters of human skin. and on this basis, feature space transformation is described. In Sec. 3, an edge detection scheme is introduced. It is based on some features of maximal surface principal curvature. The postprocessing steps used to form closed regions are discussed in Sec. 4. In Sec. 5, an experimental result is presented, followed by the discussion about the efficiency and the advantages of the method. In Sec. 6, main conclusions are drawn at the end the paper.

2.

Feature Space Transformation

This operation is to map the original image from grayscale space to energy space. Figure 1(a) shows one original image of human skin, Fig. 1(b) illustrates the grayscale distribution curve of a typical row in Fig. 1(a), with the x coordinate denoting the pixels in a row, and the y coordinate denoting the corresponding gray values. Note from Fig. 1(b) that the skin grids are of triangle shape, like the shape of the roof-edge. This makes it impossible to use any traditional step-edge detection methods to extract the roof-edged grids in human skin. From these two images, we can also observe that the values are not of a uniform type. They include convex ridges [shown in Fig. 2(a)] concave valleys [shown in Fig. 2(b)], and saddle shape [shown in Fig. 2(c)]. It is the effect of human vision to contrastness that combines these mussy valleys and ridges into a network. This feature is typical in the clinical images, which is in fact more like a texture image than a grid one. Nevertheless, whatever they might be, they do form a grid in relatively constant width and they do own the same feature—the graylevels of grid pixels change more abruptly than that of nongrid pixels, and the curvatures at the grid pixels are greater than that in nongrid pixels. Therefore, a scheme is needed to trim the original images before extraction.

Figure 1

The grayscale distribution curve (b) corresponding to the pixels in the 40th row of the skin image (a).

017402j.1.jpg

Figure 2

The one-dimensional roof-edge signals occur in the human skin image: (a) ridge edge, (b) valley edge, (c) saddle edge.

017402j.2.jpg

First of all, we define pixel energy as follows:

Eq. (1)

Epixel=αEkinetic+βEpotential.

It is a linear compounding of two components, The first term denotes kinetic energy, and the second term potential energy. The kinetic energy E kinetic reflects local gradient information. That is, E kinetic is directly proportional to the grayscale change in a certain domain of the pixel. The more rapidly the graylevel changes, the greater the value of E kinetic is.

The potential energy E potential indicates the global spatial distribution of the gradient. It is directly proportional to the grayscale variation of the pixels in the whole image.

In Eq. (1), the coefficients α and β give, respectively, the strength of kinetic and potential forces. The choice of those two coefficients is very important. A tradeoff is needed to get a rather brilliant result. Usually, we choose parameter α to be proportional to the kinetic energy E kinetic , this is to guarantee a good ridge in energy space. In the meantime, we select parameter β to be proportional to the curvature at this point to avoid the influence caused by illumination fluctuations on transformation result, because the potential energy E potential is salient at the places where there is a big illumination fluctuation, and the import of curvature to the transformation can be used to correct this error.

The kinetic energy E kinetic and the potential energy E potential are defined, respectively, by the following forms.

Let I(s)=f(x,y), s(x,y)∈D be one curved surface of image grayscale defined on the domain D, be the average grayscale of the whole surface, for each point s(x,y)∈D, its energy components can be defined as follows:

Eq. (2)

Ekinetic=12maxs1,s2εδ[|I(S1)I(s2)|2],

Eq. (3)

Epotential=δ12A[I(s)I]¯2ds,
where δ is a subdomain which is centered at the point s(x,y), A is the area of the domain δ. S1, S2, S∈δ.

According to the definitions earlier, the feature image E pixel (x,y) corresponding to the original image f(x,y) can be calculated by

Eq. (4)

Epixel(x,y)=(x,y)Ekinetic(x,y)=β(x,y)Epotential(x,y).

When digitized, Eqs. (2) and (3) yield, respectively, the following forms:

Eq. (5)

Ekinetic(x,y)=12max2f(x,y),
where

Eq. (6)

maxf(x,y)=max(x1,y1),(xj,yj)δ[|f(xi,yi)f(xj,yj)|].

It denotes the maximal magnitude of the grayscale variation of the pixels in the window δ:

Eq. (7)

Epotential(x,y)=12N2Σxi=0N1Σyi=0N1|f(xi,yi)I|¯2,
where f(xi,yi) is the gray value for the pixel at coordinate (xi,yi), (xi,yi)∈δ, δ is a N×N window centered at the pixel with coordinate (x,y), is the average of the gray values for the whole image. the size N of window δ can be determined by the width of the values in the human skin images.

3.

Edge Detection

After feature space transformation, the original image is transformed into an energy image f [Fig. 3(a)], and the grid in original image is turned into ridges with textures among it. To avoid overconnection, a filtering is mandatory to smooth the inside texture while preserving the ridges. Here, we use a morphological filter based on reconstructions. The amount of filtering is an important parameter which determines the experiment results. The amount of filtering can be manually tuned to extract different lines. A severe filtering may produce a net of deep lines, and a slight filtering can produce a net of shallow lines, for more information about this filter, refer to Refs. 1 2 3 4 5 6.

Figure 3

A sequence of extraction (a) Energy image f of original image [Fig. 1(a)]. After feature space transformation and erosion-reconstruction morphological filtering. The net of ridge appears. (b) Edge detection using the normal curvature method.

017402j.3.jpg

The filtered image f can be considered as a gray-level surface S. The points located on the ridge centerline have a maximum normal curvature on the perpendicular direction of the ridge and a minimum normal curvature along the ridge, so the centerline can be identified by analyzing the normal curvatures of the surface S.

Now we put the surface S into a fixed coordinate system Oxyz, this surface has continuous third order partial derivatives. The function of the surface is: r=r(u,v). Let P be one point on the surface S, that is, P∈S. r/∂u,∂r/∂v and n are, respectively, tangent vectors and unit normal vector of S at point P. From the theorem of the Weingarten map, we know that, at point P, there exist two orthonormal eigenvectors of the Weingarten map {e 1,e 2} called principal directions at P, and two eigenvalues corresponding to these directions called the principal curvatures which is the extremum of the normal curvature at P, denoted by k1 (the maximum normal curvature) and k2 (the minimum normal curvature).

Point P falls on ridge centerline, if it satisfies the following conditions, k1≠0, and e 1 ⋅z=0, that is, the principal direction e 1 corresponding to the maximum normal curvature k1 is always perpendicular to the Z axis. So we can detect the ridge centerline by finding k1 and e 1.

From the theories of differential geometry, we can individually define the two fundamental forms of surface S at P as

the first fundamental form:

Eq. (8)

I=Edu2+2Fdudv+Gdv2

the second fundamental form:

Eq. (9)

II=Ldu2+2Mdudv+Ndv2
here
E=ruru,F=rurv,G=rvrv,
L=2ru2n,M=2ruvn,N=2rv2n.

Tangent vectors r/∂u,∂r/∂v span a plane [named tangent plan of S at P, denoted by Tp(S)] which comprises all tangent vectors of S at the point P. For each of the tangent directions at point P, there exists a curvature named normal curvature which represents the curvature of S in this direction. The directions to which the normal curvature take an extremum are named the principal directions, and the values of these extrema are named the principal curvatures. They are the roots of the following equations:

Eq. (10)

|LλEMλFMλFNλG|=0,

Eq. (11)

|Ldu+MdvEdu+FdvMdu+NdvFdu+Gdv|=0.
Equation (10) defines the two extrema of normal curvature λ1=k12=k2,(k1⩾k2), and Eq. (11) gives the directions to which the extrema of normal curvature are attained. Let H=1/2⋅(EN−2FM+GL)/(EG−F2) and K=(LN−M2)/(EG−F2), so the two principal curvatures k1,k2 can be expressed in terms of H,K as follows:

Eq. (12)

k1,k2=H±H2K.

From Eq. (11), we can get the two principal directions corresponding respectively to the principal curvatures λ1,2=k1,2:

Eq. (13)

dudv=Mk1,2FLk1,2E=Nk1,2GMk1,2F,

After the principal curvature k1 and the corresponding principal direction e 1 are calculated, if it satisfies k1≠0 and e 1 ⋅z=0, we think it is on the centerline. Figure 3(b) shows the centerlines superimposed on the original image.

4.

Postprocessing

Figure 3(b) is the result of edge detection. To show the effect more clearly, the background is eliminated in Fig. 4(a). Many fragmentary lines appear in this image, and there exist many discontinuous centerlines. Postprocessing is needed to remove those fragmentary lines and connect discontinuous lines into a closed grid. The cleaning process can be thought of as a tracking process. In the tracking process, the length of the lines are recorded, so that we can remove those lines whose length is shorter than a threshold.

Figure 4

The effect of postprocessing: (a) before and (b) after.

017402j.4.jpg

Finally, disconnected centerlines are connected to form closed contours. The basic strategy for connecting is summarized as follows.

  • • For each open contour, the unconnected end of the contour is identified, and the direction of the contour is recorded.

  • • Centered at the unconnected end, a searching area is located along the direction,

  • • If the nearest boundary element is found, a smooth boundary segment is generated to connect the open contour to the nearest boundary element.

The result of this postprocessing is shown in Fig. 4(b). A connected, closed grid of skin is obtained. This grid can be used to measure skin parameters.

5.

Experimental Results

To illustrate the validity of the method proposed in this paper, we use, respectively, traditional step-edge detectors and ridge detectors to extract central lines of the human skin grid, and make a comparison with the result of the proposed method. The example image is taken by a skin microscope from human arms by magnifying 30 times of the real skin, which is digitized to 256 gray-levels, 320×240 pixels. The comparison indicates that this method is more suitable for the skin analysis purpose.

There are a lot of methods on step-edge detection. Most of them are based on the idea that the grayscale gradient on the edge is more greater than that anywhere else. The Sobel operator is a representative one. Figure 5(a) shows the result of the Sobel operator performed on the human skin image. Note that this method can only get some discontinuous lines of the grid. These lines are not the centerlines of human creases, and cannot be used to characterize the human skin mathematically.

Figure 5

The results obtained, respectively, by using (a) Sobel operator, (b) zerocrossing method, (c) watershed transformation, (d) drainage pattern, and (e) the method proposed in this paper.

017402j.5.jpg

Ridge detection is of high interest in image analysis. Researchers have tried to describe it in various ways. Haralick7 and Koenderink8 identified ridges as loci of extremal height in two-dimensions, and is extended to d-D in Ref. 9 as the height condition. Gauch10 and Thirion11 identified ridges as positive maxima of the curvature of the relief’s level curves. The height and vertex of ridges are of local characterizations. In implementation, ridges can be obtained by looking for sign changes in zerocrossing functions, and labeling pixels which have a change of sign, afterwards, a thinning algorithm is needed to get 1-pixel wide line. Figure 5(b) shows the effect of the zerocrossing method. Notice that the result is unsatisfactory, it yields many spurious branches, and the obtained centerline has many discontinuities.

More recent works12 13 14 have brought global characterizations of ridges. These works think of ridges as a separatrix of objects. Watershed in Refs. 2 and 15 absorbed more attention. Figure 5(c) shows the performance of the Watershed transformation method. Although this method can yield closed, continuous regions, the result is not exciting yet. Arbitrary lines appear, and some ridges disappear. When used in skin monitoring they will cause a measure error.

There are also many other approaches on ridge detection, for example, Refs. 16 17 18 extract drainage patterns by simulating the flow of water over the Earth’s surface. Used in human skin segmentation, this method cannot divide the skin into closed regions, and yields many fine branches, Figure 5(d) shows the result. It is also unsatisfactory in measuring skin parameters.

In the experiment using the method proposed in this paper, we choose the parameters as follows: N=5 and a constant α=1, β=k1. Another parameter required in filtering is scale h which denotes the texture magnitude being filtered away. Different values of h will cause a different result. Figure 5(e) shows the result when h=7, and after postprocessing such as cleaning, connection to form some closed regions. We notice that this method can divide the skin surface into closed regions, and the centerlines are continuous, very little overdetection and underdetection appear, but some centerlines did not run through the center of the creases. This needs improving in the following work. Compared with other step-edge and ridge-edge approaches, this approach is effective in extracting human grid centerlines.

6.

Conclusions

In this paper, we have proposed a method for preprocessing the sophisticated human skin image. The characteristics of the original image have been described, and on the basis of these features, Feature space transformation is presented. The aim of the feature space transformation is to map the disordered grids into a neat net of ridges which will facilitate the subsequent operations. The morphological filter can remove the texture and spurious minima of the image to avoid overconnection. the amount of filtering is manually tuned for each application. Normal curvatures are calculated, and the centerlines are detected by analyzing the normal curvatures.

Some advantages and drawbacks have been reached through analyzing the experimental results. The method can divide the skin into continuous, closed regions, and the lines have a good localization, but sometimes it will yield arbitrary lines and lose genuine lines. Nevertheless, compared with other methods, this one is much suitable for the purpose of human skin monitoring.

REFERENCES

1. 

J. Serra, Image Analysis and Mathematical Morphology, Part 2, Academic, London (1988).

2. 

S. Beucher and F. Meyer, “The Morphological Approach to Segmentation: The Watershed Transform,” in Mathematical Morphology in Image Processing, Marcel Dekker, New York (1993).

3. 

1 J. Serra and L. Vincent , “An overview of morphology filtering,” Circuits Syst. Signal Process. , 11 (1992).

4. 

L. A. Cordeo, Differential Geometry, Pitman, New York (1985).

5. 

Chuan-chih Hsing, A First Course in Differential Geometry, Wiley, New York (1981).

6. 

M. P. do Carmo, Differential Geometry of Curves and Surfaces, Prentice-Hall, Englewood Cliffs, NJ (1976).

7. 

R. Haralick , “Ridges and valleys on digital images,” Comput. Vis. Graph. Image Process. , 22 (10), 28 –38 (1983). Google Scholar

8. 

J. Koenderink and A. Doorn , Proc. SPIE , 2031 2 –13 (1993). Google Scholar

9. 

D. Eberly , R. Gardner , B. Morse , S. Pizer , and C. Scharlach , “Ridges for image analysis,” J. Math. Imaging Vision , 4 (4), 353 –373 (1994). Google Scholar

10. 

J. Gauch and S. Pizer , “Multi-resolution analysis of ridges and valleys in grey-scale images,” IEEE Trans. Pattern Anal. Mach. Intell. , 15 (6), 635 –646 (1993). Google Scholar

11. 

J. P. Thirion and A. Gourdon , “Computing the differential characteristics of isointensity surfaces,” Comput. Vis. Graph. Image Process. , 61 (2), 190 –202 (1995). Google Scholar

12. 

L. Griffin , A. Colchedter , and G. Robinson , “Scale and segmentation of grey-level images using maximum gradient paths,” Image Vis. Comput. , 10 (6), 389 –402 (1992). Google Scholar

13. 

L. Nackman , “2-D critical point configuration graphs,” IEEE Trans. Pattern Anal. Mach. Intell. , 6 (4), 442 –450 (1984). Google Scholar

14. 

P. Rosin , “Early image representation by slope districts,” J. Visual Commun. Image Represent. , 6 (3), 228 –243 (1995). Google Scholar

15. 

L. Vincent and P. Soille , “Watersheds in digital spaces: an efficient algorithm based on immersion simulations,” IEEE Trans. Pattern Anal. Mach. Intell. , 13 (6), 583 –598 (1991). Google Scholar

16. 

F. Desmet and G. Govers , “Comparison of routing algorithms for digital elevation models and their implications for predicting ephemeral gullies,” Int. J. Geograph. Inf. Systems , 10 (3), 311 –331 (1996). Google Scholar

17. 

J. O’Callaghan and D. Mark , “The extraction of drainage networks from digital elevation data,” Comput. Vis. Graph. Image Process. , 28 (3), 323 –344 (1984). Google Scholar

18. 

P. Soille and C. Gratin , “An efficient algorithm for drainage network extraction on DEMs,” J. Visual Commun. Image Represent. , 5 (2), 181 –189 (1994). Google Scholar
©(2004) Society of Photo-Optical Instrumentation Engineers (SPIE)
Liangen Zhu, Shiyin Qin, and Fugen Zhou "Skin image segmentation based on energy transformation," Journal of Biomedical Optics 9(2), (1 March 2004). https://doi.org/10.1117/1.1646412
Published: 1 March 2004
Lens.org Logo
CITATIONS
Cited by 3 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Skin

Image filtering

Image segmentation

Edge detection

Ridge detection

Image processing

Sensors

Back to Top