Open Access
17 January 2020 Shape estimation of concave specular object from multiview polarization
Daisuke Miyazaki, Ryosuke Furuhashi, Shinsaku Hiura
Author Affiliations +
Abstract

We propose a method to estimate the surface normal of concave objects. The target object of our method has a specular surface without diffuse reflection. We solve the problem by analyzing the polarization state of the reflected light. The polarization analysis gives a constraint to the surface normal. However, polarization data from a single view has an ambiguity and cannot uniquely determine the surface normal. To solve this problem, the target object should be observed from two or more views. However, the polarization of the light should be analyzed at the same surface point through the different views. This means that both the camera parameters and the surface shape should be known. The camera parameters can be estimated a priori using known corresponding points. However, it is a contradiction that the shape should be known in order to estimate the shape. To resolve this problem, we assume that the target object is almost planar. Under this assumption, the surface normal of the object is uniquely determined. We show that the surface normal of the nonplanar part can also be estimated using the proposed method.

1.

Introduction

Factories in industrial fields have a high demand for estimating the shape of cracks since it is quite important for quality control of the products. Although there are many methods for detecting cracks,1 few methods have been proposed for estimating the shape of cracks. Therefore, there is a great demand for estimating the shape of concave objects of highly specular surfaces since it is a challenging task. This paper proposes a method that estimates the surface normal of a black specular object with a concave shape by analyzing the polarization state of the reflected light in which the target object is observed from multiple views.

Three-dimensional (3-D) modeling techniques have been intensively investigated in the field of computer vision. These techniques can be categorized into two types: the geometric approach and the photometric approach. The geometric approach uses the geometrical structure of the scene, such as time-of-flight (ToF) laser range sensor, multinocular stereo, or structured light projection. The photometric approach uses the light reflected from the scene, such as photometric stereo or shape from polarization. Shape from specularity has been extensively surveyed by Ihrke et al.2

A smooth surface normal can be obtained using a photometric approach. Polarization35 is one of the photometric clues that can be used to obtain a smooth surface normal. Koshikawa and Shirai6 used circular polarization to estimate the surface normal of a specular object. Guarnera et al.7 extended their method to determine the surface normal uniquely, by changing the lighting conditions in two configurations. Morel et al.8 also disambiguated it using multiple illumination; however, they did not solve the ambiguity of the degree of polarization (DOP)35 because they did not use circular polarization. Saito et al.9 proposed the basic theory for estimating the surface normal of a transparent object using polarization. Barbour10 approximated the relation between the surface normal and the DOP and developed a commercial sensor for shape from polarization. Kobayashi et al.11 estimated the surface normal of transparent thin objects using DOP. They also estimated the thickness by analyzing the light interference. Miyazaki et al.12 estimated the surface normal of a transparent object by analyzing the polarization state of the thermal radiation from the object. Miyazaki et al.13 attempted to estimate the surface normal of a diffuse object from a single view. Miyazaki et al.14 used a geometrical invariant to match the corresponding points from two views to estimate the surface normal of a transparent object. Miyazaki and Ikeuchi15 solved the inverse problem of polarization ray tracing to estimate the surface normal of a transparent object. These methods first calculate the polarization data from input images, while Yu et al.16 used the input images themselves to estimate the surface normal without explicitly calculating the DOP.

Wolff and Boult17 developed the basic theory for showing that polarization analysis can estimate a surface normal from two views if the corresponding points are known. Rahmann and Canterakis18 estimated the surface normal of a specular object from multiple views by iteratively finding the corresponding points of these views. Rahmann19 proved that only the quadratic surfaces are estimated if the corresponding points are searched iteratively. Atkinson and Hancock20 analyzed the local structure of an object to find the corresponding points between two view points to calculate the surface normal from the polarization of two views. Atkinson and Hancock21 also provided a detailed investigation of surface normal estimation for a diffuse object from a single view. Huynh et al.22 estimated not only the surface normal but also the refractive index.

Kadambi et al.23 combined the 3-D geometry obtained by a ToF sensor and the surface normal obtained from the DOP. Smith et al.24 combined the depth sensor and the shape from polarization.

Cui et al.25 used structure from motion, while Yang et al.26 used SLAM in addition to the shape from polarization. Miyazaki et al.27 combined the visual hull and the shape from polarization.

In this study, we propose a method for creating a 3-D model using both polarization analysis and planarity assumption. The principal target objects are smooth surfaces with high specular reflection and low diffuse reflection, which are annoying targets in conventional techniques. We first calibrate multiple cameras to calculate the geometrical relationships among them. We observe the object from multiple viewpoints using a polarization imaging camera. To determine the corresponding point among multiple views, we assume the target object as planar. However, this assumption simply produces only a planar shape; thus we also use polarization information to estimate the nonplanar part of the object. The shape-from-polarization method can estimate the shape of black objects with high specularity, which cannot be estimated using the photometric stereo method because there are no diffuse reflections. The polarization information of the object is obtained from multiple viewpoints using a polarization imaging camera. The polarization data must be analyzed at identical points on the object surface when observed from multiple viewpoints. Thus, the planarity assumption can be used for estimating the surface normal from polarization data. The target object of our method is almost planar except for a crack with a small size.

Miyazaki et al.’s method27 relies on the visual hull. It is difficult to estimate a planar shape using a visual hull; in addition, it is impossible to estimate a planar shape with infinite size [Fig. 1(a)]. Our method can also be applied to infinite plane [Fig. 1(b)]. Thus, our method overcomes the disadvantage of their method,27 which means that the proposed method is fundamentally superior to their method27 if the target object is almost planar.

Fig. 1

Our contribution: (a) previous method, which is based on visual hull that is not suited to estimating planar shapes and (b) the proposed method, which is suited to estimating the concave shapes that is almost planar.

JEI_29_4_041006_f001.png

We describe our method in Sec. 2 and present our results in Sec. 3. The theory shown in Sec. 2 assumes that the target object must be completely planar. However, Sec. 3 empirically proves that our method can successfully estimate the surface normal even if the object is not completely planar. We discuss the advantages and disadvantages of our method and conclude the paper in Sec. 4.

2.

Using Polarization in Estimating the Surface Normal of Concave Objects

2.1.

Algorithm Flow

First, we explain the flow of our method (Fig. 2).

Fig. 2

Algorithm flow.

JEI_29_4_041006_f002.png

Since we observe the target object from multiple viewpoints, we calibrate each viewpoint to obtain each camera parameter. Although any calibration pattern works well, this paper assumes that each camera is calibrated using four points marked at the vertices of a square for clarity. The area that is surrounded by these markers is the target area. Using these markers, we estimate the homography H (Sec. 2.6) and rotation R (Sec. 2.5). Figure 3 shows the homographic projection from each view to the canonical square. The canonical square is any square defined by the engineer.

Fig. 3

Transformation to canonical square.

JEI_29_4_041006_f003.png

A polarization camera captures the azimuth angle ϕ of the target object (Sec. 2.2). We denote the 90-deg rotation of ϕ as vector a, which would be orthogonal to the surface normal (Sec. 2.3). Using the vector a and rotation matrix R of the camera parameter, the surface normal n is calculated using singular value decomposition (SVD) (Sec. 2.4).

Finally, the surface normal is integrated to height field.15

2.2.

Polarization

We explain only linear polarization since circular polarization is not related to our method. Light is an electromagnetic wave, and an electromagnetic wave oscillating in only one direction is said to have perfectly linear polarization, while an electromagnetic wave oscillating isotropically in all directions is called unpolarized light. The intermediate state of such light is called partially polarized light. DOP35 is one of the metrics used to represent the polarization state of light. Its value varies from 0 to 1, with 1 representing perfectly polarized light and 0 representing unpolarized light.

The maximum light observed while rotating the polarizer is denoted as Imax, and the minimum light is denoted as Imin. In this paper, the polarizer angle at which Imin is observed is called the azimuth angle ϕ. The surface normal is represented in polar coordinates, where the azimuth angle is denoted as ϕ and the zenith angle is denoted as θ (Fig. 4). The azimuth angle calculated from the polarization has 180-deg ambiguity since a linear polarizer has a 180-deg cycle. Thus, the azimuth angle of the surface normal will be either ϕ or ϕ+180  deg. The plane consisting of the incident light and surface normal vectors is called the reflection plane. The reflected light vector is also coplanar with the reflection plane since the surface is optically smooth. The orientation of the reflection plane is the same as the azimuth angle ϕ and ϕ+180  deg, which is defined on a certain xy plane and is defined as an angle between the x axis and the reflection plane projected on the xy plane. Since we capture images with a camera, the x axis and the y axis of the “image” coordinates is used.

Fig. 4

Polar coordinates of surface normal.

JEI_29_4_041006_f004.png

2.3.

Calculating the Surface Normal from Two Viewpoints

Section 2.2 described the relationship between the surface normal and the azimuth angle obtained from polarization. However, we cannot determine the surface normal uniquely because only the orientation of the reflection plane including the surface normal is obtained. We must observe the object from two viewpoints to solve this problem.

Figure 5 represents the situation of our problem. A camera has its coordinate system x axis, y axis, and z axis. The camera’s z axis is along the optical axis. The reflection plane angle ϕ is the angle between the x axis of the camera coordinate system and the line caused by the intersection between the reflection plane and the xy plane.

Fig. 5

Relationship between the surface normal and the reflection plane when observed from a single viewpoint.

JEI_29_4_041006_f005.png

We analyze the two reflection plane angles at the same surface point, corresponding to the known 3-D geometry. Our method assumes that the 3-D geometry of the target object is almost a plane. The relationship between the surface normal vector and the azimuth angle is shown in Fig. 6. The relationship between the azimuth angles for each of the cameras, represented as ϕ1 and ϕ2, and the normal vector of the reflection plane, represented as a1 and a2, is shown in Eq. (1).

Eq. (1)

a1=(cos(ϕ1+90  deg)sin(ϕ1+90  deg)0),a2=(cos(ϕ2+90  deg)sin(ϕ2+90  deg)0).

Fig. 6

Relationship between the surface normal and the reflection plane when observed from two viewpoints.

JEI_29_4_041006_f006.png

As shown in Fig. 6, the surface normal n is orthogonal to the vectors a1 and a2. After projecting the vectors a1 and a2 to the world coordinate system, we calculate the surface normal n. The rotation matrix projecting the world coordinate system to each camera coordinate system is represented as R1 and R2. The inverse of each of these rotation matrices is its transpose, and they project back from the camera coordinate system to the world coordinate system. Thus, this situation is represented as Eq. (2).

Eq. (2)

(a1R1a2R20)(nxnynz)=(000).
Namely, the world coordinate of a1 and a2 are R1a1 and R2a2. Since R1a1 and R2a2 are orthogonal to the surface normal n, (R1a1)·n=0 and (R2a2)·n=0 hold. These formulae can be expressed, in another form, as a1R1n=0 and a2R2n=0 [Eq. (2)].

2.4.

Calculating the Surface Normal from Multiple Viewpoints

This section explains the estimation process for the surface normal from the azimuth angle obtained from multiple viewpoints.

Figure 7 shows the relationship between the surface normal n of the surface point p and the azimuth angle obtained from K viewpoints. In Fig. 7, ϕk represents the azimuth angle of the surface point p observed by the camera k=(1,2,,K), and ak represents the vector orthogonal to the reflection plane under the coordinate system of the camera k.

Fig. 7

Relationship between the surface normal and the azimuth angle observed from multiple viewpoints.

JEI_29_4_041006_f007.png

The rotation matrix Rk represents the transformation from the world coordinate system to the local coordinate system of the camera indicated by k. Similar to Eq. (2), Eqs. (3) or (4) holds.

Eq. (3)

(a1R1a2R2aKRK)(nxnynz)=(000),
Or, in an other form,

Eq. (4)

An=0.
The surface normal n, which satisfies Eq. (4) in the least-squares sense, is estimated using SVD.28 The K×3 matrix A is decomposed by SVD as follows:

Eq. (5)

(a1R1a2R2aKRK)=UWVT=U(w1w20)(v1v2v3).
Here, U is a K×3 orthogonal matrix, W is a 3×3 diagonal matrix with non-negative values, and V is a 3×3 orthogonal matrix. The diagonal element wi of matrix W is the singular value of matrix A, and the singular vector corresponding to wi is vi. Owing to the relationship between the surface normal and the reflection planes, the rank of matrix A is at most 2; thus, one of the three singular values becomes 0 (see Miyazaki et al.,27 for the proof). The surface normal n is represented as Eq. (6),28 which is calculated from the singular vector that has the smallest singular value, namely, the third row of V in Eq. (5).

Eq. (6)

n=sv3.
In general, s is an arbitrary scalar coefficient; however, since the surface normal and the singular vectors are normalized vectors, s would be either +1 or 1. Whether s is positive or negative is determined so that the surface normal faces toward the camera. The surface normal estimated by Eq. (6) is the optimal value that minimizes the squared error of Eq. (4) formulated by K equations. The input data must be obtained from two or more viewpoints since the rank of matrix A is 2.

2.5.

Camera Parameters

Equation (3) or Eq. (4) calculates the surface normal from the azimuth angle under multiple viewpoints. To solve Eq. (4), the azimuth angle should be analyzed at corresponding points among multiple viewpoints. The corresponding points are determined by homography, as shown in Sec. 2.6. Equation (4) also requires the rotation matrices of each camera, namely, the extrinsic parameter of each camera should be known.

Our paper represents the projection from 3-D vertex (X,Y,Z) to two-dimensional vertex (x,y) as Eq. (7).29

Eq. (7)

(xy1)(f000f0001)(r11r12r13t1r21r22r23t2r31r32r33t3)(XYZ1).
In Eq. (7), we do not describe the camera center parameter (Cx,Cy) for clarity since we assume the pinhole camera model. We do not explain the detailed implementation to estimate these parameters f,t1,t2,t3,r11,r12,,r33.

2.6.

Homography Transform

Homography is a projection from a certain quadrangle to another certain quadrangle represented under the homographic projection. Homography represents one-to-one correspondence between two planes without redundancy or lack of information. Therefore, it is natural to use homography in our work since the target object is almost planar.

The homogeneous coordinate is defined as follows using (ξ1,ξ2,ξ3) (ξ30), where one element is added to the coordinates (x,y).

Eq. (8)

x=ξ1ξ3,y=ξ2ξ3.

Homographic projection from a certain quadrangle (x,y) to another certain quandrangle (x,y) is represented as follows:

Eq. (9)

(xy1)(ξ1ξ2ξ3)=(h11h12h13h21h22h23h31h32h33)(xy1).
Namely, homographic projection is represented by homography matrix h11,h12,,h33. Point (x,y) is projected to the point (x,y)=(ξ1/ξ3,ξ2/ξ3) by this homography matrix.

Eq. (10)

x=h11x+h12y+h13h31x+h32y+h33,

Eq. (11)

y=h21x+h22y+h23h31x+h32y+h33.

Figure 8 is an example in which vertices of quadrangle (x1,y1), (x2,y2), (x3,y3), and (x4,y4) correspond to vertices of quadrangle (x1,y1), (x2,y2), (x3,y3), and (x4,y4).

Fig. 8

Homographic projection from a certain quadrangle to another certain quadrangle.

JEI_29_4_041006_f008.png

Scaling the 3×3 homography matrix h11,h12,,h33 results in the same transformation. Thus, we fix one element as follows to uniquely determine the homography matrix:

Eq. (12)

h33=1.
Substituting the above equation into Eqs. (10) and (11) results in Eqs. (13) and (14).

Eq. (13)

xh11+yh12+h13xxh31yxh32=x,

Eq. (14)

xh21+yh22+h23xyh31yyh32=y.

Concatenating Eqs. (13) and (14) for four vertices results in Eq. (15).

Eq. (15)

(x1y11000x1x1y1x1000x1y11x1y1y1y1x2y21000x2x2y2x2000x2y21x2y2y2y2x3y31000x3x3y3x3000x3y31x3y3y3y3x4y41000x4x4y4x4000x4y41x4y4y4y4)(h11h12h13h21h22h23h31h32)=(x1y1x2y2x3y3x4y4).
Since we have eight unknowns (h11,h12,,h32) and eight equations [eight rows of the leftmost matrix in Eq. (15)], a closed-form solution exists. Solving this results in the homography matrix shown below:

Eq. (16)

H=(h11h12h13h21h22h23h31h321).
Using the homography matrix H [Eq. (16)], the corresponding points between two quadrangles (Fig. 8) are expressed by Eq. (9). Suppose that the homography of camera 1 is H1 and that of camera 2 is H2. Figure 3 shows the homographic projection from each view to the canonical square. The canonical square is any square defined by the engineer. Suppose that the pixel position of the canonical square is (x,y). The corresponding points of camera 1 (x1,y1) and camera 2 (x2,y2) are calculated as follows:

Eq. (17)

(x1y1)H1(xy),(x2y2)H2(xy).
Namely, the two points (x1,y1) and (x2,y2) correspond while the point (x,y) acted as a mediator.

3.

Experiment

3.1.

Experimental Setup

As shown in Fig. 9, the target object is surrounded by white material. This white environment acts as a light source and illuminates the target object from every direction. Cheap foaming polystyrene is used in our experiment, and it is located not strictly but roughly. Once we set this white enclosure, we do not need to move it like with photometric stereo, which needs to move the light sources. The white board is illuminated by ordinary room light which is set in an ordinary room. Often, the white board is unnecessary since wall, floor, and ceiling act as an illuminator.18

Fig. 9

Experimental environment.

JEI_29_4_041006_f009.png

The camera used is shown in Fig. 10 and Table 1. Since we have only one camera (because a polarization camera is expensive), we rotate the target object instead of rotating the camera. Note that observing a target object with multiple cameras and observing the target object rotated in multiple angles with a single camera are mathematically the same.

Fig. 10

Polarization camera.

JEI_29_4_041006_f010.png

Table 1

Specification of the camera.

ManufacturerFluxData Inc., New York
Product nameFD-1665P
SensorSony ICX414
Resolution659×494
Pixel size9.9×9.9  μm
Configuration0, 45, 90 linear polarizer
Frame rate74 fps
InterfaceIEEE-1394b

3.2.

Pseudo-Color Representation of the Result

The following sections show some results of our method. For visualization, the azimuth angle and the surface normal are represented by pseudo-color. Figures 11(a) and 11(b) show the pseudo-color representation of the azimuth angle and the surface normal of an ideal hemisphere, respectively.

Fig. 11

Pseudo-color representation of an ideal sphere: (a) azimuth angle and (b) surface normal.

JEI_29_4_041006_f011.png

3.3.

Result of Ellipsoid

The target object is shown in Fig. 12. We generated the object using a 3D printer, so we can compare the result with the ground truth, which is the digital data input to the 3D printer. The size of the square is 10×10  cm, the diameter of the long axis of the ellipse is 7.5 cm, the diameter of the short axis of the ellipse is 2.5 cm, and the maximum deepness of the concave part is 0.625 cm. The unique characteristic of our method is that we can estimate the shape of the cracks. First, we evaluate the performance of the proposed method. To guarantee the statistical reliability, we need to estimate the surface normal with wide variety and wide area. This is why we first measure the concave ellipsoid.

Fig. 12

Target object (ellipsoid).

JEI_29_4_041006_f012.png

We took one image from each of 15 different directions (Fig. 13). Pseudo-color representation of the surface normal of our method is shown in Fig. 14 and that of the ground truth is shown in Fig. 15. Note that our method successfully estimated the shape, which is almost the same as the true shape. The estimated shape is shown in Figs. 16 and 18(c), while the ground truth is shown in Figs. 17 and 18(a). The error is calculated as the angle between two surface normals of the estimated and the ground truth. Error is shown in Fig. 19(b); the average error was 4.49 deg.

Fig. 13

Input image (ellipsoid).

JEI_29_4_041006_f013.png

Fig. 14

Estimated surface normal (ellipsoid).

JEI_29_4_041006_f014.png

Fig. 15

Ground truth of surface normal (ellipsoid).

JEI_29_4_041006_f015.png

Fig. 16

Estimated shape (ellipsoid).

JEI_29_4_041006_f016.png

Fig. 17

Ground truth of shape (ellipsoid).

JEI_29_4_041006_f017.png

Fig. 18

Intersection shape (ellipsoid): (a) ground truth, (b) photometric stereo, and (c) proposed method.

JEI_29_4_041006_f018.png

Fig. 19

Estimation error (ellipsoid): (a) photometric stereo and (b) proposed method.

JEI_29_4_041006_f019.png

3.4.

Comparison to Photometric Stereo

To prove the effectiveness of our method, we compare our method with the result of the photometric stereo method.30

Photometric stereo from 15 lights is applied to the object shown in Fig. 12, and the input images are shown in Fig. 20. The surface normal of the photometric stereo method is shown in Fig. 21, and the estimated shape is shown in Figs. 18(b) and 22. Photometric stereo assumes Lambertian reflection, though the actual reflection is specular reflection, and thus, the shape is distorted. The error is shown in Fig. 19(a), and the average error was 42.3 deg. Since our error is 4.49 deg, the performance of our method is better.

Fig. 20

Input image of photometric stereo (ellipsoid).

JEI_29_4_041006_f020.png

Fig. 21

Surface normal of photometric stereo (ellipsoid).

JEI_29_4_041006_f021.png

Fig. 22

Shape of photometric stereo (ellipsoid).

JEI_29_4_041006_f022.png

3.5.

Result of Convex Object

Our method can be applied not only to concave objects but also to convex objects. To prove the wide applicability of our method, we measure a convex object. The target object is shown in Fig. 23. The input images of our method are shown in Fig. 24, while those of photometric stereo are shown in Fig. 25. The surface normal of the ground truth, the photometric stereo method, and the proposed method is shown in Figs. 26Fig. 2728. The shape of the ground truth, the photometric stereo method, and the proposed method is shown in Figs. 29Fig. 3031. The cross section of the shape of the ground truth, the photometric stereo method, and the proposed method is shown in Figs. 32(a)32(c). The error of the photometric stereo method and the proposed method is shown in Figs. 33(a) and 33(b). The average error of the photometric stereo method was 47.0 deg, while that of the proposed method was 12.9 deg.

Fig. 23

Photograph of target object (convex).

JEI_29_4_041006_f023.png

Fig. 24

Input data of our method (convex).

JEI_29_4_041006_f024.png

Fig. 25

Input images of photometric stereo (convex).

JEI_29_4_041006_f025.png

Fig. 26

Surface normal of ground truth (convex).

JEI_29_4_041006_f026.png

Fig. 27

Surface normal of photometric stereo (convex).

JEI_29_4_041006_f027.png

Fig. 28

Surface normal of proposed method (convex).

JEI_29_4_041006_f028.png

Fig. 29

Shape of ground truth (convex).

JEI_29_4_041006_f029.png

Fig. 30

Shape of photometric stereo (convex).

JEI_29_4_041006_f030.png

Fig. 31

Shape of proposed method (convex).

JEI_29_4_041006_f031.png

Fig. 32

Intersection shape (convex): (a) ground truth, (b) photometric stereo, and (c) proposed method.

JEI_29_4_041006_f032.png

Fig. 33

Estimation error (convex): (a) photometric stereo and (b) proposed method.

JEI_29_4_041006_f033.png

3.6.

Result of Stripes

To evaluate the performance of our method depending on the width of cracks, three different concave shapes with different widths are measured. Figure 34 shows the target object, and Fig. 35 shows the input images. Also, 15 images are taken, one for each direction. The surface normal of our method is shown in Fig. 36, and that of ground truth is shown in Fig. 37. The estimated shape is shown in Figs. 38 and 40(b), while ground truth is shown in Figs. 39 and 40(a).

Fig. 34

Target object (stripe).

JEI_29_4_041006_f034.png

Fig. 35

Input images (stripe).

JEI_29_4_041006_f035.png

Fig. 36

Estimated surface normal (stripe).

JEI_29_4_041006_f036.png

Fig. 37

Ground truth of surface normal (stripe).

JEI_29_4_041006_f037.png

Fig. 38

Estimated shape (stripe).

JEI_29_4_041006_f038.png

Fig. 39

Ground truth of shape (stripe).

JEI_29_4_041006_f039.png

Fig. 40

Intersection shape (stripe): (a) ground truth and (b) estimated shape.

JEI_29_4_041006_f040.png

The error map is shown in Fig. 41; the average error was 7.18 deg.

Fig. 41

Error of our method (stripe).

JEI_29_4_041006_f041.png

3.7.

Result of Worm

To simulate an acutal situation, we applied our method to a crack that is not shaped in a straight line. Figs. 42 and 43 show the target object and the input images, respectively, and Figs. 44 and 45 show the estimated surface normal and the estimated shape, respectively.

Fig. 42

Target object (worm).

JEI_29_4_041006_f042.png

Fig. 43

Input images (worm).

JEI_29_4_041006_f043.png

Fig. 44

Estimated surface normal (worm).

JEI_29_4_041006_f044.png

Fig. 45

Estimated shape (worm).

JEI_29_4_041006_f045.png

3.8.

Discussion

As shown in Fig. 41, the narrower the concave part is, the worse the result is. This is because the light is not illuminated satisfactorily in the narrow concave part. In addition, inter-reflection becomes strong at the narrow concave part.

4.

Conclusion

We propose a shape estimation method from polarization images obtained from multiple viewpoints. The proposed method computes the surface normal using SVD to minimize the least-squared error. It can estimate the shapes of the concave part of planar objects that are black and have high specularity. It is usually difficult to estimate the shape of a planar object with small details; however, our algorithm fully utilizes the property that the target object is almost planar. What is interesting in our method is that, even if we assume that the object is planar, the shape of the concave part is also successfully determined.

The experiments show that our method can estimate the shape of the crack. This property demonstrates that our method is useful for investigating product inspection in factories, damage inspection in architecture, age estimation from skin wrinkles, and so on. For example, factories want to know the reason for the defect of the product since they want to fix the problem and decrease the defects. To analyze the reason, the shape of the defects is necessary, and our method is useful for this purpose.

The disadvantage of our method is that the shape where the light has not reached cannot be estimated. However, this disadvantage does not apply only to our method but also to any method in the image processing field since an image cannot be observed if the scene is not illuminated. Our future work will develop a measurement system that illuminates the target object from any directions.

Acknowledgments

This research was supported in part by the Ministry of Education, Culture, Sports, Science, and Technology KAKENHI Grant Nos. 15H05925 and 15H02727 and in part by the Japan Society for the Promotion of Science KAKENHI Grant No. 24700176.

References

1. 

A. Mohan and S. Poobal, “Crack detection using image processing: a critical review and analysis,” Alexandria Eng. J., 57 787 –798 https://doi.org/10.1016/j.aej.2017.01.020 AEJAEB (2018). Google Scholar

2. 

I. Ihrke et al., “Transparent and specular object reconstruction,” Comput. Graphics Forum, 29 (8), 2400 –2426 https://doi.org/10.1111/cgf.2010.29.issue-8 CGFODY 0167-7055 (2010). Google Scholar

3. 

M. Born and E. Wolf, Principles of Optics, Pergamon Press, London (1959). Google Scholar

4. 

E. Hecht, Optics, Addison-Wesley, Reading, Massachusetts (2002). Google Scholar

5. 

W. A. Shurcliff, Polarized Light: Production and Use, Harvard University Press, Cambridge, Massachusetts (1962). Google Scholar

6. 

K. Koshikawa and Y. Shirai, “A model-based recognition of glossy objects using their polarimetrical properties,” Adv. Rob., 2 (2), 137 –147 https://doi.org/10.1163/156855387X00129 (1987). Google Scholar

7. 

G. C. Guarnera et al., “Estimating surface normals from spherical Stokes reflectance fields,” Lect. Notes Comput. Sci., 7584 340 –349 https://doi.org/10.1007/978-3-642-33868-7 LNCSD9 0302-9743 (2012). Google Scholar

8. 

O. Morel et al., “Active lighting applied to three-dimensional reconstruction of specular metallic surface by polarization imaging,” Appl. Opt., 45 (17), 4062 –4068 https://doi.org/10.1364/AO.45.004062 APOPAI 0003-6935 (2006). Google Scholar

9. 

M. Saito et al., “Measurement of surface orientations of transparent objects by use of polarization in highlight,” J. Opt. Soc. Am. A, 16 (9), 2286 –2293 https://doi.org/10.1364/JOSAA.16.002286 JOAOD6 0740-3232 (1999). Google Scholar

10. 

B. A. Barbour, “Apparatus and method for extracting information from electromagnetic energy including target 3D structure and materials,” U S. Patent, 8320661 (2012).

11. 

Y. Kobayashi et al., “Reconstructing shapes and appearances of thin film objects using RGB images,” in IEEE Conf. Comput. Vision and Pattern Recognit., 3774 –3782 (2016). https://doi.org/10.1109/CVPR.2016.410 Google Scholar

12. 

D. Miyazaki et al., “Determining surface orientations of transparent objects based on polarization degrees in visible and infrared wavelengths,” J. Opt. Soc. Am. A, 19 (4), 687 –694 https://doi.org/10.1364/JOSAA.19.000687 JOAOD6 0740-3232 (2002). Google Scholar

13. 

D. Miyazaki et al., “Polarization-based inverse rendering from a single view,” in Proc. IEEE Int. Conf. Comput. Vision, 982 –987 (2003). https://doi.org/10.1109/ICCV.2003.1238455 Google Scholar

14. 

D. Miyazaki, M. Kagesawa and K. Ikeuchi, “Transparent surface modeling from a pair of polarization images,” IEEE Trans. Pattern Anal. Mach. Intell., 26 (1), 73 –82 https://doi.org/10.1109/TPAMI.2004.1261080 ITPIDJ 0162-8828 (2004). Google Scholar

15. 

D. Miyazaki and K. Ikeuchi, “Shape estimation of transparent objects by using inverse polarization raytracing,” IEEE Trans. Pattern Anal. Mach. Intell., 29 (11), 2018 –2030 https://doi.org/10.1109/TPAMI.2007.1117 ITPIDJ 0162-8828 (2007). Google Scholar

16. 

Y. Yu, D. Zhu and W. A. Smith, “Shape-from-polarisation: a nonlinear least squares approach,” in Proc. IEEE Int. Conf. Comput. Vision, 2969 –2976 (2017). https://doi.org/10.1109/ICCVW.2017.350 Google Scholar

17. 

L. B. Wolff and T. E. Boult, “Constraining object features using a polarization reflectance model,” IEEE Trans. Pattern Anal. Mach. Intell., 13 (7), 635 –657 https://doi.org/10.1109/34.85655 ITPIDJ 0162-8828 (1991). Google Scholar

18. 

S. Rahmann and N. Canterakis, “Reconstruction of specular surfaces using polarization imaging,” in Proc. IEEE Comput. Soc. Conf. Comput. Vision and Pattern Recognit., 149 –155 (2001). https://doi.org/10.1109/CVPR.2001.990468 Google Scholar

19. 

S. Rahmann, “Reconstruction of quadrics from two polarization views,” in Proc. Iberian Conf. Pattern Recognit. and Image Anal., 810 –820 (2003). https://doi.org/10.1007/978-3-540-44871-6_94 Google Scholar

20. 

G. A. Atkinson and E. R. Hancock, “Shape estimation using polarization and shading from two views,” IEEE Trans. Pattern Anal. Mach. Intell., 29 (11), 2001 –2017 https://doi.org/10.1109/TPAMI.2007.1099 ITPIDJ 0162-8828 (2007). Google Scholar

21. 

G. A. Atkinson and E. R. Hancock, “Recovery of surface orientation from diffuse polarization,” IEEE Trans. Image Process., 15 (6), 1653 –1664 https://doi.org/10.1109/TIP.2006.871114 IIPRE4 1057-7149 (2006). Google Scholar

22. 

C. P. Huynh, A. Robles-Kelly and E. R. Hancock, “Shape and refractive index from single-view spectro-polarimetric images,” Int. J. Comput. Vision, 101 64 –94 https://doi.org/10.1007/s11263-012-0546-3 IJCVEQ 0920-5691 (2013). Google Scholar

23. 

A. Kadambi et al., “Polarized 3D: high-quality depth sensing with polarization cues,” in Proc. IEEE Int. Conf. Comput. Vision, 3370 –3378 (2015). https://doi.org/10.1109/ICCV.2015.385 Google Scholar

24. 

W. Smith, R. Ramamoorthi and S. Tozza, “Height-from-polarisation with unknown lighting or albedo,” IEEE Trans. Pattern Anal. Mach. Intell., 41 (12), 2875 –2888 https://doi.org/10.1109/TPAMI.2018.2868065 ITPIDJ 0162-8828 (2018). Google Scholar

25. 

Z. Cui et al., “Polarimetric multi-view stereo,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit., 1558 –1567 (2017). https://doi.org/10.1109/CVPR.2017.47 Google Scholar

26. 

L. Yang et al., “Polarimetric dense monocular slam,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit., 3857 –3866 (2018). https://doi.org/10.1109/CVPR.2018.00406 Google Scholar

27. 

D. Miyazaki et al., “Surface normal estimation of black specular objects from multiview polarization images,” Proc. SPIE, 56 (4), 041303 https://doi.org/10.1117/1.OE.56.4.041303 PSISDG 0277-786X (2016). Google Scholar

28. 

W. H. Press et al., Numerical Recipes in C: The Art of Scientific Computing, Cambridge University Press, Cambridge (1997). Google Scholar

29. 

R. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, Cambridge University Press, Cambridge (2004). Google Scholar

30. 

R. J. Woodham, “Photometric method for determining surface orientation from multiple images,” Opt. Eng., 19 (1), 139 –144 https://doi.org/10.1117/12.7972479 (1980). Google Scholar

Biography

Daisuke Miyazaki is an associate professor at Hiroshima City University. He received his BS and MS degrees in science from the University of Tokyo in 2000 and 2002, respectively, and his PhD in information science and technology in 2005. His research interests include physics-based vision. He is a member of ACM and IEEE.

Biographies of the other authors are not available.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Daisuke Miyazaki, Ryosuke Furuhashi, and Shinsaku Hiura "Shape estimation of concave specular object from multiview polarization," Journal of Electronic Imaging 29(4), 041006 (17 January 2020). https://doi.org/10.1117/1.JEI.29.4.041006
Received: 27 September 2019; Accepted: 27 December 2019; Published: 17 January 2020
Lens.org Logo
CITATIONS
Cited by 6 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Cameras

Polarization

Matrices

Specular reflections

Reflection

Imaging systems

Polarized light

Back to Top