Open Access
5 July 2018 Pose estimation method for planar mirror based on one-dimensional target
Wenlei Liu, Sentang Wu, Xiaolong Wu
Author Affiliations +
Abstract
This paper presents a pose estimation method for planar mirror based on a one-dimensional (1-D) target. First, a 1-D target is maneuvered in front of a camera so as to allow the camera to capture several images of both the original and virtual feature points on the target. Next, the normal vector of the mirror and the distance between the mirror and the camera are calculated by a closed-form method. Last, a nonlinear optimization is used to refine the closed-form solution by minimizing the reprojection error of reflection projection. Synthetic data and real experiments show that the proposed method is accurate and reliable. The proposed method is simple and flexible. No prior information about the target motion and camera pose is assumed. A minimum of two nonparallel target positions are required. It is suitable for the pose estimation of planar mirrors, especially when they are in fixed positions. We can find the posture of the fixed camera according to the normal vector of the plane mirror.

1.

Introduction

Vision sensors have advantages of flexibility and noncontact measurement. Planar mirrors are utilized in a wide variety of applications, such as fringe-reflection measurement,1 planar catadioptric vision sensors,2 stereovision by a single camera, and calibration of multiple cameras with nonoverlapping fields of view (FOVs).3,4 A common major challenge of these applications is the pose estimation of the planar mirror. It is important and necessary to develop accurate and flexible pose estimation methods for planar mirrors.

There has been some research on the pose estimation of planar mirrors. Lebraly et al.5 used a planar mirror to create an overlap between cameras by attaching markers on the mirror’s surface. However, pasting markers on the mirror was inconvenient and troublesome. Sturm et al.,6 Rodrigues et al.,7 Hesch et al.,8 and Kosuke Takahashi et al.9 proposed different pose estimation methods based on the fact that successive reflections in two-mirror planes were a fixed-axis rotation around the line of intersection of two mirror planes. Reference 10 proposed a practical visual inertial calibration method, which is based on visual observations taken in front of a planar mirror. Reference 11 presented a design and construction of a new system for stereovision with a single CCD camera. Reference 12 proposed a method for the calibration, which is used the orthogonal phase-shifted sinusoidal patterns during the calibration process.

Xiao et al.1 calculated the transformation between the camera and the LCD screen by a markerless flat mirror. Li et al.13 measured the pose of a flat mirror by capturing the reflected structured light pattern. The translation vector was calculated by the mirror motion without any tilt. However, pure translations can hardly be achieved without external equipment. Ren et al.14 calibrated a stereodeflectometry system, where two cameras captured the fringe patterns produced by an LCD screen reflected by a flat mirror. The above methods require that the mirror is placed at several positions. However, these methods fail when the camera and the mirror are in fixed positions.

In Ref. 15, an algorithm is based on an orthogonality constraint, which should be satisfied by all families of reflections of a single reference object. It can return a unique solution of three P3P problems from three mirrored images. Reference 16 presents a method that utilizes reasonable assumptions about geometric features such as a plane or an intersecting line of two planes. It obtains an initial solution via singular value decomposition (SVD), which is then refined via nonlinear optimization, which does not require any assumption about structures but estimates the parameters of geometric features. In Ref. 17, an accurate closed-form solution with explicit geometric meanings is proposed. It minimizes an error metric based on all reflections of rotation, and then derives the optimal solution that gives the real camera rotation directly by SVD of the sum of mirrored camera rotations. Then, it can minimize object-space collinearity error.

This paper is also inspired by Shah18 and Li et al.17 Shah found the best rotation matrix with determinant 1. In contrast, we seek the best orthogonal matrix with determinant 1, and the best reflection matrix can be constructed by a matrix with determinant 1. Li et al.17 calculated the camera rotation by minimizing the error of all rotation reflections. In this paper, we initialize the normal vector by minimizing the error of reflections of the unit direction vectors.

In this paper, we focus on the pose estimation of the planar mirror based on a one-dimensional (1-D) target. 1-D target has advantages of simple structure and flexibility. First, instead of moving the mirror, we move the target in front of the camera to provide multiple (at least two) views of both the original and virtual targets. The camera and the mirror are fixed during the process. Next, the normal vector of the planar mirror is estimated based on the direction vectors of the original and virtual targets. The distance between the camera and the mirror is calculated based on reflection equations. Last, the closed-form solution thus obtained is refined using nonlinear optimization by minimizing the reprojection error of reflection projection. Compared with conventional methods, the proposed method is suitable for the pose estimation of planar mirrors, especially when camera and the mirror are in fixed positions. After obtaining the position information in the plane mirror, the camera can be observed the carrier virtual image through plane mirror, then the rotation matrix, and translation vector of the image machine coordinate system {C} and the carrier coordinate system {B} can be calculated. The calibration of the installation parameters of the image machine is completed.

The rest of this paper is organized as follows: preliminary work is introduced in Sec. 2. The principle of the proposed method is described in Sec. 3. Synthetic data and real-data experiments are performed in Sec. 4. The conclusions are given in Sec. 5.

2.

Preliminaries

2.1.

Camera Model

Perspective projection is used as the camera model. The parameters and coordinates of the perspective projection model are shown in Fig. 1. The relation between a spatial point P=[X,Y,Z]T in {C} and the corresponding image point p=[u,v]T can be described as follows:

Eq. (1)

s[p1]=KP,K=[fx0u00fyv0001],
where s is a scale factor, K is the intrinsic matrix in camera. fx and fy are the equivalent focal lengths in horizontal and vertical directions, respectively. (u0,v0) denotes the principal point.

Fig. 1

The perspective projection model and coordinate relation.

OE_57_7_073101_f001.png

As shown in Fig. 2(a), both the original and the virtual targets are in the camera’s FOV. Ocxcyczc is the coordinate frame of the camera, denoted as {C}.

Fig. 2

The 1-D target is placed in front of the camera. (a) Projections of the original and virtual targets at the i’th position and (b) reflection through the planar mirror.

OE_57_7_073101_f002.png

2.2.

Pose Estimation of One-Dimensional Target

As shown in Fig. 2, Pi,j and P˜i,j denote the spatial coordinates of the j’th original and the j’th virtual feature points at the i’th position in {C}, respectively. pi,j and p˜i,j are the corresponding image coordinates, i=1,,M, j=1,,N. M is the number of target positions and N is the number of feature points on the target.

The original target and the virtual target at i’th position are uniquely characterized by Pi,1, di and P˜i,1,d˜i, respectively. di and d˜i are the unit direction vectors of the original and virtual targets at ith position, respectively.

For the ith position, Pi,1 and di can be obtained by the method proposed by Liu et al.19,20 It requires that at least three collinear feature points on the target. Suppose Pi,1(x1,y1,z1), Pi,2(x2,y2,z2), and Pi,3(x3,y3,z3), then

Eq. (2)

x3x1x2x1=y3y1y2y1=z3z1z2z1=ρ,
where ρ=Pi3Pi1/Pi2Pi1 can be computed from the distances between the points Pi,1, Pi,2, Pi,3.

Suppose that in the image coordinate system, pi,1(u1,v1), pi,2(u2,v2), and pi,3(u3,v3). Combining Eqs. (1) and (2), we can obtain

Eq. (3)

{(u1u0)z1=x1fx(v1v0)z1=y1fy(u2u0)z2=x2fx(v2v0)z2=y2fy(u3u0)z3=x3fx(v3v0)z3=y3fy(ρ1)x1=ρx2x3(ρ1)y1=ρy2y3(ρ1)z1=ρz2z3.

By solving Eq. (3), the coordinates of pi,1(x1,y1,z1), pi,2(x2,y2,z2), and pi,3(x3,y3,z3) can be retrieved and the initial value of the normal vector di can be obtained. Due to the image noise, if there are more than three feature points, results of different point triplets, e.g., (Pi,1,Pi,2,Pi,3) and (Pi,1,Pi,2,Pi,4), are not equivalent.

Let p^i,j be the image coordinates of the reprojection feature point

Eq. (4)

s[p^i,j1]=K[Pi,1+(j1)Ldi],
where s is a scale factor, L is the known interval of two adjacent feature points on the target, and K denotes the intrinsic matrix in camera.

To utilize the redundant information of feature points, Pi,1 and di are refined by minimizing the following reprojection error using Levenberg–Marquardt algorithm:21

Eq. (5)

f(Ω)=j=1Nd2(pi,j,p^i,j),
where Ω=(Pi,1,di). di=[cosθicosφi,sinφi,cosθisinφi]T. θi and φi are the pitch angle and yaw angle of the i’th unit direction vector with respect to {C}, respectively. d(·) is the distance function between two points. The initial value of (Pi,1,di) can be obtained by Liu’s method.19

The pose of the i’th virtual target (P˜i,1,d˜i) can also be obtained similarly.

2.3.

Planar Mirror Reflection

As shown in Fig. 2(b), camera {C} is in front of a planar mirror Π. The mirror is uniquely determined by its unit normal vector n and the distance l with respect to {C}.2,17 A spatial point P lies on the mirror plane if and only if it satisfies

Eq. (6)

nTP=l.

The Householder matrix H22,23 is widely used in the mirror reflection model, and its equation is as follows:

Eq. (7)

H=I2nnT.

The Householder transformation is a linear transformation that transforms a vector into a mirror image reflected by a hyperplane. The Householder transformation can zero some elements of a vector and keep the norm of the vector constant.

The schematic diagram of a mirror surface emission model is shown in Fig. 3. The equation of the plane Π is as follows:

Eq. (8)

nTx=0  xΠ.

Fig. 3

The schematic diagram of a mirror surface emission model.

OE_57_7_073101_f003.png

If xΠ, then

Eq. (9)

Hx=(I2nnT)x=x.

If yΠ, then

Eq. (10)

Hy=(I2nnT)(x+kn)=x+kn2  knnTn=xkn=y˜.

The reflection in Π can be represented by the following matrix:

Eq. (11)

S=[H2ln01].

Through the conversion of S can obtain the position of the virtual mirror P˜ corresponding to the space point P in the camera coordinate system. The space point P and the corresponding virtual point are related as follows:

Eq. (12)

[P˜1]=S[P1].

3.

Principle of the Mirror Pose Estimation

3.1.

Analysis of the Unique Solution

The calculation of the normal vector n is equivalent to solving the Householder matrix H. According to Eqs. (11) and (12), di and d˜i are related by the Householder matrix as follows:

Eq. (13)

d˜i=Hdi.

From Eq. (13), one position is not enough to obtain a unique solution. For two positions, we have

Eq. (14)

d˜1=Hd1,d˜2=Hd2.

According to the principle of mirror image, assume that target real image is positioned for the right-handed coordinate system, and in the target virtual image, the left and right, upside down, and up and down directions remain the same with respect to the left-hand target image coordinate system.4 Therefore, the relationship between real image d1×d2 and virtual image d˜1×d˜2 is as follows:

Eq. (15)

d˜1×d˜2=H(d1×d2).

Combining Eqs. (14) and (15), we have

Eq. (16)

[d˜1d˜2d˜×d˜2]=H[d1d2d1×d2].

If d1 is not parallel to d2, matrix [d1d2d1×d2] is invertible. Thus, H can be uniquely determined as follows:

Eq. (17)

H=[d˜1d˜2d˜×d˜2][d1d2d1×d2]1.

If d1 is parallel to d2, the target motion is a pure translation, det [d1d2d1×d2]=0, thus a unique solution cannot be found.

3.2.

Computing the Normal Vector n

If the target is placed more than two times, the best Householder matrix can be computed from the solution to the orthogonal Procrustes problem.21 According to Eq. (13), the sum of the normal vector error is

Eq. (18)

E=i=1Mei=i=1Md˜iHdiF2,
where ·F denotes the Frobenius norm, xF2=tr(xTx)=tr(xxT). If measurements are noiseless, E=0.

The best approximation of matrix H is equivalent to the matrix that solves the minimization problem of Eq. (18). According to the properties of the Householder matrix, Eq. (18) is expanded as

Eq. (19)

E=i=1Mtr[(d˜iTdiTHT)(d˜iHdi)]=2M2tr(HΔ),
where Δ=i=1Mdid˜iT.

It can be seen that minimization of E is equivalent to maximize tr(HΔ). Therefore, the matrix H that solves the minimization problem of Eq. (18) is equivalent to H that solves

Eq. (20)

maxHtr(HΔ).

There are some researches18,22 on finding the best rotation matrix with determinant +1. However, the Householder matrix is not a rotation but actually a reflection with determinant 1. To construct the best reflection, a lemma is introduced.

Lemma 1

For a given 3×3 matrix Δ and a Householder matrix H

Eq. (21)

tr(HΔ)tr(DΣ),
where UΣVT is the full SVD of Δ, the singular values are in decreasing order, and

Eq. (22)

D=diag[1,1,det(VUT)].

Proof

Using tr(AB)=tr(BA) and the SVD result of matrix Δ, we have

Eq. (23)

tr(HΔ)=tr(HUΣVT)=(VTHUDDΣ),
where D2=I is the identity matrix. Let H^=VTHUD.

Eq. (24)

H^H^T=(VTHUD)(VTHUD)T=I,

Eq. (25)

det(H^)=det(VTHUD)=det(VT)det(H)det(U)det(D)=[det(VT)2][det(U)]2det(H)=1.

According to Eqs. (24) and (25), we can come to the conclusion that H^ is an orthogonal matrix with determinant 1 and hence a rotation matrix.

Suppose that H^=[h11h1jhi1hij], Σ=diag(σ1,σ2,σ3), we have

Eq. (26)

tr(HΔ)=tr(H^DΣ)=σ1h11+σ2h22σ3h33det(VUT).

Therefore, the trace is maximized if hii=1. As H^ is an orthogonal matrix, this means that H^ would have to be the identity matrix. Then

Eq. (27)

tr(HΔ)σ1+σ2σ3det(VUT)=tr(DΣ).

According to two sets of the original and virtual direction vectors, the best reflection matrix can be obtained by Theorem 1.

Theorem 1.

The solution to the maximization problem of Eq. (20) is as follows:

Eq. (28)

H=VDUT,
where UΣVT is the full SVD of Δ with decreasing singular values and D is subject to Eq. (22).

Proof.

From Lemma 1, the maximization problem of Eq. (20) is achieved if a Householder matrix H can be constructed such that

Eq. (29)

tr(HΔ)=tr(DΣ).
Let H=VDUT, then tr(HΔ)=tr(VDUTUΣVT)=tr(DΣ).

When the Householder matrix H is determined, the unit normal vector n is obtained.

3.3.

Computing the Distance l

From Eq. (12), we have

Eq. (30)

2ln=P˜i,jHPi,j.

Therefore, the distance l can be computed as

Eq. (31)

l=12MNi=1Mj=1N[nT(P˜i,jHPi,j)],
where Pi,j and P˜i,j can be calculated according to the Eq. (3). H can be calculated according to Eq. (17).

3.4.

Nonlinear Optimization

The results obtained in Sec. 3.2 and Sec. 3.3, which were called as the closed-form solution. The closed-form calculation is fast, but it is sensitive to the noise. In this subsection, the closed-form solution is refined with nonlinear optimization by minimizing the reprojection error of reflection. According to the camera model, we have

Eq. (32)

s[pi,j1]=K[I0][I2nnT2ln01][pi,j1],
where s is a scale factor, K denotes the intrinsic matrix in camera, I is a 3×3 identity matrix, and 0 is a 1×3 zero vector. pi,j is the reprojected image coordinates.

The mirror normal vector n and the distance l are refined by minimizing the following function using the Levenberg–Marquardt algorithm21

Eq. (33)

f(Ω)=i=1Mj=1Nω2(p˜i,j,p͡i,j),
where Ω=(n,l), n=[cosθcosφ,sinθ,cosθsinφ]T. θ and φ are the pitch angle and yaw angle of the normal vector n with respect to {C}, respectively. ω(·) is the distance between image points and p˜ij is the image coordinate corresponding to the imaginary point pi,j. After optimization, the optimal value of the normal vector n and the distance l is obtained.

3.5.

Calculating the Relative Position of the Image Machine {C} and the Carrier {B}

Figure 4 shows the relationship diagram of the vector coordinate system {B}, the vector image frame {B}, and the machine coordinate system {C}, where the RBC and RBC stand for the rotation vector, the tBC and tBC represent the translation vector, and the S mean mirror reflection matrix.

Fig.4

The relationship diagram of the vector coordinate system {B}, the vector image frame {B}, and the machine coordinate system {C}.

OE_57_7_073101_f004.png

The image of the mounting carrier in the mirror is observed by the plane mirror, so as to obtain the transformation relationship between the plane frame {C} and the vector virtual frame {B}. According to the above results, the transformation relationship between the image coordinate system {C} and the carrier image coordinates {B} can be obtained directly. Thus, the calibration of the installation parameters of the imager can be completed.

Eq. (34)

[RBCtBC01]=[RBCtBC01][I2nnT2ln01],
where R is the rotation matrix, t is the translation vector, subscripts BC and BC, respectively, represent the conversion relationship between the camera coordinate system {C} and the virtual carrier coordinate {B}, and the transformation relationship between the camera coordinate system {C} and the carrier image coordinates {B}.

3.6.

Method Summary

The calibration method of the installation parameters of the image machine based on the plane mirror is proposed in this section, which keeps the relative position of the image machine and the plane mirror unchanged. The main processes of the proposed method are as follows:

  • Step 1: The intrinsic parameters of the camera are calibrated using the Bouguet Camera Calibration Toolbox based on Zhang’s method.24,25 The internal parameters of the camera mainly include the following: fx,fy,u0,v0.

  • Step 2: The 1-D target with N feature points is placed in front of the camera for M times. For each position, one image is taken by the camera. Both the original and virtual targets are in the camera’s FOV. All the images are undistorted according to the intrinsic calibration results. Thus, image coordinates are distortionless.

  • Step 3: The optimal estimation value of matrix H can be obtained by Orthogonal Procrustes problem. For the ith position, compute the value of Pi,1,di and P˜i,1,d˜i, respectively. Repeat this step for i=1,,M.

  • Step 4: By obtaining the minimum value of the reprojection error function, the optimal value of the mirror normal vector n and the image machine-mirror distance l are obtained.

  • Step 5: The rotation matrix RBC and the translation vector tBC between the image machine coordinate system and the carrier coordinate system are calculated, and the calibration of the installation parameters of the image machine is completed.

4.

Experimental Results

Through the previous derivation, the mirror position (the mirror vector n and the image machine-mirror distance l) are the key parameters in the calibration process of the installation parameters of the image machine. Therefore, this section takes the pose error of plane mirror as the accuracy evaluation index, analyzes the related factors that affect the accuracy through simulation data and physical experiments, and compares it with other calibration methods.

4.1.

Simulation Data

To evaluate the accuracy, the estimation errors are defined as follows:

  • (a) Error of the mirror normal vector n against the truth value ng is defined as follows:

    Eq. (35)

    En=arccos(nTng).

  • (b) Error of the distance l against the truth value lg is defined as follows:

    Eq. (36)

    El=llg.

  • (c) Reprojection error is defined as follows:

    Eq. (37)

    Eq=i=1Mj=1Np˜i,jp^i,j22MN.

In this experiment, one simulated camera and a planar mirror are used to construct the system. The intrinsic parameters of the camera are fx=fy=995.556, u0=512, v0=384. The image resolution is 1024  pixel×768  pixel. The mirror normal vector n=[0.5,0,0.866]T, i.e., θ=0, φ=120  deg. The distance l=400  mm, the interval L=25  mm.

The positions of target are randomly generated. Only the original and virtual targets in the camera’s FOV are used. Gaussian noise with zero mean and different standard deviation σ is added to the image coordinates of feature points. The effects of several factors on the accuracy of the pose estimation are evaluated quantitatively by synthetic data as follows.

4.1.1.

Influence of the characteristic point noise on the accuracy

In this experiment, M=10, N=9. Gaussian noise is added to the image coordinates of the feature points. The noise level varies from 0 to 1.0 pixel. For each level, 100 independent trials are performed to compute the relative errors of En, El and RMS errors of Eq.

From Fig. 5, the error increases linearly with the noise level. It indicates that the closed-form solution can be effectively adjusted by the proposed method. If the noise level of a system is <0.5  pixel, relative errors of the normal vector n and the distance l are <0.06% and 0.5%, respectively, which are applicable for common applications.

Fig. 5

Error of pose estimation versus the noise level: (a) the relative error of En versus noise level, (b) the relative error of El versus noise level, and (c) the RMS error of Eq versus noise level.

OE_57_7_073101_f005.png

4.1.2.

Influence of the number of target placement on the accuracy

In this experiment, Gaussian noise with σ=0.2  pixel is added to the image coordinates of the feature points. M varies from 2 to 12, N is set to 9. For each level, 100 independent trials are performed and are used to compute relative errors.

From Fig. 6, the error decreases with the increasing number of target positions. When the target is placed only twice, the pose estimation error is rather high. The estimation error decreases rapidly as the number of positions increases. When there are enough positions, the contribution of additional positions decreases.

Fig. 6

Error of pose estimation vs. the number of target positions M (a) the relative error of En versus the number of target positions, (b) the relative error of El versus the number of target positions, and (c) the RMS error of Eq versus the number of target positions.

OE_57_7_073101_f006.png

4.1.3.

Influence of the number of target feature points on the accuracy

In this experiment, Gaussian noise with σ=0.2  pixel is added to the image coordinates of the feature points. M is set to 10, N varies from 3 to 16. For each level, 100 independent trials are performed and are used to evaluate relative errors.

From Fig. 7, increasing number of feature points benefits the estimation accuracy. When there are only three feature points on the target, the estimation error reaches maximum. The accuracy improves sharply when N increases from 3 to 8. When N>10, the accuracy stables at the certain level.

Fig. 7

Error of pose estimation versus the number of feature points N (a) the relative error of En versus the number of feature points, (b) the relative error of El versus the number of feature points, and (c) the RMS error of Eq versus the number of feature points.

OE_57_7_073101_f007.png

4.2.

Physical Experiments

As shown in Fig. 8(a), a Canon 60D digital camera and a planar mirror are in fixed positions. The image resolution of the camera is 1920  pixel×1280  pixel. The intrinsic parameters of the vision sensor are calibrated using the Bouguet’s calibration toolbox,15 as shown in Table 1. We use the intersection of two rows of checkerboards equivalently to the 1-D target. The distance between neighboring feature points in the checkerboard is 20 mm, i.e., L=20  mm. The two-row checkerboard is used to define 1-D target for feature extraction.

Fig. 8

(a) An image captured by the camera. Both the original and virtual 1-D target are in the camera’s FOV and (b) a checkerboard is used to evaluate the accuracy.

OE_57_7_073101_f008.png

Table 1

Intrinsic parameters of the camera.

fxfyu0v0
Camera15481548960640
Bouguet1561.2931560.153972.497597.061
Error (%)0.860.791.36.7

As shown in Fig. 8(b), a checkerboard is placed between the camera and the mirror. Both the original and virtual checkerboards are in the camera’s FOV. The spatial coordinates of the corner points of the checkerboards can be obtained by the Bouguet’s toolbox. In this experiment, reflective error Es and the reprojection error Eq are used to evaluate the accuracy

Eq. (38)

Es=1(nx+1)(ny+1)i=1nx+1j=1ny+1Q˜i,j(HQi,j+2ln),
where Qi,j and Q˜i,j are the spatial coordinates of the original and virtual corner points of the checkerboards, with respect to {C}. nx and ny are the number of squares along two orthogonal directions, respectively.

To verify the effect of several factors shown in the synthetic experiment, pose estimations of different position numbers and different feature points are implemented.

4.2.1.

Influence of the number of target placement on the accuracy

In this experiment, the number of target positions M is changed from 2 to 12, the number of feature points N is set to 9. From Fig. 9, the result is similar to the result of the synthetic experiment in Sec. 4.1.2. An increasing number of target positions benefits the accuracy, especially when the number of target positions is relatively small.

Fig. 9

Error of pose estimation versus the number of target positions M (a) the reflective error Es versus target position number and (b) the reprojection error Eq versus target position number.

OE_57_7_073101_f009.png

4.2.2.

Influence of the number of target feature points on the accuracy

In this experiment, the number of target positions M is set to 10, the number of feature points N is changed from 3 to 12. From Fig. 10, it also indicates a similar result of the synthetic experiment in Sec. 4.1.3. The estimation error peaks at N=3, which corresponds to the minimum quantity of feature points. Then, the estimation error decreases sharply with the increasing number of feature points. When N>8, the estimation error flattens out gradually.

Fig. 10

Error of pose estimation versus the number of feature points N (a) the reflective error ES versus the number of feature points and (b) the reprojection error Eq versus the number of feature points.

OE_57_7_073101_f010.png

4.2.3.

Discussions

Evidently, there exists a trade-off between the accuracy and the workload. According to the effect, analysis of the number of target positions M and the number of feature points N, M=10 and N=9 is a balanced choice. When the error of an image point is 0.5 pixel, the synthetic experiment shows that relative errors of the normal vector n and the distance l are 0.06% and 0.45%, respectively. It is acceptable for common applications such as the camera-to-base calibration.8

In addition, there is a gap between the results of synthetic data and real-data experiments. There may be some reasons for this. First, the feature point extraction method is a commonly used method with good generality and acceptable accuracy. Extraction algorithms with higher accuracy improve the accuracy, which will be further investigated. Second, the mirror used in this paper is a domestic appliance, not a precision optical device. Precision mirrors can be used to improve the accuracy.

4.3.

Comparisons

There are many methods for the pose estimation. Here, three typical methods are described as follows:

  • Method 1: This method is similar to the proposed method. As shown in Fig. 8(b), a checkerboard is used as the calibration target, instead of the 1-D target. The corner points are extracted and calibrated by the Bouguet’s toolbox. The pose of the mirror is also refined using nonlinear optimization. We use method 1 as the criterion for evaluating the merits of the proposed method.

  • Method 2: As shown in Fig. 11, a reference object is placed on the ground and remains fixed. Then, the pose of the mirror can be calculated by Li’s method.17

Fig. 11

A virtual image of the reference object.

OE_57_7_073101_f011.png

For each method, 10 images are captured to estimate the mirror pose. The reflective error Es is used to evaluate the accuracy. The results of different methods are shown in Table 2.

Table 2

The estimation errors of different methods.

Proposed method (M=10, N=9)Closed-form solution (M=10, N=9)Method 1Method 2
Es/mm1.416.381.375.71

From Table 2, the comparisons show that the proposed method achieves a good accuracy. Compared with the closed-form solution, the closed solution is simple and fast, but it is more sensitive to noise and larger in error. The same line determined by the selection of different points is different. Therefore, the normal vectors determined by these lines will also have certain differences, which will produce certain errors. Compared with the checkerboard of method 1, the errors of the two methods are relatively small. However, the number of feature points of 1-D target is reduced a lot, under the condition that the computation accuracy is not decreased. Therefore, according to the 1-D target of the known size, the calibration of the imaging machine can be achieved without reducing the precision, and the limitation of the complex calibration tool can be removed, so that the calibration is more flexible and convenient. The accuracy of method 3 is roughly the same as the closed-form solution. However, method 3 fails when the camera and the mirror cannot be moved.

5.

Conclusion

In this paper, a pose estimation method for planar mirrors is presented. The method only needs a simple 1-D target and a digital camera. First, the camera captures the target in different positions. The pose of the target in each position is obtained. Then, the normal vector and the distance between the camera and the mirror are calculated by a closed-form method, followed by a nonlinear optimization.

Our method is simple and flexible. A minimum of two nonparallel target positions and three collinear feature points on the target are needed. No prior information about the target motion and the camera’s pose is required. The proposed method is accurate and robust to noises. Both synthetic data and real experiments demonstrate the effectiveness of our method. In addition, both the camera and the mirror do not need to move during the process, so it is suitable for application where the camera and the mirror need to remain fixed.

In some applications, the transformation between the body frame and the camera frame need to be accurately calibrated. However, cameras mounted on the body usually cannot observe the body directly. We plan to apply our method for the camera-to-body calibration in the future.

Acknowledgments

The authors acknowledge the support by the Industrial Technology Development Program under Grant No. B1120131046.

References

1. 

Y. Xiao, X. Su and W. Chen, “Flexible geometrical calibration for fringe-reflection 3D measurement,” Opt. Lett., 37 (4), 620 –622 (2012). https://doi.org/10.1364/OL.37.000620 OPLEDP 0146-9592 Google Scholar

2. 

G. L. Mariottini et al., “Planar mirrors for image-based robot localization and 3-D reconstruction,” Mechatronics, 22 398 –409 (2012). https://doi.org/10.1016/j.mechatronics.2011.09.004 MECHER 0957-4158 Google Scholar

3. 

X. Feng and D. Pan, “Research on the application of single camera stereo vision sensor in three-dimensional point measurement,” J. Mod. Opt., 62 (15), 1204 –1210 (2015). https://doi.org/10.1080/09500340.2015.1024775 JMOPEW 0950-0340 Google Scholar

4. 

R. K. Kumar et al., “Simple calibration of non-overlapping cameras with a mirror,” in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition, 2602 –2608 (2008). Google Scholar

5. 

P. Lebraly et al., “Flexible extrinsic calibration of non-overlapping cameras using a planar mirror: Application to vision-based robotics,” in Proc. of IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, (IROS, 2010), 5640 –5647 (2010). Google Scholar

6. 

P. Sturm and T. Bonfort, “How to compute the pose of an object without a direct view?,” Lect. Notes Comput. Sci., 3852 21 –31 (2006). https://doi.org/10.1007/11612704_3 Google Scholar

7. 

R. Rodrigues, J. P. Barreto and U. Nunes, “Camera pose estimation using images of planar mirror reflections,” Lect. Notes Comput. Sci., 6314 382 –395 (2010). https://doi.org/10.1007/978-3-642-15561-1_28 Google Scholar

8. 

J. A. Hesch, A. I. Mourikis, S. I. Roumeliotis, “Mirror-based extrinsic camera calibration,” Algorithmic Foundations of Robotics VIII, 57 Springer, Berlin, Heidelberg (2009). https://doi.org/10.1007/978-3-642-00312-7_18 Google Scholar

9. 

K. Takahashi, S. Nobuhara and T. Matsuyama, “Mirror-based camera pose estimation using an orthogonality constraint,” IPSJ Trans. Comput. Vision Appl., 8 11 –19 (2016). https://doi.org/10.2197/ipsjtcva.8.11 Google Scholar

10. 

G. Panahandeh, M. Jansson and P. Händel, “Calibration of an IMU-camera cluster using planar mirror reflection and its observability analysis,” IEEE Trans. Instrum. Meas., 64 (1), 75 –88 (2015). https://doi.org/10.1109/TIM.2014.2329388 Google Scholar

11. 

T. P. Pachidis and J. N. Lygouras, “Pseudostereo-vision system: a monocular stereo-vision system as a sensor for real-time robot applications,” IEEE Trans. Instrum. Meas., 56 (6), 2547 –2560 (2007). https://doi.org/10.1109/TIM.2007.908231 Google Scholar

12. 

Y. Xu et al., “A calibration method for non-overlapping cameras based on mirrored absolute phase target,” Int. J. Adv. Manuf. Technol., (2018). https://doi.org/10.1007/s00170-018-1704-8 Google Scholar

13. 

L. Li et al., “Flat mirror tilt and piston measurement based on structured light reflection,” Opt. Express, 22 (22), 27707 –27716 (2014). https://doi.org/10.1364/OE.22.027707 OPEXFF 1094-4087 Google Scholar

14. 

H. Ren, F. Gao and X. Jiang, “Iterative optimization calibration method for stereo deflectometry,” Opt. Express, 23 (17), 22060 –22068 (2015). https://doi.org/10.1364/OE.23.022060 OPEXFF 1094-4087 Google Scholar

15. 

J. Y. Bouguet, “Camera calibration toolbox for Matlab,” (2013) http://www.vision.caltech.edu/bouguetj Google Scholar

16. 

T. Matsuyama, S. Nobuhara and K. Takahashi, “A new mirror-based extrinsic camera calibration using an orthogonality constraint,” in IEEE Conf. on Computer Vision and Pattern Recognition, 1051 –1058 (2012). Google Scholar

17. 

X. Li et al., “Accurate mirror-based camera pose estimation with explicit geometric meanings,” Sci. China-Technol. Sci., 57 (12), 2504 –2513 (2014). https://doi.org/10.1007/s11431-014-5713-z Google Scholar

18. 

M. Shah, “Comparing two sets of corresponding six degree of freedom data,” Comput. Vis. Image Understanding, 115 (10), 1355 –1362 (2011). https://doi.org/10.1016/j.cviu.2011.05.007 CVIUF4 1077-3142 Google Scholar

19. 

Z. Liu et al., “Novel calibration method for non-overlapping multiple vision sensors based on 1D target,” Opt. Lasers Eng., 49 (4), 570 –577 (2011). https://doi.org/10.1016/j.optlaseng.2010.11.002 Google Scholar

20. 

Z. Liu, F. J. Li and G. J. Zhang, “An external parameter calibration method for multiple cameras based on laser rangefinder,” Measurement, 47 954 –962 (2014). https://doi.org/10.1016/j.measurement.2013.10.029 0263-2241 Google Scholar

21. 

J. J. Moré, “The Levenberg–Marquardt algorithm: implementation and theory,” Numerical Analysis, 105 –116 Springer(1978). Google Scholar

22. 

F. Zhang, “Matrix analysis and applications,” Int. J. Inf. Syst. Sci., 309 (1), i (2004). Google Scholar

23. 

K. Kanatani, “Analysis of 3-D rotation fitting,” IEEE Trans. Pattern Anal. Mach. Intell., 16 (5), 543 –549 (1994). https://doi.org/10.1109/34.291441 ITPIDJ 0162-8828 Google Scholar

24. 

Z. Y. Zhang, “A flexible new technique for camera calibration,” IEEE Trans. Pattern Anal. Mach. Intell., 22 (11), 1330 –1334 (2000). https://doi.org/10.1109/34.888718 ITPIDJ 0162-8828 Google Scholar

25. 

Y. Bok et al., “Extrinsic calibration of non-overlapping camera-laser system using structured environment,” in IEEE/RSJ Int. Conf. on Intelligent Robots & Systems, 436 –443 (2014). Google Scholar

Biography

Wenlei Liu received his BS degree at Shandong University of Science and Technology, Qingdao, in 2012 and his MS degree at Beihang University, Beijing, in 2016. Currently, he is pursuing his PhD at Beihang University, Beijing, China. He mainly studies the high-precision visual relative navigation, and also studies the data fusion of multi data sensors to achieve the purpose of cooperative detection and guidance.

Sentang Wu received his PhD in dynamics, ballistics, and aircraft motion control systems from National Aviation University, Ukraine, in 1992. Currently, he is a professor of automation science and electrical engineering, and a PhD tutor at Beihang University, Beijing, China. He is also the navy missile expert in the National Defense Basic Research Institute and a member of the academic committee. His research interests include the theory and application of nonlinear stochastic systems, computer information processing and control, aircraft cooperative control, precision, and guidance.

Xiaolong Wu received his BS degree in Sichuan University, Chengdu, in 2007 and his PhD in Beihang University, Beijing, in 2017. He mainly studies the information acquisition system of the missile’s autonomous formation, including the relative navigation method of the missile's autonomous formation, the calibration of the vision sensor’s installation parameters and the optimization of the field of view, and the cooperative detection and guidance of the missile's autonomous formation.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Wenlei Liu, Sentang Wu, and Xiaolong Wu "Pose estimation method for planar mirror based on one-dimensional target," Optical Engineering 57(7), 073101 (5 July 2018). https://doi.org/10.1117/1.OE.57.7.073101
Received: 28 March 2018; Accepted: 12 June 2018; Published: 5 July 2018
Lens.org Logo
CITATIONS
Cited by 2 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Mirrors

Cameras

Calibration

Error analysis

Imaging systems

Optical engineering

Fiber Bragg gratings

RELATED CONTENT


Back to Top