Instrumentation, Techniques, and Measurement

Improving the accuracy of phase-shifting techniques

[+] Author Affiliations
William Cruz-Santos, Lourdes López-García

CU-UAEM Valle de Chalco, Hermenegildo Galeana 3, Valle de Chalco, Estado de México 56615, Mexico

Arturo Redondo-Galvan

CU-UAEM Zumpango, Camino Viejo a Jilotzingo Valle Hermoso Zumpango, Estado de México 55600, Mexico

Opt. Eng. 54(5), 054102 (May 13, 2015). doi:10.1117/1.OE.54.5.054102
History: Received January 15, 2015; Accepted April 23, 2015
Text Size: A A A

Open Access Open Access

Abstract.  The traditional phase-shifting profilometry technique is based on the projection of digital interference patterns and computation of the absolute phase map. Recently, a method was proposed that used phase interpolation to the corner detection, at subpixel accuracy in the projector image for improving the camera–projector calibration. We propose a general strategy to improve the accuracy in the search for correspondence that can be used to obtain high precision three-dimensional reconstruction. Experimental results show that our strategy can outperform the precision of the phase-shifting method.

Figures in this Article

Phase-shifting profilometry (PSP) has been used for accurate three-dimensional (3-D) measurement of objects in many applications such as biometric, entertainment, and industrial inspection, among others. An important stage in PSP is the camera–projector calibration. Thus, improved phase-shifting algorithms have been proposed. Yu et al.1 propose an algorithm based on interpolating absolute phase values at known features points with subpixel accuracy in the image data. A similar idea was proposed by Li et al.2 and used to improve the camera and projector calibration. This method, however, has only been applied when a set of feature points can be determined with high accuracy, usually the detected corners of a chessboard calibration pattern. In this paper, we propose a general strategy to enhance the accuracy of the 3-D measurement of objects using phase-shifting techniques by improving the search for correspondence. It is based on the interpolation of absolute phase values over a set of feature points in the image data. We show that for textured objects, this method produces high precision 3-D coordinates surface for the set of feature points. For textureless objects, a texture image can be projected in order to use the proposed method. The application of our method was two-fold, for improving the accuracy of the camera–projector calibration and for improving the accuracy of the phase-to-height conversion algorithm. We validate our method for real data and we compare the obtained object profiles with the traditional phase-shifting algorithm without phase interpolation.

The intensities of the phase-shifted3 images at every pixel (x,y) are given by Display Formula

Ik(x,y)=I+Icos[Φ+δk],(1)
where I(x,y) is the average intensity, I(x,y) is the intensity modulation, δk is the introduced shift, and Φ(x,y) is the phase to be determined for k=1,,N. A least-squares solution4 for the phase with δk=2π(k1)/N is given by Display Formula
Φ(x,y)=tan1k=1NIksin(2πkN)k=1NIkcos(2πkN).(2)

The phase Φ is called the wrapped phase which is characterized by the module 2π discontinuities and a continuous phase Φ(x,y) can be obtained by using a unwrapping algorithm.5 Zhang and Huang6 proposed a method for projector calibration based on absolute phase maps which can be obtained if a set of reference points are known between the camera and projector. The absolute phase is computed as Φ(x,y)=Φ(x,y)Φ0, where Φ0 is the average of phase values of the set of reference points. An one-to-one correspondence between the camera and the projector pixels was obtained considering vertical and horizontal absolute phase maps, Φv and Φh, respectively. Thus, for every camera pixel (x,y), the corresponding projector pixel (x,y) is given as Display Formula

x=p·Φh(x,y)πH2,y=p·Φv(x,y)πW2,(3)
where W, H, and p are given values corresponding to the width, height, and pitch of the fringe patterns.

Our proposal consists of two main stages: camera–projector calibration and 3-D reconstruction. All steps to obtain 3-D reconstruction are shown in Fig. 1. The traditional output of the first stage is the set of intrinsic and extrinsic calibration parameters. Instead, we generate another set of calibration parameters by using the last two steps as follows. Corner detection and interpolation is the detection of corners (at subpixel accuracy) for each image of the calibration pattern in the camera. Given the list of corner positions, the corresponding projector pixels are calculated through Eq. (3) and phase interpolation.7 Then, camera–projector calibration8 is performed using this list of corners.

Graphic Jump Location
Fig. 1
F1 :

Fundamental stages for the camera–projector calibration and three-dimensional reconstruction.

The second stage is the 3-D reconstruction which consists of (a) projection of six vertical and horizontal interference patterns over an object into the scene, also vertical and horizontal central lines as reference images; (b) computation of the continuous phase and absolute phase of the object; (c) detection of feature points in the captured image data of the object, that will be detailed briefly; and (d) computation of 3-D coordinates using a phase-to-height algorithm.

The interpolation of feature points is taken as special step that is necessary to describe. First, an image of the object is captured onto the scene. Second, an initial set of corners is detected using a smallest univalue segment assimilating nucleus9 corner detector. Third, a refinement algorithm10 is applied to the initial set of corners, and obtain a set of corners at subpixel accuracy. Finally, the corresponding projector positions are computed for the set of detected corners using phase interpolation. With these steps, for objects with regular texture many corners are detected. However, the strategy fails to find significant set of corners in textureless objects. An alternative step, for this kind of objects is to project a texture image on the object in order to provide it with a regular texture.

In order to validate the calibration results, we compare two cases: in the first case projector images were generated to calibrate the projector and in the second case we use the detected corners in the camera calibration patterns to interpolate the corresponding corners in the projector. Table 1 shows standard deviation of the reprojection corners error (in pixel) in both x and y directions for the projector calibration. Row 1 in Table 1 shows the pixel error of the projector calibration using projector images, while rows 2 to 5 show the pixel errors using nearest, linear, cubic, and spline interpolation of the set of corners in the camera. We found that linear interpolation of the detected corners in the camera gives the best result for the calibration of the projector, whereas nearest interpolation of the detected corners in the camera for projector calibration gives the worst result. Figure 2 shows the reprojection corner error for the camera calibration and projector calibration using linear interpolation of the camera corners.

Table Grahic Jump Location
Table 1Pixel error in the x and y directions of the projector calibration.
Graphic Jump Location
Fig. 2
F2 :

Reprojection corner error of the calibration for the (a) camera and (b) projector using linear interpolation of corners on the captured calibration patterns.

Figure 3(a) shows a steel plane surface on which a texture image was projected in order to provide it with a regular texture. Figure 3(b) shows the detected corners over Fig. 3(a) and Fig. 3(c) shows the detected corners at subpixel precision. Figure 3(d) shows the reconstruction profile of the plane surface using the traditional phase-shifting method and Fig. 3(e) shows the reconstruction using interpolation of feature points obtained through phase interpolation shown in Fig. 3(c). The reconstruction results shown in Figs. 3(d) and 3(e) were obtained using the calibration parameters that uses linear interpolation in Table 1. We fit an ideal plane to point clouds shown in Fig. 3(d) and Fig. 3(e) for the traditional and proposed method, respectively. Table 2 show quantitative results for both methods and in the last, nearest, linear, cubic, and spline interpolation of feature points were used. The mean, sum of squared errors (SSE), and standard deviation of the orthonormal distances of every point to the fitted plane are shown. These quantities were obtained by averaging 100 iterations of different samples of 1000 randomly chosen points. We found that the proposed method using linear interpolation of feature points has better accuracy than the traditional method, whereas using nearest-neighbor interpolation of feature points has the worst accuracy over all the cases.

Graphic Jump Location
Fig. 3
F3 :

(a) High precision steel surface, (b) projected texture, (c) subpixel detected corners on image shown in (b), (d) reconstruction using the classic phase-shifting method, and (e) reconstruction using the phase interpolation method for the detected corners shown in (c).

Table Grahic Jump Location
Table 2The sum of squared errors (SSE), the mean distance, and the standard deviation of the orthogonal distances of 1000 randomly chosen points to a fitted plane are shown. The units of the results are given in millimeters.

A method is proposed to improve the accuracy of the traditional phase-shifting techniques. It is based on the detection of a set of feature points in the image space at subpixel accuracy, and phase interpolation in order to generate its corresponding projector positions. The set of feature points is computed for objects with enough texture, and for a textureless object, an image is projected over the scene. We found by experimentation that the proposed method improves the accuracy of the traditional phase-shifting technique.

Yu  H.  et al., “3D profilometry system based on absolute phase calibration,” Proc. SPIE. 6357, , 63570M  (2006). 0277-786X CrossRef
Li  Z.  et al., “Accurate calibration method for a structured light system,” Opt. Eng.. 47, (5 ), 053604  (2008). 0091-3286 CrossRef
Geng  J., “Structured-light 3D surface imaging: a tutorial,” Adv. Opt. Photonics. 3, (2 ), 128 –160 (2011). 1943-8206 CrossRef
Malacara  Z., and Servín  M., Eds., Interferogram Analysis For Optical Testing. , 2nd ed.,  CRC Press, Taylor & Francis ,  Boca Raton, Florida  (2010).
Ghiglia  D., and Pritt  M., Two-Dimensional Phase Unwrapping: Theory, Algorithms, and Software. ,  John Wiley and Sons ,  New York  (1998).
Zhang  S., and Huang  P. S., “Novel method for structured light system calibration,” Opt. Eng.. 45, (8 ), 083601  (2006). 0091-3286 CrossRef
Press  W. H.  et al., Eds., Numerical Recipes in C: The Art of Scientific Computing. , 2nd ed.,  Cambridge University Press ,  New York  (1992).
Bouguet  J. Y., “Camera calibration toolbox for MATLAB,” www.vision.caltech.edu/bouguetj (1995).
Smith  S. M., and Brady  J. M., “SUSAN–a new approach to low level image processing,” Int. J. Comput. Vision. 23, (1 ), 45 –78 (1995). 0920-5691 CrossRef
Shi  J., and Tomasi  C., “Good features to track,” in  Proc. 1994 IEEE Computer Society Conf. on Computer Vision and Pattern Recognition (CVPR’94) , pp. 593 –600 (1994).
© The Authors. Published by SPIE under a Creative Commons Attribution 3.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.

Citation

William Cruz-Santos ; Lourdes López-García and Arturo Redondo-Galvan
"Improving the accuracy of phase-shifting techniques", Opt. Eng. 54(5), 054102 (May 13, 2015). ; http://dx.doi.org/10.1117/1.OE.54.5.054102


Figures

Graphic Jump Location
Fig. 3
F3 :

(a) High precision steel surface, (b) projected texture, (c) subpixel detected corners on image shown in (b), (d) reconstruction using the classic phase-shifting method, and (e) reconstruction using the phase interpolation method for the detected corners shown in (c).

Graphic Jump Location
Fig. 2
F2 :

Reprojection corner error of the calibration for the (a) camera and (b) projector using linear interpolation of corners on the captured calibration patterns.

Graphic Jump Location
Fig. 1
F1 :

Fundamental stages for the camera–projector calibration and three-dimensional reconstruction.

Tables

Table Grahic Jump Location
Table 2The sum of squared errors (SSE), the mean distance, and the standard deviation of the orthogonal distances of 1000 randomly chosen points to a fitted plane are shown. The units of the results are given in millimeters.
Table Grahic Jump Location
Table 1Pixel error in the x and y directions of the projector calibration.

References

Yu  H.  et al., “3D profilometry system based on absolute phase calibration,” Proc. SPIE. 6357, , 63570M  (2006). 0277-786X CrossRef
Li  Z.  et al., “Accurate calibration method for a structured light system,” Opt. Eng.. 47, (5 ), 053604  (2008). 0091-3286 CrossRef
Geng  J., “Structured-light 3D surface imaging: a tutorial,” Adv. Opt. Photonics. 3, (2 ), 128 –160 (2011). 1943-8206 CrossRef
Malacara  Z., and Servín  M., Eds., Interferogram Analysis For Optical Testing. , 2nd ed.,  CRC Press, Taylor & Francis ,  Boca Raton, Florida  (2010).
Ghiglia  D., and Pritt  M., Two-Dimensional Phase Unwrapping: Theory, Algorithms, and Software. ,  John Wiley and Sons ,  New York  (1998).
Zhang  S., and Huang  P. S., “Novel method for structured light system calibration,” Opt. Eng.. 45, (8 ), 083601  (2006). 0091-3286 CrossRef
Press  W. H.  et al., Eds., Numerical Recipes in C: The Art of Scientific Computing. , 2nd ed.,  Cambridge University Press ,  New York  (1992).
Bouguet  J. Y., “Camera calibration toolbox for MATLAB,” www.vision.caltech.edu/bouguetj (1995).
Smith  S. M., and Brady  J. M., “SUSAN–a new approach to low level image processing,” Int. J. Comput. Vision. 23, (1 ), 45 –78 (1995). 0920-5691 CrossRef
Shi  J., and Tomasi  C., “Good features to track,” in  Proc. 1994 IEEE Computer Society Conf. on Computer Vision and Pattern Recognition (CVPR’94) , pp. 593 –600 (1994).

Some tools below are only available to our subscribers or users with an online account.

Related Content

Customize your page view by dragging & repositioning the boxes below.

Related Book Chapters

Topic Collections

Advertisement
  • Don't have an account?
  • Subscribe to the SPIE Digital Library
  • Create a FREE account to sign up for Digital Library content alerts and gain access to institutional subscriptions remotely.
Access This Article
Sign in or Create a personal account to Buy this article ($20 for members, $25 for non-members).
Access This Proceeding
Sign in or Create a personal account to Buy this article ($15 for members, $18 for non-members).
Access This Chapter

Access to SPIE eBooks is limited to subscribing institutions and is not available as part of a personal subscription. Print or electronic versions of individual SPIE books may be purchased via SPIE.org.