Next Article in Journal
A Framework for Online Document Verification Using Self-Sovereign Identity Technology
Next Article in Special Issue
High-Precision Tribometer for Studies of Adhesive Contacts
Previous Article in Journal
Energy Internet Opportunities in Distributed Peer-to-Peer Energy Trading Reveal by Blockchain for Future Smart Grid 2.0
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

3D Metrology Using One Camera with Rotating Anamorphic Lenses

1
School of Mechanical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
2
School of Mechanical Engineering, Jinan University, Jinan 250022, China
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(21), 8407; https://doi.org/10.3390/s22218407
Submission received: 5 October 2022 / Revised: 26 October 2022 / Accepted: 31 October 2022 / Published: 1 November 2022
(This article belongs to the Special Issue Applications of Manufacturing and Measurement Sensors)

Abstract

:
In this paper, a novel 3D metrology method using one camera with rotating anamorphic lenses is presented based on the characteristics of double optical centers for anamorphic imaging. When the anamorphic lens rotates −90° around its optical axis, the 3D data of the measured object can be reconstructed from the two anamorphic images captured before and after the anamorphic rotation. The anamorphic lens imaging model and a polynomial anamorphic distortion model are firstly proposed. Then, a 3D reconstruction model using one camera with rotating anamorphic lenses is presented. Experiments were carried out to validate the proposed method and evaluate its measurement accuracy. Compared with stereo vision, the main advantage of the proposed 3D metrology approach is the simplicity of point matching, which makes it suitable for developing compact sensors for fast 3D measurements, such as car navigation applications.

1. Introduction

In modern industry, 3D metrology is an important technology, based on various methods. Optical metrology and non-optical metrology are the two categories of 3D metrology [1]. Non-optical 3D metrology methods may use a coordinate measuring machine (CMM), a scanning electron microscope, and a scanning probe microscope [2,3], whereas optical 3D metrology methods may employ chromatic confocal microscopy, point autofocus instruments, focus variation instruments, phase-shifting interferometry, coherence scanning interferometry, imaging confocal microscopy, and stereo vision. An overview of such methods was presented in detail [4,5,6]. Each 3D metrology method has its own properties and differs in application areas, measurement accuracy, scale, efficiency, cost, etc. Optical metrology procedures are often fast, precise, and non-destructive of the measured objects. Stereo vision, which uses two cameras with spherical lenses, is a frequently used optical metrology approach [7,8,9]. The measuring precision of stereo vision can range from several microns to several millimeters, depending on the measured area and the spherical lenses used. The most challenging aspect of stereo vision involves the extraction of the corresponding points from stereo images. Projectors are widely used to generate structured light to achieve high-quality corresponding points [10]. Stereo vision has been widely used in many areas, such as industry metrology, agriculture, daily living, etc. [11,12]
Anamorphic lenses are increasingly employed in the film industry to capture wide images for broad screens [13,14,15]. Anamorphic lenses are distinguished by two characteristics: double focal lengths and double optical centers [16,17,18]. Double focal lengths means the focal lengths differ greatly in the tangential and sagittal planes. This characteristic of anamorphic lenses is often used in the film industry, as it allows images to be compressed horizontally. Metrology, physics, and some optical imaging applications also use anamorphic lenses [19,20,21,22]. Double optical centers means there are two optical centers, one in the tangential plane and the other in the sagittal plane. The distance between the two optical centers is the anamorphic distance, which is an intrinsic quantity for anamorphic lenses. The anamorphic distance will provide anamorphic images with more 3D information, making anamorphic lenses considerably more suitable for 3D metrology. H. Durko used anamorphic lenses for measuring parts with large aspect ratios [23]. F. Blais performed an exploratory calibration of anamorphic lenses based on a pinhole imaging model [24]. In our previous work, a high-precision anamorphic lens calibration method was proposed with 3D and 2D calibration targets [16,25].
In this study, a novel 3D metrology method using one camera with rotating anamorphic lenses is proposed based on its characteristic of double optical centers. When the anamorphic lens rotates −90° around its optical axis, the 3D data of the measured object can be reconstructed from the two anamorphic images captured before and after the anamorphic rotation. The rest of this paper is organized as follows. Section 2 presents the anamorphic lens imaging model as well as the anamorphic distortion model. Section 3.1 describes the 3D reconstruction process employing rotating anamorphic lenses. Section 3.2 describes the point matching approach, and Section 3.3 depicts the rotating anamorphic stereo vision. Section 4 describes the experiments and evaluates the proposed method’s measurement accuracy. Finally, Section 5 concludes the paper.

2. Anamorphic Imaging Model

This section presents the imaging model and the distortion model of anamorphic lenses, which differ significantly from those based on spherical lenses. Anamorphic lenses are typically composed of rear spherical lenses on the back and an anamorphic attachment on the front. The anamorphic attachment consists of cylindrical lenses that have no optical power in planes parallel to the cylindrical axis but have optical power in planes perpendicular to it.

2.1. Anamorphic Imaging Model

The imaging model of anamorphic lenses differs from the traditional pinhole imaging model of spherical lenses. The anamorphic distance and the anamorphic angle are introduced as new extra intrinsic parameters [16]. Figure 1 depicts the imaging model of anamorphic lenses, where ad denotes the anamorphic distance, and aa denotes the anamorphic angle. If the anamorphic lens is precisely assembled, the anamorphic angle aa, as shown in Figure 1, can be as small as zero. The cylindrical lenses or the anamorphic attachment determine the anamorphic distance ad, which must be precisely calibrated. OW-XWYWZW are the world coordinates, Oci-XciYciZc are the camera coordinates centered in the CCD plane, and Op-XpYpZc are the pixel coordinates centered in the CCD plane. Ocx and Ocy are the optical centers in the horizontal plane and the vertical plane, and fx and fy are the focal lengths in the two planes. The imaging model for anamorphic lenses can be expressed as follows:
X C Y C Z C = R T X W Y W Z W 1 Y I f y = Y C Z C + a d X I f x = X C Z C X P Y P = cos a a sin a a - sin a a cos a a X I Y I
where [Xw; Yw; Zw] is a point in the world coordinates, [Xc; Yc; Zc] is the point expressed in the anamorphic coordinates. R and T denote the rotational and translating matrix from the world coordinates to the anamorphic coordinates. (XI, YI) are the image coordinates in the camera coordinates in the image plane, and (XP, YP) are the image coordinates in the pixel coordinates in the image plane.

2.2. Anamorphic Distortion Model

Anamorphic lenses have a distinct distortion model with respect to spherical lenses. The front anamorphic attachment introduces many more aberrations [26]. Based on the aberration theory and numerical experiments for anamorphic lenses, we proposed a polynomial distortion model for anamorphic lenses, which allows for high-precision camera calibration [25]. As shown in Equation (2), anamorphic distortions can be divided into radial distortions (Xrad, Yrad), third-order distortions (X3, Y3), and second-order distortions (X2, Y2). (xd, yd) denotes the distorted image coordinates, and (xc, yc) denotes the undistorted image coordinates. [k1, k2, n1, n2, m1, m2, x21, x12, x03, x20, x11, x02, y30, y21, y12, y03, y20, y11, y02] are the 19 distortion coefficients for anamorphic lenses that must be calibrated.
x c = x d 1 + k 1 x d 2 + y d 2 + k 2 x d 2 + y d 2 2 + m 1 a 2 x d 2 + y d 2 + m 2 a 2 x d 2 + y d 2 2 X r a d + x 21 x d 2 y d + x 12 x d y d 2 + x 03 y d 3 X 3 + x 20 x d 2 + x 11 x d y d + x 02 y d 2 X 2 y c = y d 1 + n 1 x d 2 + y d 2 + n 2 x d 2 + y d 2 2 + m 1 a 2 x d 2 + y d 2 + m 2 a 2 x d 2 + y d 2 2 Y r a d + y 30 x d 3 + y 21 x d 2 y d + y 12 x d y d 2 + y 03 y d 3 Y 3 + y 20 x d 2 + y 11 x d y d + y 02 y d 2 Y 2

3. 3D Metrology Using Rotating Anamorphic Lenses

3.1. Description of 3D Metrology Using Rotating Anamorphic Lenses

As shown in Figure 2a, the vertical position is the first anamorphic position, and the cylindrical axis of the anamorphic attachment is vertical. After a −90° rotation, the cylindrical axis of the anamorphic lens is horizontal, and this anamorphic position is known as the horizontal position. The anamorphic lens can switch between the two positions by rotating. The imaging model of anamorphic lenses shown in Section 2 can be reduced to an ideal anamorphic lens for ease of analysis, which implies that the anamorphic angle is zero, there are no distortions, and the rotation angle is precisely −90° without decentering. As shown in Figure 1 and Equation (1), the following equations describe the vertical position of an ideal anamorphic lens:
x v = X C Z C f x y v = Y C ( Z C + a d ) f y
where (xv, yv) refer to the image position of the object point PV = (XC, YC, ZC), and PV is in the anamorphic coordinates in the vertical position. After the anamorphic rotation, the coordinates of the object point in the anamorphic coordinates change as follows:
P H = R P V
If the anamorphic lens rotates alone its optical axis by −90°, R can be expressed as:
R = 0 1 0 1 0 0 0 0 1
Thus, when the anamorphic lens is rotated to its horizontal position, the object coordinates can be calculated as PH = [YC; −XC; ZC], and we have the following equations:
x h = Y C Z C f x y h = X C ( Z C + a d ) f y
Given Equations (3) and (6), the coordinates of the object point can be calculated as:
X C = a d x v y v f y x h f x y v Y C = a d x h y v f y x h f x y v Z C = a d f x y v f y x h f x y v
Thus, given the internal parameters fx, fy, and ad, as well as the image coordinates (xv, yv) and (xh, yh) of an object point in an ideal anamorphic lens, the 3D coordinates of this point can be easily reconstructed using Equation (7).

3.2. Point Matching

Point matching is crucial for stereo vision but is also challenging because the two images are captured by two different cameras. Points in different positions will have different image points in stereo images. Epipolar constraints and projectors are commonly adopted to achieve robust point matching, which requires a complex calculation. Point matching limits the applications of stereo vision in areas requiring fast or real-time 3D reconstruction, such as car navigation. Compared with stereo vision, point matching is much simpler in our proposed 3D metrology method. A simulation is provided to demonstrate its image point matching, as shown in Figure 3, Figure 4, Figure 5 and Figure 6. The simulated anamorphic lens has the following internal parameters: ad = 30 mm, aa = 0°, fx = 12 mm, fy = 16 mm, and no distortions. As shown in Figure 3, the object to be built is a collection of points on a spherical surface, with their coordinates given in the vertical position of the anamorphic lens. The simulated image points on the image planes are shown in Figure 4, where the dot points and the circle points express the image points when the anamorphic lens is in the vertical and horizontal positions, respectively. It is not easy to match the image points from these two positions in the current stage. In our method, the point matching can be greatly simplified by using a parameter of an anamorphic lens known as the anamorphic ratio AR, which is:
A R f y f x
As a result, if the dot points and circle points in Figure 4 are expanded horizontally and vertically by AR, respectively, we obtain the corresponding image points shown in Figure 5. Because the corresponding points are nearby dot points and circle points, point matching in rotating anamorphic lenses becomes a simple and straightforward task. In addition, as illustrated in Figure 6, we can trace the image positions during the anamorphic rotation from the dot points to the circle points for point matching. Point matching in anamorphic rotating 3D metrology is much easier than stereo vision with spherical lenses, as shown in Figure 5 and Figure 6. The deviations between the dot points and the circle points are due to the anamorphic distance ad. In this example, the anamorphic distance of 30 mm was much smaller than the object distance of 1500 mm. The deviation between the dot points and the circles in the image planes will increase when the object points are imaged using an anamorphic lens with a large anamorphic distance.

3.3. Stereo Vision with Anamorphic Lenses

Inevitably, there will be a misalignment between the optical axis of the anamorphic lens and the rotation axis of the rotary table, which will be magnified if the object is far away from the lens. As a result, large errors will occur if Equation (7) is applied directly for 3D reconstruction. A rational approach for high-precision 3D metrology is to treat the rotating anamorphic lenses as two distinct anamorphic lenses, which constitute a stereo vision with anamorphic lenses. The main differences between stereo vision and stereo vision with anamorphic lenses are that spherical lenses are replaced by anamorphic lenses and the baseline distance between the two anamorphic lenses is extremely small.
For 3D metrology using rotating anamorphic lenses, we should first calibrate the internal parameters of the anamorphic lenses according to the anamorphic imaging model and the anamorphic distortion model. A 3D calibration target can be used to calibrate the anamorphic lens [16]. As can be seen in Equation (2), the distortion model of the anamorphic lens has 19 unknown distortion coefficients. If only one image of a 3D calibration target is used for anamorphic lens calibration, local convergence is likely to occur for the distortion coefficients. Commonly, the 3D calibration target cannot cover the entire imaging area, which will lead to a poor calibration result for other imaging areas. Therefore, to achieve a stable calibration result for the entire imaging area, mixed calibration targets are adopted, which means 3D calibration targets and 2D calibration targets are both required [25]. The initial values of the internal parameters are calibrated using the 3D calibration target, and those parameters are refined using 2D calibration targets. After anamorphic calibration using mixed calibration targets, the internal anamorphic lens parameters [fx, fy, u0, v0, ad, aa] and the 19 distortion coefficients [k1, k2, n1, n2, m1, m2, x21, x12, x03, x20, x11, x02, y30, y21, y12, y03, y20, y11, y02] can be determined.
After anamorphic lens calibration, we should calibrate the relative position of the anamorphic lens in the vertical position and the horizontal position. Two images of the 3D calibration target are captured, one for the vertical position and one for the horizontal position. Based on the calibrated internal parameters of the anamorphic lens and the image of the 3D calibration target, the relative position [tx, ty, tz, a, b, r] between the 3D calibration target and the anamorphic lens can be easily calibrated:
T = t x t y t z
R = cos ( b ) cos ( r ) sin ( a ) sin ( b ) cos ( r ) cos ( a ) sin ( r ) sin ( a ) sin ( r ) + cos ( a ) sin ( b ) cos ( r ) cos ( b ) sin ( r ) cos ( a ) cos ( r ) + sin ( a ) sin ( b ) sin ( r ) cos ( a ) sin ( b ) sin ( r ) sin ( a ) cos ( r ) sin ( b ) cos ( b ) sin ( a ) cos ( a ) cos ( b )
where T and R refer to the translation vector and the rotation matrix. For the two anamorphic positions, there are the following equations between the coordinates of the 3D calibration target and the two anamorphic coordinates in the two positions:
P C V = R C V W P W + T C V W P C H = R C H W P W + T C H W
where the superscript W refers to the coordinates of the 3D calibration target, and the subscripts CV and CH refer to the coordinates of the anamorphic lens in the vertical and horizontal positions. Given Equation (11), the relationship between the vertical and the horizontal anamorphic coordinates can be determined as follows:
P C H = R C H C V P C V + T C H C V
where R C H C V and T C H C V are determined as follows:
R C H C V = R C H W R C V W 1 T C H C V = T C H W R C H C V T C V W
The horizontal position is rotated from the vertical position by −90° along the optical axis; thus, the relative position between the two anamorphic positions is similar to [tx, ty, tz, a, b, r] = [0 mm, 0 mm, 0 mm, 0°, 0°, 90°]. After the anamorphic lens calibration and the stereo calibration for the rotating anamorphic lenses, the 3D reconstruction using rotating anamorphic stereo vision is similar to that in stereo vision with spherical lenses.

4. Experiments

4.1. Experiments

An experiment was performed to validate the proposed 3D metrology method using one camera with rotating anamorphic lenses. The design of the experiment is shown in Figure 7. The camera was MER-504-10Gx-P from Daheng China, and its resolution was 2448 × 2448 with a pixel size of 3.45 μm × 3.45 μm. The anamorphic lens was mounted on the rotary table HK130-10. The servo motor was SGM7J-02A7C6E, and the servo controller was SGD7S-1R6A00A002, both from YASKAWA, Japan. As shown in Figure 7, the 3D calibration target was composed of two 2D planar calibration targets, and the relative position of the two planar calibration targets was determined by the 3D calibration target camera, which was previously calibrated. As shown in Figure 8, the anamorphic lens was composed of a front anamorphic attachment and a 16 mm Computar spherical lens. As shown in Figure 9, the anamorphic attachment comprised three cylindrical lens elements, and the effective focal length in the YOZ plane was −2489 mm. A paraxial lens design for anamorphic lenses can be seen in [27,28]. The lens parameters for the anamorphic attachment are shown in Table 1.
The anamorphic lens was calibrated using 3D and 2D calibration targets, and the internal parameters are shown in Table 2. And as shown in Figure 7, the relative position of the vertical position with respect to the horizontal position was calibrated using the 3D calibration target, and the results were [tx, ty, tz, a, b, r] = [−0.8943 mm, 0.3591 mm, −0.1712 mm, −0.5586°, −0.8074°, 90.0059°]. After the anamorphic lens calibration and the rotating anamorphic stereo calibration, we conducted the 3D metrology using the rotating anamorphic lens. As shown in Figure 10, to achieve dense 3D points, a checkerboard was adopted with a square’s length of 5 mm. The original anamorphic images are shown in Figure 10. Figure 10a shows the image when the anamorphic lens was in the vertical position, and Figure 10b shows the image when the anamorphic lens was in the horizontal position. Figure 10b was achieved by rotating the anamorphic lens by −90° anticlockwise along the optical axis. Once the corresponding corners in Figure 10a,b were determined, the 3D coordinates of the corners could be reconstructed from the rotating anamorphic stereo vision.
One possible method to determine the corresponding corners between Figure 10a,b is to trace the corners while rotating the anamorphic lens, similar to Figure 6. In order to trace the corners correctly, the rotating speed must be limited. Another method is to rectify the two anamorphic images by the anamorphic ratio AR, as shown in Equation (6), after which the corresponding points are very close, as in Figure 5. Figure 11a was obtained by expanding Figure 10a horizontally by the AR. Rotating Figure 10b by −90° anticlockwise and expanding the height of the rotated image by the AR, we obtained Figure 11b. As can be seen in Figure 11a,b, the two images are very similar after AR rectification, and the corresponding points are nearby, for the overlapping areas in Figure 11a,b.
The chessboard corners in Figure 11a were extracted and represented as red points in Figure 12; the corners in Figure 11b were represented by green circles in Figure 12. The red points and green circles in Figure 12a are in their original positions. It can be seen that the corresponding corners were not adjacent and were separated from one another. This is because there was a misalignment between the optical axis and the rotation axis of the rotary table. This misalignment was further amplified by the large object distance. If Equation (7) is directly applied for 3D reconstruction, there will be significant errors. Thus, the rotating anamorphic stereo vision was adopted for the 3D reconstruction. As can be seen in Figure 11, the two images are similar, the only difference between Figure 11a,b being a decentering. If we shift Figure 11b by (du, dv) = (105.0792 pixels, −28.8297 pixels), the two images will overlap. In Figure 12b, the corresponding points are very close, as shown in Figure 5.
Given the internal parameters of the anamorphic lens, the relative positions between the vertical anamorphic lens and the horizontal anamorphic lens, and the corresponding image points, the 3D coordinates of the corners could be reconstructed using the rotating anamorphic stereo vision. As shown in Figure 13 and Figure 14, the 3D positions of the corners in checkerboards were constructed. Figure 13 shows the images of the checkerboards when the anamorphic lens was in the vertical position, and Figure 14 shows the images of the calibration targets when the anamorphic lens was in the horizontal position. When constructing the eight objects, the first two objects were 3D objects, and the other six objects were 2D objects. The constructed 3D points are shown in Figure 15, while not all corners in the checkerboard were reconstructed. Some points were no constructed for they lay in the anamorphic gas where the construction accuracy was very low. The anamorphic gas is discussed in detail in Section 4.2.2. As can be seen in Figure 15, the reconstruction accuracy was low compared with that of the stereo vision. The accuracy of this metrology method is also discussed in Section 4.2.

4.2. Accuracy Analysis

As shown in Equation (14), the pixel error was the main source of error for this metrology, though other parameters such as ad, fx, and fy are also very important. This method’s measuring principle is based primarily on the anamorphic distance, which changes the image position compared to spherical lenses. The anamorphic distance ad shown in Section 4.1 was small in comparison to the object distance; thus, any small pixel position error would have a significant impact on the measurement results. From Equation (7), we have:
d X C = X C x h δ x h + X C x v δ x v + X C y v δ y v d Y C = Y C x h δ x h + Y C y v δ y v d Z C = Z C x h δ x h + Z C y v δ y v
where δxh, δxv, and δyv are the pixel errors. If we substitute the pixel errors in Equation (14) with δ, the point error can be deduced from Equation (14) as follows:
Δ x y z = d X C 2 + d Y C 2 + d Z C 2 = a d δ x 2 y 2 f x 2 f y 2 + x 2 f x 2 + x 1 y 1 f x f y + y 2 f y 2 f y x h f x y v 2
where the coefficients x2y2, x2, x1y1, and y2 in Equation (15) are given by:
x 2 y 2 = x h y v 2 x 2 = 2 y v 4 x 1 y 1 = 2 y v 2 x h 2 + x v + y v x h x v y v y 2 = x h 4 + x v + y v 2 x h 2 2 x v y v x v + y v x h + x v 2 y v 2
To further evaluate the measurement accuracy, we conducted two simulation experiments. The simulation anamorphic lens was the same as the anamorphic lens shown in Section 3.2: fx = 12 mm, fy = 16 mm, aa = 0°, and ad = 30 mm.

4.2.1. Accuracy Analysis for a Point

First, we simulated an object point with the coordinates [XC, YC, ZC] = [500 mm, 500 mm, 1500 mm] in anamorphic coordinates for the vertical position. The pixel coordinates of the point in the vertical and the horizontal anamorphic positions were calculated using Equations (3)–(6). Given the pixel coordinates and the calibrated parameters of the rotating anamorphic stereo vision, the 3D coordinates of the point could be calculated. The pixel errors in the image plane were assumed to have a Gaussian normal distribution with a standard deviation of 0.2 pixels, which was the calibration result of the anamorphic lens. Then, 5000 3D points were reconstructed with varying pixel errors, and the reconstructed 3D points are shown in Figure 16. The standard deviation of the reconstructed 3D point was 17.3378 mm.

4.2.2. Accuracy Analysis for a Surface

We also simulated a set of object points in a plane. The XC and YC coordinates ranged from −1000 mm to 1000 mm, with an interval of 50 mm, and all the ZC were 1500 mm. The pixel errors in the image plane satisfied a Gauss normal distribution, and the standard deviation of the pixel error was set to 0.2 pixels. The standard deviation for each reconstructed 3D point is shown in Figure 16. As can be seen in Figure 17, the measuring accuracy for the points away from the XC axis was high, while it was very low if the object points were near the XC axis. The points near the XC axis were removed, which left a gap in Figure 17. Thus, there was a measuring gap if rotating anamorphic lenses were used for 3D construction. We named this gap the anamorphic gap.
If we individually increase the anamorphic distance from 30 mm to 100 mm, the measuring accuracy will increase correspondingly, as shown in Figure 18. The lens parameters ad, fx, and fy are closely related because they are determined by the anamorphic lens structure, and it is not possible to simply change one parameter independently [25]. We proposed paraxial lens designs for anamorphic lenses with zero anamorphic distance [26], but designing an anamorphic attachment with an extremely large anamorphic distance appeared to be difficult.
The measuring precision of this 3D metrology method appeared to be lower than that of stereo vision with spherical lenses. The fundamental reason for this is that, in comparison to stereo vision, the baseline distance between the two anamorphic positions was quite small. The measuring precision will be as good as that of stereo vision with spherical lenses if the two anamorphic lenses are sufficiently apart from each other. Because the two images are captured with the same revolving anamorphic lens, the main advantage of this 3D metrology is its ease of point matching. The appropriate matching points are very close if the two images are rectified with the anamorphic ratio, as shown in Figure 5 and Figure 12. Alternatively, as illustrated in Figure 6, it is possible to trace the matching points by rotating the anamorphic lens. When the anamorphic lens is rotated at a rapid speed, it can measure 3D data four times in one circle, which is ideal for a quick 3D reconstruction with low measurement accuracy. One potential application of this would be car navigation.

5. Conclusions

A 3D metrology method using one camera with rotating anamorphic lenses is proposed in this paper. The anamorphic lens can rotate by −90° along the optical axis, capturing two anamorphic images of an object before and after the rotation. This setup is similar to that of stereo vision; therefore, it can be considered a stereo vision with anamorphic lenses. The main differences between stereo vision and rotating anamorphic lenses are that anamorphic lenses replace spherical lenses, and the baseline distance between the two anamorphic lenses is very small. This feature will make rotating anamorphic lenses a compact sensor suitable for 3D metrology in a constrained space. The feasibility of the proposed method in 3D reconstruction is based on one internal parameter of anamorphic lenses, i.e., the anamorphic distance, which will slightly change the image position of the object. The corresponding points between the two anamorphic images are close if the two anamorphic images are rectified by the anamorphic ratio, which substantially simplifies the corresponding point matching and reduces the calculation amount. When compared to stereo vision, the measurement accuracy of the rotating anamorphic stereo vision is lower, especially for the image points near the XC axis of the anamorphic lens in the vertical position. These characteristics make the rotating anamorphic stereo vision suitable for a fast 3D reconstruction without a high demand for measurement accuracy, such as car navigation applications. Further research might include high precision anamorphic lens calibration, error compensation, point matching for rotating anamorphic lenses, and anamorphic lens designs with a large anamorphic distance.

Author Contributions

Conceptualization, J.Z., X.C. and J.X.; methodology, X.C. and J.Z.; software, X.C and J.Z.; validation, J.Z.; data curation, J.Z.; writing—original draft preparation, J.Z.; writing—review and editing, X.C. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (52175478), the Defense Industrial Technology Development Program (JCKY2020203B039), the Science and Technology Commission of Shanghai Municipality (21511102602), the Doctoral Foundation of the University of Jinan (XBS1641), and the Joint Special Project of Shandong Province (ZR2016EL14).

Data Availability Statement

The data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Leach, R. Optical Measurement of Surface Topography; China Science Publishing & Media Ltd.: Beijing, China, 2012; pp. 154–196. [Google Scholar]
  2. Gaha, R.; Durupt, A.; Eynard, B. Towards the implementation of the Digital Twin in CMM inspection process: Opportunities, challenges and proposals. Procedia Manuf. 2021, 54, 216–221. [Google Scholar] [CrossRef]
  3. Sutton, M.A.; Li, N.; Joy, D.C.; Reynolds, A.P.; Li, X. Scanning electron microscopy for quantitative small and large deformation measurements part I: SEM imaging at magnifications from 200 to 10,000. Exp. Mech. 2007, 47, 775–787. [Google Scholar] [CrossRef]
  4. Harding, K. Challenges and opportunities for 3D optical metrology: What is needed today from an industry perspective. Two-Three-Dimens. Methods Insp. Metrol. VI 2008, 7066, 112–119. [Google Scholar]
  5. Marrugo, A.G.; Gao, F.; Zhang, S. State-of-the-art active optical techniques for three-dimensional surface metrology: A review. JOSA A 2020, 37, B60–B77. [Google Scholar] [CrossRef] [PubMed]
  6. Soid, S.N.; Zainal, Z.A. Spray and combustion characterization for internal combustion engines using optical measuring techniques—A review. Energy 2011, 36, 724–741. [Google Scholar] [CrossRef]
  7. O’Riordan, A.; Newe, T.; Dooly, G.; Toal, D. December. Stereo vision sensing: Review of existing systems. In Proceedings of the 2018 12th International Conference on Sensing Technology (ICST), Limerick, Ireland, 4–6 December 2018. [Google Scholar]
  8. Lazaros, N.; Sirakoulis, G.C.; Gasteratos, A. Review of stereo vision algorithms: From software to hardware. Int. J. Optomech. 2008, 2, 435–462. [Google Scholar] [CrossRef]
  9. Hussmann, S.; Ringbeck, T.; Hagebeuker, B. A performance review of 3D TOF vision systems in comparison to stereo vision systems. In Stereo Vision; Intechopen: London, UK, 2008. [Google Scholar]
  10. Zhang, S. High-speed 3D shape measurement with structured light methods: A review. Opt. Lasers Eng. 2018, 106, 119–131. [Google Scholar] [CrossRef]
  11. Chen, M.; Tang, Y.; Zou, X.; Huang, Z.; Zhou, H.; Chen, S. 3D global mapping of large-scale un-structured orchard integrating eye-in-hand stereo vision and SLAM. Comput. Electron. Agric. 2021, 187, 106237. [Google Scholar] [CrossRef]
  12. Tang, Y.; Zhou, H.; Wang, H.; Zhang, Y. Fruit detection and positioning technology for a Camellia oleifera C. Abel orchard based on improved YOLOv4-tiny model and binocular stereo vision. Expert Syst. Appl. 2023, 211, 118573. [Google Scholar] [CrossRef]
  13. Navarro, A.V.; Navarro, A.V.; Garrido, C.A. Anamorphic Lens. US 9063321 B2, 23 June 2015. [Google Scholar]
  14. Lippman, D.H.; Teverovsky, D.S.; Bentley, J.L. Monte Carlo first-order design method for anamorphic cinema zoom lenses. Opt. Eng. 2021, 60, 051203. [Google Scholar] [CrossRef]
  15. Xu, C.; Song, W.; Wang, Y. Design of a miniature anamorphic lens with a freeform front group and an aspheric rear group. Opt. Eng. 2021, 60, 065104. [Google Scholar] [CrossRef]
  16. Zhang, J.; Wang, X.; Ma, M.; Li, F.; Liu, H.; Cui, H. Camera calibration for anamorphic lenses with three-dimensional targets. Appl. Opt. 2020, 59, 324–332. [Google Scholar] [CrossRef] [PubMed]
  17. Dodoc, A. Anamorphic prime and zoom lenses. In Proceedings of the Zoom Lenses VI, SPIE, San Diego, CA, USA, 9 September 2019; Volume 11106, pp. 21–40. [Google Scholar]
  18. Zhang, J.; Chen, X.; Liu, H.; Li, F.; Sun, X. Thin lens aberrations for anamorphic lenses. Appl. Opt. 2019, 58, 182–188. [Google Scholar] [CrossRef] [PubMed]
  19. Soskind, M.; Soskind, Y.G. Propagation invariant laser beams for optical metrology applications. In Proceedings of the Modeling Aspects in Optical Metrology V, SPIE, Munich, Germany, 21 June 2015; Volume 9526, pp. 392–398. [Google Scholar]
  20. Ma, M.; Shao, H.; Zhang, J.; Wang, X.; Li, G. A Calibration Method of Anamorphic Lens Camera Based on Virtual 3D Target. In Proceedings of the 2019 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), Hong Kong, China, 8 July 2019. [Google Scholar]
  21. Towers, C.E.; Towers, D.P.; Campbell, H.I. Three-dimensional particle imaging by wavefront sensing. Opt. Lett. 2006, 31, 1220–1222. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Cao, Z.; Wang, K.; Wu, Q. Aspherical anamorphic lens for shaping laser diode beam. Opt. Commun. 2013, 305, 53–56. [Google Scholar] [CrossRef]
  23. Durko, H.L.; Barrett, H.H.; Furenlid, L.R. High-Resolution Anamorphic SPECT Imaging. IEEE Trans. Nucl. Sci. 2014, 61, 1126–1135. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Blais, F.; Beraldin, J.-A. Calibration of an anamorphic laser based 3-D range sensor. In Proceedings of the SPIE Videometrics V., San Diego, CA, USA, 27 July–1 August 1997; Volume 3174, pp. 113–122. [Google Scholar]
  25. Chen, X.; Zhang, J. High-precision anamorphic lens calibration with 3D and 2D calibration targets. Appl. Opt. 2022, 61, 6062–6075. [Google Scholar] [CrossRef] [PubMed]
  26. Yuan, S.; Sasian, J. Aberrations of anamorphic optical systems. II. Primary aberration theory for cylindrical anamorphic systems. Appl. Opt. 2009, 48, 2836–2841. [Google Scholar] [CrossRef] [PubMed]
  27. Chen, X.; Zhang, J. Lens design for parallel cylindrical anamorphic attachments with finite object distance. Appl. Opt. 2022, 61, 4610–4619. [Google Scholar] [CrossRef] [PubMed]
  28. Jinkai, Z.; Chen, X. Paraxial lens design of anamorphic lenses with a fixed anamorphic ratio. OSA Contin. 2019, 2, 1430–1454. [Google Scholar]
Figure 1. Anamorphic imaging model. (a) Imaging rays in the Yc-Ocy-Zc plane; (b) imaging rays in the Xc-Ocx-Zc plane.
Figure 1. Anamorphic imaging model. (a) Imaging rays in the Yc-Ocy-Zc plane; (b) imaging rays in the Xc-Ocx-Zc plane.
Sensors 22 08407 g001
Figure 2. The two anamorphic positions. (a) Vertical position and (b) horizontal position which is achieved by a rotation of the anamorphic lens in the vertical position by −90°.
Figure 2. The two anamorphic positions. (a) Vertical position and (b) horizontal position which is achieved by a rotation of the anamorphic lens in the vertical position by −90°.
Sensors 22 08407 g002
Figure 3. Simulated object points on a spherical surface in anamorphic coordinates of the vertical position.
Figure 3. Simulated object points on a spherical surface in anamorphic coordinates of the vertical position.
Sensors 22 08407 g003
Figure 4. Simulated image points on the image plane. The dot points refer to the image points when the anamorphic lens is in the vertical position, and the circle points refer to the image points when the anamorphic lens is in the horizontal position.
Figure 4. Simulated image points on the image plane. The dot points refer to the image points when the anamorphic lens is in the vertical position, and the circle points refer to the image points when the anamorphic lens is in the horizontal position.
Sensors 22 08407 g004
Figure 5. Simulated image points after anamorphic ration (AR) expansion. The dot points are rectified horizontally by AR, and the circle points are rectified vertically by AR.
Figure 5. Simulated image points after anamorphic ration (AR) expansion. The dot points are rectified horizontally by AR, and the circle points are rectified vertically by AR.
Sensors 22 08407 g005
Figure 6. Simulated image points tracing from the vertical position to the horizontal position.
Figure 6. Simulated image points tracing from the vertical position to the horizontal position.
Sensors 22 08407 g006
Figure 7. Experiment of 3D metrology using rotation anamorphic lenses.
Figure 7. Experiment of 3D metrology using rotation anamorphic lenses.
Sensors 22 08407 g007
Figure 8. Anamorphic lens composed of a front anamorphic attachment and a rear spherical lens. The anamorphic lens was mounted on a rotary table which could rotate the anamorphic lens by −90° along the optical axis. (a) Side view; (b) Front view.
Figure 8. Anamorphic lens composed of a front anamorphic attachment and a rear spherical lens. The anamorphic lens was mounted on a rotary table which could rotate the anamorphic lens by −90° along the optical axis. (a) Side view; (b) Front view.
Sensors 22 08407 g008
Figure 9. Anamorphic attachment for a Computar 16 mm spherical lens in the YOZ plane.
Figure 9. Anamorphic attachment for a Computar 16 mm spherical lens in the YOZ plane.
Sensors 22 08407 g009
Figure 10. Anamorphic images. (a) Image when the anamorphic lens was in the vertical position, and (b) image when the anamorphic lens was in the horizontal position.
Figure 10. Anamorphic images. (a) Image when the anamorphic lens was in the vertical position, and (b) image when the anamorphic lens was in the horizontal position.
Sensors 22 08407 g010
Figure 11. Anamorphic images after anamorphic ratio (AR) rectification. (a) Rectified image when the anamorphic lens was in the vertical position, and (b) rectified image when the anamorphic lens was in the horizontal position.
Figure 11. Anamorphic images after anamorphic ratio (AR) rectification. (a) Rectified image when the anamorphic lens was in the vertical position, and (b) rectified image when the anamorphic lens was in the horizontal position.
Sensors 22 08407 g011
Figure 12. Corners after anamorphic ratio (AR) rectification. The red points indicate the corners in Figure 11a, and the green points indicate the corners in Figure 11b. (a) Image points in their original position, (b) green points shift their positions entirely, after which the two images almost overlap. This pixel decentering was due to the deviation between the optical axis and the axis of the rotary table.
Figure 12. Corners after anamorphic ratio (AR) rectification. The red points indicate the corners in Figure 11a, and the green points indicate the corners in Figure 11b. (a) Image points in their original position, (b) green points shift their positions entirely, after which the two images almost overlap. This pixel decentering was due to the deviation between the optical axis and the axis of the rotary table.
Sensors 22 08407 g012
Figure 13. Images for 3D construction when the anamorphic lens was in the vertical position. (a,b) refer to images of 3D targets and (ch) refer to the images of a 2D target.
Figure 13. Images for 3D construction when the anamorphic lens was in the vertical position. (a,b) refer to images of 3D targets and (ch) refer to the images of a 2D target.
Sensors 22 08407 g013
Figure 14. Images for 3D construction when the anamorphic lens was in the horizontal position. (a,b) refer to images of 3D targets and (ch) refer to the images of a 2D target.
Figure 14. Images for 3D construction when the anamorphic lens was in the horizontal position. (a,b) refer to images of 3D targets and (ch) refer to the images of a 2D target.
Sensors 22 08407 g014
Figure 15. Constructed 3D points for images (ah) in Figure 13 and Figure 14.
Figure 15. Constructed 3D points for images (ah) in Figure 13 and Figure 14.
Sensors 22 08407 g015
Figure 16. Image showing 5000 constructed 3D points for an object point in [500 mm, 500 mm, 1500 mm] with different pixel errors. The standard deviation of the constructed 3D points was 17.3378 mm.
Figure 16. Image showing 5000 constructed 3D points for an object point in [500 mm, 500 mm, 1500 mm] with different pixel errors. The standard deviation of the constructed 3D points was 17.3378 mm.
Sensors 22 08407 g016
Figure 17. Standard deviation for object points in a plane. The standard deviations for object points with small YC coordinates were removed for large reconstruction errors. The standard deviation of the pixel position was 0.2 pixels, the ad was 30 mm, and the ZC was 1500 mm.
Figure 17. Standard deviation for object points in a plane. The standard deviations for object points with small YC coordinates were removed for large reconstruction errors. The standard deviation of the pixel position was 0.2 pixels, the ad was 30 mm, and the ZC was 1500 mm.
Sensors 22 08407 g017
Figure 18. Standard deviation for object points in a plane. The standard deviations for object points with small YC coordinates were removed for large reconstruction errors. The standard deviation of the pixel position was 0.2 pixels, the ad was 100 mm, and the ZC was 1500 mm.
Figure 18. Standard deviation for object points in a plane. The standard deviations for object points with small YC coordinates were removed for large reconstruction errors. The standard deviation of the pixel position was 0.2 pixels, the ad was 100 mm, and the ZC was 1500 mm.
Sensors 22 08407 g018
Table 1. Parameters of the anamorphic attachment used for the anamorphic lens.
Table 1. Parameters of the anamorphic attachment used for the anamorphic lens.
e0 (mm)e1 (mm)e2 (mm)T11 (mm)T12 (mm)T2 (mm)
16.1630623040444
f (mm)R11 (mm)R12 (mm)R13 (mm)R21 (mm)R22 (mm)
16155156.4−857.6−174.8124.5
n1V1n2V2n3V3
1.51679764.2123511.67270232.178881.51679764.212351
Table 2. Internal parameters of the calibrated anamorphic lens.
Table 2. Internal parameters of the calibrated anamorphic lens.
fx (mm)fy (mm)u0 (pixel)v0 (pixel)ad (mm)
12.052016.10261.2790 × 1031.0023 × 10326.5516
aa (°)k1k2n1n2
−0.57890.0304−2.7151 × 10−50.0183−3.3372 × 10−5
m1m2x21x12x03
−0.01551.277e-59.1986 × 10−5−0.01391.8928 × 10−5
y30y21y12y03x20
1.6727 × 10−50.0111.0611 × 10−4−0.00231.3802 × 10−4
x11x02y20y11y02
−4.7071 × 10−55.8709 × 10−59.3657 × 10−5−3.9944 × 10−4−2.9851 × 10−4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, X.; Zhang, J.; Xi, J. 3D Metrology Using One Camera with Rotating Anamorphic Lenses. Sensors 2022, 22, 8407. https://doi.org/10.3390/s22218407

AMA Style

Chen X, Zhang J, Xi J. 3D Metrology Using One Camera with Rotating Anamorphic Lenses. Sensors. 2022; 22(21):8407. https://doi.org/10.3390/s22218407

Chicago/Turabian Style

Chen, Xiaobo, Jinkai Zhang, and Juntong Xi. 2022. "3D Metrology Using One Camera with Rotating Anamorphic Lenses" Sensors 22, no. 21: 8407. https://doi.org/10.3390/s22218407

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop