1. Introduction
When we want to use the manipulator combined with vision to grasp, we usually obtain the pose information of the object in space through the camera. However, the pose information is based on the camera coordinate system and cannot be used directly. If we want the robot end effector to reach the target position, we need to know the coordinate information of the target position in the robot coordinate system. Robot hand–eye calibration is the key technology to solve this problem. It refers to the establishment of the relationship between the manipulator coordinate system and the image coordinate system so that the manipulator can accurately reach the specified position under the guidance of the camera.
At present, many experts and scholars have proposed many hand–eye calibration methods, which are mainly divided into the linear calibration method and the iterative calibration method. The iterative method is mainly used to improve accuracy and robustness. The linear method is an efficient on-line hand–eye calibration method [1].
Shiu and Ahmad [2] first introduced the dynamic equation AX = XB into hand–eye calibration and provided the minimum configuration for the unique solution. Sai and Lenz [3] proposed another highly linear method. Firstly, the translation of hand–eye transformation is estimated, and then its rotation is estimated. This method has been widely used in many systems. The traditional hand–eye calibration method is mainly based on [4,5] the two-step method. Firstly, the rotation matrix of hand–eye parameters is calculated, and then the translation vector is solved by using the rotation matrix, and the parameters of hand–eye matrix are obtained by combining the rotation and translation parameters. This method is simple and easy to understand, but its inherent defect is error transmission [6]. Jianfeng Jiang et al. [7] summarized the hand–eye calibration method into four steps: camera pose, mechanical claw pose, mathematical model, and error measurement. Huajian Song et al. [8] proposed a new analytical solution based on cost function to estimate the hand–eye matrix with measurement error. In order to obtain the nonsingular analytical solution based on the modified Rodrigues parameters, a new additional rotation theory must be introduced. Haihua Cui et al. [9] solved the rotation relationship by twice transforming the motion of the target carried by the robot and then solved the transformation relationship through multiple rotating motions in the robot tool coordinate system.
To solve this problem, some scholars have proposed a synchronous calibration method of rotation and translation parameters. Yuan Zhang et al. [10] proposed [10] a synchronous optimization method for the calibration and measurement of a typical hand–eye positioning system. By establishing the measurement calibration model under nonlinear constraints, an iterative method and closed solution method were proposed to effectively suppress the influence of calibration error on measurement accuracy. Zeng Jinsong et al. [11] proposed a calculation method of nonlinear robot vision guidance calibration parameters based on maximum likelihood estimation on the basis of analyzing the transformation relationship of coordinate systems in robot vision guidance. This method constructs the matrix transformation measure function in the calculation process according to the maximum likelihood estimation and uses the Levenberg Marquardt algorithm to solve the nonlinear least squares problem. The camera calibration parameters and hand–eye calibration parameters can be solved at one time.
In addition, for the establishment of the model, many scholars have discussed the modeling and solution methods of hand–eye calibration. Common modeling methods include quaternion, dual quaternion, and Denavit–Hartenberg (D–H). Zhongtao Fu et al. [12] proposed a collaborative motion coordinate calibration method based on dual quaternion, which effectively solved the problem of synchronous calibration. Deng Shi Chao and Mei Feng put forward the hand–eye calibration algorithm of octet [13] on the basis of dual quaternion and realized the synchronous optimal solution of rotation and translation matrix through the model of AX = XB. Xiao Wang et al. [14] proposed a new model based on a pair of dual equations. When the homogeneous matrix is decoupled or not, the simultaneous solutions and separable solutions of the dual equations are given. Simulation and experiments verify the feasibility and superiority of the proposed method in terms of accuracy and robustness. Kenji Koide and Emanuele Menegatti [15] proposed a hand–eye calibration technology based on minimizing the re-projection error. This method directly obtains the images of the calibration mode without explicitly estimating the camera attitude for each input image, effectively solves the estimation problem, and is easy to extend to different projection models. It can deal with different camera models by changing the projection model.
Some scholars have also combined big data with deep learning [16,17] to explore a new method of hand–eye system calibration. This method does not need to establish an accurate system mathematical model. Based on a large number of experimental data and neural network learning modes, the parameters of the hand–eye system are obtained through training. However, there are some disadvantages. In order to ensure the accuracy of training, tens of thousands of data samples are usually needed, and it takes a lot of time and energy to collect samples.
In hand–eye calibration, the vision system in which the camera is installed in a fixed position outside the manipulator body and the camera does not move with the manipulator is called an eye-to-hand hand–eye system [18]. This installation method has the advantages of convenient installation, simple calculation, and a low rate of measurement errors and is the preferred scheme in machine vision projects [19], as shown in Figure 1 [20].
The defect is that when the camera is imaging, there will be a phenomenon of near—large and far—small. Therefore, one hand–eye calibration is only suitable for grasping objects of the same height. If you want to grab objects of different heights, you need to calibrate several times, and the calibration results are only applicable to grabbing objects with the calibrated height, which is discrete. Once a new product is involved in the system, you need to recalibrate the hand–eye transformation matrix at the height of the product. Moreover, each recalibration requires the disassembly of the robot end effector. Considering the mechanical installation error, the previously calibrated height also needs to be recalibrated. In this way, with the increase in product models, the calibration work becomes more and more arduous.
In order to solve the above problems, this paper first calibrates the camera parameters to eliminate the effect of camera distortion on imaging quality. It ignores the effect of the calibration height, simplifies the hand–eye transformation into a plane-rigid transformation, and then introduces the calibration height parameters to establish a mathematical model that has a highly linear relationship between the parameters of the hand–eye rigid transformation matrix and the calibration and verifies the correctness of the model through experiments. According to the hand–eye calibration model, in any eye-to-hand-type hand–eye system, only a few key height hand–eye relationships need to be calibrated, and the hand–eye relationships of objects of any height in the camera field of view can be calculated based on this linear relationship. The number of hand–eye calibrations at different heights is reduced, and the efficiency of hand–eye calibration is improved.
The rest of this paper is organized as follows: in the second section, the pinhole camera model is introduced. The third section introduces the rigid transformation of the coordinate system. The fourth section is experimental verification and analysis, including camera parameter calibration, rigid transformation from the calibration plate coordinate system to robot coordinate system, and analysis of the experimental results. The fifth part is the conclusion. The sixth part is the acknowledgement of the project.
2. Pinhole Camera Model
A pinhole camera model refers to a model in which all the scenes that can be captured by the camera are displayed upside down on the camera imaging plane through the center point of the camera’s optical axis [21]. The optical center of a camera refers to the axis of the camera coordinate system with the optical axis of the camera. Usually, this axis is called the optical center of the camera. As shown in Figure 2, O represents the center point of the optical axis in the camera; this is the projection center point of the camera. The image plane O′-{u, v} in Figure 2 is a virtual plane, and the true image plane is plane O-{u, v}, which is symmetrical about the center of the projection. Only the image displayed on this plane is upside down. In order to facilitate the calculation, the orientation and size of the image must be transformed so that the transformed image is consistent with the orientation of the original object. In this way, the imaging plane is equivalently converted to the image plane O′-{u, v}.
When the projection center and optical axis of the camera are parallel to the Z axis, the distance from the projection center to the image plane O′-{u, v} is the focal length f of the camera. Assume that there is a point P(XB,YB,ZB) on the real-world calibration plate whose projection on the image plane is Pc (u, v), according to the similar triangle theory:
(1)
Scilicet:
(2)
When the origin of the two-dimensional image coordinate system and the intersection point O′ of the optical axis and the imaging plane do not coincide, the origin of the two-dimensional image coordinate system needs to be translated to the O′ position. Remember this transformation as (tu, tv); therefore
(3)
It can be seen that in the standard pinhole camera model, the coordinate value of point (u, v) on the two-dimensional image coordinate system has a linear relationship with the coordinate value of point (XB, YB) in the world coordinate system. Rewritten into homogeneous coordinate form, it is expressed as
(4)
The matrix M is called the camera’s internal parameter matrix, and it determines how the points in the real world are projected onto the camera.
The ideal perspective model is a pinhole imaging model, and the object and image will meet the relationship of similar triangles. However, in fact, due to the processing and assembly errors of the camera optical system, the lens cannot meet the relationship in which the object and image form a similar triangle, so there will be distortion between the actual image on the camera image plane and the ideal image. Distortion belongs to the geometric distortion of imaging, which is the phenomenon of picture distortion and deformation caused by different magnification of the image in different areas on the image plane. The degree of this deformation increases successively from the center of the picture to the edge of the picture, which is mainly reflected at the edge of the picture. The two main factors causing distortion are (1) lens shape (radial distortion) and (2) the lens is not parallel to the imaging plane (tangential distortion). The two distortions are shown in the Figure 3 and Figure 4.
In order to reduce distortion, try to avoid taking pictures with the widest angle end or the farthest end of the lens focal length. Due to the manufacturing or installation error of the camera, the lens of the camera has a certain distortion. We usually need to correct the distortion of the lens first, and the imaging model of the camera can meet the standard small-hole imaging model [22,23].
3. Rigid Transformation of the Coordinate System
An affine transformation is a transformation model of a two-dimensional plane coordinate system to another two-dimensional plane coordinate system [24]. When the objects in the two coordinate systems only have rotation and translation transformations, it is said that there is a rigid transformation relationship between the two coordinate systems [25]. In Euclidean space, suppose a point A (x, y) obtains a point B (x′, y′) through a rotation R transform and a translation t transform, and the two transformations are combined to make
(5)
where f (x, y) is the plane before transformation and f (x′, y′) is the plane after transformation. Let the rotation angle of the plane about the Z axis be θ, and thenIf point B (x′, y′) is subjected to a rotation R′ transformation and a translation t’ transformation again to obtain point C (x″, y″), then the transformation process from point A to point C can be expressed as
(6)
This is obviously not a linear transformation process. As the combination of basic transformation factors increases, the transformation process will become more and more complicated. Considering the convenience of calculation, here introduce the homogeneous coordinates to represent the points in the plane, and then the transformation from point A to point B can be expressed as
(7)
Equation (6) can be expressed as
(8)
In this way, no matter how many times the basic transformation factors are combined, the original coordinate points and the transformed coordinate points are always linear. In terms of mechanical installation of the entire system, it is necessary to ensure that the robot, the plane of the conveyor belt and the optical axis of the camera are parallel to each other. When the Z axes of the two coordinate systems are parallel, the amount of displacement in the Z direction is temporarily ignored so that the three-dimensional space coordinate system of the robot coordinate system is compressed to a two-dimensional plane coordinate system, which is convenient for subsequent calculations. As shown in Figure 5, represents the image coordinate system and represents the robot coordinate system. Since the motion of the robot is a rigid body, and the camera is fixed at the same time, when the Z axes of the two coordinate systems are parallel, there are only a few rotation transformations and translation transformations in the two spatial coordinate systems (the linear transformation model, according to Equation (8)). Therefore, the transformation model of the image coordinate system to the robot coordinate system can be expressed as
(9)
where n is the number of transformations from the image coordinate system to the robot coordinate system, represents the synthetic rotation matrix, represents the synthetic translation matrix, and represents the final transformation matrix.From Equations (1) and (4), the linear relationship between the coordinate value of the image coordinate system and the coordinate value of the world coordinate system is only affected by the height of the sampling point from the camera optical center. Because the robot coordinate system is equivalent to the world coordinate system, the transformation matrix of the robot coordinate system and the image coordinate system should also have a certain relationship with the sampling height:
Suppose that
(10)
where H is the sampling height. Therefore, Equation (9) can be expressed asThe above formula establishes a rigid transformation model of the image coordinate system to the robot coordinate system at an arbitrary height H. The following is an experiment to verify the correctness of the model.
4. Experimental Verification and Analysis
In the second section, we propose and establish a rigid transformation mathematical model from image coordinate system to the robot coordinate system at any height h through the standard small hole imaging model in the first section. Next, we mainly verify the correctness of the rigid transformation model proposed in Section 2 through experiments.
First, in the standard pinhole camera model established in Section 1, the matrix m in Equation (4) is called the camera’s internal parameter matrix, which determines how points in the real world are projected to the camera. Through the Zhang Zhengyou calibration method, we obtain the internal parameter matrix M of the camera.
Second, by eliminating the barrel distortion of the camera lens, the four-point calibration method is used to calibrate the four-point coordinates of the calibration plate at different heights in the machine coordinate system and the image coordinate system. The matrix parameters at different heights H are calculated.
Finally, we made a broken line diagram of matrix parameters and calibration height H in the three-dimensional coordinate system, which clearly showed the distribution of matrix parameters at different heights. There was an obvious linear relationship between matrix parameters and calibration height, and it further fitted the linear relationship between matrix parameters and calibration height H. Through the experimental results, the random error in the calibration system is analyzed, the linear relationship between the calibration height and the camera pixel density is studied, and the systematic error in the experimental process is deeply studied. Through verification, the positioning error of the proposed arbitrary height rigid transformation model in practical application is less than 0.08%. The specific experiment and analysis process are as follows.
4.1. Camera Parameter Calibration
The internal parameter matrix and distortion coefficient of the camera are the content of the camera calibration. The knowledge of domestic and foreign scholars has been perfected. Dainis and Juberts [26] give the parameter matrix of the camera by a linear transformation; Ganapathy [27] for the first time gives a method to obtain the camera parameter matrix through the perspective transformation matrix, but the simplest and most practical is Zhang Zhengyou’s calibration method [28,29]. This calibration method does not require high-precision calibration equipment, and the calibration accuracy is higher than the traditional calibration method. Therefore, this paper chooses Zhang Zhengyou’s calibration method as a method to obtain camera parameters. The calibration samples are shown in Figure 6 below.
This paper uses a high-precision alumina calibration plate with a low thermal expansion coefficient. The specific parameters are shown in the following Table 1 and Table 2.
We used the Camera Calibrator toolbox in MATLAB to import the samples in Figure 6 and enter the parameters of the calibration plate. The output is shown in Figure 7.
The final calibration results are shown in the table below:
Table 2Calibration results of camera internal parameters.
Parameter Type | Calibration Results |
---|---|
Equivalent focal length | [5010.81789826726, 5011.24572785725] |
Principal point coordinates | [2768.70959640510, 1806.34169717229] |
Radial distortion | k1 = −0.0625967952090845, k2 = 0.133984194777852 |
Tangential distortion | p1 = −0.000122713140590104, p2 = 0.00160031845139996 |
Scaling factor | = 0.168076608984161 |
Mean reprojection error | Mean reprojection error = 0.08 |
From the Table 2 calibration results, the camera imaging quality is much less affected by tangential distortion than by radial distortion. Therefore, when the accuracy requirements of the project are low, in order to improve the calculation speed, the influence of tangential distortion on the imaging quality can be ignored. The internal parameter matrix of the camera can be derived from Table 2:
4.2. Rigid Transformation from Calibration Plate Coordinate System to Robot Coordinate System
The calibration plate selected in this experimental step is a rectangular plate of 2000 mm × 1200 mm. The smallest enclosing rectangle of the original image is shown in Figure 8a. Due to the effects of lens distortion, the entire image is barrel-shaped [30]. The image effect after distortion correction is shown in Figure 8b. The image exactly matches the standard rectangular frame, indicating that the image distortion has been significantly corrected.
The calibration plate selected in this paper is a rectangular calibration plate. We calibrated the four vertices on the upper surface of the rectangular calibration plate, moved the calibration plate to different heights, calibrated the four-point coordinates in turn, obtained the four-point coordinates of the workpiece in the image coordinate system and the machine coordinate system at different heights, and calculated the different matrix parameters.
For workpieces of different sizes and heights, we know that three non-collinear points can determine a plane. If the accuracy of the four-point calibration results is enough, there is no need for more point calibration. For a quadrilateral workpiece, you can take the four vertices of the upper surface of the workpiece, move the workpiece to different heights, calibrate the four-point coordinates of the upper surface in turn, obtain the four-point coordinates of the workpiece in the image coordinate system and the machine coordinate system at different heights, and calculate the matrix parameters at different heights. For irregular workpieces, we can take the non-collinear key four-point positions on the upper surface of the workpieces to fit into a quadrilateral, determine the key quadrilateral shape of the workpiece, move the workpiece to different heights, calibrate the four-point coordinates of the upper surface in turn, obtain the four-point coordinates of the workpiece in the image coordinate system and the machine coordinate system at different heights, and calculate the matrix parameters at different heights.
In the actual grasping work, the transformation from the image coordinate system to the robot coordinate system is realized by the matrix parameters at different heights, the four-point positions obtained by the camera and the matrix so that the shape and position of the workpiece can be determined, and then the target grasping work of the robot can be realized.
We used the robot’s tip tool (TCP) to establish a tool coordinate system, sequentially aligned the four corner points of the calibration board, and read the corresponding coordinate values on the teaching pendant, as shown in Figure 9.
Due to the large variety of products and the different sizes of products at different heights, a hand–eye calibration is required at each product height position. However, in order to adapt to new height products, hand–eye calibration is generally required at a sufficiently small height step. According to the four-point calibration method, the calibration is performed for every 10 mm height, and some calibration results are shown in Table 3.
According to the principle of the four-point calibration method [31,32]. The rigid transformation matrix of the image coordinate system and the robot coordinate system at each height is obtained, as shown in Table 4.
4.3. Analysis of Experimental Results
The robot grasping process is the inverse process of the camera imaging process, that is, the transformation process from the image coordinate system to the world coordinate system. By calibrating the calibration plates at different heights, we can obtain the rigid transformation matrix parameters of each height, which can realize the transformation from the image coordinate system to the robot coordinate system and realize the target grasping of the robot. However, in this conversion, an important parameter—the height of the workpiece—is lost. In this way, there will be the problem that the same image coordinate points may represent different world coordinate points, resulting in the robot’s “inaccurate grasp.” In addition, due to the wide variety of products and different imaging dimensions of products at different heights, the height position of each product should be calibrated once.
Secondly, if there is no corresponding calibration for the different heights of a new product, we can only use the transformation matrix of the adjacent height calibration for calculation. Although many calibration experiments are carried out in a small step, there will still be errors between the transformation matrix generated by the discrete calibration height and the transformation matrix under the actual product height. In order to solve this problem, we know from the rigid transformation model established in Section 2 that the transformation matrix between robot coordinate system and image coordinate system should also have a certain relationship with the sampling height. Through this linear relationship, we can solve this kind of problem by fitting the rigid change matrix at the partial height obtained in Section 4.2 of the experiment to the rigid change matrix at any height. According to Table 4, we connect the relationship between the parameter values of and height with broken lines in the same three-dimensional coordinates, as shown in Figure 10.
It can be seen from Figure 10 that matrix parameters show an obvious linear rlationship with the broken line diagram of H. Tx and Ty reflect the change in the translation matrix. A11, A12, A13, and A14 are all near the z = 0 plane, which is also fully consistent with the rotation matrix parameters proposed by us. According to Table 4, we further analyze the relationship between the calibration height and various parameters of . The results are shown in Figure 11.
As can be seen from Figure 11, except for A11 and A22, except for the two parameters, there are some points that are under-fitted. Considering A11 and A22, the two parameters are relatively small, and the approximate linear processing will not have a great impact on the results. In practical application, the points with high dispersion can be eliminated and refitted.
We further analyze the relationship between the calibration height and various parameters of , as shown in Table 5.
As can be seen from Table 5, except for the A11 and A22 two parameters, the other parameters have a strict linear relationship with the calibration height, and the correlation coefficient R2 > 0.995, |Pearsons’s R| > 0.997, adj R-Square > 0.995. Therefore, the proposed linear result is shown in Equation (11).
(11)
Considering the error caused by the experiment, we further analyzed the experimental results. We can see from the calibration results in Table 3 that the calibration coordinates are in the camera and robot coordinate systems at different heights, as shown in Figure 12.
It can be found that when H is 105 mm, XcB’ has great errors, and errors will inevitably occur in reading data in the actual calibration process. Considering A11 and A22, the two parameters are relatively small, and the approximate linear processing will not have a great impact on the results. In practical application, one-step data preprocessing can be carried out first, and then the points of matrix parameters with high dispersion can be eliminated for refitting. In order to further study the experimental results, we found that there is also a linear relationship between pixel density and camera calibration height H. The research is as follows:
Pixels per inch (PPI) indicates the number of pixels per unit length of content. The larger the PPI, the higher the resolution and fidelity of the representative image. The size of the research object in this paper is in mm, so the pixel density is defined as the number of pixels contained per mm, using Pr as an indicator.
As can be seen from Figure 4
(12)
when the focal length and position of the camera are fixed, the greater the height of the workpiece, the smaller the distance H′ between the optical center of the camera and the upper surface of the workpiece, and the smaller the field of view of the camera. The equation is(13)
Set the resolution of the camera as M × n. If M > N, the calculation formula of PPI is
(14)
Visible pixel density Pr is inversely proportional to the distance H′ between the optical center of the camera and the upper surface of the workpiece and is directly proportional to the height h of the workpiece.
From Figure 11 and Table 3, the corresponding relationship is established to calculate the corresponding length pixel ratio (Pr_l) and width pixel ratio (Pr_w), as shown in Equation (15). Where l1 and l2 are the length dimensions of the image, w1 and w2 are the width dimensions of the image, and l and W are the workpiece dimensions.
(15)
The table of pixel ratio relationship under the height of Equation (15) is shown in Table 6.
Through Table 6, we fit the relationship between the length–width pixel ratio and the workpiece height, as well as the correlation coefficient, as shown in Figure 13.
As shown in Figure 13. It can be seen that the fitting data in the length direction and the fitting data in the width direction are basically the same, but there is also a deviation. Ideally, it should be a straight line. This is due to the camera installation error, unevenness, experimental operation and other errors, resulting in the inconsistency of density.
Further analysis shows that the correlation coefficients of the fitting results in the length and width directions are also inconsistent. This is because the fixed focus lens is used in this paper. When the workpiece changes, the impact of “defocus blur” will inevitably appear in the imaging delay of the camera. However, the error caused by this reason is within the tolerance range of the system since the workpiece is always captured in a certain height range in practical application. Moreover, the lenses used now have a certain depth of field (DOF), so there is no need to make up for the error caused by defocus blur in practical application.
From Figure 13, the correlation coefficient of the fitting result in the width direction is large, and the pixel density in the width direction is greater than that in the length direction. In order to better reflect the relationship between pixel density and calibration height, this paper takes the fitting result in the width direction as the pixel density Pr of the workpiece at height H. Namely
It can be seen from Equation (11) that there is still a highly linear relationship between the parameters of the rigid transformation matrix from the image coordinate system to the robot coordinate system and the calibration height under multiple errors. In the eye-to-hand hand–eye system, the parameters of matrix in the calibration system at any height can be calculated quickly according to the fitting results. The measured result is the display data of the teaching pendant at a certain height of a point on the calibration board measured according to the four-point calibration method shown in Figure 9. The result is to calculate the matrix under the calibration according to Equation (11) and substitute it into Equation (10) to calculate the coordinate value. The verification results are shown in Table 7.
As can be seen from Table 7, firstly, the rigid transformation matrix is calculated according to the linear relationship between the transformation matrix parameters and the calibration height, and then the transformation matrix is used to calculate the position of the robot’s target point under the visual guidance. The error is less than 0.08%, so this linear relationship is reasonable and effective in the calibration system.
5. Conclusions
Through the introduction, this paper introduces the specific working mode of the hand–eye calibration method and the current research status of the hand–eye calibration method. It also discusses the defect in hand–eye calibration: due to the small phenomenon of camera imaging, it cannot be applied to machine grasping at any height. Based on the imaging model of a standard pinhole camera, the rigid transformation matrix is introduced to establish the rigid transformation mathematical model from image coordinate system to robot coordinate system at any height h, and the hand–eye transformation is simplified into plane rigid transformation. In the process of the experiment, the internal parameters and distortion parameters of the camera were obtained according to the Zhang Zhengyou calibration method, which eliminates the influence of the camera barrel distortion on the imaging quality and ignores the influence of the calibration height. Then, the machine coordinate system and image coordinate system of the calibration plate at different heights were calibrated using the four-point calibration method, and the parameter matrix at different heights h was obtained. From the obvious linear relationship between the distribution of the parameter matrix in the three-dimensional coordinate system, the linear relationship between the parameters of the rigid transformation matrix from the image coordinate system to the robot coordinate system and the calibration height was further fitted. The random error of the experimental results was analyzed, and the linear relationship between the calibration height and the pixel density of the image was further studied. Through the experimental results, the systematic errors in the experimental process were deeply analyzed. The experimental results show that the hand–eye relationship of objects at any height in the camera’s field of view is calculated according to this linear relationship, which is accurate and suitable for the positioning of products at any height, and the positioning error is less than 0.08%.
According to this conclusion, the eye-to-hand type hand–eye calibration system only needs to calibrate and calculate the rigid transformation matrix of several sampling heights to calculate the hand–eye rigid transformation matrix of the system at any height, and the result is accurate and effective. The calibration of the coordination system brings great convenience.
Resources, S.S.; Writing—original draft, D.Z. and W.W.; Writing—review & editing, S.G., D.Z. and W.W. All authors have read and agreed to the published version of the manuscript.
The author of this article declares that there is no conflict of interest related to this manuscript.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figure 8. Image processing results of the calibration plate before and after distortion correction: (a) before distortion correction and (b) after distortion correction.
Figure 10. Broken line diagram of rigid transformation matrix parameters and calibrated height parameters: (a) The Z−axis is the parameter value, and (b) Z−axis is the calibration height.
Figure 11. Relationship between rigid transformation parameters and calibration height.
Figure 12. Coordinate analysis diagram: (a) three−dimensional display of calibration plate coordinates and (b) actual camera robot coordinates.
Figure 13. Relationship between length–width pixel ratio and calibration plate height.
Alumina calibration plate parameters.
Model | Dimensions (mm) | Checker Side Length (mm) | Pattern Array | Accuracy (mm) |
---|---|---|---|---|
LGP500-300 | 500 × 500 | 30 | 13 × 12 | ±0.02 |
Four-point calibration results.
Height (mm) | Corner Position | Robot Coordinate System (mm) | Image Coordinate System (Pixel) | ||
---|---|---|---|---|---|
|
|
|
|
||
15 | A | −1449.7 | 1585.93 | 1284.804 | 432.0025 |
B | 549.17 | 1604.26 | 1290.65 | 3296.744 | |
C | 536.96 | 2804.99 | 3015.627 | 3293.224 | |
D | −1462.22 | 2785.86 | 3009.781 | 428.4819 | |
45 | A | −1414.15 | 1530.26 | 1191.118 | 470.0939 |
B | 584.55 | 1561.77 | 1218.606 | 3365.488 | |
C | 564.22 | 2762.3 | 2958.563 | 3348.97 | |
D | −1434.56 | 2729.52 | 2931.075 | 453.5752 | |
75 | A | −1417.77 | 1522.52 | 1165.935 | 453.0028 |
B | 581.1 | 1536.81 | 1170.013 | 3373.189 | |
C | 571.3 | 2737.35 | 2925.071 | 3370.738 | |
D | −1427.86 | 2722.06 | 2920.992 | 450.5518 | |
105 | A | −1429.65 | 1529.60 | 1163.014 | 423.642 |
B | 569.51 | 1530.36 | 145.789 | 3368.993 | |
C | 568.25 | 2730.58 | 2916.097 | 3379.346 | |
D | −1431.17 | 2729.40 | 2933.322 | 433.995 |
Parameters of rigid transformation matrix
Height (mm) |
|
|
|
|
|
|
---|---|---|---|---|---|---|
15 | −0.00574 | 0.697815001 | −1743.855103 | 0.695863008 | 0.005118 | 689.4714355 |
30 | −0.00531 | 0.694389999 | −1738.144653 | 0.692277014 | 0.004842 | 699.4268799 |
45 | −0.00515 | 0.69036603 | −1732.568604 | 0.689655006 | 0.004555 | 706.3405762 |
55 | −0.00428 | 0.688233972 | −1729.682251 | 0.687260985 | 0.003596 | 713.5334473 |
75 | −0.00471 | 0.684557021 | −1722.456787 | 0.683767021 | 0.00411 | 723.1809692 |
85 | −0.00460 | 0.682471991 | −1718.754883 | 0.681801021 | 0.003916 | 728.6638184 |
95 | −0.00443 | 0.680664003 | −1716.002808 | 0.679901004 | 0.003839 | 733.9542236 |
105 | −0.00475 | 0.678767979 | −1711.740601 | 0.677829027 | 0.004294 | 739.3514404 |
Rigid transform matrix parameters fitting table.
A11 | A12 | Tx | A21 | A22 | Ty | |
---|---|---|---|---|---|---|
Linear Fitting | Y = A + B × X | |||||
Plot | A11 values | A12 Values | Tx values | A21 Values | A22 Values | Ty values |
Weight | No Weighting | |||||
Intercept | −0.00561 |
0.70041 |
−1748.81438 |
0.69845 |
0.005 |
682.24144 |
Slope | 1.16124 × 10−5 |
−2.09957× 10−4 |
0.35111 |
−1.96476× 10−4 |
−1.13693 × 10−5 |
0.54652 |
Residual Sum |
7.12351 × 10−7 | 1.35738 × 10−6 | 0.72949 | 3.75039 × 10−7 | 9.87247 × 10−7 | 3.44764 |
Pearson’s r | 0.75941 | −0.99787 | 0.99959 | −0.99933 | −0.69652 | 0.9992 |
R−Square (COD) | 0.5767 | 0.99574 | 0.99918 | 0.99865 | 0.48515 | 0.9984 |
Adj.R−Square | 0.50615 | 0.99503 | 0.99904 | 0.99843 | 0.39934 | 0.99813 |
Length and width pixel ratio at different heights.
Height (mm) | Workpiece Size (mm) | Image Size (Pixel) |
|
|
---|---|---|---|---|
15 | 2000 × 1200 | 2865 × 1724 | 1.432374 | 1.437484 |
30 | 2879 × 1734 | 1.439345 | 1.444716 | |
45 | 2896 × 1740 | 1.447762 | 1.45003 | |
55 | 2906 × 1746 | 1.452295 | 1.454739 | |
75 | 2920 × 1755 | 1.460094 | 1.46255 | |
85 | 2929 × 1760 | 1.464606 | 1.466992 | |
95 | 2937 × 1765 | 1.4685 | 1.470833 | |
105 | 2945 × 1770 | 1.4727 | 1.475282 |
Verification of the applicability of the rigid transformation matrix.
Height |
Results of Testing (mm) | Calculation Results (mm) | ||||
---|---|---|---|---|---|---|
X | Y |
|
Error |
|
Error | |
20 | 691.07 | 1983.25 | 691.2215 | 0.0219% | 1982.9732 | 0.0139% |
55 | −1487.95 | 1901.81 | −1488.0752 | 0.0084% | 1901.3576 | 0.0238% |
88 | −1419.21 | 2504.49 | −1419.7325 | 0.0368% | 2503.8941 | 0.0238% |
105 | 537.27 | 2542.82 | 537.6739 | 0.0752% | 2543.1803 | 0.0142% |
References
1. Liu, J.; Wu, J.; Li, X. Robust and accurate hand–eye calibration method based on schur matric decomposition. Sensors; 2019; 19, 4490. [DOI: https://dx.doi.org/10.3390/s19204490] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31623249]
2. Shiu, Y.C.; Ahmad, S. Calibration of Wrist-Mounted Robotic Sensors by Solving Homogeneous Transform Equations of the Form AX = XB. IEEE Trans. Robot. Autom.; 1989; 5, pp. 16-29. [DOI: https://dx.doi.org/10.1109/70.88014]
3. Tsai, R.Y.; Lenz, R.K. A new technique for fully autonomous and efficient 3d robotics hand-eye calibration. Proceedings of the 4th International Symposium on Robotics Research; Santa Clara, CA, USA, 1 May 1988; pp. 287-297.
4. Huang, C.; Chen, D.; Tang, X. Robotic hand-eye calibration based on active vision. Proceedings of the 2015 8th International Symposium on Computational Intelligence and Design (ISCID); Hangzhou, China, 12–13 December 2015; pp. 55-59.
5. Wu, J.; Liu, M.; Qi, Y. Computationally Efficient Robust Algorithm for Generalized Sensor Calibration Problem AR=RB. IEEE Sens. J.; 2019; 19, pp. 9512-9521. [DOI: https://dx.doi.org/10.1109/JSEN.2019.2924668]
6. Ali, I.; Suominen, O.; Gotchev, A.; Morales, E.R. Methods for simultaneous robot-world-hand–eye calibration: A comparative study. Sensors; 2019; 19, 2837. [DOI: https://dx.doi.org/10.3390/s19122837]
7. Jiang, J.; Luo, X.; Luo, Q.; Qiao, L.; Li, M. An overview of hand-eye calibration. Int. J. Adv. Manuf. Technol.; 2021; 119, pp. 77-97. [DOI: https://dx.doi.org/10.1007/s00170-021-08233-6]
8. Song, H.; Du, Z.; Wang, W.; Sun, L. Singularity analysis for the existing closed-form solutions of the hand-eye calibration. IEEE Access; 2018; 6, pp. 75407-75421. [DOI: https://dx.doi.org/10.1109/ACCESS.2018.2882183]
9. Cui, H.; Sun, R.; Fang, Z.; Lou, H.; Tian, W.; Liao, W. A novel flexible two-step method for eye-to-hand calibration for robot assembly system. Meas. Control; 2020; 53, pp. 2020-2029. [DOI: https://dx.doi.org/10.1177/0020294020964842]
10. Zhang, Y.; Qiu, Z.; Zhang, X. A Simultaneous Optimization Method of Calibration and Measurement for a Typical Hand–Eye Positioning System. IEEE Trans. Instrum. Meas.; 2020; 70, pp. 1-11. [DOI: https://dx.doi.org/10.1109/TIM.2020.3013308]
11. Zeng, J.; Xue, W.; Zhai, X. A Synchronous Solution Method of Calibration Parameters for Visual Guidance of Robot. Mach. Tool Hydraul.; 2019; 47, pp. 37-40.
12. Fu, Z.; Pan, J.; Spyrakos-Papastavridis, E.; Chen, X.; Li, M. A dual quaternion-based approach for coordinate calibration of dual robots in collaborative motion. IEEE Robot. Autom. Lett.; 2020; 5, pp. 4086-4093. [DOI: https://dx.doi.org/10.1109/LRA.2020.2988407]
13. Deng, S.; Feng, M. Research on Hand-eye Calibration of Monocular Robot. Modul. Macine Autom. Manuf. Tech.; 2021; 12, pp. 53-57.
14. Wang, X.; Huang, J.; Song, H. Simultaneous robot–world and hand–eye calibration based on a pair of dual equations. Measurement; 2021; 181, 109623. [DOI: https://dx.doi.org/10.1016/j.measurement.2021.109623]
15. Koide, K.; Menegatti, E. General hand–eye calibration based on reprojection error minimization. IEEE Robot. Autom. Lett.; 2019; 4, pp. 1021-1028. [DOI: https://dx.doi.org/10.1109/LRA.2019.2893612]
16. Levine, S.; Pastor, P.; Krizhevsky, A.; Ibarz, J.; Quillen, D. Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. Int. J. Robot. Res.; 2018; 37, pp. 421-436. [DOI: https://dx.doi.org/10.1177/0278364917710318]
17. Pinto, L.; Gupta, A. Supersizing self-supervision: Learning to grasp from 50k tries and 700 robot hours. Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA); Stockholm, Sweden, 16–21 May 2016; pp. 3406-3413.
18. Wang, L.; Qian, L. Research on Robot Hand-Eye Calibration Method Based on Three-Dimensional Calibration Block. Laser Optoelectron. Prog.; 2021; 58, pp. 539-547.
19. Muis, A.; Ohnishi, K. Eye-to-hand approach on eye-in-hand configuration within real-time visual servoing. IEEE/ASME Trans. Mechatron.; 2005; 10, pp. 404-410. [DOI: https://dx.doi.org/10.1109/TMECH.2005.852397]
20. Liang, P.; Lin, W.; Luo, G.; Zhang, C. Research of Hand–Eye System with 3D Vision towards Flexible Assembly Application. Electronics; 2022; 11, 354. [DOI: https://dx.doi.org/10.3390/electronics11030354]
21. Juarez-Salazar, R.; Zheng, J.; Diaz-Ramirez, V.H. Distorted pinhole camera modeling and calibration. Appl. Opt.; 2020; 59, pp. 11310-11318. [DOI: https://dx.doi.org/10.1364/AO.412159]
22. Zhang, Z.; Luo, B. Research and implementation of robot vision positioning. J. Harbin Inst. Technol.; 1997; 29, pp. 85-89.
23. Tang, W. Design and Implementation of Visual Localization Algorithm for Demo System of Construction Robots. Master’s Thesis; Harbin Institute of Technology: Harbin, China, 2020.
24. Ying, J.; Chen, W.; Yang, H. Research on Parking spaces recognization and counting algorithm based on affine transformation and template matching. Appl. Res. Comput.; 2022; 39, pp. 919-924.
25. Han, J. Higher Institutional Science; Mechanical Industry Press: Beijing, China, 2004.
26. Dainis, A.; Juberts, M. Accurate remote measurement of robot trajectory motion. Proceedings of the 1985 IEEE International Conference on Robotics and Automation; St. Louis, MO, USA, 25–28 March 1985; pp. 92-99.
27. Ganapathy, S. Decomposition of transformation matrices for robot vision. Pattern Recognit. Lett.; 1984; 2, pp. 401-412. [DOI: https://dx.doi.org/10.1016/0167-8655(84)90007-2]
28. Ma, S.; Zhang, Z. Computer Vision: Basics of Computing Theory and Algorithms; Science Press: Beijing, China, 1998.
29. Wang, T.; Wang, L.; Zhang, W.; Duan, X.; Wang, W. Design of infrared target system with Zhang Zhenyou calibration method. Opt. Precis. Eng.; 2019; 27, pp. 1828-1835. [DOI: https://dx.doi.org/10.3788/OPE.20192708.1828]
30. Feng, W. Research and Implementation of Image Barrel Distortion Correction. Master’s Thesis; North China University of Technology: Beijing, China, 2011.
31. Zhang, Y.; Qiu, Z.; Zhang, X. Calibration method for hand-eye system with rotation and translation couplings. Appl. Opt.; 2019; 58, pp. 5375-5387. [DOI: https://dx.doi.org/10.1364/AO.58.005375]
32. Wu, A.; He, W.; Ouyang, X. A method of hand-eye calibration for palletizing robot based on openCV. Manuf. Technol. Mach. Tool; 2018; 6, pp. 45-49.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
In view of the phenomenon that camera imaging will appear large up close and small from afar in the eye-to-hand hand-calibration system, one hand–eye calibration is carried out. The manipulator is only suitable for grasping objects of the same height, and the calibration results cannot be applied to grasping products with variable height. Based on the study of the pinhole camera model and the rigid transformation model between coordinate systems, the introduction of the calibration height parameters, the relationship between parameters of the rigid transformation matrix between image the coordinate system and the robot coordinate system, and sampling height are established. In the experiment, firstly, through the calibration of camera parameters, the influence of camera distortion on imaging quality is eliminated, and the influence of calibration height is ignored. Then, the machine coordinate system and image coordinate system of the calibration plate at different heights are calibrated using the four-point calibration method. The parameters of the rigid transformation matrix at different heights (H) are calculated. Finally, through experimental analysis, the high linear relationship between the parameters of the rigid transformation matrix from the image coordinate system to the robot coordinate system and the calibration height is fitted. By analyzing the random error of the experiment, the linear relationship between calibration height and pixel density is further established, and the systematic error of the experimental process is deeply analyzed. The experimental results show that the hand–eye calibration system based on this linear relationship is precise and suitable for grabbing products of any height, and the positioning error is less than 0.08%.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer