1. Introduction
Depth estimation using stereo cameras involves estimating distance by analyzing the disparity between left and right images, a process essential for applications like the autonomous driving of robots or cars and object detection. Traditionally, the Semi-Global Matching (SGM) method, known for its speed and accuracy, has been widely used [1,2,3,4]. Recently, with the advancement of on-device AI technology, research on disparity estimation using neural networks has become increasingly active [5,6,7,8,9,10,11]. The core challenge of stereo matching lies in identifying the corresponding point in the right (R) image that matches a point in the left (L) image.
Image rectification is a technique that places L and R images on the same 3D plane and aligns matching points horizontally, simplifying the stereo matching problem from a 2D search to a 1D search. This technique is crucial for real-time performance in stereo camera matching, which demands high frame rates. However, rectification is challenging due to ambiguities in feature matching, lens distortion, sensitivity to initial conditions, and other factors. Notably, the rectification problem does not have a unique solution, leading to the development of various methods tailored to specific needs.
One solution to facilitate the detection of reference points for feature matching is to use images taken with a checkerboard chart, consisting of squares arranged in a pattern that is easily detectable within the camera’s field of view. Twists and misalignments often occur during the manufacturing assembly process for various reasons. Based on the known 3D pattern information, calibration to obtain intrinsic parameters such as focal length, camera center, and lens distortion is performed first. This is followed by rectification to find the common intrinsic parameters and rotation matrix for both cameras. Once the output parameters of the calibration and rectification are recorded in permanent memory, the captured images from the L and R cameras can be remapped on the fly.
Bouguet’s method [12], provided as a function in image processing tools such as MATLAB and OpenCV, is the most popular rectification algorithm. It rotates and scales the left and right images to align detected features on the same horizontal line. In most cases, the resulting images are well aligned, making it easy to compare reference points between the two images. However, while this method is suitable for two cameras with very different fields of view, it is not ideal for rectifying stereo cameras physically designed in parallel. Figure 1 shows a before-and-after rectification of L and R of a stereo camera using Bouguet’s method, where the resulting images tend to be excessively rotated compared to the original. This unintentional rotation can be problematic in applications such as autonomous driving, where the position or orientation of the camera apparatus is critical.
Recently, depth estimation cameras have started incorporating a third camera alongside the stereo pair. This additional camera, equipped with a filter that absorbs specific wavelengths, adds color information to the disparity map, which is then used to create a 3D point cloud [13]. This setup is also beneficial for multi-view stereo vision or high-precision 3D scanning systems, as it provides a broader field of view. However, there can be cases where the third camera and the depth information are not aligned, leading to poor 3D quality. This misalignment occurs because existing alignment methods do not consider the third camera, necessitating an additional calibration and rectification process to align the third camera with the stereo cameras.
In this paper, we introduce a novel rectification technique that simultaneously aligns the left, right, and third cameras without causing unintended image rotation, using just a single captured image. Unlike existing methods that necessitate multiple checkerboard images, our approach aligns and remaps all three cameras in one unified process. This method offers a significant advantage by eliminating the need for separate rectification and remapping for the third camera. Moreover, the rectified image used for disparity creation remains unrotated, ensuring a stable display. The proposed technique achieves stereo-matching performance comparable to Bouguet’s algorithm, even when aligning three cameras simultaneously, as evidenced by tests on 19 camera samples from the depth camera manufacturing process. The aligned images demonstrate superior disparity matching quality. This system is highly effective in correcting mechanical alignment errors in depth cameras and can significantly improve the accuracy of disparity estimation using deep learning.
We review previous work in Section 2, introduce the proposed rectification method in Section 3, and provide experimental results and analysis in Section 4. Further discussion is provided in Section 5. Section 6 concludes the paper.
2. Related Work
The Hartley [14] and the Bouguet rectifications [12] are widely recognized for their simplicity and performance, making them popular choices in image-processing toolkits like MATLAB and OpenCV. The Hartley algorithm focuses on aligning the epipolar lines of two images by finding and matching corresponding features between the left and right images. Although straightforward, it cannot accurately determine the size or distance of objects because it does not account for the camera’s intrinsic parameters.
In contrast, Bouguet’s method [12] is effective when camera calibration parameters are known. It aims to minimize reprojection distortion while maximizing the common viewing area. This method divides the rotation matrix into two separate matrices for the left and right cameras. Each camera rotates half a rotation to align their principal rays parallel to the vector sum of their original directions, achieving coplanar alignment but not row alignment. To horizontally align the epipolar lines, a rotation matrix is computed starting from the epipole’s direction, which is along the translation vector between the two cameras’ centers of projection.
Monasse [15] introduced a three-step image rectification method that improves stereo rectification by minimizing distortion through geometric transformations. This process involves aligning epipolar lines to make them parallel and horizontal, using a robust algorithm that splits the camera rotation into three steps, thereby reducing computational complexity and improving accuracy. Unlike traditional methods, it minimizes a single parameter, the focal length, making it more resilient to initial epipolar line misalignment.
Kang [16] presented a stereo image rectification method using a horizontal baseline to address geometric errors caused by manual camera arrangements and internal camera characteristic differences. Unlike traditional calibration-based methods that often result in visual distortions like image skewness, this method calculates a baseline parallel to the real-world horizontal line, estimates camera parameters, and applies a rectification transform to the original images to produce rectified stereo images with minimal distortion.
Isgrb [17] proposed a novel algorithm for projective rectification that bypasses the explicit computation of epipolar geometry or the fundamental matrix. Instead, it leverages the known form of the fundamental matrix for rectified image pairs to establish a minimization problem that directly yields the rectifying homographies from image correspondences. This approach simplifies the correspondence problem by aligning corresponding points along horizontal scanlines, thus reducing the problem from 2D to 1D.
Pollefeys [18] introduced a rectification method for stereo images capable of handling all possible camera motions. This method employs a polar parameterization of the image around the epipole, transferring information between images through the fundamental matrix. This approach avoids the issues encountered by traditional and cylindrical rectification methods, which can result in excessively large images or fail to rectify altogether.
Fusiello [19] presented a linear rectification method for general, unconstrained stereo rigs, which computes rectifying projection matrices from the original camera’s perspective projection matrices. This method computes rectifying projection matrices from the original camera’s perspective projection matrices, providing a compact and easily reproducible solution.
Lafiosca [20] introduced a rectification method to compute a homography that reduces perspective distortion, offering an improvement over the rectification method by Loop and Zhang [21]. This approach provides a closed-form solution that avoids the need for numerical optimization.
While existing methods are general-purpose algorithms designed to align images taken from various angles and primarily focus on reducing alignment errors, the proposed method is specifically tailored for mass-producing stereo cameras. It simultaneously aligns three cameras, ensuring that the reference image is not unintentionally rotated. To facilitate this process, we also propose an image-capturing environment that includes a specially designed checkerboard and its specific configuration.
3. Proposed Method
Figure 2 visualizes the existing calibration and rectification methods for aligning the left, right, and RGB cameras, as well as the proposed triple-camera alignment system. In existing systems, depth estimation typically involves the left and right cameras, with an additional RGB camera for generating the 3D point cloud. These systems require two separate calibration and rectification processes: one for estimating depth using the left and right cameras, and another for encoding color information using the left and RGB cameras. The rectification outputs a rotation matrix in a common 3D plane, considering the relative relationship between the two cameras. Although the left camera is common to both rectification processes, the rotation matrix and common plane for each are different. Additionally, the existing rectification process is time-consuming because multiple checkerboard images are required to extract more features and determine the intrinsic and extrinsic parameters of the cameras. In contrast, the proposed method uses a single image for each camera and requires only one unified calibration and rectification step, where both the right and RGB cameras are calibrated and rectified relative to the left camera, which serves as the main reference for depth estimation and 3D point-cloud generation.
3.1. Checkerboard Design and Lab Setting
To perform camera calibration and rectification, it is necessary to capture multiple chart images from different positions to detect feature points. The proposed method utilizes four checkerboards for a single shot, as illustrated in Figure 3A, each placed in a different quadrant. Each checkerboard consists of a 19 × 14 grid pattern, with squares measuring 24 mm on each side. The top-left checkerboard faces straight ahead, while the other three are rotated approximately 30 degrees around different axes, facilitating the generation of 3D multi-view geometry points. Each checkerboard is positioned to include as much of the corner areas of the image as possible, ensuring that distortion correction and rectification are effectively applied to the edges of the image. With this setup, a total of 228 × 4 = 912 reference points are obtained.
3.2. Checkerboard Corner Detection
The captured image in Section 3.1 is divided into four sections to utilize the existing corner-detection methods, and corner detection is performed in each section. To facilitate this, the remaining area except the approximate chart area considering the margin is first masked in black as described in Figure 3B, and then corner detection is performed. Various feature detection methods such as Harris corner detection [22], SIFT [23], SURF [24], ORB [25], FAST [26], BRIEF [27], HOG [28], and MSER [29] often show poor performance near the out-of-focus area (in our case, it mostly happens in the off-center regions), due to lens-assembly defects that often occur in the early stage of manufacturing. We adopt Geiger’s method [30], which accurately and reliably detects corner locations even in chart images with distorted or blurred areas. The results of detecting corners in the checkerboard image using the proposed method are shown in Figure 4. The red dots in each grid pattern represent the corners detected in the left image, and the same corner-detection process is applied to the right and RGB images.
3.3. Calibration and Rectification
For triple-camera rectification, the process begins with single camera calibration for the left, right, and RGB images to determine each camera’s intrinsic/extrinsic parameters and lens-distortion coefficients. This involves obtaining extrinsic parameters, which represent the transformation from the 3D world to the camera coordinate system for the checkerboard reference point, and intrinsic parameters, which map 3D camera coordinates to 2D image coordinates [31,32]. Figure 5 illustrates the coordinates and transformation matrices in 2D and 3D during the rectification process, depicting the transition from observed to rectified points for a single camera.
3.3.1. Single-Camera Calibration
The calibration process involves determining the camera matrix , called an intrinsic matrix and extrinsic matrix . The intrinsic matrix comprises the focal lengths and along the and axes, and the principal points and , as depicted in Equation (1).
(1)
The extrinsic matrix consists of rotation and translation parameters that transform the 3D corners in world coordinates into the 3D coordinates of the camera, as shown in Equation (2). Since the proposed system uses four checkerboards, there are four extrinsic matrices (, , , ). Each external matrix corresponding a single checkerboard consists of a 3 × 3 rotation matrix and a 3 × 1 translation matrix along the three axes.
(2)
In Figure 5, and represent the corner points in the 2D raw image and the 3D raw plane, respectively. Additionally, and are the 2D and 3D coordinates after lens-distortion correction, and and are the corresponding coordinates after rectification.
(3)
Equation (3) describes the lens-distortion model, where (, ) and (, ) are the coordinates of the corner points in the distorted (observed) and lens-corrected image, respectively [31]. and in Equation (3) represent the first and second lens-distortion coefficients. The distance between the principal point and the distortion-corrected coordinates determines the degree of radial distortion.
and from the 3D to 2D transformation matrix are used for both the distorted and corrected images, as shown in Equation (3). Likewise, the conversion from 2D image coordinates to 3D homogenous coordinates is achieved by multiplying to and . Given the distorted image points and the corresponding 3D corner points in world coordinates, obtaining the undistorted function, in Figure 6 is an inverse problem. This can typically be estimated using iterative methods or approximate algorithms, such as the Newton–Raphson method [33] or numerical optimization [34].
In the calibration process, the intrinsic matrix is obtained first by singular value decomposition (SVD). The initial extrinsic matrix is subsequently derived [31]. The 3D world coordinates of checkboard corners are transformed to camera 3D coordinate by multiplying . The initial and are then obtained using Equation (3). Once the initial parameters are established, optimization is performed to minimize the Euclidean error between the observed corner points and the estimated corner points using Equation (4).
(4)
where and are the number of checkerboards in the chart image and the total number of corner points in each checkerboard, respectively ( = 4, = 228). While the intrinsic matrix is unaffected by external environmental factors, vary depending on the camera or checkboard position. Thus, is required only for the optimization process and is not documented for calibration purposes. The 3D points in world coordinates depend on each checkerboard’s locations. However, even if we assume that the world coordinates of the four checkerboards are the same, as long as only changes, it is acceptable to use instead of . In Equation (4), the intrinsic/extrinsic matrices and lens-distortion coefficients are optimized to minimize the difference between the observed points and estimated points in 2D image coordinates.3.3.2. Triple-Camera Calibration
Once the intrinsic and extrinsic parameters and the lens-distortion coefficients for each camera are obtained, the relative relationship of the three cameras is considered. As mentioned in [12], the relative rotation and translation of the right to the left camera are expressed in Equation (5). Similarly, the relative rotation and translation of the RGB to the left camera are expressed in Equation (6).
(5)
(6)
where , , and are the left, right and RGB camera’s external rotation matrices, respectively, and , , and are the translation matrices. Although the relative rotation and translation matrices should theoretically be the same regardless of the checkerboard location, slight variations occur. A single set of , , , and is determined, and the intrinsic matrices of each camera are adjusted for consistency. The loss function in Equation (7) minimizes the difference between the observed and the projected 2D points in each camera using Equations (5) and (6). For the extrinsic matrices in Equation (7), only are optimized, while the extrinsic matrices of the right camera and the RGB camera are computed based on the relative rotation and translation, as described in Equation (8).(7)
(8)
For initial values of these relative matrices, median matrix values across all checkboards are used.
3.3.3. Triple-Camera Rectification
The triple-camera calibration performed in the previous section does not ensure that the epipolar lines of all three cameras are horizontally aligned, indicating that the image planes of the three cameras in 3D space are not equivalent. Rectification aims to find a common intrinsic matrix and rotation matrices for each camera, creating a common plane for all and making the epipolar lines parallel. To achieve this, the corrected coordinates of the observed points in 2D image space is first estimated using the function mentioned in Section 3.3.1. The distortion-corrected point is transformed to 3D coordinates by multiplying the inverse of , and then rotated again by to become rectified points in 3D camera coordinates. Lastly, distortion-free and rectified is obtained by multiplying the common intrinsic matrix . In summary, the rectified 2D image point from the observed point in a camera is calculated as follows:
(9)
The resulting rotation matrix for each camera necessarily rotates the original image to place 3D points in a common plane. However, there can be multiple common planes to satisfy the condition. The misalignment in the manufacturing process among different cameras is not usually severe. Disparity computation is performed in various ways from the rectified images, and the resulting disparity map should be visible to the external display. The rectified, randomly rotated field of view in the disparity map is not only unsuitable for visual display, but also inadequate for distance estimation in autonomous driving. Therefore, the proposed method is designed to minimize the vertical location error of the rectified points subjected to as follows:
(10)
where represents the -coordinate of the rectified 2D point in Equation (9) for each camera. is a minimum threshold value for the ratio of the focal length in the rectified image, to the focal length of the left camera . In the process of finding , , that minimize Equation (10), is fixed as an identity matrix. Consequently, if there is significant misalignment among three cameras, reducing to decrease the size of the rectified image naturally helps minimize the vertical difference between rectified points. However, since reducing the image size is not desirable, a threshold value is used to ensure a minimum size for , thereby preventing the rectified image from being excessively reduced. Once , , are obtained, the relationship between and is established and stored as a mapping table for each camera. This mapping is then used to remap and for each frame, resulting in a rectified image. The optimizations in Equations (4), (7) and (10) uses the Levenberg–Marquardt (LM) algorithm [35]. This method combines the Gradient descent method and the Gaussian–Newton method, offering both speed and stability. The Trust-Region-Reflective algorithm [36] can also be selected as an optimization method. We present a summary of the proposed algorithm below as Algorithm 1, along with an example illustrating the estimated parameters and errors produced at each step.Algorithm 1. Require: three checkerboard images captured by cameras arranged in a row [30,31]. |
4. Experimental Results
To evaluate the performance of the proposed method, a newly designed stereoscopic camera with a wide field of view was constructed, as shown in Figure 6A. During the manufacturing process, 19 pairs of left, right, and RGB raw images with a resolution of 1280 × 800 were captured from 19 different camera modules. These images were used to capture the displayed checkerboard pattern. The left/right images are in 8-bit format, and the RGB images are in 24-bit JPG format. During the calibration and rectification process, the images are resized to an output size of 848 × 480 to optimize real-time performance in stereo matching. The evaluation metric for the vertical alignment error between two rectified cameras A and B is based on the Mean Absolute Error (MAE), calculated as follows:
(11)
Table 1 shows and from the rectified images for the existing methods and the proposed method, across 19 samples. All methods used the same corner-detection results before calibration, and for the Fusiello [19] and Lafiosca [20] methods, which do not provide calibration, we used the intrinsic parameters from the triple-calibration results of the proposed method. All comparison methods were tested separately for the L/R and L/RGB cases, whereas only the proposed method performed the simultaneous calibration and rectification of L/R/RGB. of 0.98 was used in Equation (10) unless otherwise specified. The proposed method achieved an average alignment error of 0.112 for L/R and 0.085 for L/RGB, ranking third among all compared methods. Bouguet and Lafiosca’s methods, which are based on calibrated scenarios, achieved the best scores, while Hartley’s method, designed for an uncalibrated scenario, ranked fourth. Fusiello’s method, although relatively simple and aimed at calibrated cameras, lacks lens-distortion modeling, resulting in the worst performance.
To scrutinize the rectification results in Table 1, we listed the relative angles and locations among the L/R/RGB cameras from the triple-camera calibration in Table 2. Since the RGB camera images are converted to grayscale before processing, it is assumed that there are no distinct differences between L/R and L/RGB rectification, except for the baseline, which reflects the relative y locations in Table 2. In all methods, the average error for L/RGB in Table 1 is approximately 30% lower than that for L/R, which can be attributed to the difference in baselines. This is because, for camera pairs with the same angular deviation around the XYZ axes, a longer baseline makes the rectification process more sensitive to rotations around these axes. Any small misalignment in rotation can cause significant misalignment in the rectified images, increasing the overall error.
The goal of rectification is to simplify stereo matching by reducing a 2D search to a 1D search. To evaluate how the error difference between Bouguet’s method and the proposed method affects stereo-matching results, we computed the disparity using the SGM method [1] on left and right images before and after rectification. Figure 7 compares the resulting depth maps from 19 camera samples. The first to third columns show the original left images, the corresponding rectified images from Bouguet’s method, and the rectified images from the proposed method, respectively. The fourth to last columns display the disparity map without rectification, the disparity maps obtained using Bouguet’s method, and those generated using the proposed method. The disparity maps from both Bouguet’s and the proposed rectified images are quite similar, with both enhancing the disparity map to a level where the 3D geometry of the calibration chart can be reliably estimated. However, the rectified left images from Bouguet’s method in the second column show random rotations of the reference checkerboards, which were vertical in the original images, whereas the proposed method preserves the original checkerboard orientation in all rectified images. This tendency is also evident in the disparity maps in the fifth and sixth columns of Figure 7.
To determine the extent of undesired perspective rotation in the rectified images, we obtained the positions of the three red-marked points on the front-facing chart in the top left corner of Figure 3A and calculated the horizontal and vertical angle rotations after rectification as indicated by the red arrows. Figure 8, Figure 9 and Figure 10 illustrate the results for the left, right, and RGB images for each method. Existing methods, such as [12,19,20], exhibit significant rotations from the original images, with a maximum deviation of approximately 3 degrees in both horizontal and vertical directions, regardless of whether the images are left, right, or RGB. In contrast, the proposed method shows almost no rotation for the reference left image, with a maximum deviation of only 0.17 degrees, which is likely due to the scaling of the rectified image. For the right and RGB-rectified images in the proposed method, there is an average rotation of about 0.4 degrees, which is still significantly lower compared to the existing methods.
The perspective rotation in the rectification process appears to mainly result from the vertical misalignment of the stereo cameras. The samples with severe rotations have relatively large vertical misalignments of more than 1 mm (y-axis), as indicated in Table 2: modules 1, 8, 10, 11, 13, 15, 17, and 18 for L/R cameras, and modules 6, 9, 15, 16, and 18 for L/RGB cameras. Since the alignment error is correlated with the baseline, it is observed that a physical vertical misalignment between two closely positioned cameras can also lead to significant rotations in the rectified images, resulting in a similar vertical alignment between L/R and L/RGB, but with a larger angle of rotation in L/RGB than in L/R. If existing methods are used, the modules with larger rotation images may need to be discarded during the manufacturing process due to the substantial perspective rotations observed after rectification. Moreover, the resulting left images from L/R and L/RGB rectification in the existing methods are independent, and therefore cannot be used simultaneously, as in the proposed method. Existing methods focus on minimizing the alignment error of stereo pairs. In contrast, the proposed algorithm rectifies all three cameras simultaneously while keeping the reference camera’s movement fixed, permitting minor alignment compromises as long as they do not impact stereo matching. Consequently, the performance gap between Bouguet’s method and the proposed method in Table 1 appears to stem from the reduced degrees of freedom caused by fixing the left camera’s rotation and the concurrent rectification of the third camera.
Figure 11, Figure 12, Figure 13 and Figure 14 provide a comparison of the results before and after rectification for real-world scenes. The first row shows the original images from the left, right, and RGB cameras, while the second row offers an enlarged view of the dotted box area in each image. The third row presents the rectified images after applying the proposed method, and the fourth row displays the matching points identified using SURF. Note that all images are shown without noise reduction or other ISP tuning, leading to noticeable noise in all images and a purple tint in the RGB images. Before rectification, the images are not vertically aligned, but after rectification, the left, right, and RGB images are nearly perfectly aligned, with no unwanted rotation in the left image, which serves as the reference for disparity estimation throughout the study. The alignment errors for each process and the number of corners used in the calculation are detailed in the figure captions.
5. Discussion
5.1. Rectification with Multiple Images
Although the proposed method uses a single image for calibration and rectification, this image includes four checkerboard sub-images. Theoretically, using four separate images, each containing a single checkerboard, would not impact the accuracy of rectification. However, if each checkerboard is positioned arbitrarily and overlaps in areas of the viewing angle, it could affect the accuracy scores. To assess the accuracy of rectification with multiple captures, we created three different capturing environments, as illustrated in Figure 15, and compared them with the single-capture scenario. Table 3 summarizes the vertical alignment errors across different settings. Multi-I achieved the highest accuracy in both L/R and L/RGB rectifications, followed by Multi-II for L/R, and then the single-image rectification. Multi-III resulted in the highest error. This variation is primarily due to the coverage of the four checkerboards; overlapping checkerboard locations, as seen in Multi-I and Multi-II, help to minimize errors, while dispersed placements, as in Multi-III, lead to increased alignment errors. We also compared the rectification accuracy for the real-world scenes in Figure 11, Figure 12, Figure 13 and Figure 14 using SURF between the single-capture and Multi-I scenarios, as summarized in Table 4. However, assuming that using SURF to detect feature points in real-world images is numerically less accurate than corner detection on the checkerboard, no significant differences were observed, and both methods were able to find stable matching points.
5.2. The Affect of in Optimization
Unlike existing methods, where the left and right images are freely rotated to align the epipolar lines horizontally, the proposed method only rotates and scales the right or RGB cameras, keeping the rotation matrix of the left camera fixed as the identity matrix. Consequently, the degrees of freedom in the optimization process described in Equation (10) are reduced compared to cases where no additional constraints are applied. In general, as the focal length of the rectified image decreases, the image size also reduces, leading to a decrease in the vertical alignment error between the two cameras. An incorrect choice of value could prevent the optimization from converging and impact the accuracy of the rectification. Therefore, we tested the alignment error and with varying values in (10). Figure 16 summaries the results of vertical alignment errors and focal length changes. As shown in Figure 16A,B, with the exception of module 12, most of the modules exhibit a decrease in alignment error as becomes lowered. This is because the focal length of the rectified images becomes smaller, as seen in Figure 16C. Smaller focal lengths result in smaller image sizes, which is advantageous for calculating vertical errors. Therefore, in the optimization of (10), the process tends to reduce the image size down to the lower limit of the focal length to minimize errors. Module 12, however, has a very small relational location in the y-axis for both L/R and L/RGB, as shown in Table 2, and also exhibits minimal angle differences. This indicates that the optimization for module 12 was primarily carried out using the rotation matrices, with little to no change due to the focal length constraint.
6. Conclusions and Future Work
In this paper, we propose a method for simultaneously rectifying left, right, and RGB cameras. We comprehensively outline the entire rectification process of the proposed method using the intrinsic/extrinsic calibration and lens-distortion correction formulas. Unlike existing methods that require multiple images to rectify two cameras, our approach enables the simultaneous rectification of three camera images with a single shot, ensuring that the images do not undergo unnecessary rotation after rectification. This advantage makes the proposed technique highly useful for the production of stereo cameras for distance estimation. In our experiment, we showed that the vertical misalignment between two cameras plays a key role in perspective rotation images in the rectification. Therefore, including the camera’s translation in the optimization process for rectification is anticipated to be a topic for future research.
Methodology, S.W.; Software, M.J. and S.W.; Validation, M.J.; Formal analysis, J.-W.K.; Investigation, M.J.; Writing – original draft, M.J.; Writing – review & editing, S.W.; Supervision, S.W.; Project administration, J.P. and J.-W.K.; Funding acquisition, J.P. and J.-W.K. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.
The authors declare no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 1. Results of Bouguet’s rectification for L and R images. (Top) Before rectification, (Bottom) after rectification, (Left) Images from left camera, (Right) Images from right camera. Please note that, although the resulting images are well-rectified, they still exhibit rotation even with minor L/R misalignment inputs.
Figure 2. (A) A diagram of the existing calibration and rectification system based on stereo cameras and (B) the proposed simultaneous triple-camera calibration and rectification system.
Figure 3. (A) Checkerboard configuration designed for single-shot calibration and rectification for three cameras. (B) Four segmented masked images for corner detection.
Figure 5. Calibration, lens-distortion correction, and rectification process diagram for a single camera.
Figure 6. (A) Our stereoscopic camera with a wide field of view for distance measurement; (B) camera configuration and XYZ coordinates.
Figure 7. Rectification performance comparisons for each of the 19 modules. (A) Original left image, (B) rectified image using Bouguet’s method, (C) rectified image using the proposed method, (D) SGM disparity map using unrectified inputs, (E) SGM disparity map using Bouguet’s method, (F) SGM disparity map using the proposed method.
Figure 7. Rectification performance comparisons for each of the 19 modules. (A) Original left image, (B) rectified image using Bouguet’s method, (C) rectified image using the proposed method, (D) SGM disparity map using unrectified inputs, (E) SGM disparity map using Bouguet’s method, (F) SGM disparity map using the proposed method.
Figure 8. Rotated angle distributions in the rectified left images in the horizontal and vertical directions.
Figure 9. Rotated angle distributions in the rectified right images in the horizontal and vertical directions.
Figure 10. Rotated angle distributions in the rectified RGB images in the horizontal and vertical directions.
Figure 11. Comparison before and after rectification using the proposed method—scene 1 (left/right/RGB). Alignment errors for original images: L/R: 2.96, L/RGB: 6.75; for rectified images: L/R: 0.95, L/RGB: 0.89. Number of matching points: 33.
Figure 12. Comparison before and after rectification using the proposed method—scene 2 (left/right/RGB). Alignment errors for original images: L/R: 4.51, L/RGB: 8.21; for rectified images: L/R: 0.99, L/RGB: 1.08. Number of matching points: 34.
Figure 13. Comparison before and after rectification using the proposed method—scene 3 (left/right/RGB). Alignment errors for original images: L/R: 2.64, L/RGB: 6.36; for rectified images: L/R: 0.89, L/RGB: 0.85. Number of matching points: 46.
Figure 14. Comparison before and after rectification using the proposed method—scene 4 (left/right/RGB). Alignment errors for original images: L/R: 3.05, L/RGB: 7.49; for rectified images: L/R: 0.83, L/RGB: 1.05. Number of matching points: 53.
Figure 15. Four calibration settings for comparing multiple and single captures: (A) Multi-I: the first multiple capture, (B) Multi-II: the second multiple capture, (C) Multi-III: the third multiple capture, and (D) represents a single capture.
Figure 16. Vertical alignment errors by [Forumla omitted. See PDF.] for each sample: (A) L/R, (B) L/RGB, and changes in focal length of rectified image (C).
Comparison of
Num. | Original | Hartley [ | Fusiello [ | Lafiosca [ | Bouguet [ | Proposed | ||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
L/R | L/RGB | L/R | L/RGB | L/R | L/RGB | L/R | L/RGB | L/R | L/RGB | L/R | L/RGB | |
1 | 3.502 | 5.139 | 0.147 | 0.112 | 0.143 | 0.141 | 0.107 | 0.070 | 0.107 | 0.077 | 0.127 | 0.090 |
2 | 3.693 | 2.650 | 0.142 | 0.101 | 0.216 | 0.159 | 0.097 | 0.045 | 0.088 | 0.046 | 0.115 | 0.043 |
3 | 1.314 | 6.671 | 0.105 | 0.097 | 0.193 | 0.161 | 0.051 | 0.080 | 0.059 | 0.069 | 0.077 | 0.090 |
4 | 5.161 | 6.009 | 0.132 | 0.124 | 0.181 | 0.169 | 0.067 | 0.078 | 0.067 | 0.071 | 0.135 | 0.069 |
5 | 2.368 | 1.880 | 0.122 | 0.100 | 0.192 | 0.147 | 0.044 | 0.050 | 0.056 | 0.053 | 0.074 | 0.067 |
6 | 4.007 | 1.020 | 0.092 | 0.101 | 0.144 | 0.130 | 0.053 | 0.084 | 0.060 | 0.068 | 0.094 | 0.092 |
7 | 0.445 | 8.102 | 0.142 | 0.077 | 0.198 | 0.133 | 0.070 | 0.034 | 0.071 | 0.039 | 0.139 | 0.059 |
8 | 4.052 | 6.075 | 0.195 | 0.122 | 0.266 | 0.194 | 0.105 | 0.054 | 0.101 | 0.060 | 0.138 | 0.062 |
9 | 1.572 | 1.959 | 0.110 | 0.080 | 0.204 | 0.155 | 0.063 | 0.042 | 0.056 | 0.052 | 0.109 | 0.096 |
10 | 10.372 | 4.878 | 0.136 | 0.096 | 0.195 | 0.164 | 0.093 | 0.095 | 0.084 | 0.070 | 0.101 | 0.071 |
11 | 4.902 | 13.379 | 0.121 | 0.083 | 0.171 | 0.190 | 0.044 | 0.027 | 0.062 | 0.038 | 0.081 | 0.100 |
12 | 0.446 | 2.356 | 0.100 | 0.085 | 0.216 | 0.146 | 0.086 | 0.044 | 0.064 | 0.043 | 0.063 | 0.087 |
13 | 6.517 | 4.865 | 0.162 | 0.100 | 0.239 | 0.176 | 0.181 | 0.062 | 0.140 | 0.054 | 0.161 | 0.065 |
14 | 1.622 | 1.115 | 0.103 | 0.077 | 0.161 | 0.113 | 0.037 | 0.027 | 0.045 | 0.041 | 0.056 | 0.038 |
15 | 5.822 | 1.365 | 0.126 | 0.107 | 0.167 | 0.147 | 0.072 | 0.071 | 0.073 | 0.064 | 0.174 | 0.102 |
16 | 1.111 | 0.752 | 0.132 | 0.112 | 0.205 | 0.145 | 0.111 | 0.067 | 0.088 | 0.081 | 0.138 | 0.208 |
17 | 6.145 | 3.775 | 0.170 | 0.130 | 0.289 | 0.157 | 0.056 | 0.060 | 0.076 | 0.074 | 0.118 | 0.084 |
18 | 5.750 | 3.420 | 0.120 | 0.104 | 0.161 | 0.149 | 0.040 | 0.050 | 0.051 | 0.061 | 0.135 | 0.112 |
19 | 0.898 | 0.911 | 0.116 | 0.103 | 0.194 | 0.133 | 0.056 | 0.064 | 0.050 | 0.060 | 0.091 | 0.074 |
min | 0.445 | 0.752 | 0.092 | 0.077 | 0.143 | 0.113 | 0.037 | 0.027 | 0.045 | 0.038 | 0.056 | 0.038 |
max | 10.372 | 13.379 | 0.195 | 0.130 | 0.289 | 0.194 | 0.181 | 0.095 | 0.140 | 0.081 | 0.174 | 0.208 |
avg | 3.668 | 4.017 | 0.130 | 0.101 | 0.197 | 0.153 | 0.075 | 0.058 | 0.074 | 0.059 | 0.112 | 0.085 |
Relational angles and locations between L/R and L/RGB camera from calibration. Note that applying this value to the R or RGB image will transform their coordinates to match those of the L camera. Misalignments exceeding ±1 mm in the y axes are marked in red.
Num. | L/R | L/RGB | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Relational Angle (Degree) | Relational Location (mm) | Relational Angle (Degree) | Relational Location (mm) | |||||||||
x | y | z | x | y | z | x | y | z | x | y | z | |
1 | 0.24 | 0.42 | −1.16 | −49.10 | −0.92 | 0.06 | 0.111 | 0.084 | −0.490 | −36.322 | −0.966 | −0.095 |
2 | −0.19 | 0.02 | −0.22 | −50.61 | | −0.31 | −0.006 | −0.203 | −0.217 | −38.061 | −0.187 | −0.104 |
3 | −0.04 | −0.62 | −0.36 | −50.82 | −0.65 | −0.44 | −0.169 | −0.440 | −0.426 | −37.803 | 0.572 | 0.081 |
4 | −1.81 | 0.38 | 0.04 | −49.35 | −0.85 | −0.38 | −1.407 | 0.367 | −0.103 | −36.140 | −0.617 | 0.042 |
5 | −0.14 | −0.20 | −0.60 | −50.68 | 0.18 | 0.11 | −0.160 | −0.256 | −0.465 | −37.325 | 0.732 | 0.082 |
6 | −0.01 | 0.28 | −0.42 | −49.95 | −0.39 | −0.21 | 0.000 | 0.171 | −0.262 | −36.943 | | −0.176 |
7 | −0.16 | 0.32 | −0.03 | −50.45 | 0.66 | −0.38 | −0.068 | 0.560 | −0.502 | −37.787 | −0.303 | 0.602 |
8 | −0.09 | −0.38 | 0.10 | −49.78 | | −0.66 | 0.020 | −0.628 | 0.246 | −37.083 | 0.152 | −0.308 |
9 | 0.03 | −0.14 | −0.20 | −50.34 | −0.47 | −0.64 | 0.411 | −0.422 | −0.402 | −37.850 | | −0.312 |
10 | −0.08 | −0.12 | −0.14 | −50.03 | | −0.41 | −0.064 | −0.296 | −0.417 | −37.759 | 0.737 | −0.338 |
11 | −0.04 | 0.17 | 0.36 | −50.30 | | 0.34 | 0.037 | 0.013 | −0.076 | −38.085 | −0.517 | 0.492 |
12 | 0.07 | −0.05 | −0.02 | −51.69 | −0.32 | 0.10 | −0.107 | 0.094 | 0.125 | −37.796 | −0.111 | −0.350 |
13 | −0.22 | −0.19 | 0.26 | −50.94 | | −0.17 | −0.156 | −0.078 | 0.299 | −37.358 | 0.419 | −0.226 |
14 | −0.03 | −0.19 | 0.01 | −50.69 | −0.43 | −0.35 | 0.059 | −0.208 | −0.393 | −37.260 | 0.105 | −0.241 |
15 | −0.13 | 0.20 | −0.72 | −49.18 | | −0.69 | −0.078 | 0.094 | −0.211 | −36.592 | | −0.378 |
16 | −0.07 | −0.16 | −0.34 | −49.76 | | −0.62 | −0.151 | −0.026 | −0.071 | −35.715 | | −1.003 |
17 | 0.11 | −0.44 | −0.71 | −50.66 | | −0.35 | −0.007 | −0.396 | −0.854 | −37.227 | −0.766 | −0.397 |
18 | −0.11 | −0.06 | −0.08 | −50.33 | −0.91 | −0.73 | −0.017 | −0.148 | −0.120 | −37.484 | | −0.031 |
19 | −0.06 | −0.02 | −0.37 | −50.57 | −0.02 | −0.41 | −0.143 | −0.001 | −0.343 | −36.673 | 0.060 | −0.234 |
min | −0.01 | −0.02 | 0.01 | −49.10 | −0.02 | 0.06 | 0.000 | −0.001 | −0.071 | −35.715 | 0.060 | −0.031 |
max | −1.81 | −0.62 | −1.16 | −51.69 | 2.04 | −0.73 | −1.407 | −0.628 | −0.854 | −38.085 | −1.993 | −1.003 |
avg | −0.14 | −0.04 | −0.24 | −50.28 | 0.12 | −0.32 | −0.100 | −0.090 | −0.246 | −37.224 | −0.015 | −0.152 |
Summary of rectification accuracy of checkboard corners for multiple and single captures.
Multiple Capture | Single Capture | ||||||
---|---|---|---|---|---|---|---|
Multi-I | Multi-II | Multi-III | |||||
L/R | L/RGB | L/R | L/RGB | L/R | L/RGB | L/R | L/RGB |
0.047 | 0.037 | 0.070 | 0.102 | 0.195 | 0.201 | 0.106 | 0.094 |
Summary of rectification accuracy in
| | | | |||||
---|---|---|---|---|---|---|---|---|
L/R | L/RGB | L/R | L/RGB | L/R | L/RGB | L/R | L/RGB | |
Multi-I | 0.269 | 0.338 | 0.427 | 0.442 | 0.462 | 0.484 | 0.320 | 0.432 |
Single | 0.236 | 0.374 | 0.667 | 0.615 | 0.338 | 0.388 | 0.328 | 0.437 |
References
1. Hirschmuller, H. Stereo processing by semiglobal matching and mutual information. IEEE Trans. Pattern Anal. Mach. Intell.; 2007; 30, pp. 328-341. [DOI: https://dx.doi.org/10.1109/TPAMI.2007.1166] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/18084062]
2. Facciolo, G.; De Franchis, C.; Meinhardt, E. MGM: A significantly more global matching for stereovision. Proceedings of the British Machine Vision Conference (BMVC); Swansea, UK, 7–10 September 2015.
3. Bethmann, F.; Luhmann, T. Semi-global matching in object space. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci.; 2015; 40, pp. 23-30. [DOI: https://dx.doi.org/10.5194/isprsarchives-XL-3-W2-23-2015]
4. Schauwecker, K. Real-time stereo vision on FPGAs with SceneScan. Proceedings of the Forum Bildverarbeitung; Karlsruhe, Germany, 29–30 November 2018; Volume 339.
5. Chang, J.R.; Chen, Y.S. Pyramid stereo matching network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Salt Lake City, UT, USA, 18–23 June 2018; pp. 5410-5418.
6. Chuah, W.; Tennakoon, R.; Hoseinnezhad, R.; Bab-Hadiashar, A. Deep learning-based incorporation of planar constraints for robust stereo depth estimation in autonomous vehicle applications. IEEE Trans. Intell. Transp. Syst.; 2021; 23, pp. 6654-6665. [DOI: https://dx.doi.org/10.1109/TITS.2021.3060001]
7. Xu, G.; Wang, X.; Ding, X.; Yang, X. Iterative geometry encoding volume for stereo matching. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Vancouver, BC, Canada, 14–24 June 2023; pp. 21919-21928.
8. Kendall, A.; Martirosyan, H.; Dasgupta, S.; Henry, P.; Kennedy, R.; Bachrach, A.; Bry, A. End-to-end learning of geometry and context for deep stereo regression. Proceedings of the IEEE International Conference on Computer Vision; Venice, Italy, 22–29 October 2017; pp. 66-75.
9. Tankovich, V.; Hane, C.; Zhang, Y.; Kowdle, A.; Fanello, S.; Bouaziz, S. Hitnet: Hierarchical iterative tile refinement network for real-time stereo matching. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Nashville, TN, USA, 20–25 June 2021; pp. 14362-14372.
10. Guo, X.; Yang, K.; Yang, W.; Wang, X.; Li, H. Group-wise correlation stereo network. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Long Beach, CA, USA, 15–20 June 2019; pp. 3273-3282.
11. Huang, B.; Zheng, J.Q.; Giannarou, S.; Elson, D.S. H-net: Unsupervised attention-based stereo depth estimation leveraging epipolar geometry. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; New Orleans, LA, USA, 18–24 June 2022; pp. 4460-4467.
12. Bradski, G.; Kaehler, A. Learning OpenCV: Computer Vision with the OpenCV Library; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2008.
13. Alshawabkeh, Y. Linear feature extraction from point cloud using color information. Herit. Sci.; 2020; 8, 28. [DOI: https://dx.doi.org/10.1186/s40494-020-00371-6]
14. Hartley, R.I. Theory and practice of projective rectification. Int. J. Comput. Vis.; 1999; 35, pp. 115-127. [DOI: https://dx.doi.org/10.1023/A:1008115206617]
15. Monasse, P.; Morel, J.M.; Tang, Z. Three-step image rectification. Proceedings of the BMVC 2010-British Machine Vision Conference; Aberystwyth, UK, 31 August–3 September 2010; BMVA Press: Durham, UK, 2010; pp. 89-91.
16. Kang, Y.S.; Ho, Y.S. Efficient stereo image rectification method using horizontal baseline. Advances in Image and Video Technology: 5th Pacific Rim Symposium, PSIVT 2011, Gwangju, Republic of Korea, 20–23 November 2011, Proceedings, Part I; Springer: Berlin/Heidelberg, Germany, 2012; pp. 301-310.
17. Isgro, F.; Trucco, E. Projective rectification without epipolar geometry. Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149); Fort Collins, CO, USA, 23–25 June 1999; IEEE: Piscataway, NJ, USA, 1999; Volume 1, pp. 94-99.
18. Pollefeys, M.; Koch, R.; Van Gool, L. A simple and efficient rectification method for general motion. Proceedings of the Seventh IEEE International Conference on Computer Vision; Kerkyra, Greece, 20–27 September 1999; IEEE: Piscataway, NJ, USA, 1999; Volume 1, pp. 496-501.
19. Fusiello, A.; Trucco, E.; Verri, A. A compact algorithm for rectification of stereo pairs. Mach. Vis. Appl.; 2000; 12, pp. 16-22. [DOI: https://dx.doi.org/10.1007/s001380050120]
20. Lafiosca, P.; Ceccaroni, M. Rectifying homographies for stereo vision: Analytical solution for minimal distortion. Proceedings of the Science and Information Conference; London, UK, 14–15 July 2022; Springer: Cham, Switzerland, 2022; pp. 484-503.
21. Loop, C.; Zhang, Z. Computing rectifying homographies for stereo vision. Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149); Fort Collins, CO, USA, 23–25 June 1999; IEEE: Piscataway, NJ, USA, 1999; Volume 1, pp. 125-131.
22. Harris, C.; Stephens, M. A combined corner and edge detector. Proceedings of the Alvey Vision Conference; Manchester, UK, 31 August–2 September 1988; Volume 15, pp. 10-5244.
23. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis.; 2004; 60, pp. 91-110. [DOI: https://dx.doi.org/10.1023/B:VISI.0000029664.99615.94]
24. Bay, H.; Tuytelaars, T.; Van Gool, L. Surf: Speeded up robust features. Computer Vision–ECCV 2006: 9th European Conference on Computer Vision, Graz, Austria, 7–13 May 2006, Proceedings, Part I; Springer: Berlin/Heidelberg, Germany, 2006; pp. 404-417.
25. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. Proceedings of the 2011 International Conference on Computer Vision; Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2564-2571.
26. Rosten, E.; Drummond, T. Machine learning for high-speed corner detection. Computer Vision–ECCV 2006: 9th European Conference on Computer Vision, Graz, Austria, 7–13 May 2006, Proceedings, Part I; Springer: Berlin/Heidelberg, Germany, 2006; pp. 430-443.
27. Calonder, M.; Lepetit, V.; Strecha, C.; Fua, P. Brief: Binary robust independent elementary features. Computer Vision–ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010, Proceedings, Part IV; Springer: Berlin/Heidelberg, Germany, 2010; pp. 778-792.
28. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05); San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 1, pp. 886-893.
29. Matas, J.; Chum, O.; Urban, M.; Pajdla, T. Robust wide-baseline stereo from maximally stable extremal regions. Image Vis. Comput.; 2004; 22, pp. 761-767. [DOI: https://dx.doi.org/10.1016/j.imavis.2004.02.006]
30. Geiger, A.; Moosmann, F.; Car, Ö.; Schuster, B. Automatic camera and range sensor calibration using a single shot. Proceedings of the 2012 IEEE International Conference on Robotics and Automation; Saint Paul, MN, USA, 14–18 May 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 3936-3943.
31. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell.; 2000; 22, pp. 1330-1334. [DOI: https://dx.doi.org/10.1109/34.888718]
32. Tsai, R. A versatile camera calibration technique for high-accuracy 3D machine vision metrology using off-the-shelf TV cameras and lenses. IEEE J. Robot. Autom.; 1987; 3, pp. 323-344. [DOI: https://dx.doi.org/10.1109/JRA.1987.1087109]
33. Galántai, A. The theory of Newton’s method. J. Comput. Appl. Math.; 2000; 124, pp. 25-44. [DOI: https://dx.doi.org/10.1016/S0377-0427(00)00435-0]
34. Nocedal, J.; Wright, S.J. Numerical Optimization; Springer: New York, NY, USA, 2006.
35. Duc-Hung, L.; Cong-Kha, P.; Trang, N.T.T.; Tu, B.T. Parameter extraction and optimization using Levenberg-Marquardt algorithm. Proceedings of the 2012 Fourth International Conference on Communications and Electronics (ICCE); Hue, Vietnam, 1–3 August 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 434-437.
36. Coleman, T.F.; Li, Y. An interior trust region approach for nonlinear minimization subject to bounds. SIAM J. Optim.; 1996; 6, pp. 418-445. [DOI: https://dx.doi.org/10.1137/0806023]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
In this study, we propose a novel rectification method for three cameras using a single image for depth estimation. Stereo rectification serves as a fundamental preprocessing step for disparity estimation in stereoscopic cameras. However, off-the-shelf depth cameras often include an additional RGB camera for creating 3D point clouds. Existing rectification methods only align two cameras, necessitating an additional rectification and remapping process to align the third camera. Moreover, these methods require multiple reference checkerboard images for calibration and aim to minimize alignment errors, but often result in rotated images when there is significant misalignment between two cameras. In contrast, the proposed method simultaneously rectifies three cameras in a single shot without unnecessary rotation. To achieve this, we designed a lab environment with checkerboard settings and obtained multiple sample images from the cameras. The optimization function, designed specifically for rectification in stereo matching, enables the simultaneous alignment of all three cameras while ensuring performance comparable to traditional methods. Experimental results with real camera samples demonstrate the benefits of the proposed method and provide a detailed analysis of unnecessary rotations in the rectified images.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 Department of Information and Communication Engineering, Korea University of Technology and Education (KOREATECH), Cheonan-si 31253, Republic of Korea;
2 R&D Center, VisioNext, Seongnam-si 13488, Republic of Korea;