1. Introduction
In recent years, panoramic images have been extensively employed in ground remote sensing applications, including street image acquisition, traffic monitoring, virtual reality, robot navigation and mobile mapping applications, etc. [1,2,3,4,5]. At present, there are mainly two types of panoramic imaging methods on the market, namely catadioptric imaging and multi-camera imaging methods. The catadioptric panoramic method uses a plane mirror to refract the surrounding light to a single camera [6,7,8,9]. Due to their complex imaging mechanisms and manufacturing processes, catadioptric panoramic systems have been less commonly used compared to the second methods. With the development of sensors and computing abilities, many panoramic camera systems have been developed by a combination of multiple cameras [10,11,12,13,14,15,16,17,18,19,20,21,22,23,24], which can be divided into two categories: monocular panorama systems and binocular panorama systems. Although a binocular panorama system is able to provide the 3D scene information needed for seamless panorama generation and 3D reconstruction, it requires higher complexity of system design and processing than that of the monocular panorama system [25]. Therefore, the development of monocular panorama systems using multiple cameras and the generation of accurate panoramic images remain crucial in the fields of remote sensing and computer vision applications.
In recent years, several monocular panoramic imaging systems based on multiple cameras have been developed. Among these, professional panoramic systems often incorporate multiple cameras, such as Point Grey’s Ladybug series cameras [10,11], Google’s vehicle NCTech [12], Leica’s mobile panoramic cameras [13], and more. To achieve precise panorama generation with these integrated systems, relevant research can be categorized into two main areas. The first approach is feature point-based algorithms [19,20,21,22,23]. Various tools have been developed in the multimedia industry for panorama creation, such as Kolor Autopano and Autostitch [19], among others.
Panoramas are generated based on the second method, which relies on a panoramic imaging model derived from the pre-calibrated geometric relationships between cameras. Both the ideal and rigorous panoramic spherical and cylindrical imaging models have been proposed [24,25,26,27,28] and successfully applied for vehicle-based remote sensing mapping and panoramic SLAM 3D reconstruction, particularly with the professional panoramic camera LadyBug3 [25]. However, it is important to mention that both the ideal and rigorous panoramic imaging models are developed under the assumption of constant object-distance information in the surroundings.
In order to meet the requirements of accurate panoramic mapping and 3D reconstruction applications, the accuracy and robustness of virtual panorama still need to be improved by embodying the varied object-distance information into a mosaicking imaging scheme. In this case, it is a vital challenge to accurately estimate the object-distance information of the surrounding scene. In addition, problems such as empty pixels and pixel aggregation in the virtual panoramas generated by the direct panoramic imaging models may still exist.
This paper presents a panorama generation scheme and the advanced technologies used for a monocular panoramic system. The primary contributions of this work lie in three sides:
(1). In contrast to expensive panoramic camera systems that rely on professional cameras, a low-cost panoramic camera system is developed by a combination of eight low-cost web cameras in this paper.
(2). An inverse rigorous panoramic imaging model is derived completely based on the pre-calibration of the inner orientation elements, distortion parameters of each camera, and the relative orientation relationships between cameras.
(3). A scheme is proposed to estimate object-distance information within a surrounding scene. By expending the traditional Vertical Line Locus (VLL) object-based image matching algorithm [29], a panoramic Vertical Line Locus algorithm called PVLL is proposed to derive the optimized object-distance grids.
The rest of the paper is organized as follows: Section 2 introduces a panoramic camera system integrated with multiple cameras, developed by our team. Moreover, the inverse rigorous panoramic imaging model is derived completely based on pre-calibration of the panoramic camera system. In Section 3, we propose the panoramic generation method, with a specific focus on the object-distance estimation algorithm. The detailed experiment and analysis are provided in Section 4. Section 5 summarizes this research and provides recommendations related to future research works.
2. Related Work
2.1. Feature Point-Based Panoramic Imaging Method
For a set of overlapping images acquired with a multi-camera panoramic system, it is possible to generate a panoramic image relying on feature point-based methods. In the approach, the SIFT feature points are detected and matched on the overlapping regions of each pair of images from adjacent cameras [19,20,21]. Assuming that the scene is flat, the homographic estimations [22,23] are performed based on corresponding feature-points for each two adjacent cameras. At this point, all the transformations are modified relative to the master image, which is required for the generation of the panoramic images. The panoramic imaging approach is an extension of feature-based image registration and stitching of a single pair of images, which can be applied in scenes with rich texture features. However, this method heavily depends on achieving a sufficient distribution and accurate matching of feature points, as well as requiring substantial overlap between adjacent cameras. Additionally, it requires less variation in object-distance information from the surroundings, as the homographic transformation only exists between two planes.
2.2. Traditional Panoramic Imaging Model Based Method
The panoramic imaging model-based method, which departs from the feature-based method, requires accurate inner orientation elements of each camera and the relative orientation parameters between cameras to develop an accurate projection transformation relationship from the coordinate system of a perspective camera to that of the virtual panoramic camera.
Traditionally, the ideal panoramic spherical and cylindrical imaging model are formulated by disregarding the eccentric error between the center of the virtual panorama camera and that of an actual camera [26,27]. The rigorous panoramic imaging model is further proposed by incorporating the eccentric vectors among the optical centers of the cameras into the panoramic imaging model. The rigorous model is theoretically more accurate than the ideal model [25]. However, both the ideal and rigorous panoramic imaging models are developed without consideration of the varied object-distance information in a scene. Lin [30] tries to estimate a unified cylinder projection radius based on a loss function to minimize the projection error of the whole scene. Due to difficulties in fitting a regular cylinder surface for a real environment, stitching errors may still appear where the projection radius departs from the object-distance information. In addition, traditional panoramic models describe the transformation relationship from perspective imaging to panoramic imaging, which can easily lead to pixel aggregation or empty pixels in panoramic images.
Aiming at accurate and seamless panoramic imaging with a multi-camera system, an inverse panoramic imaging method is proposed by estimation of object-distance information of the surrounding scene, construction of the inverse transformation model and introduction of the eccentric errors.
3. Proposed Method
3.1. Overview of the Panoramic System
In this paper, we have designed a panoramic camera system that incorporates eight low-cost, high-definition web cameras mounted on an octagonal platform, as illustrated in Figure 1a. The panoramic camera system has a view field in the horizontal direction and view field in the vertical direction. Each camera has 2952 × 1944 pixels and is equipped with a lens of 5 mm focal length, with view field in the horizontal direction and view field in the vertical direction.
3.2. Proposed Inverse Panoramic Imaging Model
3.2.1. Relative Orientation between a Camera and the Panoramic Camera
According to the structure of our designed system in Figure 1, the relative orientation relationships are derived in this section. Let be the optical central point of the camera , and be the image space coordinate system. The panoramic cylindrical coordinate system is defined with its origin, set as O, located at the optical center shared by all cameras. The X, Y, and Z axes of this coordinate system are aligned parallel to the y, z, and x axes, respectively, of the image space coordinate system of the first camera, as illustrated in Figure 2. Thus, the orientation rotation matrix of the camera can be derived by Equation (1).
(1)
where is the rotation matrix from the panoramic cylindrical coordinate system to the camera coordinate system of the camera. is the rotation matrix from the image space coordinate system of the first camera to that of the camera, and is a matrix, .In addition, it is required to calculate the relative translation vector, set as , between point O and point by Equation (2).
(2)
where is the pre-calibrated relative translation vector between the optical centre point of the first camera and that of the camera.Aiming at removing empty pixels and pixel aggregation in virtual panoramas, an inverse panoramic imaging model is further derived in this paper.
3.2.2. Inverse Panoramic Imaging Model
Let point be projected on the cylinder surface, which is captured from the camera and mapped to point in the real image . The ideal image model can be derived based on a traditional collinear equation [11], as shown in Equation (3).
(3)
where is the scale factor, is the inner orientation matrix of the camera, is the undistorted coordinate of . and are the cylindrical coordinates of points and , respectively. In addition, the cylindrical projection point satisfies Equation (4):(4)
where represents the cylindrical radius.The panoramic cylindrical surface is unwrapped onto a panoramic plane, with being taken as its panoramic planar coordinate system, as illustrated in Figure 2. Point P is mapped to point based on Equation (5), which is derived from Equations (3) and (4).
(5)
where stand for the coordinate of point in the virtual image in the x- and y- directions, respectively.3.3. Object-Distance Estimation Algorithm
A panoramic image generated from the multi-camera system can be divided into the overlapping areas projected from two adjacent cameras and non-overlapping areas projected from only one camera, as shown in Figure 3. Traditionally, the panoramic images can be generated with a single projection radius under the assumption of a constant object-distance in the surroundings. While projection errors may occur due to variations in object-distance information, these errors are typically not easily discernible in the non-overlapping areas of a virtual panoramic image. However, stitching seams and geometric inconsistencies may become apparent and visible in the overlapping regions of a panoramic image. For example, the object points , , and are respectively projected onto distinct overlapping regions of a virtual cylinder surface with the radius by individual cameras, as shown in Figure 4. In this case, the displacements between corresponding projection points and (k = 1, 2, 3) from the same object point are contingent upon the difference between the object-distance and the radius
In this case, a crucial challenge lies in estimating the object-distance information of the surrounding scene captured within the overlapping regions of each pair of adjacent cameras. To solve the problem, this paper presents a three-step algorithm for determining object-distance information in the surroundings to create seamless panoramic images.
In the first step, we estimate the scope of overlapping regions in a panoramic image and generate pyramid grids. In the second step, object-distance information is determined using an improved VLL (Vertical Locus Line Method for the surroundings captured in the overlapping regions of the panoramic image). Finally, we employ an interpolation method to estimate object-distance information for each non-overlapping region.
3.3.1. Generation Algorithm of Pyramid Grids in Overlapping Regions
In this paper, each overlapping region of a panoramic image is first generated around the central line, which is defined as the angular bisector vector between two adjacent cameras.
As illustrated in Figure 3, let be a point on the cylinder surface, and be the vector in the same direction as the angular bisector vector between and . The vector can be derived by Equation (6). In addition, is then projected onto the point within the panoramic image. The -coordinate of can be computed by Equation (7). Further, the overlapping region , with a width of and setting as the central line, is generated, expanding pixels away from the y-axis.
(6)
(7)
In Equations (6) and (7), is set as where represent the , and cylindrical coordinates of point , respectively. is the scale factor, and is the x-coordinate of in the panoramic image coordinate system. is the width of the virtual panorama image. The parameters and take on the following values: .
By adopting a strategy of progressive refinement from coarse to fine, eight object-distance maps (ODMs) are constructed for each overlapping region of a virtual panoramic image. Assuming that the resolution of is pixels by pixels, a pyramid grid is established with levels for each overlapping region as follows:
First, the th level of the pyramid grid is segmented into cells with dimensions by . Each cell at this level has a resolution of pixels by pixels, and the pixels within the cell share the same object-distance.Second, the pyramid grids from the to the top layer of the level can be generated in a coarse-to-fine manner. At the level, the grid has dimensions of by cells, and each cell has a resolution of (of pixels by pixels), where takes values from N−1 down to .
Considering the balance of efficiency and accuracy, the parameters , , and are given empirical values, namely 256, 2560, 8, and 4, respectively.
3.3.2. Estimation of Object-Distance Pyramid Maps in Overlapping Regions
The traditional object space-based matching method, which was called the Vertical Locus Line Method (VLL) [29], is often used to estimate depth information by accurately matching a pair of image patches without directly calculating the three-dimensional computation for object points. Previous approaches of the VLL matching algorithm are mainly constrained to perspective projection images and require priori depth information. As a result, the VLL matching algorithm may exhibit reduced efficiency or failure due to the absence of prior object-space information. This paper proposes the Panoramic Vertical Locus Line (PVLL) matching algorithm to estimate the object-distance of panoramic surroundings based on pyramid grids ranging from coarse to fine.
-
(1). Estimation of Candidate Object-Distance
An object-distance set is estimated priorly to increase the efficiency of the search-space size for PVLL matching. Supposing that the minimum and maximum object-distances of surroundings are defined as and , the reprojected locations of the object point (where ) will be restricted between point to point in image , and between point and point in , respectively, as shown in Figure 5.
Given the object-distance of point is , the horizontal displacement between the projection location and point in image can be approximately computed by Equation (8).
(8)
where is the width of the image of the sub-camera, is the horizontal field of view of the sub-camera, and is the displacement between the optical center and the origin of the cylindrical projection coordinate system.Further, the list of candidate object-distances, set as , can be acquired using Algorithm 1 based on Equation (8), where is the displacement steps in pixels.
Algorithm 1: Solution of Projected Radius List |
Input: , , , , , |
As the object-distance changes continuously from to , the back-projection location of an object point imaged in an overlapping region approximately moves along a straight line by ε pixels in a real image, as shown in Figure 6. Here, has been set to 1 pixel to provide an optimal balance between the object-distance resolution and computational efficiency.
(2). Panoramic Vertical Line Locus Algorithm
Definition 1. The corresponding image patches are represented as and at the th pyramid level, respectively, where be the grid cell located at the row and column of the pyramid grid in , be the object-distance in this cell, denotes the index of in , and here takes values from N down to t.
We can assume that the pixel point in the grid cell is reprojected to locations of in image and in image using as the cylinder radius based on Equation (9) derived from Equation (5), respectively. Thus, a pair of grey values can be acquired based on the grey bilinear interpolation algorithm using four-pixel points around the points and , respectively. In this way, and can be generated with a resolution of pixels by pixels, respectively.
(9)
Definition 2. Let be the Normalized Cross-Correlation image matching parameter between image blocks and
The value of can be computed by Equation (10).
(10)
where and are the greyscales located at in the image block and , respectively. , are the average grey values of the two image patches, respectively, where .Object-distance estimation based on the above definitions: Different from previous VLL-based methods that predict object-distance information directly, the proposed PVLL object-space matching method computes the index of the required object-distance in the estimated list from coarse to fine.
First, an objective function is generated at the pyramid layer in the overlapping region to optimize the object-distance information of each grid cell. is proposed based on the similarity of the image intensity between a pair of corresponding image patches and the smoothness of object-distance between adjacent grid cells, as depicted in Equation (11).
(11)
where and are the weight coefficients of similarity and smoothness, respectively. is the objective function; is the count of rows for the pyramid gird. Further, a group of optimum object-distance indices, denoted by , can be estimated by iteratively solving Equation (12) using a dynamic programming algorithm [31]. Then, an object-distance map (ODM) is generated for the overlapping region at a resolution of by in the level, based on the optimal index parameters. Additionally, a corresponding index map is also created.(12)
Second, from the to the top pyramid layer , we employ an iterative image matching procedure as follows:
For each grid cell , a group of candidate object-distances are determined by Equation (13).
(13)
where are set as the optimum estimation of the object-distance index of the grid cell at the row and column in the previous layer, where takes values from N − 1 down to t.Then, these candidates are chosen with regular index increments ζ, changing between the minimum object-distance and the maximum object-distance . The value of K has been determined through experimentation to achieve the optimal balance between candidate performance and computational load. The optimum value of the parameter k, represented by , is therefore extracted by searching the maximum NCC value using Equation (14), which is derived from Equations (9) and (10).
(14)
where is computed by a pair of corresponding image patches based on Equation (10), and the image patches are generated by reprojection of each pixel in onto and based on Equation (9), respectively.Therefore, the optimal object-distance is set as for the grid cell .
In this way, the resolution and structure of the object-distance maps, along with the index maps, are updated with the optimum object-distance for each grid cell, until reaching the top pyramid level.
3.4. Object-Distance Interpolation Method in the Non-Overlapping Regions
An interpolation method is adopted to derive object-distance information in the non-overlapping regions. For a pixel point located in the non-overlapping region between two overlapping regions and , its object-distance can be interpolated from the known object-distance information of two points and using Equation(15), as shown in Figure 7.
(15)
where , and are the object-distances of pixel point , and , respectively. The object-distance information can be calculated using the object-distance maps in the top-level of the aforementioned pyramid grids.The values of and are determined by and , respectively, where represents the width of an overlapping region. In this context, takes values from 1 to 7, j takes values from 2 to 7, and k takes values from 3 to 8, including 0. The parameters of and represent the displacements of the centre lines and from the y-axis of the panoramic image system, respectively. In the same way, the object-distances can be estimated for all the pixel points in the non-overlapping regions.
Finally, a panoramic image can be generated based on Equation (9) with the estimated object-distance information on the overlapping regions and non-overlapping regions.
4. Experimental Results
In this section, we conducted detailed experiments and carried out a comprehensive analysis for generating panoramic images using the panoramic camera system developed in this paper. The algorithm for the panorama generation was implemented in C++ using the OpenCV 4.3 development package on a computer equipped with an AMD Ryzen 7 5800H, 3.2 GHz processor, and 16 GB main memory.
4.1. Calibration of the Panoramic Camera
In this paper, the multi-camera system was fixed on an octagonal structure to ensure the constant relationships among cameras. A two-step method is conducted to calibrate the internal parameters of each camera and ROPs (relative orientation parameters). Firstly, more than 30 images of a high-precision chessboard were taken with each camera at different poses and positions, as shown in Figure 8. The inner orientation elements and the distortion coefficients of each camera were calibrated by Zhang’s calibration method [32]. The interior orientation elements for each camera are shown in Table 1.
Secondly, a professional three-dimensional calibrate field was used to calibrate the ROPs. Ten groups of images were taken at different poses and positions, where each group of images are captured at the same time, as shown in Figure 9.
The external parameters of the camera at the exposure epoch were calibrated based on the traditional PnP method [33]. Let the first camera be the master camera, while the other seven cameras are slave cameras. Let and represent the rotation matrix and position of the camera relative to the world coordinate system at the kth exposure epoch, where k = 1, 2, …, K. In this experiment, the value of K is 10. The ROPs between the master camera and any slave camera can be computed by Equations (16) and (17). The optimum relative orientation parameters can be seen in Table 2.
(16)
(17)
4.2. Visualized Analysis
We conducted a performance comparison of our proposed method (Method C) with AutoStitch [19] (Method A) and the traditional cylindrical panoramic stitching method [32] (Method B) using both indoor and outdoor scenes, as depicted in Figure 10.
Figure 11 displayed the panoramas for an outdoor scene. In Figure 11a, stitching truncation (resulting in structural discontinuity) can be observed in the roadway, highlighted with an orange box. In Figure 11b, the repetition of road lamps is evident, highlighted with the blue box. In contrast, the panoramic image generated by our algorithm (Method C) demonstrates visual consistency within the highlighted overlapping regions, as depicted in Figure 11c.
Figure 12 displays the panoramas for an indoor scene. In Figure 12a, more truncation and ghosting of the ceiling can be noticed, as highlighted with the orange box. Furthermore, the panoramic image shows significant gaps, particularly in areas with a low-texture environment lacking distinct feature points, which are highlighted with red boxes in the overlapping regions. In Figure 12b, you can observe that the ceilings are repeatedly generated, highlighted with an orange box. In contrast, the panoramic image generated by our algorithm (Method C) is complete and seamless in the area highlighted within the same orange box, as depicted in Figure 12c.
Therefore, it becomes evident that when compared to Method A and Method B, our approach (Method C) maintains visual consistency in both indoor and outdoor scenes and effectively eliminates stitching gaps, especially in scenes with low-texture regions.
4.3. Quantitative Analysis
Our quantitative analysis was conducted based on the following steps:
(1). Each pixel within every overlapping region of a panoramic image was individually projected onto two overlapping original images captured by adjacent cameras, and the grayscale value of the projection points was obtained. In this way, two panoramic image patches were generated for each overlapping region, as shown in Figure 13.
(2). The SIFT [34] feature points were then extracted and matched for each pair of panoramic image patches. The average Euclidean distance was then computed between corresponding feature points in every pair of panoramic image patches. In addition, the parameters of the Structural Similarity Index (SSIM) [31], Peak Signal-to-Noise Ratio (PSNR) [35], Normalized Cross-Correlation (NCC) coefficients, and Stitched Image Quality Evaluator (SIQE) [36] were determined to quantify the dissimilarities between each pair of panoramic image patches. The Root Mean Square Error (RMSE), average SSIM, average PSNR, and average NCC are further computed based on the average Euclidean distances, SSIM values, PSNR values, NCC coefficients, and SIQE respectively, obtained from the dataset of 160 panoramas.
(3). Moreover, using the open-source library functions [37], we calculated the Stitched Panoramic Image Quality Assessment (SPIQA) [38] parameter for each group of images from eight cameras and the corresponding generated panoramic images. Then, we calculated the average of 160 SPIQAs.
An image pair generated from adjacent cameras.
[Figure omitted. See PDF]
This comprehensive analysis allowed us to evaluate the quality and dissimilarities based on various metrics and parameters, encompassing both structural and pixel-wise differences, as depicted in Table 3.
Because the object distance in the outdoor scene is significantly larger than the eccentric distance between the optical center of a sub-camera and that of the virtual panoramic camera, the stitching accuracy of a panorama is less affected by inconsistencies in object-distance information within the surrounding scene. Thus, the proposed method only has a small improvement compared to Methods A and B. However, due to the restricted range of object-distances in indoor scenes, the eccentricity errors between cameras cannot be disregarded and may lead to substantial stitching errors when employing both Method A and Method B, especially when there are variations in the object-distance. Our proposed algorithm generates a panorama by utilizing the multiple projection radiuses from the estimation of object-distance information, which demonstrates far higher accuracy than Method A and B. In addition, it can be seen clearly that the stitching time of Method C is much higher than that of Method A and B.
5. Conclusions and Future Work
Aiming at remote sensing and mapping applications, we have designed a low-cost image acquisition system through the integration of eight web cameras, which were securely mounted on a circular rig. To solve the mosaicking problems caused by inconstant and inconsistent object-distance information within a scene, a panoramic image generation algorithm was proposed to improve the existing direct mono-cylinder imaging model. The key aspect of the algorithm is the Panoramic Vertical Line Locus matching algorithm called the PVLL algorithm, which is proposed by extending the traditional VLL algorithm. Moreover, the object-distance grids can be generated based on the PVLL algorithm in the overlapping regions of a panorama. Thus, the panorama can be generated by utilizing the multiple projection radiuses from the estimation of object-distance information. The experiments show that the proposed imaging method successfully improves the stitching accuracy and visual effects of virtual panoramas.
In summary, this paper discusses the problems of geometric mosaicking errors in multi-camera panoramic systems and conducts research on the combination of accurate calibration, depth estimation, and post-processing techniques to ensure a seamless and accurate panoramic view, despite variations in object distances and eccentric errors between camera optical centers. The method described in this paper is suitable for a multi-camera system with an annular structure, and we plan to extend it to a panoramic system with a 720-degree spherical design in our future work for ground remote sensing applications in both outdoor and indoor environments.
Methodology, H.C. and Z.Z.; Software, Z.Z. and F.Z.; Writing—original draft, H.C. and Z.Z.; Writing—review and editing, H.C. and Z.Z.; Visualization, Z.Z.; Translation Check, F.Z.; Project administration, H.C.; Funding acquisition, H.C. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
Data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy (General Data Protection Regulation possible issues).
The authors declare no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 9. Images for computing relative orientation parameters (2 groups are displayed).
Figure 10. The image sequences of an outdoor and indoor scene (the numbers in the picture indicate which camera numbered from 1 to 8 obtained the sub-images).
Figure 11. Panorama of an outdoor scene. (a) Method A; (b) Method B; (c) Method C.
Inner orientation elements of each camera.
Cam |
|
|
|
|
|
|
---|---|---|---|---|---|---|
1 | 2.2782 | 2.4294 | 1.3259 | 0.9473 | 0.0281 | 0.0098 |
2 | 2.2684 | 2.4194 | 1.1535 | 1.0435 | 0.0255 | 0.0036 |
3 | 2.2729 | 2.4238 | 1.3974 | 1.0620 | 0.0257 | 0.0045 |
4 | 2.2759 | 2.4256 | 1.2523 | 1.0469 | 0.0285 | −0.0028 |
5 | 2.2601 | 2.4104 | 1.2447 | 9.4271 | 0.02718 | 0.0095 |
6 | 2.2744 | 2.4254 | 1.1629 | 1.0865 | 0.0253 | 0.0013 |
7 | 2.2761 | 2.4268 | 1.2567 | 1.0027 | 0.0253 | 0.0089 |
8 | 2.2731 | 2.4244 | 1.5161 | 1.0635 | 0.0306 | −0.0094 |
Relative orientation parameters of the combined cameras.
Cam | α/rad | β/rad | γ/rad | X/cm | Y/cm | Z/cm |
---|---|---|---|---|---|---|
1–2 | 0.0174 | 0.7814 | −0.0029 | 4.8081 | 0.0295 | −1.9565 |
1–3 | 0.0091 | 1.5850 | 0.0057 | 6.5616 | −0.0294 | 6.6810 |
1–4 | 0.0046 | 2.3532 | 0.0038 | 4.8170 | 0.0683 | −11.5291 |
1–5 | −0.0205 | −3.1253 | −0.0108 | −0.2132 | 0.1015 | −13.3953 |
1–6 | −0.0140 | −2.3357 | −0.0212 | −4.7553 | 0.0871 | −11.2954 |
1–7 | −0.0011 | −1.5729 | −0.0238 | −6.8198 | 0.0953 | −6.7039 |
1–8 | 0.0105 | −0.7723 | −0.0197 | −4.6815 | 0.0800 | −1.8884 |
Stitching accuracy and efficiency.
Scene | Method | RMSE | SSIM | PSNR | NCC | SPIQA | SIQE | Time(s) |
---|---|---|---|---|---|---|---|---|
Method A | 6.3341 | 0.5530 | 16.2321 | 0.7076 | 0.6131 | 60.0379 | 8.9112 | |
Outdoor | Method B | 8.2422 | 0.5575 | 15.5963 | 0.5650 | 0.5839 | 54.7586 | 0.1251 |
Method C | 0.7498 | 0.6003 | 20.6502 | 0.8157 | 0.8586 | 68.4837 | 0.7539 | |
Method A | 16.0997 | 0.7237 | 20.0779 | 0.8399 | 0.7582 | 51.6073 | 8.7745 | |
Indoor | Method B | 34.9983 | 0.7579 | 17.4994 | 0.3803 | 0.7938 | 58.7872 | 0.1268 |
Method C | 0.9837 | 0.8197 | 27.0731 | 0.8593 | 0.9047 | 63.0523 | 0.7473 |
References
1. Kao, S.T.; Ho, M.T. Ball-Catching System Using Image Processing and an Omni-Directional Wheeled Mobile Robot. Sensors; 2021; 21, 3208. [DOI: https://dx.doi.org/10.3390/s21093208] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34063130]
2. Wu, F.; Song, H.; Dai, Z.; Wang, W.; Li, J. Multi-camera traffic scene mosaic based on camera calibration. IET Comput. Vis.; 2021; 15, pp. 47-59. [DOI: https://dx.doi.org/10.1049/cvi2.12009]
3. Krishnakumar, K.; Gandhi, S.I. Video stitching using interacting multiple model based feature tracking. Multimed. Tools Appl.; 2018; 78, pp. 1375-1397. [DOI: https://dx.doi.org/10.1007/s11042-018-6116-0]
4. Qu, Z.; Wang, T.; An, S.; Liu, L. Image seamless stitching and straightening based on the image block. IET Image Process.; 2018; 12, pp. 1361-1369. [DOI: https://dx.doi.org/10.1049/iet-ipr.2017.1064]
5. Li, L.; Yao, J.; Xie, R.; Xia, M.; Zhang, W. A Unified Framework for Street-View Panorama Stitching. Sensors; 2016; 17, 1. [DOI: https://dx.doi.org/10.3390/s17010001]
6. Wenbo, J.; Xuefeng, G.; Bingkun, H.; Shitao, L.; Hongyang, Y. Expansion of Conical Catadioptric Panoramic Image of Inner Surface of Cylindrical Objects. Acta Opt. Sin.; 2021; 41, 0311002. [DOI: https://dx.doi.org/10.3788/AOS202141.0311002]
7. Amani, A.; Bai, J.; Huang, X. Dual-view catadioptric panoramic system based on even aspheric elements. Appl. Opt.; 2020; 59, 7630. [DOI: https://dx.doi.org/10.1364/AO.400023]
8. Baskurt, D.O.; Bastanlar, Y.; Cetin, Y.Y. Catadioptric hyperspectral imaging, an unmixing approach. IET Comput. Vis.; 2020; 14, pp. 493-504. [DOI: https://dx.doi.org/10.1049/iet-cvi.2019.0784]
9. Ko, Y.J.; Yi, S.Y. Catadioptric Imaging System with a Hybrid Hyperbolic Reflector for Vehicle Around-View Monitoring. J. Math. Imaging Vis.; 2017; 60, pp. 503-511. [DOI: https://dx.doi.org/10.1007/s10851-017-0770-0]
10. Khoramshahi, E.; Campos, M.; Tommaselli, A.; Vilijanen, N.; Mielonen, T.; Kaartinen, H.; Kukko, A.; Honkavaara, E. Accurate Calibration Scheme for a Multi-Camera Mobile Mapping System. Remote Sens.; 2019; 11, 2778. [DOI: https://dx.doi.org/10.3390/rs11232778]
11. Zhang, Y.; Huang, F. Panoramic Visual SLAM Technology for Spherical Images. Sensors; 2021; 21, 705. [DOI: https://dx.doi.org/10.3390/s21030705] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33494160]
12. Buyuksalih, G.; Baskaraca, P.; Bayburt, S.; Buyuksalih, I.; Rahman, A.A. 3D city modelling of istanbul based on lidar 333 data and panoramic images—Issues and challenges. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci.; 2019; XLII-4/W12, pp. 51-60. [DOI: https://dx.doi.org/10.5194/isprs-archives-XLII-4-W12-51-2019]
13. Nespeca, R. Towards a 3D digital model for management and fruition of Ducal Palace at Urbino. An integrated survey with mobile mapping. Sci. Res. Inf. Technol.; 2019; 8, pp. 1-14. [DOI: https://dx.doi.org/10.2423/i22394303v8n2p1]
14. Afifi, A.; Takada, C.; Yoshimura, Y.; Nakaguchi, T. Real-Time Expanded Field-of-View for Minimally Invasive Surgery Using Multi-Camera Visual Simultaneous Localization and Mapping. Sensors; 2021; 21, 2106. [DOI: https://dx.doi.org/10.3390/s21062106] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33802766]
15. Hongxia, C.; Lijun, C.; Ning, W.; Tingting, L. Calibration Method with Implicit Constraints for Multi-View Combined Camera Using Automatic Coding of Marker Points. Chin. J. Lasers; 2020; 47, 0110003. [DOI: https://dx.doi.org/10.3788/CJL202047.0110003]
16. Ke, X.; Huang, F.; Zhang, Y.; Tu, Z.; Song, W. 3D Scene Localization and Mapping Based on Omnidirectional SLAM. IOP Conf. Ser. Earth Environ. Sci.; 2021; 783, 012143. [DOI: https://dx.doi.org/10.1088/1755-1315/783/1/012143]
17. Ullah, H.; Zia, O.; Kim, J.H.; Han, K.; Lee, J.W. Automatic 360° Mono-Stereo Panorama Generation Using a Cost-Effective Multi-Camera System. Sensors; 2020; 20, 3097. [DOI: https://dx.doi.org/10.3390/s20113097]
18. Qu, Z.; Lin, S.P.; Ju, F.R.; Liu, L. The Improved Algorithm of Fast Panorama Stitching for Image Sequence and Reducing the Distortion Errors. Math. Probl. Eng.; 2015; 2015, 428076. [DOI: https://dx.doi.org/10.1155/2015/428076]
19. Brown, M.; Lowe, D.G. Automatic Panoramic Image Stitching using Invariant Features. Int. J. Comput. Vis.; 2006; 74, pp. 59-73. [DOI: https://dx.doi.org/10.1007/s11263-006-0002-3]
20. Alwan, M.G.; AL-Brazinji, S.M.; Mosslah, A.A. Automatic panoramic medical image stitching improvement based on feature-based approach. Period. Eng. Nat. Sci.; 2022; 10, 155. [DOI: https://dx.doi.org/10.21533/pen.v10i2.2902]
21. Zhu, J.T.; Gong, C.F.; Zhao, M.X.; Wang, L.; Luo, Y. Image mosaic algorithm based on pca-orb feature matching. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci.; 2020; XLII-3/W10, pp. 83-89. [DOI: https://dx.doi.org/10.5194/isprs-archives-XLII-3-W10-83-2020]
22. Hoang, V.D.; Tran, D.P.; Nhu, N.G.; Pham, T.A.; Pham, V.H. Deep Feature Extraction for Panoramic Image Stitching. Intelligent Information and Database Systems; Springer: Berlin/Heidelberg, Germany, 2020; pp. 141-151. [DOI: https://dx.doi.org/10.1007/978-3-030-42058-1_12]
23. Woo Park, K.; Shim, Y.J.; Jin Lee, M.; Ahn, H. Multi-Frame Based Homography Estimation for Video Stitching in Static Camera Environments. Sensors; 2019; 20, 92. [DOI: https://dx.doi.org/10.3390/s20010092] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31877903]
24. Ji, S.; Shi, Y. Image matching and bundle adjustment using vehicle-based panoramic camera. Cehui Xuebao Acta Geod. Cartogr. Sin.; 2013; 42, pp. 94–100+107. Available online: https://api.semanticscholar.org/CorpusID:130900916 (accessed on 6 July 2021).
25. Wang, X.; Li, D.; Zhang, G. Panoramic Stereo Imaging of a Bionic Compound-Eye Based on Binocular Vision. Sensors; 2021; 21, 1944. [DOI: https://dx.doi.org/10.3390/s21061944] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33802076]
26. Sato, T.; Ikeda, S.; Yokoya, N. Extrinsic Camera Parameter Recovery from Multiple Image Sequences Captured by an Omni-Directional Multi-camera System. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2004; pp. 326-340. [DOI: https://dx.doi.org/10.1007/978-3-540-24671-8_26]
27. Lemaire, T.; Lacroix, S. SLAM with Panoramic Vision. J. Field Robot.; 2007; 24, pp. 91-111. [DOI: https://dx.doi.org/10.1002/rob.20175]
28. Shi, Y.; Ji, S.; Shi, Z.; Duan, Y.; Shibasaki, R. GPS-Supported Visual SLAM with a Rigorous Sensor Model for a Panoramic Camera in Outdoor Environments. Sensors; 2012; 13, pp. 119-136. [DOI: https://dx.doi.org/10.3390/s130100119]
29. Linder, W. Digital Photogrammetry—A Practical Course; Springer: Berlin/Heidelberg, Germany, 2006.
30. Lin, M.; Xu, G.; Ren, X.; Xu, K. Cylindrical panoramic image stitching method based on multi-cameras. Proceedings of the 2015 IEEE International Conference on Cyber Technology in Automation, Control, and Intelligent Systems (CYBER); Shenyang, China, 8–12 June 2015; IEEE: New York, NY, USA, 2015; pp. 1091-1096. [DOI: https://dx.doi.org/10.1109/cyber.2015.7288097]
31. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process.; 2004; 13, pp. 600-612. [DOI: https://dx.doi.org/10.1109/TIP.2003.819861]
32. Lepetit, V.; Moreno-Noguer, F.; Fua, P. EPnP: An Accurate O(n) Solution to the PnP Problem. Int. J. Comput. Vis.; 2008; 81, pp. 155-166. [DOI: https://dx.doi.org/10.1007/s11263-008-0152-6]
33. Zhang, Z.Y. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell.; 2000; 22, pp. 1330-1334. [DOI: https://dx.doi.org/10.1109/34.888718]
34. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis.; 2004; 60, pp. 91-110. [DOI: https://dx.doi.org/10.1023/B:VISI.0000029664.99615.94]
35. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. Proceedings of the 2010 20th International Conference on Pattern Recognition; Istanbul, Turkey, 23–26 August 2010; pp. 2366-2369. [DOI: https://dx.doi.org/10.1109/ICPR.2010.579]
36. Madhusudana, P.C.; Soundararajan, R. Subjective and Objective Quality Assessment of Stitched Images for Virtual Reality. IEEE Trans. Image Process.; 2019; 28, pp. 5620-5635. [DOI: https://dx.doi.org/10.1109/TIP.2019.2921858] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31217105]
37. Madhusudana, P.C.; Soundararajan, R. Official Implementation of Stitched Image Quality Evaluator (SIQE). Available online: https://github.com/pavancm/Stitched-Image-Quality-Evaluator (accessed on 10 October 2023).
38. Cheung, G.; Yang, L.; Tan, Z.; Huang, Z. A Content-aware Metric for Stitched Panoramic Image Quality Assessment. Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW); Venice, Italy, 22–29 October 2017; pp. 2487-2494. [DOI: https://dx.doi.org/10.1109/ICCVW.2017.293]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Panoramic imagery from multi-camera systems often suffers the problem of geometric mosaicking errors due to eccentric errors between the optical centers of cameras and variations in object-distances within the panoramic environment. In this paper, an inverse rigorous panoramic imaging model was derived completely for a panoramic multi-camera system. Additionally, we present an estimation scheme aimed at extracting object-distance information to enhance the seamlessness of panoramic image stitching. The essence of the scheme centers around our proposed object-space-based image matching algorithm called the Panoramic Vertical Line Locus (PVLL). As a result, panoramas were generated using the proposed inverse multi-cylinder projection method, utilizing the estimated object-distance information. The experiments conducted on our developed multi-camera system demonstrate that the root mean square errors (RMSEs) in the overlapping areas of panoramic images are no more than 1.0 pixel. In contrast, the RMSEs of the conventional traditional methods are typically more than 6 pixels, and in some cases, even exceed 30 pixels. Moreover, the inverse imaging model has successfully addressed the issue of empty pixels. The proposed method can effectively meet the accurate panoramic imaging requirements for complex surroundings with varied object-distance information.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
1 Department of Information and Technology, Bohai University, Jinzhou 121013, China;
2 Beijing Business School, Beijing 102209, China