This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
1. Introduction
Unstructured environments, such as parking lots or alleyways, contain intersections, and there is a need for autonomous vehicles to navigate intersections in these environments. The conventional autonomous navigation studies [1–3] use a global path with localization data. The global path consists of multiple waypoints that the vehicle can pass through intersections. The localization data are obtained through the global positioning system or by the simultaneous localization and mapping technique [4–6]. However, in an unstructured environment with narrow roads or complex obstacles, the localization data can be inaccurate, which increases the possibility of collision with obstacles.
Driving methods by detecting road and obstacles with local sensor data have been studied rather than tracking the global path with inaccurate localization data [7–10]. Among the local sensors, vision has been widely used due to its low price and low memory usage. For motion-planning at the intersection, it is necessary to recognize the intersection and detect the branch road existing at the intersection. Model and learning-based methods have been studied for branch road detection.
The model-based detection methods recognize and distinguish the branch road existing in an intersection according to a pattern of its shape, size, and direction. In [7–9], a road is segmented and contours of the road are recognized. The branch roads are distinguished according to the difference in position and direction between the vehicle and the road boundary, and entry points are recognized at the widest area of the branch road. Yi et al. [10] proposed a method of converting the road distance from the vehicle into a histogram using the road segmentation image. Then, the branch road is classified according to the histogram distribution. However, model-based detection methods do not accurately detect branch roads, especially in environments with changes in the width or curvature of the branch road, or in the presence of obstacles [11]. These methods are also sensitive to sensor noise in nonaccurately distinguished road boundaries.
Deep learning-based methods have been studied to address challenges in the model-based method. Object detection methods using the bounding box are used for autonomous driving, and recent studies [12–14] propose improved methods to obtain robust results in a variety of weather and light conditions. Methods for improving detection accuracy [15] and computing efficiency [16] have been proposed. In addition, a method for recognizing roads in various environments (structured, unstructured, lane/line-based, and curb) has been proposed [17]. However, the mentioned studies are difficult to be applied to the environments with intersections.
Recognition methods related to intersections and branch roads using deep learning are as follows. In [18], vehicle kinematic information, point cloud acquired by 3D LiDAR, and OpenStreetMap (OSM) are used to find branch roads at an intersection using machine learning. However, this method [18] uses global data, OSM, and a high-cost 3D LiDAR sensor. The camera image is passed through a long-term recurrent convolutional network to recognize vehicles passing through an intersection [19]. Through the drivable area matching method and model generation method, intersections classified into 7 types are detected [20, 21]. However, these methods [19–21] do not distinguish branch roads at the intersection. A method in [22] distinguishes branch roads in a form of region of interests on a front view image to determine whether a vehicle is driving at an intersection. However, this method is applied only in a structured environment where the width and curvature of the branch road are constant, not in unstructured environments. Further, detected branch roads are not applied to the intersection driving method, and a separate intersection driving algorithm is used.
To address the limitations, this paper proposes a vision-based branch road detection method using deep learning that can learn various shapes of branch roads in unstructured environments and situations with obstacles in the branch road. Branch roads consist of straight, left turn, and right turn roads, which are detected and are represented as a shape of rotated bounding boxes. Compared to using the unrotated bounding box, the rotated version can detect the branch road more accurately even when the vehicle is turning at an intersection. The number of branch roads is recognized differently according to the nonintersection road, and three-way and four-way intersections. At the intersection, one branch road is selected according to a global plan, and the inside of the selected road is regarded as the drivable area of an occupancy grid map. Furthermore, for safe navigating, obstacles inside the selected road are considered in this map by using the segmented drivable area image. This occupancy grid map is used as an input to existing motion-planning algorithms to enable intersection navigation. In addition, drivable area segmentation and branch road detection consist of a single multi-task deep neural network, which improves the learning performance of each task and reduces overall network memory usage.
2. Method
This study proposes a method to detect branch roads existing at an intersection in unstructured environments using deep learning. In addition, a method to obtain inputs for the use of existing motion-planning algorithms at intersections is proposed by combining the data from detected branch road, global planning, and drivable area segmentation. The overall system architecture of the proposed method is shown in Figure 1. The method uses vision, road distance information, vehicle velocity, and navigation information. It does not use the global path and localization data. The navigation information gives the command to vehicles going straight, turning left, or turning right when a vehicle is passing through the intersection. This information is obtained through global planning using a topological map consisting of intersections and roads, before starting the navigation [23].
[figure(s) omitted; refer to PDF]
A multi-task network is proposed to perform two tasks: segmenting the drivable area and detecting the branch roads as rotated bounding boxes. The drivable area segmentation image is used to obtain an occupancy grid map
2.1. Bird’s Eye View Image Transform
A camera sensor consists of two lenses and is used to detect drivable/nondrivable areas as well as branch roads at an intersection. The front view camera image is transformed to the bird’s eye view image which is depicted in Figure 2. The world coordinates ([A, B, C, D], as shown in Figure 2(d)), can be calculated to obtain an 11 m
[figure(s) omitted; refer to PDF]
2.2. Multi-Task Perception Network for Intersection Navigation
The perception network receives the transformed image of the bird’s eye view. As shown in Figure 3, the perception network shares one encoder, which is used in YOLOP [24], DLT-Net [25], and Multinet [26]. Then, the output of the encoder is passed on to two decoders. Each decoder performs the drivable area segmentation and rotated road bounding box detection tasks, which receives the same abstracted features for the drivable area from the one encoder. Therefore, the network size, computation time, and GPU usage can be reduced. It also decreases the possibility of overfitting by learning more generalized shared expressions to simultaneously fit multiple tasks.
[figure(s) omitted; refer to PDF]
The structure of the encoder and the segmentation decoder is the same as the Multinet [26]; however, the detection decoder is different. The encoder is based on the VGG16 network [27] which is widely used for training 2D data and shows high accuracy while ensuring real-time computation due to its simple structure. Each decoder and its output are described in detail in below subsections. The sizes of the encoder layers are set according to the 200
2.2.1. Drivable Area Segmentation (Occupancy Grid Map
The structure of the segmentation decoder is shown in the upper left side of Figure 3. Features abstracted by the encoder have a low resolution 7
The output of the segmentation decoder is a probability of the drivable at each pixel in the input image, which is shown in Figure 4(b). The closer the pixel is to black, the more likely it is drivable. The Otsu algorithm [31] is used to calculate the threshold value (Figure 4(c)). This algorithm divides the pixels into two classes by randomly setting a boundary value and repeatedly obtains the intensity distribution of the two classes. Then, it selects the boundary value that makes the distribution of the values of the two classes most uniform. In other words, an optimal threshold value at which the ratio difference between binary-classified pixels can be smallest is obtained. The size of the segmented image is 200
[figure(s) omitted; refer to PDF]
The data labeling criteria for training the drivable area segmentation image are in the image as follows: roads, road marks, a stop line, and crosswalk are labeled as the drivable area. The other area except the drivable area is designated as the nondrivable area. Sidewalks, parking spaces (including parking line), road boundary lines, pedestrian walkways, and vehicles are regarded as the nondrivable area.
2.2.2. Rotated Road Bounding Box Detection
In this study, the rotated bounding box is used to detect and distinguish the branch roads. In general, the bounding box is recognized as an unrotated form in the image frame. In this study, the rotated bounding box is used because, when a vehicle is driving at an intersection, the direction of the road differs from that of the vehicle. In this case, if the unrotated bounding box is used, it is not possible to cover all the drivable area without including obstacles inside the box (see Figure 5(c)). Moreover, an area that is not an actual branch road can be mistaken as a branch road (see the third (center box), fourth (right box), fifth (center box) images in Figure 5(c)). The rotated bounding box can detect one branch road as one box and cover the drivable area as much as possible (see Figure 5(b)), and the detection network can accurately find the box with a feature of the branch road.
[figure(s) omitted; refer to PDF]
A structure of the detection decoder is illustrated at the bottom right side of Figure 3 and is like the YOLO network [32]. The abstracted features, and the output of the encoder, pass a 1
Each grid
Through the 7
[figure(s) omitted; refer to PDF]
The data labeling criteria for training the rotated road bounding box are as follows. (1) The drivable area between the front of the vehicle’s bonnet and the front end of the image is labeled as a rectangle. In this case, the width of the rectangle should not exceed the vehicle’s width. (2) Additional labeling is needed if there are unlabeled drivable areas on the side of the image. (3) The maximum number of boxes is three. (4) The size of the box should exceed 7
2.3. Intersection Decision
When two or more boxes are recognized, it is regarded that the vehicle is driving at an intersection. As shown in Figures 7(a) and 8(d), multiple boxes can be detected at a nonintersection road, which is a false-positive case. To deal with this, a road distance between the center of intersections
[figure(s) omitted; refer to PDF]
If two or more boxes are recognized, and a difference between
2.3.1. Road Occupancy Grid Map
When it is determined that the vehicle is passing the intersection, one box is selected according to the navigation information, such as going straight, turning left, and turning right. This information is obtained through a global plan to visit all the roads in the topological map [23]. This process is shown in Figure 1 and Figure 8. The criteria for selecting one box are as follows.
Going straight: the box closest to the middle of the image.
Turning left: the leftmost box in the image.
Turning right: the rightmost box in the image.
Through the selected box, a road occupancy grid map
2.4. Merged Occupancy Grid Map
To navigate intersections, using existing motion-planning algorithms that do not take the intersection navigation into account, a merged occupancy grid map
The drivable area on each map is defined as the “true” value (the black area (grid) in Figure 8), and the nondrivable area as the “false” value (the white area (grid) in Figure 8). Each grid in
3. Experimental Setup
3.1. Vehicle and Camera Setup
As shown in Figure 9, an autonomous vehicle, Hyundai HG 240, was used in experiments. One laptop computer was used to implement the proposed method. For training and executing the deep neural network, Nvidia GTX 1080 TI 8G GPU was used. The CPU was 3.9 GHz Intel i9-8950HK 2,9 GHz with a memory size of 16 gigabytes.
[figure(s) omitted; refer to PDF]
A camera was attached 1.55 m above the ground and 0.25 m away from the center of the vehicle. Moreover, it was rotated about 20
3.2. Multi-Task Perception Network Training Details
A method for training a multi-task-based perception network followed a classic fine-tuning pipeline [26]. The weights of each decoder were calculated and individually updated with the different loss values. To update the weights of the encoder, two-loss values were added with different weightings (segmentation: 25%; detection: 75%).
The loss function of the drivable area segmentation
The Adam optimizer with a
The dataset was collected in three parking lots having 18 intersections (see Figure 10). One bird’s eye view image per second was collected while driving, and 1,069 images were collected. In each task, 80% of the dataset was used for training, and the remaining portion of the dataset was used for validation. The entire image was used to segment the drivable and nondrivable areas by using a pixel annotation tool [35]. Among the collected images, half of the nonintersection data were excluded, and the road bounding box was trained using 772 images. The reason not to use the entire collected data is to balance the ratio between the nonintersection and intersection data by excluding a similar situation from the nonintersection data. Therefore, the detection accuracy could be higher at the intersection than when using the entire data.
[figure(s) omitted; refer to PDF]
3.3. Model-Based Branch Road Detection Method
The intersection scan model (ISM) method [10] is used to compare the performance of detecting branch roads with the proposed method. The ISM method recognizes branch roads at the intersection using the segmented image, without the global information. ISM [10] defines 21 traversable lines from a scan center point to the end of the image according to a traversable direction [13°, 21°, …, 165°, and 173°]. The traversable directions are divided into the left turn (6 traversable lines), straight (9 traversable lines), and right turn (6 traversable lines) on the horizontal axis of the histogram.
If there are obstacles on the traversable line, an obstacle distance from the scan center point to the closest obstacle on the traversable line is calculated (blue lines in Figure 11). The traversable distance ratio of the traversable line’s length to the obstacle distance is obtained. These ratios according to the traversable direction are used to obtain an ISM histogram (see the red dashed box in Figure 11). If more than half of the traversable distance ratio exceeding the threshold (0.7) exists in the traversable direction section, it can be determined that the branch road exists. Here, the distance ratio is the vertical axis of the ISM histogram.
[figure(s) omitted; refer to PDF]
4. Experimental Results
4.1. Quantitative Analysis of Multi-Task Network
The perception network was tested by validation dataset. The pixel accuracy metric was used to evaluate the performance of the drivable area segmentation:
Detecting result of the rotated road bounding boxes was used over union, the intersection over union (IoU) metric:
Table 1
Results of the perception network.
Parking lots | ||||
Figure 10(a) | Figure 10(b) | Figure 10(c) | ||
Drivable area segmentation (pixel accuracy, (6)) | 95.81% | 94.15% | 93.73% | |
Branch road detection (IoU, (7)) | Rotated box (Proposed) | 98.58% | 97.31% | 97.10% |
Unrotated box (Figure 5(c)) | 97.73% | 95.17% | 94.79% |
As shown in Figures 12 and 13, the resulting images of the segmentation task are indicated as the green area in the bird’s eye view image or the black area in
[figure(s) omitted; refer to PDF]
4.2. Comparison of Branch Road Detection Methods
To evaluate the intersection recognition performance, the proposed rotated road bounding box detection algorithm and ISM [10] algorithm were tested at the same image. A branch road detection accuracy was defined as follows:
Table 2
Intersection detection results.
Rotated box (protected) | Unrotated box (%) | ISM algorithm [10] | |
Intersection accuracy (8) | 97.2 | 95.4% | 82.7% |
“#n” in Figure 12(a) and “#
Cases #3’, #4’, and #5’ of Figure 12(b) are cases where the branch road was not detected in ISM since the area behind the obstacle was recognized as the nondrivable area due to the shadow of the camera. The traversable distance ratios of the left (#3’ and #4’) and right (#5’) turn sections were short, and each branch road was not detected. In #6’, the straight branch road was too wide, and the drivable distance ratio of the left turn was recognized as too long and was mistaken as the left branch road. Situations #3’, #4’, #5’, and #6’ may be addressed by properly tuning the model parameters of ISM, although it is unclear whether these parameters can be applied to other situations. However, the proposed learning-based method was able to accurately find the branch roads without finding the model parameters.
In addition, the computation time between the two methods was additionally compared. The proposed method, multi-task network, showed 27.3 fps, and ISM showed 29.5 fps. Although ISM showed a higher calculation speed, the difference between the two methods is not much different. These methods are faster than the driving control period, 20 fps.
4.3. Results of Applying Existing Motion-Planning Algorithms
The model-based motion-planning algorithms (Tentacle [36] and VVF (velocity vector field) [37]) and learning-based motion-planning algorithm (DAgger (data aggregation) [38]) were used for the intersection driving test. These algorithms utilized
Three motion-planning algorithms were tested in three parking lots (see Figure 10), and the driving results are shown in Figure 13. They were able to calculate an action toward the branch road and could avoid obstacles existing at the branch road because obstacles were regarded as the nondrivable area in
5. Conclusions
This study proposed a method that detects branch roads at an intersection using vision and deep learning, which can be used alongside the existing motion-planning algorithm for navigating in an unstructured environment. The proposed multi-task network distinguished the branch roads at an intersection as the rotated bounding box. At the intersection, the inner area of a box selected through the navigation information was regarded as the drivable area while considering obstacles. Testing in the parking lot, the proposed method detected the branch roads more robustly than the model-based method using the distance and direction histogram of the branch road in the cases where branch roads varied in size and shape, or the drivable area was detected noisy. In addition, the vehicle successfully navigated the intersection by applying the proposed perception method to the existing motion-planning algorithms such as the tentacle, field, imitation learning algorithms without using global information. In the future, the experiment will be conducted in more diverse environments.
[1] M. Hentschel, B. Wagner, "Autonomous robot navigation based on openstreetmap geodata," pp. 1645-1650, DOI: 10.1109/ITSC.2010.5625092, .
[2] B. B. K. Ayawli, R. Chellali, A. Y. Appiah, F. Kyeremeh, "An overview of nature-inspired, conventional, and hybrid methods of autonomous vehicle path planning," Journal of Advanced Transportation, vol. 2018,DOI: 10.1155/2018/8269698, 2018.
[3] T. Ort, K. Murthy, R. Banerjee, S. K. Gottipati, D. Bhatt, I. Gilitschenski, L. Paull, D. Rus, "Maplite: autonomous intersection navigation without a detailed prior map," IEEE Robotics and Automation Letters, vol. 5 no. 2, pp. 556-563, DOI: 10.1109/lra.2019.2961051, 2020.
[4] M. Á. de Miguel, F. García, J. M. Armingol, "Improved lidar probabilistic localization for autonomous vehicles using gnss," Sensors, vol. 20 no. 11,DOI: 10.3390/s20113145, 2020.
[5] Z. Wang, A. Lambert, "A low-cost consistent vehicle localization based on interval constraint propagation," Journal of Advanced Transportation, vol. 2018,DOI: 10.1155/2018/2713729, 2018.
[6] M. Osman, A. Hussein, A. Al-Kaff, "Intelligent vehicles localization approaches between estimation and information: a review," ,DOI: 10.1109/ICVES.2019.8906426, .
[7] E. C. Pereira, D. A. Lima, A. C. Victorino, "Autonomous vehicle global navigation approach associating sensor based control and digital maps," pp. 2404-2409, DOI: 10.1109/ROBIO.2014.7090699, .
[8] D. A. De Lima, A. C. Victorino, "Sensor-based control with digital maps association for global navigation: a real application for autonomous vehicles," pp. 1791-1796, DOI: 10.1109/ITSC.2015.291, .
[9] Q. Li, L. Chen, Q. Zhu, M. Li, Q. Zhang, S. S. Ge, "Intersection detection and recognition for autonomous urban driving using a virtual cylindrical scanner," IET Intelligent Transport Systems, vol. 8 no. 3, pp. 244-254, DOI: 10.1049/iet-its.2012.0202, 2014.
[10] Y. Yi, L. Hao, Z. Hao, S. Songtian, L. Ningyi, S. Wenjie, "Intersection scan model and probability inference for vision based small-scale urban intersection detection," pp. 1393-1398, DOI: 10.1109/IVS.2017.7995905, .
[11] A. L. Ballardini, H. Saz, S. Carrasco Limeros, J. Lorenzo, I. Parra Alonso, N. Hernández Parra, I. García Daza, M. Á. Sotelo, "Urban intersection classification: a comparative analysis," Sensors, vol. 21 no. 18,DOI: 10.3390/s21186269, 2021.
[12] M. Hnewa, H. Radha, "Object detection under rainy conditions for autonomous vehicles: a review of state-of-the-art and emerging techniques," IEEE Signal Processing Magazine, vol. 38 no. 1, pp. 53-67, DOI: 10.1109/msp.2020.2984801, 2021.
[13] Z. Liu, Y. Cai, H. Wang, L. Chen, H. Gao, Y. Jia, Y. Li, "Robust target recognition and tracking of self-driving cars with radar and camera information fusion under severe weather conditions," IEEE Transactions on Intelligent Transportation Systems, vol. 23 no. 7, pp. 6640-6653, DOI: 10.1109/TITS.2021.3059674, 2021.
[14] A. Bell, T. Mantecón, C. Díaz, C. R. del Blanco, F. Jaureguizar, N. García, "A novel system for nighttime vehicle detection based on foveal classifiers with real-time performance," IEEE Transactions on Intelligent Transportation Systems, vol. 23 no. 6, 2022.
[15] Y. Cai, T. Luan, H. Gao, H. Wang, L. Chen, Y. Li, M. A. Sotelo, Z. Li, "Yolov4-5d: an effective and efficient object detector for autonomous driving," IEEE Transactions on Instrumentation and Measurement, vol. 70,DOI: 10.1109/tim.2021.3065438, 2021.
[16] Z. Yao, X. Chen, "Efficient lane detection technique based on lightweight attention deep neural network," Journal of Advanced Transportation, vol. 2022,DOI: 10.1155/2022/5134437, .
[17] T. Almeida, B. Lourenço, V. Santos, "Road detection based on simultaneous deep learning approaches," Robotics and Autonomous Systems, vol. 133,DOI: 10.1016/j.robot.2020.103605, 2020.
[18] L. Wang, J. Wang, X. Wang, Y. Zhang, "3d-lidar based branch estimation and intersection location for autonomous vehicles," pp. 1440-1445, DOI: 10.1109/IVS.2017.7995912, .
[19] D. Bhatt, D. Sodhi, A. Pal, V. Balasubramanian, M. Krishna, "Have i reached the intersection: a deep learning-based approach for intersection detection from monocular cameras," pp. 4495-4500, DOI: 10.1109/IROS.2017.8206317, .
[20] A. L. Ballardini, A. H. Saz, M. Á. Sotelo, "Model guided road intersection classification," pp. 703-709, DOI: 10.1109/IV48863.2021.9575605, .
[21] F. Yan, K. Wang, B. Zou, L. Tang, W. Li, C. Lv, "Lidar-based multi-task road perception network for autonomous vehicles," IEEE Access, vol. 8, pp. 86753-86764, DOI: 10.1109/access.2020.2993578, 2020.
[22] J. B. Chipka, M. Campbell, "Estimation and navigation methods with limited information for autonomous urban driving," Engineering Reports, vol. 1 no. 4,DOI: 10.1002/eng2.12054, 2019.
[23] M. Kim, J. Ahn, J. Park, "Global planning method for visiting roads with parking spaces in priority using rural postman problem," pp. 4184-4189, .
[24] D. Wu, M. Liao, W. Zhang, X. Wang, "Yolop: you only look once for panoptic driving perception," 2021. http://arxiv.org/abs/2108/11250
[25] Y. Qian, J. M. Dolan, M. Yang, "Dlt-net: joint detection of drivable areas, lane lines, and traffic objects," IEEE Transactions on Intelligent Transportation Systems, vol. 21 no. 11, pp. 4670-4679, DOI: 10.1109/tits.2019.2943777, 2020.
[26] M. Teichmann, M. Weber, M. Zoellner, R. Cipolla, R. Urtasun, "Multinet: real-time joint semantic reasoning for autonomous driving," pp. 1013-1020, DOI: 10.1109/IVS.2018.8500504, .
[27] K. Simonyan, A. Zisserman, "Very deep convolutional networks for large-scale image recognition," 2014. http://arxiv.org/abs/1409.1556
[28] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, "Imagenet: a large-scale hierarchical image database," pp. 248-255, DOI: 10.1109/CVPR.2009.5206848, .
[29] J. Long, E. Shelhamer, T. Darrell, "Fully convolutional networks for semantic segmentation," Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3431-3440, DOI: 10.1109/CVPR.1997.609286, .
[30] O. Ronneberger, P. Fischer, T. Brox, "U-net: convolutional networks for biomedical image segmentation," pp. 234-241, .
[31] N. Otsu, "A threshold selection method from gray-level histograms," IEEE transactions on systems, man, and cybernetics, vol. 9 no. 1, pp. 62-66, DOI: 10.1109/tsmc.1979.4310076, 1979.
[32] J. Redmon, A. Farhadi, "Yolo9000: better, faster, stronger," Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7263-7271, DOI: 10.1109/CVPR.2017.690, .
[33] A. Neubeck, L. Van Gool, "Efficient non-maximum suppression," vol. 3, pp. 850-855, DOI: 10.1109/ICPR.2006.479, .
[34] J. Zhang, S. Singh, "Loam: lidar odometry and mapping in real-time robotics: science and systems," Robotics, vol. 2 no. 9, 2014.
[35] A. Bréhéret, "Pixel annotation tool," August 2017. https://github.com/abreheret/PixelAnnotationTool
[36] H. Mouhagir, R. Talj, V. Cherfaoui, F. Aioun, F. Guillemard, "Evidential-based approach for trajectory planning with tentacles, for autonomous vehicles," IEEE Transactions on Intelligent Transportation Systems, vol. 21 no. 8, pp. 3485-3496, DOI: 10.1109/tits.2019.2930035, 2020.
[37] V. Olunloyo, M. Ayomoh, "Autonomous mobile robot navigation using hybrid virtual force field concept," European Journal of Scientific Research, vol. 31 no. 2, pp. 204-228, 2009.
[38] J. Ahn, M. Kim, J. Park, "Vision-based autonomous driving for unstructured environments using imitation learning," 2022. http://arxiv.org/abs/2202.10002
[39] J. Ahn, S. Shin, M. Kim, J. Park, "Accurate path tracking by adjusting look-ahead point in pure pursuit method," International Journal of Automotive Technology, vol. 22 no. 1, pp. 119-129, DOI: 10.1007/s12239-021-0013-7, 2021.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2022 Joonwoo Ahn et al. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Autonomous vehicles need a driving method to be less dependent on localization data to navigate intersections in unstructured environments because these data may not be accurate in such environments. Methods of distinguishing branch roads existing at intersections using vision and applying them to intersection navigation have been studied. Model-based detection methods recognize patterns of the branch roads, but are sensitive to sensor noise and difficult to apply to various complex situations. Therefore, this study proposes a method for detecting the branch roads at the intersection using deep learning. The proposed multi-task deep neural network distinguishes the branch road into a shape of rotated bounding boxes and also recognizes the drivable area to consider obstacles inside the box. Through the output of the network, an occupancy grid map consisting of one branch road at an intersection is obtained, which can be used as an input to the existing motion-planning algorithms that do not consider intersections. Experiments in real environments show that the proposed method detected the branch roads more accurately than the model-based detection method, and the vehicle drove safely at the intersection.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details




1 Dynamic Robotic Systems (DYROS) Lab., Graduate School of Convergence Science and Technology, Seoul National University, 1, Gwanak-ro, Seoul 08826, Republic of Korea
2 Dynamic Robotic Systems (DYROS) Lab., Graduate School of Convergence Science and Technology, Seoul National University, 1, Gwanak-ro, Seoul 08826, Republic of Korea; ASRI, RICS, Seoul National University, Seoul, Republic of Korea; Advanced Institutes of Convergence Technology, Suwon 443-270, Republic of Korea