DOI QR코드

DOI QR Code

3D LIDAR Based Vehicle Localization Using Synthetic Reflectivity Map for Road and Wall in Tunnel

  • Im, Jun-Hyuck (Department of Electronics Engineering, Konkuk University) ;
  • Im, Sung-Hyuck (Satellite Navigation Team, Korea Aerospace Research Institute) ;
  • Song, Jong-Hwa (AESA Radar R&D Center, Hanwha Systems) ;
  • Jee, Gyu-In (Department of Electronics Engineering, Konkuk University)
  • Received : 2017.08.17
  • Accepted : 2017.10.18
  • Published : 2017.12.15

Abstract

The position of autonomous driving vehicle is basically acquired through the global positioning system (GPS). However, GPS signals cannot be received in tunnels. Due to this limitation, localization of autonomous driving vehicles can be made through sensors mounted on them. In particular, a 3D Light Detection and Ranging (LIDAR) system is used for longitudinal position error correction. Few feature points and structures that can be used for localization of vehicles are available in tunnels. Since lanes in the road are normally marked by solid line, it cannot be used to recognize a longitudinal position. In addition, only a small number of structures that are separated from the tunnel walls such as sign boards or jet fans are available. Thus, it is necessary to extract usable information from tunnels to recognize a longitudinal position. In this paper, fire hydrants and evacuation guide lights attached at both sides of tunnel walls were used to recognize a longitudinal position. These structures have highly distinctive reflectivity from the surrounding walls, which can be distinguished using LIDAR reflectivity data. Furthermore, reflectivity information of tunnel walls was fused with the road surface reflectivity map to generate a synthetic reflectivity map. When the synthetic reflectivity map was used, localization of vehicles was able through correlation matching with the local maps generated from the current LIDAR data. The experiments were conducted at an expressway including Maseong Tunnel (approximately 1.5 km long). The experiment results showed that the root mean square (RMS) position errors in lateral and longitudinal directions were 0.19 m and 0.35 m, respectively, exhibiting precise localization accuracy.

Keywords

1. INTRODUCTION

Precise vehicle localization is essential for safe driving for autonomous vehicles. In general, position of autonomous driving vehicles is obtained through the global positioning system (GPS). However, GPS signals cannot be received in tunnels. Localization at a road section of short distance can be estimated through dead reckoning (DR) method. However, as the driving distance becomes longer, the position error becomes larger due to the accumulation of error. In this regard, the accumulated position error of DR needs to be corrected using a variety of sensors used in autonomous driving vehicles. When cameras are used, a lateral position error in tunnels can be corrected using lane detection. However, a longitudinal position error is difficult to be corrected as only few feature points can be extracted from tunnels. On the contrary, Light Detection and Ranging (LIDAR) system can measure a distance to a structure separated from the tunnel walls such as signboards or jet fans accurately, and it is much more effective at longitudinal position error correction than using cameras. Moreover, feature points of lanes and walls can be extracted using the LIDAR reflectivity information. Thus, this study conducted vehicle localization inside a tunnel using three-dimensional (3D) LIDAR.

Localization inside a tunnel requires detection of structures that are separated from the tunnel wall basically. However, few structures are found in a tunnel and detection of structures using LIDAR distance information is difficult. Thus, this study used LIDAR reflectivity information for vehicle localization. Normally, fire hydrants and evacuation guide lights are installed in both sides of the tunnel wall in every certain distance. The reflectivity characteristic of these structures is very different from that of surrounding walls. Using this characteristic, a reflectivity map of walls at the same height with that of this structure can be created to perform correlation matching. This is similar to the localization technique using a road surface reflectivity map in the study of Levinson et al. (2007, 2010). When a road surface reflectivity map is used, accurate positioning can be achieved in a general road where road line is marked with a dashed line. The reflectivity information of tunnel walls can be fused in the same manner used in the existing road surface reflectivity map. When such fused synthetic reflectivity map is used, localization can be done consecutively in general road and tunnel via single correlation matching for road surface and tunnel wall. In this study, accumulative position error of DR was corrected through correlation matching using 3D LIDAR. The experiments were conducted at an expressway including Maseong Tunnel (approximately 1.5 km long) and the driving speed was approximately 90 km/h. For the reference vehicle location, post-processing results of RTK/INS (NovAtel SPAN System, using the RINEX data in the continuously operating reference station in Suwon, baseline length: approximately 12 km) were used. From this, localization performance was analyzed.

The present paper is organized as follows: In Section 2, studies related to localization inside tunnels are described; in Section 3, a method to create a synthetic reflectivity map is presented; in Section 4, the configuration of the localization system is explained; in Section 5, localization results are analyzed; and lastly, in Section 6, conclusions are presented.

2. STUDIES RELATED TO LOCALIZATION INSIDE TUNNELS

Table 1 presents the requirements of localization accuracy for autonomous driving (Ballingall 2013). The lateral localization accuracy within 70 cm is required at the confidence level of 95% when the map is used (V2I) as specified in Table 1. The accuracy refers to a root mean square (RMS) position error within approximately 35 cm. The requirements of the lateral localization accuracy are not yet officially documented. However, the accuracy requires a RMS position error within approximately 50 cm. Generally, autonomous driving vehicles are operated in environments where GPS signals can be received. In an open space, the position accuracy within a few tens of centimeters can be achieved using a precision positioning technique through the GPS such as Real Time Kinematic (RTK). However, GPS signals cannot be received in tunnels and the requirements of localization accuracy should be satisfied using only sensors mounted on vehicles.

Table. 1. Localization accuracy requirement for autonomous driving.

Type Level Accuracy requirement Communication
Latency (seconds)
95% confidence
level (m)
Root mean
square (order)
V2I
(absolute)
 
V2V
(relative)
 
Road-level
Lane-level
Where-in-lane-level
Road-level
Lane-level
Where-in-lane-level
5.0
1.1
0.7
5.0
1.5
1.0
Meter
Sub-meter
Decimeter
Meter
Sub-meter
Decimeter
1-5
1.0
0.1
0.1
0.1
0.01-0.1
*Austroads Research Report (2013), Vehicle Positioning for C-ITS in Australia (Background
Document)

 

The LIDAR system has been widely used for localization in tunnels. A study by Duff et al. (2003) used 2D LIDAR for localization in mines. However, 2D LIDAR at that time was mainly used for the purpose of prevention of collision by recognizing a distance between vehicle and both sides of the tunnel wall rather than using the concept of localization. Since then, 3D LIDAR systems mountable to vehicles have been developed and LIDAR has been widely utilized as a core sensor for the localization of mobile vehicles. Tardioli & Villarroel (2014) published a paper about localization at a tunnel-like environment. In their study, a robot's position was estimated by extracting walls using 2D LIDAR mounted horizontally to a robot in the corridor inside the building. If a robot moved in the corridor in parallel with the wall, a longitudinal position cannot be estimated using the extracted wall information. To overcome this, their study used the distance from the wall through zig-zag moving of a robot to enable a longitudinal position estimation. If the odometer information of the robot was additionally used, both of lateral and longitudinal position errors were able to be reduced. However, since vehicles cannot move in zig-zags in a tunnel, this method cannot be applied to real road tunnels.

Larson et al. (2014) studied robot localization at the cylindrical drain pipe. Their study rotated 2D LIDAR for the use of 3D LIDAR thereby performing Iterative Closest Point (ICP) scan matching. They also used visual odometry using stereo cameras. In their study, no quantitative values about localization results were presented, but only the position error up to 90 cm during 900 m movement occurred in unofficial tests was mentioned. More recently, a study on autonomous flight using a drone at a hydraulic pipe was published (Ozaslan et al. 2017). In the study, localization of drone mounted with 3D LIDAR was conducted in a hydraulic pipe, which was similar to a tunnel environment. The normal distribution transform (NDT) scan matching technique based on a point cloud was used for localization. The localization result showed that a position error in lateral and height directions was less than approximately 5% of 5.5 m diameter hydraulic pipe, and the longitudinal position error was not considered. As presented in the previous two papers, ICP or NDT scan matching techniques using LIDAR distance information show poor performance in longitudinal matching in tunnels. Thus, they are not applicable to actual road tunnels.

Daoust et al. (2016) published a paper about localization at tunnels where subway trains were operated. In their study, a global map was created by accumulating point clouds over the entire test section after collecting data through 2D LIDAR mounted in front of the subway train. Later, local maps were created in every update cycle of LIDAR data thereby performing point cloud matching with the created global map. Furthermore, since subway trains run over the railway, only longitudinal positions were considered. In the tunnel where the test was conducted, signal light posts were installed in every certain distance for safe operation of subway trains. These posts were distinguishable from surrounding structures and by using these, it was possible to correct the longitudinal position error. The localization result showed that a position error of up to 1.8 m occurred in the tunnel section. However, these posts are not present in road tunnels where vehicles pass through.

As described above, studies on localization using LIDAR in tunnels have been conducted and published previously, but all of them used LIDAR distance information. On the contrary, few studies have been conducted on localization at road tunnels where vehicles run. When only the LIDAR distance information is used in road tunnels, localization accuracy and reliability are poor because only few structures other than tunnel walls are available. However, as mentioned in Section 1, since structures of which reflectivity is different from each other are installed on road tunnel walls at every certain distance, vehicle localization can be done using these structures. Currently, no other studies have been found on vehicle localization inside tunnels using LIDAR reflectivity information.

3. CREATION OF SYNTHETIC REFLECTIVITY MAP

Fig. 1 shows the camera image inside the tunnel and the road reflectivity map of the section seen through the image. As shown in the camera image in Fig. 1, no significant structures separated from the wall are present inside the tunnel. Also, no special features are displayed in the road reflectivity map in the same section for longitudinal position estimation. However, a fire hydrant as marked with No. 1 in Fig. 1 and evacuation guide light as marked with No. 2 are installed at every certain distance. These safety facilities have a highly different reflectivity characteristic than that of surrounding walls. In the present study, Velodyne HDL-32E 3D LIDAR sensor was used. Fig. 2 shows the mounting location of 3D LIDAR. Figs. 3 and 4 depict the LIDAR reflectivity data of fire hydrant and evacuation guide light, respectively.

Fig. 1.PNG 이미지
Fig. 1. Camera image (left) and road reflectivity map (right) in tunnel.

Fig. 2.png 이미지
Fig. 2. Mounting position of 3D LIDAR.

Fig. 3.PNG 이미지
Fig. 3. Reflectivity image for fire hydrant in tunnel.

As shown in Figs. 3 and 4, the LIDAR reflectivities of hydrant and evacuation guide light show clear difference from that of the surrounding wall. Using this difference in reflectivity, longitudinal position errors of vehicles can be corrected. Here, a map with the same form of the road reflectivity map in Fig. 1 can be created using the reflectivity information as shown in Figs. 3 and 4. Furthermore, a synthetic reflectivity map that is present at a single 2D plane can be created by projecting the reflectivity map of the vertical wall to the road plane. Fig. 5 explains the creation principle of the synthetic reflectivity map.

Fig. 4.PNG 이미지
Fig. 4. Reflectivity image for evacuation guidance light in tunnel.

Fig. 5.PNG 이미지
Fig. 5. The principle of generating synthetic reflectivity map.

As shown in Fig. 5, a certain height over the hydrant and evacuation guide light is removed from the vertical cross-section of the tunnel. After this, as shown in the planar figure of three-dimensional figure, reflectivity data of the wall are projected to both sides of the road reflectivity map plane. A synthetic reflectivity map is created as shown in the lower left end of the Fig. 5 through this process. Fig. 6 shows the synthetic reflectivity map created through this process. As shown in Fig. 6, the areas where (1) fire hydrant and (2) evacuation guide light are located are clearly distinguished from surrounding walls in the map. Using the synthetic reflectivity map, vehicle localization can be done through a single correlation matching even in a tunnel.

Fig. 6.PNG 이미지
Fig. 6. Generated synthetic reflectivity map.

4. DESIGN OF LOCALIZATION SYSTEM

In this study, the correlation matching result with the synthetic reflectivity map is used as a measurement in the extended Kalman Filter (EKF). The EKF is designed as follows:

4.1 Time Update

Time update uses the output value of GPS/DR. For the GPS/DR sensor, CruizCore DS6200 from Microinfinity was used, and the azimuth accuracy was within 5° in open space. The state variable of the filter can be expressed as presented in Eq. (1).

\(q=\left(\delta x,\delta y\right)^T\)                                                                                                            (1)

where \(\delta x\) and \(\delta y\) refer to 2D horizontal position error of vehicle in the ENU coordinate system. The state equation is presented in Eq. (2).

\({\widetilde{q}}_k=F_K\cdot{\hat{q}}_{k-1}\)                                                                                                              
\(F_k=\left[\begin{matrix}1&0\\0&1\\\end{matrix}\right]\)                                                                                                            (2)

As presented in Eq. (2), no change in the error of GPS/DR for a short period of time was assumed.

4.2 Measurement Update

The measurement equation is presented in Eq. (3).

\(z_k=\left[\begin{matrix}\Delta x\\\Delta y\\\end{matrix}\right]=H_k\cdot{\widetilde{q}}_k\)                                                                                                   
\(H_k=\left[\begin{matrix}1&0\\0&1\\\end{matrix}\right]\)                                                                                                             (3)

where \(\Delta x\) and \(\Delta y\) refer to the result values of correlation matching. The measurement update is conducted as presented in Eq. (4).

\({\hat{q}}_k={\widetilde{q}}_k+K(z_k-H\cdot{\widetilde{q}}_k)\)                                                                                                 (4)

where \(K\) refers to the Kalman gain.

5. LOCALIZATION RESULT AND ANALYSIS

Fig. 7 shows the experimental section. As shown in Fig. 7, the experiment was conducted at approximately 3 km-long section (red line) including Maseong Tunnel on Yeongdong Expressway, and data obtained from 4pm to 6pm during weekdays when traffic volumes were large were used. The RTK/INS post-processing results provided by the NovAtel SPAN system were used as the reference trajectory of the driving vehicle. The vehicle ran the same section twice. Then, a synthetic reflectivity map was created using the first driving data, and vehicle localization was performed using the second driving data. However, the RTK/INS post-processing results had a slight position error in a tunnel. Fig. 8 shows the overlap of the maps created using data at two different times around the RTK/INS location.

Fig. 7.PNG 이미지
Fig. 7. Vehicle trajectory including Masung tunnel.

Fig. 8.PNG 이미지
Fig. 8. The result overlapping two maps at different points in time.

Fig. 8a shows the outside section of the tunnel while Fig. 8b and 8c show the inside section of the tunnel. In Fig. 8a, two maps at two different times are exactly overlapped as the road lines were matched. This match indicates that the RTK/INS post-processing results are very accurate. However, as shown in Fig. 8b, some road lines are not exactly matched. Approximately 20 cm of error in the lateral direction of vehicle driving trajectory occurred considering that the grid size of the map used in this study was 10 cm. A relatively smaller position error is shown in Fig. 8c than in Fig. 8b. Furthermore, as the emergency exit area was matched in the longitudinal direction in Figs. 8b and 8c, a longitudinal error was rarely revealed compared to the lateral error. Since the tunnel was nearly in a straight line, it was expected to have smaller longitudinal position error than the lateral error once correction was made considering the accurate positions of the start and end points of the tunnel. As shown above, a small position error in the lateral direction was also found in the post-processing result using expensive RTK/INS equipment inside of the tunnel. This error was represented as an error in the map, and ultimately is reflected in the localization error.

Fig. 9 shows the localization result using the road reflectivity map. The red line in Fig. 9 indicates the localization result through correlation matching, and the blue line represents the result of GPS/DR. The vertical pink lines represent the start and end in the tunnel section. As shown in the figure, road lines are also present inside of the tunnel, and thus the lateral position error was minimal. However, the longitudinal position error was very large because correlation matching was rarely done. The RMS position error in the tunnel section was 0.19 m and 1.39 m in lateral and longitudinal directions, respectively. As described above, localization cannot be done in the tunnel section using only the road reflectivity map.

f9.jpg 이미지
Fig. 9. Lateral (left) and longitudinal (right) position error (road reflectivity map),

Fig. 10 shows the localization result using the synthetic reflectivity map. As shown in Fig. 10, the lateral position error was also minimal, as it was when the road reflectivity map was used. Furthermore, the longitudinal position error was also limited to a certain level of error. This was due to the improvements on correlation matching performance in the longitudinal direction as a result of the effect of fire hydrant and evacuation guide light. It was found that the longitudinal position error of GPS/DR was suddenly changed drastically at the end of the tunnel. This was due to the sudden correction of accumulative DR errors caused by the GPS signal re-reception as the vehicle exited the tunnel. As a result, the GPS/DR position input was drastically changed during time update, and thereby affecting correlation matching. Conclusively, a position error estimated in the filter was drastically changed, and the changed error was directly expressed as the position error. However, this drastically changed error can be resolved to some extent through residual analysis. The localization result showed that the RMS position error in the tunnel section was 0.19 m and 0.35 m in lateral and longitudinal directions, respectively. Furthermore, the drift error of GPS/DR was well corrected in both of the lateral and longitudinal directions.

f10.jpg 이미지
Fig. 10. Lateral (left) and longitudinal (right) position error (synthetic reflectivity map).

As mentioned above, a small position error is included in the map used in this study. Approximately 20 cm of lateral error was included in the map as shown in Fig. 8b. The lateral position error in Figs. 9 and 10 was also biased approximately 20 cm in the same section as well. Thus, if the map was more accurate, the lateral RMS position error would be further reduced. However, even if these errors in the map were included, localization results using the synthetic reflectivity map were verified as highly accurate.

6. CONCLUSION

In this paper, a synthetic reflectivity map was created using 3D LIDAR in tunnels, and vehicle localization was conducted through correlation matching with the reflectivity map. The experiment results showed that 19 cm of lateral and 35 cm of longitudinal RMS position errors occurred. Considering the created map's error, the lateral position error would be further reduced if the map was accurate. As presented in Table 1, localization accuracies required for autonomous driving were 35 cm and 50 cm of lateral and longitudinal RMS position errors, respectively. Thus, the localization technique using the synthetic reflectivity map satisfied the position accuracy for autonomous driving in tunnels.

In this paper, a synthetic reflectivity map using reflectivity of road surface and tunnel walls simultaneously was proposed. However, the grid-based map has a drawback of its large data size. Thus, a study to reduce the large amount of map data will be performed in the future.

ACKNOWLEDGMENTS

This research was supported by a grant from Development of Cooperative Automated Driving Roadway Systems based on LDM and V2X (17TLRP-B101406-03) in Transportation Logistics Research Program by the Ministry of Land, Infrastructure and Transport of Korean government.

References

  1. Ballingall, S. 2013, Vehicle Positioning for C-ITS in Australia (Background Document), Austroads, Sydney. https://www.onlinepublications.austroads.com.au/items/APR431-13
  2. Daoust, T., Pomerleau, F., & Barfoot, T. D. 2016, Light at the end of the tunnel: high-speed lidar-based train localization in challenging underground environments, Conference on Computer and Robot Vision, 1-3 June 2016, Victoria, British Columbia, Canada. https://doi.org/10.1109/CRV.2016.54
  3. Duff, E. S., Roberts, J. M., & Corke, P. I. 2003, Automation of an underground mining vehicle using reactive navigation and opportunistic localization, In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, 27-31, Oct 2003, Las Vegas, Nevada, USA. https://doi.org/10.1109/IROS.2003.1249742
  4. Larson, J., Okorn, B., Pastore, T., Hooper, D., & Edwards, J. 2014, Counter Tunnel Exploration, Mapping, and Localization with an Unmanned Ground Vehicle, Proc. SPIE 9084, 6-8, May 2014, Baltimore, MD, USA. https://www.hsdl.org/?view&did=762738
  5. Levinson, J., Montemerlo, M., & Thrun, S. 2007, Map-based precision vehicle localization in urban environments, In Robotics Science and Systems (Cambridge, MA: MIT Press). https://doi.org/10.15607/RSS.2007.III.016
  6. Levinson, J. & Thrun, S. 2010, Robust vehicle localization in urban environments using probabilistic maps, In Proceedings of the 2010 IEEE International Conference on Robotics and Automations, May 3-8, 2010, Anchorage, Alaska, USA. https://doi.org/10.1109/ROBOT.2010.5509700
  7. Ozaslan, T., Loianno, G., Keller, J., Taylor, C. J., Kumar, V., et al. 2017, Autonomous Navigation and Mapping for Inspection of Penstocks and Tunnels With MAVs, IEEE Robotics and Automation Letters, 2, 1740-1747. https://doi.org/10.1109/LRA.2017.2699790
  8. Tardioli, D. & Villarroel, J. L. 2014, Odometry-less localization in tunnel-like environments, 2014 IEEE International Conference on Autonomous Robot Systems and Competitions, 14-15 May 2014, Espinho, Portugal. https://doi.org/10.1109/ICARSC.2014.6849764