ArticlePDF Available

3D LiDAR-Based Precision Vehicle Localization with Movable Region Constraints

Authors:

Abstract and Figures

This paper discusses a high-performance similarity measurement method based on known map information named the cross mean absolute difference (CMAD) method. Applying the conventional normalized cross-correlation (NCC) feature registration method requires sufficient numbers of feature points, which must also exhibit near-normal distribution. However, Light Detection and Ranging (LiDAR) ranging point cloud data scanned and collected on-site are scarce and do not fulfill near-normal distribution. Consequently, considerable localization errors occur when NCC features are registered with map features. Thus, the CMAD method was proposed to effectively improve the NCC algorithm and localization accuracy. Because uncertainties in localization sensors cause deviations in the localization processes, drivable moving regions (DMRs) were established to restrict the range of location searches, filter out unreasonable trajectories, and improve localization speed and performance. An error comparison was conducted between the localization results of the window-based, DMR–CMAD, and DMR–NCC methods, as well as those of the simultaneous localization and mapping methods. The DMR–CMAD method did not differ considerably from the window-based method in its accuracy: the root mean square error in the indoor experiment was no higher than 10 cm, and that of the outdoor experiment was 10–30 cm. Additionally, the DMR–CMAD method was the least time-consuming of the three methods, and the DMR–NCC generated more localization errors and required more localization time than the other two methods. Finally, the DMR–CMAD algorithm was employed for the successful on-site instant localization of a car.
Content may be subject to copyright.
Sensors 2019, 19, 942; doi:10.3390/s19040942 www.mdpi.com/journal/sensors
Article
3D LiDAR-Based Precision Vehicle Localization with
Movable Region Constraints
Chih-Ming Hsu* and Chung-Wei Shiu
Graduate Institute of Mechatronic Engineering, National Taipei University of Technology,
Taipei 106, Taiwan; jmshiu@ntut.edu.tw
* Correspondence: jmshiu@ntut.edu.tw; Tel.: +886-912-827-171
Received: 31 January 2019; Accepted: 18 February 2019; Published: 22 February 2019
Abstract: This paper discusses a high-performance similarity measurement method based on
known map information named the cross mean absolute difference (CMAD) method. Applying the
conventional normalized cross-correlation (NCC) feature registration method requires sufficient
numbers of feature points, which must also exhibit near-normal distribution. However, Light
Detection and Ranging (LiDAR) ranging point cloud data scanned and collected on-site are scarce
and do not fulfill near-normal distribution. Consequently, considerable localization errors occur
when NCC features are registered with map features. Thus, the CMAD method was proposed to
effectively improve the NCC algorithm and localization accuracy. Because uncertainties in
localization sensors cause deviations in the localization processes, drivable moving regions (DMRs)
were established to restrict the range of location searches, filter out unreasonable trajectories, and
improve localization speed and performance. An error comparison was conducted between the
localization results of the window-based, DMRCMAD, and DMRNCC methods, as well as those
of the simultaneous localization and mapping methods. The DMRCMAD method did not differ
considerably from the window-based method in its accuracy: the root mean square error in the
indoor experiment was no higher than 10 cm, and that of the outdoor experiment was 1030 cm.
Additionally, the DMRCMAD method was the least time-consuming of the three methods, and the
DMRNCC generated more localization errors and required more localization time than the other
two methods. Finally, the DMRCMAD algorithm was employed for the successful on-site instant
localization of a car.
Keywords: localization; normalized cross-correlation; 3D LiDAR
1. Introduction
Autopilot technology was developed to substantially improve driving safety and convenience
and thereby mitigate the burden of drivers. In the future, fully automated vehicles will likely
constitute the main body of the smart transportation system and replace human drivers entirely;
however, in its current form, this technology is more moderately implemented in advanced driver
assistance systems. One mandatory function of autopilot technology is environmental perception,
which prevents collision. Similarly crucial is accurate localization, particularly in urban environments
where vehicles are operated on city roads; like human drivers, automated vehicles must adhere to
traffic rules.
The existing self-localization systems in vehicles are dependent on sensor assistance, and they
are categorized into passive or active sensor systems according to the types of sensors used. Passive
sensor systems are further divided into global navigation satellite systems, inertial navigation
systems, and stereo visions. The accuracy of global navigation satellite systems is affected by non-
line-of-sight reception and multipath interference [1]; when operating in indoor environments, they
Sensors 2019, 19, 942 2 of 25
also may not receive consistent signals, hampering their ability to provide location information.
Inertial navigation systems, which provide accurate relative location of vehicles instantly, are subject
to deteriorating accuracy over time [2]. Stereo visions involve the use of vision-based lane detection
and provide location information by detecting stop lines [3], curbs [4,5], arrows [6,7], and traffic
signals [8]. However, in poorly lit environments, such as indoor parking lots, the localization
accuracy of stereo visions may decrease; moreover, these systems cannot locate vehicles on the road
without the aforementioned signs.
The existing active sensors comprise laser range finders and light detection and ranging (LiDAR)
systems, the latter of which are more commonly used. Prior research on smart vehicles has
predominately discussed the successful implementation of two-dimensional (2D) LiDAR and
Velodyne active sensor systems. Overall, active sensors are widely favored over passive sensors
because they simplify the distance estimation processes of basic distances and generate desirable
localization results.
Simultaneous localization and mapping (SLAM) systems simultaneously illustrate and update
maps of unknown environments and locate agents. SLAM is a primary component of robot
navigation systems, and has considerably evolved over the past two decades. As its name implies,
SLAM systems provide both localization and mapping functions. The localization functions comprise
curb-based, road mark-based, and landmark-and-building-based localization. Hata et al. [9]
developed a curb-based localization technique using Velodyne sensors, wherein curbs were
identified as obstacles and detected through the multilayer laser ring compression in the LiDAR.
Subsequently, Hata et al. [10] proposed a road mark-based localization technique, again using
Velodyne sensors; in this method, all the road signs on road surfaces were captured using the
intensity information in the LiDAR for localization. A landmark-and-building-based localization
technique, also using Velodyne sensors, was then established by Choi [11]. He developed a mixed-
map SLAM system to illustrate environments through the simultaneous employment of grid and
feature maps, which consist of both 2D and three-dimensional (3D) mapping. The GraphSLAM
algorithm is used for 2D mapping and was regarded as a least-square problem by Thrun et al. [12].
In large-scale mapping, GraphSLAM can process a massive number of features and integrate global
positioning system information to its mapping processes. For example, Levinson [13] combined a
global positioning system, an inertial measurement unit (IMU), an odometer, and LiDAR data to
generate high-resolution 2D surface maps. However, 3D mapping is more reliable and accurate than
2D mapping. Particularly, featureless roads benefit the most when preprepared maps are used for
navigation because they mitigate the cumulative errors in SLAM, whose instant localization results
may be undesirable for these roads.
Previous studies have incorporated various estimation technologies to solve the problems in
SLAM. Smith and Cheeseman [14,15] developed the extended Kalman filter (EKF) to solve these
problems. However, when the road mark information in the EKF increases, its covariance matrix may
expand and aggravate calculation load; in short, road mark localization errors can escalate into
substantial cumulative errors in the EKF. Moreover, the EKF is only suitable for solving linear
systems; when used to solve nonlinear systems, the EKF may lead to slow convergence or undesirable
divergence. Subsequently, Montermerlo et al. [16,17] developed FastSLAM, which is based on
particle filters. FastSLAM 1.0 employed only basic odometer information to estimate the location of
a robot, and thus the estimation accuracy of the system decreased following an increase in the
cumulative errors in the odometer. By contrast, FastSLAM 2.0 applied the concept of EKF updates
and the linearization of nonlinear environmental information to improve its localization accuracy.
However, using the EKF to update the location of the robot increased the quantity of the
environmental information and the calculation cost.
For a 3D point cloud map, existing approaches, including Normal Distribution Transform (NDT)
[1821], Iterative Closest Point (ICP) [2224], and Monte Carlo localization (MCL) [2531] can be
adopted as the map matching module. NDT divides the point cloud space into several grids and
calculates their normal distribution, after which it analyzes the optimal solution of the transfer by
calculating the probability distribution function. NDT requires more point cloud data than other
Sensors 2019, 19, 942 3 of 25
methods do because NDT analyzes the transfer relationship through their distribution, and the large
amount of data results in a slow calculation speed. ICP is currently the most commonly used
algorithm in scan matching; it searches for the nearest points between two point clouds and analyzes
their transfer relationship through singular value decomposition. However, ICP is used to find locally
optimal solutions, and both favorable point cloud data and initial values are required to yield a
relatively satisfactory convergence. Three problems with the conventional Monte Carlo localization
(MCL) algorithm [25,26] still need to be addressed. Specifically, when the number of particles is
difficult to determine, or when the particles are assigned excessive weights, the algorithm may
generate only local optimal solutions, and the robot cannot have its location restored after it is
entrapped in such solutions. Improvements on MCL, such as the KullbackLeibler divergence MCL
[2728] and self-adaptive MCL [30,31], have been created to mitigate these problems; however, the
problems surrounding direction estimation remain. Furthermore, MCL involves the use of random
particles and is inapplicable for instant localization. To solve the aforementioned problems,
preprepared maps can be used for instant mapping. In this study, 3D LiDAR was employed for 3D
environmental perception, a map database was preprepared, and an algorithm was designed for
instant and accurate indoor and outdoor localization. Moreover, there exist a number of range sensor
models in the literature that measure cross-correlations [3233] between a measurement and the map.
A common technique is known as map matching. Map matching techniques provide the ability to
transform scans into maps. Once both maps are in the same reference frame, they can be compared
using the map cross-correlation function. Applying the conventional normalized cross-correlation
(NCC) feature registration method requires sufficient numbers of feature points, which must also
exhibit near-normal distribution. However, LiDAR ranging point cloud data scanned and collected
on-site are scarce and do not fulfill near-normal distribution. Consequently, considerable localization
errors occur when NCC features are registered with map features.
Earlier, Chong et al. [34] proposed the synthetic LiDAR, in which synthetic 2D information was
scanned according to 3D features and the problems in localization and mapping were solved through
2D methods. A 3D rolling window was used to reestablish the 3D environmental information and
the surface normal vector was calculated. Subsequently, the 3D point cloud was divided into blocks.
A threshold value was determined to preserve the feature points perpendicular to the ground and
project them to a virtual 2D plane, thereby completing the construction of the synthetic LiDAR model.
The present study employed the synthetic LiDAR as its basic framework, proposed a high-
robustness similarity measurement method measuring cross mean absolute difference (CMAD), and
integrated the CMAD in the drivable moving regions (DMRs) for instant localization. This
localization method detected moving objects more satisfactorily than did the conventional
normalized cross-correlation (NCC) method. Notably, only the 3D LiDAR was employed herein, and
no additional odometer or IMU data were incorporated for support. Therefore, the search method
did not feature a motion model, and the vehicle travel information could not be accurately identified.
Furthermore, when the window-based method was employed for localization, the estimated location
would suddenly shift sideways or backwards; hence, DMRs were added to restrict unreasonable
movement regions, filter inappropriate localization trajectories, and improve the speed and
performance of the localization system. According to the results, although the DMR method did not
markedly differ from the window-based method in its localization accuracy, it required less time to
locate the vehicle than did the window-based method. Additionally, when a particle filter was used
for localization, the particles were not required to be spread across the entire window region; rather,
they were only required to be spread to the farthest possible drivable region, thereby shortening the
convergence time and attaining instant localization.
Sensors 2019, 19, 942 4 of 25
2. Method
2.1.Procedural Structure of the Localization Algorithm
Figure 1 illustrates the structure of the map localization algorithm. A 3D point cloud was
mapped, calibrated, and segmented to obtain the required point cloud information, and the
information was then transformed into a grid map. Finally, a virtual LiDAR scanning method was
employed to extract environmental features and establish a map database.
The aforementioned procedures were also employed for the LiDAR on-site scanning process.
After all the features were extracted, the initial location and direction were estimated through the
mean energy method, and the DMR was incorporated in the feature registration process to identify
the current location and direction of the vehicle.
Figure 1. Flow chart of the localization algorithm.
2.2. Calibration
In the experiment, the position of the LiDAR was not calibrated and thus diverged from the road
surface, leading to the point cloud information pattern depicted in Figure 2. Therefore, four points
were selected from the XZ plane of the point cloud, portrayed in Figure 2a, and two individual
points were selected for the slope calculation in (1). The inverse trigonometric function  was
then employed to calculate the angles and , as shown in (2). Subsequently, the mean between
the two angles was obtained to identify the angle of rotation , as shown in (3), and the entire 3D
point cloud was calibrated for easy map database construction. Figure 2b depicts the divergence
between the LiDAR heading angle and the head direction of the vehicle.




 (1)
  (2)
  
(3)
Map
Calibration
Segmentation
Transform into
Grid Map
Feature
Extraction
Drivable Moving Search Region
and Feature Registration
Find the best position and
orientation of vehicle
Mean Energy and Feature
Registration
Offline
LIDAR
Calibration
Segmentation
Transform into
Grid Map
Feature
Extraction
Online
Sensors 2019, 19, 942 5 of 25
(a)
(b)
Figure 2. Calibration of the 3D point cloud. (a) Divergence in the Light Detection and
Ranging (LiDAR) pitch angle. (b) Divergence between the LiDAR heading angle and the
head direction of the vehicle.
2.3. Segmentation
The calibrated point cloud was segmented to capture the features orthogonal to the road surface.
Figure 3a illustrates the schematic of the original point cloud, in which the green, blue, red, and
purple points represent walls, roads, vehicles, and pipelines, respectively. Only the environmental
outline was needed in the experiment; the vehicle and pipelines were not required. Therefore, the
cloud was segmented in a range larger than the height of the vehicle () but smaller than that
of the pipelines ( ), as outlined in Figure 3b. However, the desired outline may not be
thoroughly presented after segmentation if miscellaneous points, such as the purple points depicted
in Figure 3b, are still present.
The goal of segmentation was to obtain the inner product between the normal vectors of the
segmented point cloud and those of the road point cloud. Therefore, the features perpendicular to
the road must be retained. The inner product was within the range between the threshold values and
represented the desired feature point. Calculating the inner product required segmenting the point
cloud information of the road (Figure 3d), and calculating the normal vectors of the segmented point
cloud and road point cloud. The vectors were calculated by fitting the least squares plane to the local
neighboring points [35].
After the normal vectors of the segmented point cloud and road point cloud were calculated
(Figure 3c,e) and the mean of the normal vectors of the road point cloud was obtained. Subsequently,
the inner product between the normal vectors of the segmented cloud and the aforementioned mean
was calculated. Finally, the threshold values were established for the inner product, and the point
cloud information between −0.005 and 0.005 were retained, as indicated in Figure 3f,g.
Y
X
 
 
 
 
Campus Outline
Calibration
Y
X
Sensors 2019, 19, 942 6 of 25
(b)
(d)
(f)
(g)
Figure 3. Segmentation of the 3D point cloud and its inner product. (a) Original point cloud
information. (b) Point cloud information between  and . (c) Vector information of
the segmented point cloud. (d) Point cloud information of the road. (e) Vector information of the road.
(f) Vector information of the inner product. (g) Point cloud information of the inner product.
2.4. Transforming the Point Cloud into a Grid Map
The 3D point cloud was transformed into a 2D grid map for extracting features and constructing
a map database following its segmentation and calibration. Figure 4a presents the segmented 3D
point cloud information, which was compressed into a grid map through transformation (Figure 4b).
Environmental information about the walls and columns was contained in the grid map, where each
grid was 10 10 cm2. Additionally, the grid map was incorporated to establish a map database as
illustrated in Figure 4c, in which the green area represents the possible location of the vehicle; notably,
this database featured x-axis, y-axis, and features information. The aforementioned procedures were
also employed for the LiDAR on-site scanning process to capture the environmental features on-site,
which were then registered with the feature data in the map database to identify the optimal location
and direction.
Sensors 2019, 19, 942 7 of 25
(a)
(b)
(c)
Figure 4. Transformation of the 3D point cloud into a 2D grid map. (a) Segmented point cloud. (b)
Grid map. (c) Map database.
2.5. Feature Extraction
The environmental features were extracted from the grid map for their registration with the
LiDAR data. Notably, virtual LiDAR scanning, rather than global scanning, was used to extract the
features because locally scanned features are more accurate than those scanned globally. The features
were scanned from within the circle with radius R (30% of the maximum scanning range), as depicted
in Figure 5a; the initial scanning point (i.e., 0°) was on the left of the virtual LiDAR, and the data were
scanned counterclockwise. The features extracted were expressed as distanceangle data (Figure 5b),
in which the x-axis represents the angles and the y-axis represents the distances. Therefore, each cycle
of LiDAR scan yielded 360 degrees of distanceangle data. Assuming that the grid database exhibited
N possible vehicle locations, indicated by the green area in Figure 5a, each cycle of LiDAR scanning
yielded 360 degrees of distanceangle data. Moreover, because each of the N locations exhibited
different environmental features, the database featured a total of N 360 feature data. These data
were then incorporated into the map database containing only location information.
X
Y
X
Y
X
Y
R
Degree
Distance
18 27 36
Sensors 2019, 19, 942 8 of 25
(a)
(b)
Figure 5. Extraction of the environmental features. (a) Virtual LiDAR scan. (b) Environmental features
scanned within the radius R.
2.6. Feature Registration
The map and LiDAR features were registered to obtain the optimal location and heading angle
of the vehicle in the follow-up localization calculation. The currently prevalent template matching
technique involves conducting similarity measurement through the NCC method [3638].
The reference and test signal were expressed as and s, and its angle of movement on the test
signal was displayed as .  represents a test signal with shifting angle. The NCC level of
similarity was expressed as , which was calculated as shown in (4):
 

 (4)
where and  represent the mean of and , respectively; and  represent the
standard deviation (SD) of and , respectively; and N represents the degree of the reference
signal (set as 360). This process was conducted through Fourier transform in the frequency domain
to reduce calculation cost. Specifically, when is maximized, represents the optimal matching
angle between the test and the reference signal. The value of ranges between 1 and 1: the closer
it is to 1, the more similar the test and the reference signal are indicated to be.
Applying the NCC requires a sufficient number of matching points, which must also exhibit
near-normal distribution. Herein, the point cloud information scanned through the LiDAR was
relatively scarce. Consequently, a large localization error would result when the NCC features were
matched with those of the LiDAR and the map. To enhance the NCC algorithm, the CMAD method
was applied to compare the similarities between the map and LiDAR features. The parameters
involved in this process are listed as follows
    
      
   
where  represents the on-site LiDAR features (Figure 6b),  represents the virtually
scanned LiDAR features (Figure 6a), represents the set of the obstacle points scanned in both
 and ,  represents the set of the location points in the DMR, and V represents the
optimal location and heading angle selected through the poll mechanism.
As Figure 6d reveals, the template () was overlapped with the fixed signal () for feature
registration. The angle of the template movement on the fixed signal is expressed as , and the
similarity between the template and the signal is represented with , which was calculated as
follows
    
 (5)
where M represents the degree of W and ranges from to . When was minimal, the
LiDAR and map features were the most closely matched (Figure 6c). Because (5) only involved the
feature registration of one point, but the DMR featured a total of L points, the minimal of the L
points was used to determine the optimal location, , as shown in (6):
    (6)
where the corresponding represents the optimal heading angle .
Sensors 2019, 19, 942 9 of 25
(a)
(b)
(c)
(d)
Figure 6. Registration of the map and LiDAR features. (a) Map features. (b) LiDAR
on-site features. (c) Feature registration results. (d) Feature registration process.
Degree
Distance
Degree
Distance
Degree
 
 




  is minimum
Degree
Distance
Shift
:
:
:
Sensors 2019, 19, 942 10 of 25
2.7. Estimating the Initial Location
The feature extraction was performed locally. Therefore, some angles did not show any distance
value because no obstacles were scanned in that angle or that the distance values exceeded the
scanning radius R. Next, both the LiDAR and map features were calculated through the mean energy
method, which involves adding the distance values of all the scanned energy points together and
dividing the sum by the number of energy points. Equation (7) measures the energy of the map,
where N represents the number of energy points scanned,  represents the distance value of the
ith degree, (8) measures the energy of the LiDAR, where M represents the number of energy points
scanned, and  represents the distance value of the jth degree. Thus, the energy data of the map
were established and implemented in the map database.
 


(7)
 

(8)
The mean energy method was also applied to globally search the initial location and heading
angle of the vehicle. The possible locations of the vehicle were selected through the aforementioned
average energy values, where  is the energy of the LiDAR. A search range was set up to filter
the energy of the map close to  as follows
         (9)
where t represents the tolerable error, which was smaller indoors and larger outdoors. Because the
point cloud of the map was more comprehensive than the on-site point cloud of the LiDAR (which
was relatively scarce), most of the energy values of the map were higher than those of the LiDAR.
Theoretically the initial location can be obtained by globally search the best registration of the map
and LiDAR features. To reduce the computation time, the mean energy method was also applied to
globally search the coarse initial location and heading angle of the vehicle. The possible locations of
the vehicle were selected through the aforementioned average energy values. The selection results
obtained using the mean energy method are indicated in red in Figure 7a. To acquire the optimal
initial location and heading angle, the features from these selected positions were registered by
CMAD with the LiDAR features that were scanned at the necessary instant (Figure 7b).
(a)
(b)
Figure 7. Identification of the initial location and direction of the vehicle through the mean energy
method. (a) Possible vehicle locations identified through the mean energy method. (b) Initial location
and direction identified through the feature registration.
2.8. Window-Based Localization
After the initial location of the vehicle was identified through the mean energy method, its
location in the next time point was determined through window-based localization. Unlike the mean
energy method, which was applied for a global search, window-based localization involved locally
searching for the possible locations of vehicle through the window search method, and required
X
Y
X
Y
Sensors 2019, 19, 942 11 of 25
considerably less calculation time. First, the initial location as calculated through the mean energy
method was set as the center. A w h window was created, depicted as the purple rectangular box in
Figure 8a. Second, the possible locations of the vehicle at the current time point within the window
were determined according to the coordinates of the initial location, and are displayed as the blue
points in Figure 8b. Finally, the features of these locations were registered with the concurrent LiDAR
features to obtain the current location of the vehicle, shown as the orange point in Figure 8c.
(a)
(b)
(c)
Figure 8. Locating the vehicle through window-based localization. (a) Window search
area. (b) Current possible locations of the vehicle. (c) Final estimated location.
2.9. DMR-Based Localization
Because the location of the vehicle was searched using only 3D LiDAR without additional
odometer or IMU data, and because the system did not include any motion model, the vehicle
movement information could not be identified. Therefore, DMR-based localization was also
conducted to calculate the trajectories that approximated to the actual vehicle movement. Although
the motion trajectories that were estimated through the window-based localization could result in
sideways or backwards deviations from the normal movement status, indicated by the red circles in
Figure 9a,b, the DMR was designed to resolve this problem and reduce the number of estimated
vehicle locations.
X
Y
X
Y
X
Y
Sensors 2019, 19, 942 12 of 25
(a)
(b)
Figure 9. Deviations in window-based localization. (a) Sideways deviation in the window-
based localization. (b) Backwards deviation in the window-based localization.
Similar to the window-based localization, the DMR-based localization involved estimating the
initial location and direction of the vehicle through the mean energy method. Using this initial
location as the center, the DMR with the maximal movement distance radius R and the rotation radius
r was established (represented by the purple circle in Figure 11a). The entire DMR is illustrated in
Figure 10, where represents the previously estimated location and the initial point. The distance
between (a nearby location) and should be no longer than R, and the distances between
and   and between and   must be longer than or equal to r. The that satisfied
both these conditions was a point situated within the DMR, as determined by (10). Because the
coordinate of the initial location was known, the other locations within the DMR could then be
calculated; the possible current location of the vehicle is denoted by the blue points in Figure 11b.
Unlike the window-based estimated locations, the DMR rotated according to the changes in the
heading angle of the vehicle. Therefore, the heading angle from the previous time point was used as
the angle of rotation, and the entire search area was rotated using the z-axis. Finally, the features of
the locations in the area were registered with the concurrent LiDAR features to isolate the current
location of the vehicle on the map, which is identified by the orange point in Figure 11c.
            (10)
Figure 10. Drivable moving region (DMR).
X
Y
X
Y
R
r
r
Sensors 2019, 19, 942 13 of 25
(a)
(b)
(c)
Figure 11. Locating the vehicle through DMR-based localization. (a) Search range of
the DMR. (b) Possible current locations. (c) Estimated location and direction.
3. Experiment
3.1. Equipment
Velodyne LiDAR (VLP-16) was employed to scan the surrounding environment using a laser, and
establish a 3D point cloud. Notably, if the LiDAR sensor had been set up directly on the car, then it
would have scanned the car itself because of the insufficient height. Therefore, the aluminum extrusion
support frame attached with a cloud deck was heightened to thoroughly construct the 3D point cloud
on the outlines and obstacles of the environment. Figure 12a depicts the actual installment of the sensor
on the car, and Figure 12b illustrates the point cloud scanned on-site using the LiDAR.
(a)
(b)
Figure 12. The aluminum extrusion support frame and a cloud deck of the LiDAR system.
(a) Actual setup of the LiDAR. (b) On-site LiDAR point cloud.
X
Y
X
Y
X
Y
Sensors 2019, 19, 942 14 of 25
3.2. Indoor Localization Experiment
The indoor environment selected for this study was the parking lot in Basement 5 of National
Taipei University of Technology (NTUT). Figure 13 depicts the outline of the entire parking lot, and
the motion trajectory of the car is circled in red.
Figure 13. Indoor experiment environment.
3.2.1. Window-Based Localization
The motion trajectory detected through the window-based localization was smoothed using the
EKF and compared with the trajectory identified through the SLAM method. Figure 14a depicts the
trajectories identified through window-based localization. However, because the Window trajectory
was considerably similar to the Kalman trajectory, the reversing sections of the trajectories were
locally magnified (Figure 14b, where the SLAM trajectory is indicated by the circles and the Window
trajectory is indicated by the crosses). Notably, when the car was reversing, sideways deviations
occurred in the window-based localization. Figure 14c illustrates the lateral errors between the two
calculated trajectories. Beginning at the 368th frame (i.e., when the car began reversing), the error
substantially widened to a maximum of 2.07 grids (20.7 cm). Meanwhile, the maximal longitudinal
error was approximately 1.36 grids (13.6 cm).
(a)
(b)
Start point
Shape Representation Color
Window Trajectory
SLAM Trajectory
Parking position
Pillar
Turning point
Reversing direction
Sensors 2019, 19, 942 15 of 25
(c)
Figure 14. Comparison of the Window and SLAM trajectories. (a) Window
trajectory. (b) Local magnification (when the car was reversing). (c) Lateral and
longitudinal error comparison between the Window and simultaneous localization
and mapping (SLAM) trajectories.
3.2.2. DMR-Based Localization
The motion trajectory detected through the DMR-based localization was smoothed using the
EKF and compared with the SLAM trajectory. The DMRNCC trajectory was also compared with the
SLAM trajectory. Figure 15a depicts the DMRCMAD trajectory. However, because the three
aforementioned trajectories were considerably similar, the reversing sections of the trajectories were
locally magnified (Figure 15b, where the DMRNCC trajectory is denoted by the dots, the SLAM
trajectory is denoted by the circles, and the DMRCMAD trajectory is denoted by the crosses).
Notably, when the car was reversing, the DMRCMAD trajectory was more stable than the Window
trajectory, but the DMRNCC trajectory exhibited considerable sideways errors. Figure 15c depicts
the lateral errors between the three calculated trajectories. Beginning at the 372th frame (i.e., when
the car began reversing), the errors substantially increased to a maximum of 2.63 grids (26.3 cm). The
maximal longitudinal error was approximately 1.58 grids (15.8 cm). According to the error
comparison between the DMRNCC trajectory and the SLAM trajectory, changes in the errors were
considerable during both periods when the car moved forward and when it reversed. The maximal
lateral and longitudinal errors were approximately 3.05 grids (30.5 cm) and 2.9 grids (29 cm),
respectively. The performance of the proposed DMRCMAD algorithm is verified via vehicle tests
on a parking-lot proving ground. The proposed algorithm will be useful in the implementation of
autonomous self-parking control.
Sensors 2019, 19, 942 16 of 25
(a)
(b)
(c)
Figure 15. Comparison of the DMRCMAD and DMRNCC trajectories and the
SLAM trajectory. (a) DMRcross mean absolute difference (CMAD) trajectory. (b)
Local magnification (when the car was reversing). (c) Lateral and longitudinal
errors between the DMRCMAD and DMRnormalized cross-correlation (NCC)
trajectories and the SLAM trajectory.
3.3. Outdoor Localization Experiment
The outdoor environment for this study was the NTUT campus. Figure 16 presents a 2D outline
of the campus, including the motion trajectory of the car. The car started from the front of Chiang
Kai-Shek Memorial Building, made a U-turn around the spherical landmark in front of the Sixth
Academic Building, traveled past the Second Academic Building through the original route, and
stopped back in front of Chiang Kai-Shek Memorial Building.
Start point
Shape Representation Color
DMR-CMAD Trajectory
DMR-NCC Trajectory
SLAM Trajectory
Map
Pillar
Turning point
Parking position
Shape Representation Color
DMR-CMAD vs. SLAM
DMR-NCC vs. SLAM
Shape Representation Color
DMR-CMAD vs. SLAM
DMR-NCC vs. SLAM
Sensors 2019, 19, 942 17 of 25
Figure 16. Outdoor experiment environment. 4. Experimental Results.
3.3.1. Window-Based Localization
The motion trajectory detected through the window-based localization was smoothed using the
EKF and compared with the trajectory identified through the SLAM method. Figure 17a depicts the
Window trajectory. However, because the Window trajectory was very similar to the Kalman
trajectory, three sections of the trajectories were locally magnified: Figure 17b displays the start and
end sections; Figure 17c depicts the middle section; and Figure 17d illustrates the U-turn around the
landmark. Specifically, the SLAM trajectory is denoted by circles and the Window trajectory is
denoted by crosses. Notably, when the car was reversing, sideways deviations occurred in the
window-based localization. As indicated by Figure 17e, the lateral errors from the 368th frame to the
409th frame (i.e., when the car made the U-turn) were the most substantial and the maximal lateral
error was approximately 4.71 grids (47.1 cm). By contrast, the longitudinal errors were notable before
the U-turn and decreased after the U-turn; the maximal longitudinal error was 2.99 grids (29.9 cm).
Chiang Kai-Shek
Memorial Hall
Second Academic Building
Sixth Academic Building
Sensors 2019, 19, 942 18 of 25
(a)
(b)
(c)
(d)
Shape Representation Color
Window Trajectory
SLAM Trajectory
Start point
End point
Shape
Representation Color
Window Trajectory
SLAM Trajectory
Landmark
Shape
Representation Color
Window Trajectory
SLAM Trajectory
Sensors 2019, 19, 942 19 of 25
(e)
Figure 17. Comparison of the Window and SLAM trajectories. (a) Window
trajectory. (b) Local magnification of the trajectory (start and end section). (c) Local
magnification (middle section). (d) Local magnification (U-turn around the
landmark). (e) Lateral and longitudinal errors between the Window and SLAM
trajectories.
3.3.2. DMR-Based Localization
The motion trajectory detected through the DMR-based localization was smoothed using the
EKF and compared with the trajectory identified through the SLAM method. The DMRNCC
trajectory was also compared with the SLAM trajectory. Figure 18a depicts the DMRCMAD
trajectory. However, because all three trajectories were very similar, three sections of the trajectories
were locally magnified: Figure 18b displays the start and end sections, Figure 18c depicts the middle
section, and Figure 18d illustrates the U-turn around the landmark. Specifically, the DMRNCC
trajectory is denoted by dots, the SLAM trajectory is denoted by circles, and the DMRCMAD
trajectory is denoted by crosses. Notably, when the car was reversing, sideways deviations occurred
in the window-based localization. As indicated by Figure 18e, the lateral error at the 404th frame (i.e.,
when the car was making the U-turn) was the most substantial and the maximal lateral error was
approximately 6.31 grids (63.1 cm). By contrast, the longitudinal errors were notable before the U-
turn and decreased after the U-turn; the maximal longitudinal error, which was near the end point,
was 3.34 grids (33.4 cm). Meanwhile, the variance in the localization errors of the DMRNCC
trajectory was substantial before the car turned, when the car turned, and when the car was
approaching the end point. The maximal lateral error was approximately 19.04 grids (190.4 cm); the
maximal longitudinal error was 26.9 grids (269 cm). The performance of the proposed DMRCMAD
algorithm is verified via vehicle tests on an outdoor proving ground. The proposed algorithm will be
useful in the implementation of lane-level automated driving control.
Sensors 2019, 19, 942 20 of 25
(a)
(b)
(c)
(d)
Shape Representation Color
DMR-CMAD Trajectory
DMR-NCC Trajectory
SLAM Trajectory
Map
End point
Start point
Landmark
Shape Representation Color
DMR-CMAD Trajectory
DMR-NCC Trajectory
SLAM Trajectory
Map
Shape Representation Color
DMR-CMAD Trajectory
DMR-NCC Trajectory
SLAM Trajectory
Map
Sensors 2019, 19, 942 21 of 25
(e)
Figure 18. Comparison of the DMRCMAD and DMRNCC trajectories and the SLAM
trajectory. (a) DMRCMAD trajectory. (b) Local magnification (start and end section). (c)
Local magnification (middle section). (d) Local magnification (U-turn around the
landmark). (e) Lateral and longitudinal errors between the DMRCMAD and DMRNCC
trajectories and the SLAM trajectory.
3.4. Comparison of the Indoor and Outdoor Errors
The comparison of the root mean square errors (RMSEs) and SDs of the indoor and outdoor
errors are outlined in Table 1. The accuracy of the DMRCMAD trajectories approximated that of the
Window trajectories, but the errors of the DMRNCC trajectories, particularly those of the outdoor
trajectory, were more substantial.
Table 1. Lateral and longitudinal root mean square errors (RMSEs) and standard deviations (SDs) of
the Window, DMRCMAD, and DMRNCC trajectories.
Localization
NTUT B5 Parking Lot
Method
(1 grid = 10 cm)
Window
DMR-CMAD
DMRNCC
Lateral RMSE
0.67 grid
0.61 grid
0.79 grid
Standard Deviation
0.36 grid
0.42 grid
0.47 grid
Longitudinal RMSE
0.41 grid
0.48 grid
0.78 grid
Standard Deviation
0.26 grid
0.32 grid
0.55 grid
Localization
NTUT Campus
Method
(1 grid = 10 cm)
Window
DMR-CMAD
DMRNCC
Lateral RMSE
2.16 grid
2.27 grid
3.91 grid
Standard Deviation
1.17 grid
1.33 grid
2.83 grid
Longitudinal RMSE
1.25 grid
1.31 grid
5.25 grid
Standard Deviation
0.75 grid
0.78 grid
4.63 grid
Shape Representation Color
DMR-CMAD vs. SLAM
DMR-NCC vs. SLAM
Shape Representation Color
DMR-CMAD vs. SLAM
DMR-NCC vs. SLAM
Sensors 2019, 19, 942 22 of 25
In the experiment, Velodyne LiDAR (VLP-16) was employed to scan the surrounding
environment using a laser and establish a 3D point cloud. The computer platform that was used for
vehicle localization was equipped with an Intel i5 CPU, 8 GB DDRIII. Table 2 presents a list of the
average localization time of all four methods examined in the indoor (totally 734 frames) and outdoor
(totally 1274 frames) experiments. The localization time of the DMRCMAD method was the shortest,
the localization time of the DMRNCC method was two to three times that of the Window and DMR
CMAD methods, and the localization time of the SLAM method was the longest. Notably, the
localization times also varied between the indoor and outdoor environments, in part because the area
of the indoor site was smaller and the speed the car was slower compared with the outdoor site.
Table 2. Average localization times of the Window, DMRCMAD, DMRNCC, and SLAM
methods.
Location
NTUT B5 Parking Lot (734 Frames)
Method
Window
DMRCMAD
DMRNCC
SLAM
Time/Frame (s)
0.23
0.2
0.76
4.26
Location
NTUT Campus (1274 frames)
Method
Window
DMRCMAD
DMRNCC
SLAM
Time/Frame (s)
0.7
0.48
1.03
3.45
4. Conclusions
This study employed the map-based DMR localization method to improve upon the sudden
sideways and backwards deviations in the window localization algorithm. The DMRCMAD and
window methods did not differ substantially in their localization errors according to the RMSE and
SD comparison results. However, the DMRNCC method exhibited more errors and required more
localization time than did the DMRCMAD and window methods; the DMRCMAD method was
the least time-consuming of the four employed methods. Because the features of the outdoor
experiment environment were more complicated than those of the indoor environment, the feature
registration accuracy was slightly lower in the outdoor experiment than in the indoor environment,
and the outdoor localization errors were markedly larger than the indoor localization errors.
However, the localization accuracy of the DMRCMAD method was overall ideal, and the method
was confirmed as applicable for instant localization.
This study also incorporated the LiDAR to capture the environmental features on-site, which
were then registered with the feature data in the map database to identify the optimal location and
direction. However, if the features of future experimental sites differ from those of the map database,
inaccurate localization may result. Furthermore, because Velodyne VLP-16 LiDAR was used to
extract the on-site environmental characteristics, its number of point clouds was less than those of
other 3D LiDAR systems. The point cloud densities in outdoor environments were relatively sparse,
and the great variation of the point clouds between each frame resulted in substantial matching
iteration error. Therefore, other types of sensorsGPS and IMUcan be used to reduce the problem
of matching iteration error.
Author Contributions: Chih-Ming Hsu contributed the ideas of the research and research supervision; Chung-
Wei Shiu prepared and performed the experiments; Chih-Ming Hsu contributed to writing and revising the
manuscript.
Funding: This work was supported in part by the Ministry of Science and Technology, Taiwan, under the grants:
Institute of Information Industry, Taiwan and MOST 106-2221-E-027-064, MOST 107-2221-E-027-122 and
MOST 107-2622-E-027-009 -CC3 Taiwan.
Conflicts of Interest: The authors declare no conflict of interest.
Sensors 2019, 19, 942 23 of 25
References
1. Brummelen, J.V.; O’Brien, M. Autonomous vehicle perception: The technology of today and
tomorrow. Transport. Res. C 2018, 89, 384406.
2. Kamijo, S.; Gu, Y.; Hsu, L.-T. Autonomous Vehicle Technologies: Localization and Mapping,
IEICE ESS Fundam. Rev. 2015, 9, 131141.
3. Seo, Y.W.; Rajkumar, R. A vision system for detecting and trajectorying of stop-lines. In
Proceedings of IEEE 17th International Conference on Intelligent Transportation Systems,
Qingdao, China, 811 October 2014; pp. 19701975.
4. Oniga, F.; Nedevschi, S.; Meinecke, M.M. Curb detection based on elevation maps from
dense stereo. In Proceedings of IEEE International Conference on Intelligent Computer
Communication and Processing, Cluj-Napoca, Romania, 68 September 2007; pp. 119125.
5. Enzweiler, M.; Greiner, P.; Knoppel, C.; Franke, U. Towards multicue urban curb
recognition.In Proceedings of IEEE Intelligent Vehicles Symposium 2013, Gold Coast,
Australia, 2326 June 2013; pp. 902907.
6. Nedevschi, S.; Poppescu, V.; Danescu, R.; Marita, T.; Oniga, F. Accurate ego-vehicle global
localization at intersections through alignment of visual data with digital map. IEEE Trans.
Intell. Transport. Syst. 2013, 14, 673687.
7. Wu, T.; Ranganathan, A. Vehicle localization using road markings, In Proceedings of IEEE
Intelligent Vehicles Symposium 2013, Gold Coast, Australia, 2326 June 2013; pp. 11851190.
8. Gu, Y.; Tehrani, M.P.; Yendo, T.; Fujii, T.; Tanimoto, M. Traffic Sign Recognition with
invariance to lighting in dual-focal active camera system. IEICE Trans. Info. Syst. 2012, 95,
17751790.
9. Hata, A.Y.; Osorio, F.S.; Wolf, D.F. Robust curb detection and vehicle localization in urban
environments, In Proceedings of IEEE Intelligent Vehicles Symposium 2014, Michigan, USA,
811 June 2014; pp. 12571262.
10. Hata, A.; Wolf, D. Road marking detection using LIDAR reflective intensity data and its
application to vehicle localization. In Proceedings of IEEE 17th International Conference on
Intelligent Transportation Systems, Qingdao, China, 811 October 2014; pp. 584589.
11. Choi, J. Hybrid map-based SLAM using a Velodyne laser scanner, In Proceedings of IEEE
17th International Conference on Intelligent Transportation Systems, Qingdao, China, 811
October 2014; pp. 30823087.
12. Thrun, S.; Montemerlo, M. The graph SLAM algorithm with applications to large-scale
mapping of urban structures. Int. J. Robot. Res. 2006, 25, 403429.
13. Levinson, J.; Montemerlo, M.; Thrun, S. Map-based precision vehicle localization in urban
environments, In Proceedings of Robotics: Science and Systems, Georgia, USA, 2730 June
2007.
14. Smith, R.; Cheesman, P. On the representation of spatial uncertainty. Int. J. Robot. Res. 1987,
5, 5668.
15. Smith, R.; Self, M.; Cheeseman, P. Estimating uncertain spatial relationships in robotics. In
Autonomous Robot Vehicles; Springer: New York, NY, USA, 1990; pp. 167193.
16. Montemerlo, M.; Thrun, S.; Koller, D.; Wegbreit, B. Fast-SLAM: A factored solution to the
simultaneous localization and mapping problem. In Proceedings of Innovative Applications
of Artificial Intelligence Conference 2002, Alberta, Canada, 28 July 1 August 2002; pp. 593
598.
17. Montemerlo, M.; Thrun, S.; Koller, D.; Wegbreit, B. Fast-SLAM 2.0: An improved particle
filtering algorithm for simultaneous localization and mapping that provably converges. In
Proceedings of the International Joint Conference on Artificial Intelligence 2003, Acapulco,
Mexico, 915 August 2003; pp. 11511156.
18. Biber, P.; Straßer, W. The normal distributions transform: A new approach to laser scan
matching. In International Conference on Intelligent Robots and Systems, Las Vegas, USA,
2731 October 2003; pages 2743 2748, 2003.
Sensors 2019, 19, 942 24 of 25
19. Magnusson, M.; Andreasson, H.; Nüchter, A.; Lilienthal, A.J. Automatic appearance-based
loop detection from 3D laser data using the normal distributions transform. J. Field Robot.
2009, 26, 892914.
20. Stoyanov, T.; Magnusson, M.; Andreasson, H.; Lilienthal, A.J. Fast and accurate scan
registration through minimization of the distance between compact 3d NDT
representations. Int. J. Robot. Res. 2012, 31, 13771393.
21. Akai, N. Morales, L. Y.; Takeuchi, E.; Yoshihara, Y. Robust localization using 3D NDT scan
matching with experimentally determined uncertainty and road marker matching. In
Proceedings of IEEE Intelligent Vehicles Symposium, Redondo Beach, California, USA, 11
14 June 2017; pp. 13571364.
22. Besl, P.J.; McKay, H.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal.
Mach. Intell. 1992, 14, 239256.
23. Mendes, E.; Koch, P.; Lacroix, S. ICP-based pose-graph SLAM. In IEEE International
Symposium on Safety, Security, and Rescue Robotics (SSRR), Lausanne, Switzerland, 2327
October 2016; pp. 195200.
24. Cho, H.; Kim, E.K.; Kim, S. Indoor SLAM application using geometric and ICP matching
methods based on line features. Robot. Auton. Syst. 2018, 100, 206224.
25. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; The MIT Press: Boston, Massachusetts,
USA, 2005.
26. Hsu, C.C.; Wong, C.C.; Teng, H.C.; Ho, C.Y. Localization of Mobile Robots via an Enhanced
Particle Filter Incorporating Tournament Selection and Nelder-Mead Simplex Search. Int. J.
Innov. Comput. Inf. Control 2011, 7, 37253737.
27. Fox, D. Adapting the sample size in particle filters through KLD-sampling. Int. J. Robot. Res.
2003, 22, 9851003.
28. Kwok, C.; Fox, D.; Meila, M. Adaptive real-time particle filters for robot localization. In
Proceedings of IEEE International Conference on Robotics and Automation, Taipei, Taiwan,
1419 September 2003; Volume 2, pp. 28362841.
29. Li, T.; Sun, S.; Sattar, T.P. Adapting sample size in particle filters through KLD-resampling.
Electron. Lett. 2013, 49, 740742.
30. Zhang, L.; Zapata, R.; Lepinay, P. Self-adaptive Monte Carlo Localization for Mobile Robots
Using Range Sensors. In Proceedings of 2009 IEEE/RSJ International Conference on
Intelligent Robots and Systems, St. Louis, USA, 1115 October 2009, pp.15411546.
31. Zhang, L.; Zapata, R.; Lepinay, P. Self-Adaptive Monte Carlo for Single-Robot and Multi-
Robot Localization. In Proceedings of IEEE International Conference on Automation and
Logistics (ICAL'09), Shenyang, China, 57 August 2009; pp. 19271933.
32. Briechle, K.; Hanebeck, U.D. Self-localization of a mobile robot using fast normalized cross
correlation. In Proceedings of 1999 IEEE Int. Conf. Systems, Man, Cybernetics, Tokyo, Japan,
1215 October 1999; p. 362.
33. Choi, M.; Choi, J.; Chung, W.K. Correlation-based scan matching using ultrasonic sensors
for EKF localization. Adv. Robot. J. 2012, 26, 14951519.
34. Chong, Z.J.; Qin, B.; Bandyopadhyay, T.; Ang Jr., M.H.; Frazzoli, E.; Rus, D. Synthetic 2D
LIDAR for Precise Vehicle Localization in 3D Urban Environment. In Proceedings of IEEE
International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 610
May 2013; .pp. 15541559.
35. Shakarji, C. Least-squares fitting algorithms of the NIST algorithm testing system. J. Res.
Natl. Inst. Stand. Technol. 1998, 103, 63364.
36. Guyon, I.; Gunn, S.; Nikravesh, M.; Zadeh, L. Feature Extraction, Foundations and
Applications. Ser. Stud. Fuzz. Soft Comput. 2006, 207, doi:10.1007/978-3-540-35488-8.
37. Balaguer, B.; Balakirsky, S.; Carpin, S.; Visser, A. Evaluating maps produced by urban search
and rescue robots: Lessons learned from robocup. Auton. Robot. 2009, 27, 449464.
38. Fernández-Madrigal, J.-A.; Claraco, J.L.B. Simultaneous Localization and Mapping for Mobile
Robots: Introduction and Methods; IGI Global: Hershey, PA, USA, 2013.
Sensors 2019, 19, 942 25 of 25
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open
access article distributed under the terms and conditions of the Creative
Commons Attribution (CC BY) license
(http://creativecommons.org/licenses/by/4.0/).
... However, the TPSM is not suitable for all sensors, limiting its scalability. To address the issue that the traditional normalized cross-correlation (NCC) algorithm requires sufficient feature points, a cross means absolute difference (CMAD) algorithm based on known map information using 3D LiDAR is proposed in [47]. This method includes offline and online parts. ...
Article
Full-text available
Vehicle localization plays a crucial role in ensuring the safe operation of autonomous vehicles and the development of intelligent transportation systems (ITS). However, there is insufficient effort to compare the performance and challenges of different vehicle localization algorithms. This paper aims to address this gap by analyzing the comprehensive performance of existing advanced vehicle localization techniques and discussing their challenges. Firstly, we analyze the self-localization methods based on active and passive sensors. The results show that, the light detection and ranging (LiDAR) and vision-based techniques can reach high accuracy. However, they have high computational complexity. And only using the inertial measurement unit (IMU), global positioning system (GPS), radar, and ultrasonic sensors may not realize localization result with high accuracy. Then, we discuss V2X-based cooperative methods and analyze the multi-sensor based localization techniques and compare the comprehensive performance among all methods. Although the artificial intelligence (AI) techniques can effectively enhance the efficiency of vision-based localization algorithms, the high computational complexity still should be considered. In addition, since the IMU, GPS, radar, and ultrasonic sensors have good performance in terms of the availability, scalability, computational complexity, and cost-effectiveness, they can be used as auxiliary sensors to achieve good comprehensive performance through data fusion techniques. Finally, we propose the challenges of different techniques and look forward to future work.
... detect the edge of sidewalks as landmarks and models them using straight lines. In the localization method proposed byHsu and Shiu (2019), a 2D grid map is created by segmenting surfaces that are orthogonal to the the road plane. They divide the 2D ground plane into 10 cm x 10 cm cells and determine the occupancy of the each cell based on the number of 3D points available in the vertical column corresponding to the cell. ...
Thesis
Full-text available
Intelligent vehicles are a key component in humanity’s vision for safer, efficient, and accessible transportation systems across the world. Due to the multitude of data sources and processes associated with Intelligent vehicles, the reliability of the total system is greatly dependent on the possibility of errors or poor performances observed in its components. In our work, we focus on the critical task of localization of intelligent vehicles and address the challenges in monitoring the integrity of data sources used in localization. The primary contribution of our research is the proposition of a novel protocol for integrity by combining integrity concepts from information systems with the existing integrity concepts in the field of Intelligent Transport Systems (ITS). An integrity monitoring framework based on the theorized integrity protocol that can handle multimodal localization problems is formalized. As the first step, a proof of concept for this framework is developed based on cross-consistency estimation of data sources using polynomial models. Based on the observations from the first step, a 'Feature Grid' data representation is proposed in the second step and a generalized prototype for the framework is implemented. The framework is tested in highways as well as complex urban scenarios to demonstrate that the proposed framework is capable of providing continuous integrity estimates of multimodal data sources used in intelligent vehicle localization.
... A detailed visual representation was proven to be requisite to achieve a connection to reality to increase comprehension. Light detection and ranging (LiDAR) has been applied to 3D scan technology to document as-built scenes for construction monitoring [43][44][45] or to retrieve geometric features from a distance with a combination of aerial images [46][47][48][49][50][51]. In contrast to creating 3D models from the beginning, 3D laser scanning is an effective tool for acquiring important geometric attributes, such as detailed trees and vegetation [52]. ...
Article
Full-text available
Zoning conflicts have transformed Old Street fabrics in terms of architectural style and construction material in Lukang, Taiwan. This transformation should be assessed as a contribution to digital cultural sustainability. The objective of this study was to compare the evolved façades resultant from the changes made by the development of architectural history and urban planning. A combination of 3D scan technology and a smartphone augmented reality (AR) app, Augment®, was applied to the situated comparison with direct interaction on-site. The AR application compared 20 façades in the laboratory and 18 façades in four different sites using a flexible interface. The comparisons identified the correlation of evolved façades in real sites in terms of building volumes and components, pedestrian arcades on store fronts, and previous installations. The situated comparisons were facilitated in a field study with real-time adjustments to 3D models and analyses of correlations across details and components. The application of AR was demonstrated to be effective in reinstalling scenes and differentiating diversified compositions of vocabulary in a remote site.
... However, the cumulative positional errors can quickly spiral out of control due to the long-term absence of global localization information, especially when compromised by inertial sensors with sub-par performance [13]. This shortcoming of DR is addressed by the map-matching positioning process [14], in which LiDAR is primarily used to collect maps beforehand to generate the raw point cloud maps (as is shown in Figure 1a) with dense features [14,15]; then the point cloud of the vehicular LiDAR is registered by the ICP [16] or NDT [17] method to calculate the accurate vehicle position and rotation with centimeter-level precision. While both families of localization methods-RTK GNSS and DR with LiDAR-offer accurate vehicle localization, both require high-cost in-vehicle sensors. ...
Article
Full-text available
Real-time vehicle localization (i.e., position and orientation estimation in the world coordinate system) with high accuracy is the fundamental function of an intelligent vehicle (IV) system. In the process of commercialization of IVs, many car manufacturers attempt to avoid high-cost sensor systems (e.g., RTK GNSS and LiDAR) in favor of low-cost optical sensors such as cameras. The same cost-saving strategy also gives rise to an increasing number of vehicles equipped with High Definition (HD) maps. Rooted upon these existing technologies, this article presents the concept of Monocular Localization with Vector HD Map (MLVHM), a novel camera-based map-matching method that efficiently aligns semantic-level geometric features in-camera acquired frames against the vector HD map in order to achieve high-precision vehicle absolute localization with minimal cost. The semantic features are delicately chosen for the ease of map vector alignment as well as for the resiliency against occlusion and fluctuation in illumination. The effective data association method in MLVHM serves as the basis for the camera position estimation by minimizing feature re-projection errors, and the frame-to-frame motion fusion is further introduced for reliable localization results. Experiments have shown that MLVHM can achieve high-precision vehicle localization with an RMSE of 24 cm with no cumulative error. In addition, we use low-cost on-board sensors and light-weight HD maps to achieve or even exceed the accuracy of existing map-matching algorithms.
... Although several successful studies have demonstrated the simultaneous localisation and mapping of an agent in a static environment, and some studies even implemented obstacle avoidance, e.g. (Hsu and Shiu, 2019;Zhang and Singh, 2017;Mousazadeh et al., 2018;Costa et al., 2016), it remains to be investigated which sensor set, and their minimum required specifications, is deemed necessary to achieve this goal on the inland waterways for cargo vessels. ...
Article
Full-text available
Although road-based freight transport has large external costs, it currently dominates the hinterland cargo transport sector in Europe. An increase of the automation levels of inland cargo vessels could advance their competitiveness, hence unlocking more sustainable inland cargo transport. Moreover, these improvements could pave the way for a potential future paradigm shift towards unmanned inland cargo vessels. Therefore, this study investigates the design of an experimental platform in order to study the feasibility and current or future limitations of unmanned inland cargo vessels. To explore this design, the following three questions were handled: (i) How to design an industrially relevant research vessel?, (ii) How can an unmanned inland cargo vessel interact with or perceive its environment?, and (iii) How to control an unmanned or autonomous inland cargo vessel? The answers to these design questions delivered the blueprint to construct a first prototype of an unmanned inland cargo vessel. This prototype performed a set of outdoor experiments to verify and validate the made design choices. The successful execution of these experiments demonstrates that this vessel can serve as an experimental platform for further technological research regarding more automated or unmanned inland cargo vessels and provide fruitful insights for other research domains.
... Data acquisition is faster and more comfortable and results into a broader number of possible applications. Robotics seems to be the main beneficent of the rapid development of 3D LiDARs [28][29][30] and this is in accordance with initial observations on the significant role that LiDAR-based systems are going to play for an efficient exploration of the surrounding environment, to allow human functionalities to be substituted by automatic or semiautomatic systems. Other examples of important applications for 3D LiDAR-based systems can be found in [13], [31][32][33], where 3D LiDAR data are used in combinations with cameras and radars for the indoor exploration of inside tunnels, underground mines, and caves. ...
Article
Full-text available
The aim of this paper is to highlight how the employment of Light Detection and Ranging (LiDAR) technique can enhance greatly the performance and reliability of many monitoring systems applied to the Earth Observation (EO) and Environmental Monitoring. A short presentation of LiDAR systems, underlying their peculiarities, is first given. References to some review papers are highlighted, as they can be regarded as useful guidelines for researchers interested in using LiDARs. Two case studies are then presented and discussed, based on the use of 2D and 3D LiDAR data. Some considerations are done on the performance achieved through the use of LiDAR data combined with data from other sources. The case studies show how the LiDAR-based systems, combined with optical Very High Resolution (VHR) data, succeed in improving the analysis and monitoring of specific areas of interest, specifically how LiDAR data help in exploring external environment and extracting building features from urban areas. Moreover the discussed Case Studies demonstrate that the use of the LiDAR data, even with a low density of points, allows the development of an automatic procedure for accurate building features extraction, through object-oriented classification techniques, therefore by underlying the importance that even simple LiDAR-based systems play in EO and Environmental Monitoring.
... However, the accuracy and availability of the GPS signal cannot always meet the requirements of autonomous vehicle localization, due to satellite visibility interruption and signal multipath. To address this issue, many vehicle localization systems based on robust location algorithms with advanced sensors, such as RADAR [7], LiDAR [8,9], and camera [10], have been invented and become the focus of research. ...
Article
Full-text available
With the rapid development of the Internet of Things (IoT), autonomous vehicles have been receiving more and more attention because they own many advantages compared with traditional vehicles. A robust and accurate vehicle localization system is critical to the safety and the efficiency of autonomous vehicles. The global positioning system (GPS) has been widely applied to the vehicle localization systems. However, the accuracy and the reliability of GPS have suffered in some scenarios. In this paper, we present a robust and accurate vehicle localization system consisting of a bistatic passive radar, in which the performance of localization is solely dependent on the accuracy of the proposed off-grid direction of arrival (DOA) estimation algorithm. Under the framework of sparse Bayesian learning (SBL), the source powers and the noise variance are estimated by a fast evidence maximization method, and the off-grid gap is effectively handled by an advanced grid refining strategy. Simulation results show that the proposed method exhibits better performance than the existing sparse signal representation-based algorithms, and performs well in the vehicle localization system.
Article
Image-based indoor localization provides fundamental support for applications such as indoor navigation, virtual reality, and location-based services. Most research focuses on developing methods in good lighting conditions via RGB images; while for low lighting situations, especially at night, RGB-based methods cannot perform well. Depth images are promising alternative in such conditions as they record geometrical information instead of texture information, making it possible to work in low lighting scenarios. Current depth image-based methods, either retrieval-based methods or 3D registration-based methods, are inefficient due to its high computation overhead, preventing the wide applications. To address this issue, we propose a fast coarse-to-fine localization framework for dark environment via deep learning and keypoint-based geometry alignment. In the coarse localization phase, we jointly perform the depth completion and pose regression to relieve the occlusion caused appearance variance in depth images. In the refinement phase, keypoints are used instead of whole depth image points under the ICP alignment framework to increase the localization efficiency. The keypoints are detected on depth feature maps weakly supervised with pose regression. The experiments on the open available 7Scenes dataset show that the proposed method obtain positional accuracy of 0.143 m and orientational accuracy of 5.275°in average and only cost 0.8s for a single depth image. The code for the proposed work is available at https://github.com/lqing900205/KeyPointDepthLocalization
Article
Full-text available
With the rising amount of data in the sports and health sectors, a plethora of applications using big data mining have become possible. Multiple frameworks have been proposed to mine, store, preprocess, and analyze physiological vitals data using artificial intelligence and machine learning algorithms. Comparatively, less research has been done to collect potentially high volume, high-quality ‘big data’ in an organized, time-synchronized, and holistic manner to solve similar problems in multiple fields. Although a large number of data collection devices exist in the form of sensors. They are either highly specialized, univariate and fragmented in nature or exist in a lab setting. The current study aims to propose artificial intelligence-based body sensor network framework (AIBSNF), a framework for strategic use of body sensor networks (BSN), which combines with real-time location system (RTLS) and wearable biosensors to collect multivariate, low noise, and high-fidelity data. This facilitates gathering of time-synchronized location and physiological vitals data, which allows artificial intelligence and machine learning (AI/ML)-based time series analysis. The study gives a brief overview of wearable sensor technology, RTLS, and provides use cases of AI/ML algorithms in the field of sensor fusion. The study also elaborates sample scenarios using a specific sensor network consisting of pressure sensors (insoles), accelerometers, gyroscopes, ECG, EMG, and RTLS position detectors for particular applications in the field of health care and sports. The AIBSNF may provide a solid blueprint for conducting research and development, forming a smooth end-to-end pipeline from data collection using BSN, RTLS and final stage analytics based on AI/ML algorithms.
Article
Automatic and precise spraying in orchards is currently an active area in precision agriculture research. Research on this topic mainly focuses on the recognition of fruit trees and the planning of driving paths. The dense canopy and complex background of peach tree orchards present great challenges in fruit tree recognition. In this study, an algorithm based on colour-depth vision was developed to solve the path planning problem for precise spraying in peach orchards. First, a colour-depth binocular sensor was used to acquire video images in peach orchards. Then, a colour-depth fusion segmentation method (CDFS) based on the leaf wall area (LWA) of the colour-depth images was proposed. This method combines the result of Otsu segmentation for the green layer of the colour image and the K-means layering result of the depth image to accurately detect the LWA of peach trees. Then, image erosion was used to delimit the two largest LWAs as the region of interest (ROI). Finally, the spraying path was planned by detecting the midpoint of ROI spacing as the end of the spraying path. A comparative analysis of five image segmentation algorithms showed that the proposed CDFS algorithm produced the closest LWA to that obtained by manual segmentation and thus achieved the best segmentation effect. Comparing this automatic path plan method with the drivable areas showed that the overall accuracy rate of the path planning was 97.5%. Therefore, the proposed algorithm planned paths within the drivable area and can thus be used to accurately and automatically plan spraying paths in peach orchards, except when there were gaps between rows in the fruit orchard.
Conference Paper
Full-text available
In this paper, we present a localization approach that is based on a point-cloud matching method (normal distribution transform "NDT") and road-marker matching based on the light detection and ranging intensity. Point-cloud map-based localization methods enable autonomous vehicles to accurately estimate their own positions. However, accurate localization and ``matching error'' estimations cannot be performed when the appearance of the environment changes, and this is common in rural environments. To cope with these inaccuracies, in this work, we propose to estimate the error of NDT scan matching beforehand (off-line). Then, as the vehicle navigates in the environment, the appropriate uncertainty is assigned to the scan matching. 3D NDT scan matching utilizes the uncertainty information that is estimated off-line, and is combined with a road-marker matching approach using a particle-filtering algorithm. As a result, accurate localization can be performed in areas in which 3D NDT failed. In addition, the uncertainty of the localization is reduced. Experimental results show the performance of the proposed method.
Article
Full-text available
In this paper, we describe a representation for spatial information, called the stochastic map, and associated procedures for building it, reading information from it, and revising it incrementally as new information is obtained. The map contains the estimates of relationships among objects in the map, and their uncertainties, given all the available information. The procedures provide a general solution to the problem of estimating uncertain relative spatial relationships. The estimates are probabilistic in nature, an advance over the previous, very conservative, worst-case approaches to the problem. Finally, the procedures are developed in the context of state-estimation and filtering theory, which provides a solid basis for numerous extensions.
Conference Paper
Full-text available
A correct perception of road signalizations is required for autonomous cars to follow the traffic codes. Road marking is a signalization present on road surfaces and commonly used to inform the correct lane cars must keep. Cameras have been widely used for road marking detection, however they are sensible to environment illumination. Some LIDAR sensors return infrared reflective intensity information which is insensible to illumination condition. Existing road marking detectors that analyzes reflective intensity data focus only on lane markings and ignores other types of signalization. We propose a road marking detector based on Otsu thresholding method that make possible segment LIDAR point clouds into asphalt and road marking. The results show the possibility of detecting any road marking (crosswalks, continuous lines, dashed lines). The road marking detector has also been integrated with Monte Carlo localization method so that its performance could be validated. According to the results, adding road markings onto curb maps lead to a lateral localization error of 0.3119 m.
Article
Perception system design is a vital step in the development of an autonomous vehicle (AV). With the vast selection of available off-the-shelf schemes and seemingly endless options of sensor systems implemented in research and commercial vehicles, it can be difficult to identify the optimal system for one's AV application. This article presents a comprehensive review of the state-of-the-art AV perception technology available today. It provides up-to-date information about the advantages, disadvantages, limits, and ideal applications of specific AV sensors; the most prevalent sensors in current research and commercial AVs; autonomous features currently on the market; and localization and mapping methods currently implemented in AV research. This information is useful for newcomers to the AV field to gain a greater understanding of the current AV solution landscape and to guide experienced researchers towards research areas requiring further development. Furthermore, this paper highlights future research areas and draws conclusions about the most effective methods for AV perception and its effect on localization and mapping. Topics discussed in the Perception and Automotive Sensors section focus on the sensors themselves, whereas topics discussed in the Localization and Mapping section focus on how the vehicle perceives where it is on the road, providing context for the use of the automotive sensors. By improving on current state-of-the-art perception systems, AVs will become more robust, reliable, safe, and accessible, ultimately providing greater efficiency, mobility, and safety benefits to the public.
Article
This study presents an autonomous guided vehicle (AGV) with simultaneous localization and map building (SLAM) based on matching method, and extended Kalman filter SLAM. In general, the AGV is a large mobile robot that is used in transportation to carry cargoes, and it is guided using wired or wireless guidance systems. The guidance system based AGV accounts for a majority of robots in the mobile robot industry. However, in semiconductor factories, landmarks are unavailable; hence, the existing system has not been used in the mentioned environments. Therefore, the SLAM technology is applied in the environments, and can guide the AGV without landmarks. However, the accuracy of the SLAM can be low owing to measurement error of sensors and a cumulative calculation caused by localization sensors. Therefore, the accuracy is frequently assumed to be incorrect; moreover, the accuracy of the built map is low. In order to solve the problems, this study proposes the AGV with the SLAM based on matching methods; two matching method; geometric matching method and iterative closest point algorithm. The performance of the proposed method is compared with typical methods such as singular value decomposition / RIGID transformation based technologies using feature-point-based SLAM and is compared with the aforementioned two methods using the extended Kalman filter SLAM. The proposed method is more efficient than the typical methods used in the comparison.
Conference Paper
The first step of environment perception for autonomous vehicles is to estimate the trajectories of the ego vehicle. Based on this, we can build the map of the environment and decide the behaviours of the vehicle in the future. Since a Global Navigation Satellite System (GNSS) is inaccurate and not available in all situations, the Simultaneous Localization and Mapping (SLAM) technique has been introduced to solve this problem. In this paper, we propose a hybrid map-based SLAM using Rao-Blackwellized particle filters (RBPFs). Especially, it was designed with the Velodyne 3D HDL-64 laser scanner which provides rich and accurate data of spatial information around the vehicle. Therefore, our work comprises not only the RBPFs framework but also some signal preprocessing procedures of the sensor. Unlike prior works, we describe the environment by using a grid map and a feature map together rather than using only one of them. Based on both maps, we have formulated a new proposal distribution which is an important performance factor of the algorithm. This makes the uncertainty of a predicted vehicle position decrease drastically and therefore the robustness and efficiency of the algorithm can also be improved. The presented approach was implemented on our experimental vehicle and evaluated in the complex urban scenarios. The test results prove that our approach works well even in real outdoor environments and outperforms traditional approaches.
Article
A localization method based on an enhanced particle filter incorporating tournament selection and Nelder-Mead simplex search (NM-EPF) for autonomous mobile robots navigating in a soccer robot game field is proposed in this paper. To analyze the environment, an omnidirectional vision device is mounted on top of the robot. Through detecting the white boundary lines relative to the robot in the game field, weighting for each particle representing the robot's pose can be iteratively updated via the proposed NM-EPF algorithm. Thanks to the hybridization effect of the NM-EPF, particles converge to the actual position of the robot in a responsive way while tackling uncertainties. Simulation and experiment results have confirmed that the proposed NM-EPF has better localization performance in the soccer robot game field in comparison to the conventional particle filter.
Article
Vehicle self-localization is an important and challenging issue in current driving assistance and autonomous driving research activities. This paper mainly investigates two kinds of methods for vehicle self-localization: active sensor based and passive sensor based. Active sensor based localization was firstly proposed for robot localization, and was introduced into autonomous driving recently. The Simultaneous Localization and Mapping (SLAM) techniques is the representative in active sensor based localization. The passive sensor based localization technologies are categorized and explained based on the type of sensors, Global Navigation Satellite System (GNSS), inertial sensors and cameras. In this paper, researches utilizing active sensors and passive sensors for autonomous vehicles are reviewed extensively. Finally, our challenges on self-localization in urban canyon by the system integration of passive sensors is introduced. GNSS error has been improved for the purpose of the self-localization in urban canyon. The performance of the proposed method would suggest possible solution autonomous vehicles which makes use of passive sensors more.