ArticlePDF Available

Multitarget-Tracking Method Based on the Fusion of Millimeter-Wave Radar and LiDAR Sensor Information for Autonomous Vehicles

Authors:

Abstract

Multitarget tracking based on multisensor fusion perception is one of the key technologies to realize the intelligent driving of automobiles and has become a research hotspot in the field of intelligent driving. However, most current autonomous-vehicle target-tracking methods based on the fusion of millimeter-wave radar and lidar information struggle to guarantee accuracy and reliability in the measured data, and cannot effectively solve the multitarget-tracking problem in complex scenes. In view of this, based on the distributed multisensor multitarget tracking (DMMT) system, this paper proposes a multitarget-tracking method for autonomous vehicles that comprehensively considers key technologies such as target tracking, sensor registration, track association, and data fusion based on millimeter-wave radar and lidar. First, a single-sensor multitarget-tracking method suitable for millimeter-wave radar and lidar is proposed to form the respective target tracks; second, the Kalman filter temporal registration method and the residual bias estimation spatial registration method are used to realize the temporal and spatial registration of millimeter-wave radar and lidar data; third, use the sequential m-best method based on the new target density to find the track the correlation of different sensors; and finally, the IF heterogeneous sensor fusion algorithm is used to optimally combine the track information provided by millimeter-wave radar and lidar, and finally form a stable and high-precision global track. In order to verify the proposed method, a multitarget-tracking simulation verification in a high-speed scene is carried out. The results show that the multitarget-tracking method proposed in this paper can realize the track tracking of multiple target vehicles in high-speed driving scenarios. Compared with a single-radar tracker, the position, velocity, size, and direction estimation errors of the track fusion tracker are reduced by 85.5%, 64.6%, 75.3%, and 9.5% respectively, and the average value of GOSPA indicators is reduced by 19.8%; more accurate target state information can be obtained than a single-radar tracker.
Sensors 2023, 23, 6920. https://doi.org/10.3390/s23156920 www.mdpi.com/journal/sensors
Communication
Multitarget-Tracking Method Based on the Fusion of
Millimeter-Wave Radar and LiDAR Sensor Information for
Autonomous Vehicles
Junren Shi 1,*, Yingjie Tang 1, Jun Gao 2, Changhao Piao 1,2 and Zhongquan Wang 1
1 School of Automation, Chongqing University of Posts and Telecommunications, Chongqing 400065, China;
laoauhako@gmail.com (Y.T.); piaoch@cqupt.edu.cn (C.P.); akaxi611@gmail.com (Z.W.)
2 School of Computer Science and Technology, Chongqing University of Posts and Telecommunications,
Chongqing 400065, China; gaoj315@gmail.com
* Correspondence: shijr@cqupt.edu.cn
Abstract: Multitarget tracking based on multisensor fusion perception is one of the key technologies
to realize the intelligent driving of automobiles and has become a research hotspot in the field of
intelligent driving. However, most current autonomous-vehicle target-tracking methods based on
the fusion of millimeter-wave radar and lidar information struggle to guarantee accuracy and reli-
ability in the measured data, and cannot effectively solve the multitarget-tracking problem in com-
plex scenes. In view of this, based on the distributed multisensor multitarget tracking (DMMT) sys-
tem, this paper proposes a multitarget-tracking method for autonomous vehicles that comprehen-
sively considers key technologies such as target tracking, sensor registration, track association, and
data fusion based on millimeter-wave radar and lidar. First, a single-sensor multitarget-tracking
method suitable for millimeter-wave radar and lidar is proposed to form the respective target tracks;
second, the Kalman filter temporal registration method and the residual bias estimation spatial reg-
istration method are used to realize the temporal and spatial registration of millimeter-wave radar
and lidar data; third, use the sequential m-best method based on the new target density to find the
track the correlation of different sensors; and finally, the IF heterogeneous sensor fusion algorithm
is used to optimally combine the track information provided by millimeter-wave radar and lidar,
and finally form a stable and high-precision global track. In order to verify the proposed method, a
multitarget-tracking simulation verification in a high-speed scene is carried out. The results show
that the multitarget-tracking method proposed in this paper can realize the track tracking of multi-
ple target vehicles in high-speed driving scenarios. Compared with a single-radar tracker, the posi-
tion, velocity, size, and direction estimation errors of the track fusion tracker are reduced by 85.5%,
64.6%, 75.3%, and 9.5% respectively, and the average value of GOSPA indicators is reduced by
19.8%; more accurate target state information can be obtained than a single-radar tracker.
Keywords: millimeter-wave radar; lidar; autonomous vehicles; multitarget tracking; data fusion
1. Introduction
Smart cars have great potential in reducing traffic accidents, alleviating traffic con-
gestion, and improving road and vehicle utilization. Their main technologies can be di-
vided into three parts: environmental perception [1], path planning [2], and decision-mak-
ing control [3]. As a basic work in the environment perception of intelligent vehicles, mul-
titarget tracking [4] is of great significance for ensuring the safety of autonomous vehicles
and improving the ability of intelligent vehicles to understand the environment.
Millimeter-wave radars and LiDAR have become mainstream sensors on board au-
tonomous vehicles for target tracking [5]. However, millimeter-wave radars cannot obtain
the geometric information and category information of the target, and are easily affected
by clutter, noise, and multipath, and false targets in the detection also limit its multitarget-
Citation: Shi, J.; Tang, Y.; Gao, J.;
Piao, C.; Wang, Z.
Multitarget-Tracking Method Based
on the Fusion of Millimeter-Wave
Radar and LiDAR Sensor
Information for Autonomous
Vehicles. Sensors 2023, 23, 6920.
https://doi.org/10.3390/s23156920
Academic Editor: Wei Yi
Received: 9 July 2023
Revised: 29 July 2023
Accepted: 2 August 2023
Published: 3 August 2023
Copyright: © 2023 by the author.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license
(https://creativecommons.org/licenses
/by/4.0/).
Sensors 2023, 23, 6920 2 of 20
tracking capability [6]. Compared with millimeter-wave radar, lidar can obtain three-di-
mensional information of the driving environment and can give very accurate spatial po-
sition information. It has incomparable advantages in obstacle detection and target rang-
ing, but the long-distance point cloud is sparse. As a result, its multitarget-tracking capa-
bility is limited [7]. In order to solve the multitarget-tracking problem in complex scenes,
it is imperative to extend single-sensor multitarget tracking to multisensor multitarget
tracking.
Multisensor target-tracking technology is also called multisensor information fusion
technology, which can be divided into centralized, distributed, and hybrid according to
the architecture [8]. Among them, the centralized information fusion system is to send the
original measurement information obtained by each sensor to the central processor for
time–space registration, data association, tracking, and other processing. This fusion
method has a small information loss and high data fusion accuracy, which can reach fu-
sion in the optimal sense [9–12]. However, centralized fusion requires high data quality
and large communication bandwidth, which will increase the burden on the fusion center
and result in poor real-time data processing. Each sensor in the distributed information
fusion system has its own data-processing center, which can independently process the
obtained measurement data. Each local sensor sends the processed and compressed data
to the fusion center, where they are combined and reasoned, and information fusion is
finally realized. Compared with centralized fusion, distributed fusion has lower require-
ments on channel capacity, strong fault tolerance, and easy expansion [13–16]. The hybrid
information fusion model requires each sensor to send the original measurement and the
processed local target track to the fusion center at the same time, taking the advantages of
both centralized and distributed fusion into account, but the structure of the hybrid fusion
method is more complex than the previous two fusion methods, increasing the cost of
communication and calculation [17–20]. Because the distributed fusion structure is simple
and easy to implement, this paper chooses the distributed fusion structure as the basic
framework for autonomous-vehicle multitarget tracking.
As an important research direction in multisource information fusion, target tracking
is an important way to realize sensor network detection. The change in the number of
targets and the uncertainty of measurement information bring great challenges to multi-
target tracking. In traditional multitarget-tracking algorithms based on data association
technology, such as the Joint Probability Data Association (JPDA) algorithm [21], Multiple
Hypotheses Tracking (MHT) algorithm [22], etc., as the number of targets and the number
of clutters increase, the number of calculations increases exponentially. The Probability
Hypothesis Density (PHD) filter [23] based on a Random Finite Set (RFS) can achieve tar-
get without complex data association by passing the first-order moment of the multitarget
posterior probability density estimation of the number and target state. In the process of
distributed multisensor multitarget tracking, it is necessary to transform the data from
multiple sensors into the same space–time reference frame. Since different sensors have
different transmission rates and sampling periods, and there are sensor system deviations
and measurement errors, direct conversion will reduce the accuracy of data fusion. There-
fore, sensor spatiotemporal registration is required when processing multisensor data
[24]. In the distributed multisensor multitarget tracking system, each source has an inde-
pendent information-processing system, which can independently track the surrounding
environment targets and generate corresponding target tracks. Tracks from different sys-
tems may represent the same target due to overlapping detection areas between sensors.
Therefore, track correlation between sensors is required to find the track corresponding
to the same target [25]. Distributed multisensor multitarget tracking is also called distrib-
uted multisensor data fusion. In this system, each local sensor first forms its own target
track based on the single-sensor multitarget-tracking algorithm, and then each sensor
sends the target track to the fusion center to complete the space–time registration and
track association, and then the fusion center based on the estimated fusion criterion tracks
Sensors 2023, 23, 6920 3 of 20
from the same target are estimated and fused to form a stable, high-precision global track
[26].
At present, most autonomous-vehicle target-tracking methods based on the fusion of
millimeter-wave radar and lidar information rarely consider key technologies such as tar-
get tracking, sensor registration, track correlation, and data fusion, and it is difficult to
guarantee the accuracy and reliability of the data with these methods. They cannot effec-
tively solve the multitarget-tracking problem in complex scenes.
In conclusion, this research makes the following contributions:
Taking the distributed multisensor multitarget tracking (DMMT) system as the basic
framework and based on millimeter-wave radar and lidar, a multitarget-tracking method
for autonomous vehicles is proposed that comprehensively considers key technologies
such as target tracking, sensor registration, track correlation, data fusion, etc., so as to ef-
fectively improve the target-tracking accuracy and stability in complex scenes. The main
work is as follows: (i) taking millimeter-wave radar and lidar as objects, respectively, a
single-sensor multitarget-tracking method suitable for millimeter-wave radar and lidar is
proposed to form their respective target tracks; (ii) using the Kalman filter temporal reg-
istration method and the residual bias estimation spatial registration method to realize the
space–time registration of millimeter-wave radar and lidar data; (iii) using the sequential
m-best method based on the new target density to find the track correlation of different
sensors; (iv) using the IF heterogeneous sensor fusion algorithm to optimally combine the
track information provided by millimeter-wave radar and lidar, and finally form a stable
and high-precision global track.
2. Program Framework
The research program in this paper is mainly for autonomous vehicles under high-
speed driving conditions. Its main characteristics include: high driving speed (higher than
80 km/h), many traffic participants, and complex driving scenarios. When an autonomous
vehicle is driving on the expressway, it needs to continuously pay attention to the driving
situation of the surrounding target vehicles, especially for the maneuvering targets adja-
cent to the target. There may be two or more measurements associated with this; at this
time, the target no longer satisfies the one-to-one assumption (one-to-one correspondence
between targets and measurements), and even the increase in the number of targets and
measurements leads to missed or false detections [27]. Therefore, the research plan needs
to solve the problems of many measurement interference sources in complex driving sce-
narios, which are prone to misidentification or missed detection.
Considering the autonomous-vehicle sensor configuration, sensor characteristics,
and environmental factors, millimeter-wave radar and lidar were selected for information
fusion to detect and track target characteristics. The autonomous vehicle was equipped
with four millimeter-wave radars and a 32-line lidar. Among them, the millimeter-wave
radars were respectively arranged on the front, rear, left, and right sides of the vehicle
body, and the range and azimuth resolutions of the radars were 2.5 m and 6°, respectively.
The millimeter-wave radar installed directly in front of the vehicle was a long-range radar
with a detection range of 250 m and a horizontal field of view of 30°; the millimeter-wave
radar equipped directly behind the vehicle was a medium-range radar with a detection
range of 100 m and a horizontal field of view of 90°; and the millimeter-wave radars
equipped on the left and right sides of the vehicle body were short-range radars with a
detection range of 30 m and a horizontal field of view of 160°. In addition, the lidar was
configured on the top of the vehicle body; the detection distance, horizontal field of view,
and vertical field of view were 200 m, 360°, and 40°, respectively, and the distance resolu-
tion, horizontal field of view resolution, and vertical field of view resolution were 5 cm,
0.2°, and 0.4° respectively.
According to the needs of autonomous vehicles and sensor selection, this paper pro-
poses a multitarget-tracking method for autonomous vehicles that comprehensively con-
siders key technologies such as target tracking, sensor registration, track association, and
Sensors 2023, 23, 6920 4 of 20
data fusion. The millimeter-wave radar and lidar sensors first formed their own target
tracks based on the single-sensor multitarget-tracking algorithm, and then sent the target
tracks to the fusion center to complete the space–time registration and track association;
then, the fusion center estimated and fused the tracks from the same target based on the
estimated fusion criteria, and finally formed a stable and high-precision global track. The
overall framework is shown in Figure 1, and the data fusion in the figure only represents
the final stage of multisensor information fusion.
Extended
Object
Tracker
(GM-PHD)
JPDA
Tracker
(IMM-UFK)
Radars
Lidar
Track Association
(SMBTANTD)
Data Fusion
(IF-based T2TF)
Fusion Center
Spatial
Registration
(RBER)
Time
Registration
(IMM-TF)
Target Tracking
Sensor
Sensor Registration
Figure 1. Flow chart of multisensor multitarget tracking.
The specific steps are as follows:
1. Single-sensor target tracking: The extended target tracker is constructed according to
the GM-PHD algorithm and the rectangular target model, and the millimeter-wave
radar sensor uses the extended target tracker to track multiple targets and generate
the local track of the target object. Then, a JPDA tracker configured with the Interact-
ing multiple modules–unscented Kalman filter (IMM-UKF) algorithm is built, which
is used by the lidar to track multiple targets and generate the local tracks of target
objects.
2. The spatiotemporal registration of sensors: The Kalman filter method is used to reg-
ister the asynchronous measurement information of each sensor to the same target at
the same time, so as to realize sensor time registration. The Residual Bias Estimation
Registration (RBER) method is used to estimate and compensate the detection infor-
mation of the space public target, so as to realize sensor space registration.
3. Sensor track association: using the sequential m-best track association algorithm
based on the new target density (SMBTANTD), each time in an iterative manner,
tracks from the next sensor are introduced and correlated with the previous results.
4. Sensor data fusion: the IF heterogeneous sensor fusion algorithm is used to avoid the
repeated calculation of public information and realize the optimal combination of
track information provided by millimeter-wave radar and lidar, so as to obtain more
accurate target status information.
3. Track Fusion and Management
3.1. Single-Sensor Target Tracking
The essence of target tracking is the process of estimating the state and number of
targets by using the noisy measurement data obtained by each sensor [28]. Considering
the configuration and characteristics of the millimeter-wave radar sensor and the lidar
sensor comprehensively, the corresponding target-tracking research was carried out.
3.1.1. Millimeter-Wave Radar Target Tracking
In traditional tracking methods such as Global Nearest Neighbor (multiObject-
Tracker and trackerGNN), Joint Probabilistic Data Association (trackerJPDA), and Multi-
ple Hypothesis Tracking (trackerTOMHT), the tracked object is considered to satisfy the
Sensors 2023, 23, 6920 5 of 20
one-to-one hypothesis that each scan of the sensor will return a detected target value.
However, for a millimeter-wave radar mounted on an autonomous vehicle, due to the
improvement in its resolution, there may be two or more measurements associated with
the measured target. In addition, for autonomous vehicles driving on highways, most of
the measured targets are also high-speed vehicles. In view of this, the GM-PHD algorithm
and rectangular target model were used to construct an extended target tracker.
According to the rectangular target model of the measured object, the rectangular
extended target state of the measured object was obtained; the rectangular extended target
state is expressed as:
= , ,
x X
, (1)
among them, ξ represents the state vector of the extended target of the detected object; γ
represents the measurement rate state of the extended target of the measured object,
which is a scalar subject to the gamma distribution; x represents the motion state of the
extended target of the measured object; the vector x can be modeled as x = [x, y, v, θ, ω, L,
W]T; [x, y]T represents the two-dimensional coordinate position information of the meas-
ured target; v, θ, and ω represent the speed, direction angle, and angular velocity of the
target, respectively; L and W represent the length and width of the detection target, re-
spectively; and X represents the extended state of the extended target of the measured
object, that is, the state after the expansion of a single piece of measurement information.
When the extended target measurement is distributed in a certain space in disorder, usu-
ally a rectangle can be used to approximate the extended state of the extended target.
We constructed the GM-PHD filter according to the rectangular extended target state
of the detected object; the GM-PHD filter approximates the multitarget probability hy-
pothesis density through the Gaussian component with weights, assuming that the mul-
titarget posterior PHD at k 1 time can be expressed as a Gaussian mixture:
1
1 1 1 1
1
; ,
k
J
i i i
k k k k
i
D N m P
x x , (2)
among them, Jk−1 is the number of Gaussian components at time k − 1, ωi
k−1 is the weight of
the i-th Gaussian component, N(·; m, P) represents the probability density function of
Gaussian distribution, N(x; mi
k−1, P i
k−1) is the probability density function of the i-th Gaussian
distribution at time k − 1, the mean is mi
k−1, and the covariance is P i
k−1.
Multitarget prediction PHD at time k and multitarget posterior PHD at time k can
also be expressed in Gaussian mixture form:
1
1 , 1 1 1 1
1
; ,
k k
J
i i i
k
k k S k k k k k k k k
i
D D N m P
x x x x , (3)
1
,1
1
1 ; ,
k k
k
J
j j j
D k k
k k k k k k k k
z z j
D p D z N m z P
x x x , (4)
among them, x represents the motion state of the rectangular extended target state of the
detected object; DS,k|k-1(x) represents the multitarget prediction PHD at time k; DS,k|k−1(x)
represents the PHD of the surviving Gaussian component at time k; γk(x) represents the
new target at time k, that is, the new observation point PHD acquired by the sensor; ωi
k|k−1
represents the weight of the i-th Gaussian component at time k; N(x; mi
k|k−1, P i
k|k−1) means the
Gaussian component whose mean is mi
k|k−1 and whose covariance is P i
k|k−1; Jk|k−1 means the
number of Gaussian components predicted at time k; Dk|k(x) represents the posterior mul-
titarget PHD at time k; pD,k represents the target detection probability, which means that
when there is a signal at the input end of the millimeter-wave radar, due to noise, two
judgment results may be concluded—that there is a “signal” or “no signal”—and when
there is a signal, it will provide the probability that millimeter-wave radar will make a
Sensors 2023, 23, 6920 6 of 20
correct decision of "signal"; ωj
k(z) represents the weight of the j-th update component; zk
represents the measured value of the target at time k, which refers to the physical infor-
mation value such as the target coordinates obtained by the millimeter-wave radar; N(x;
mj
k|k(z), Pj
k|k) means that the mean is mj
k|k(z); and the covariance is the Gaussian component
of Pj
k|k. The GM-PHD filter was constructed by predicting Dk|k−1(x) and updating Dk|k(x)
through the GM-PHD algorithm, and then the track tracker suitable for millimeter-wave
radar was obtained, that is, the extended target tracker.
3.1.2. LiDAR Target Tracking
The amount of measurement information obtained by lidar measurements for each
target is far more than that of a millimeter-wave radar; it will also obtain ground meas-
urement information, and the amount of data output by it is not the same order of mag-
nitude as that of a millimeter-wave radar. If the extended object tracker is directly used
for processing, the output signal will seriously lag due to the high computational com-
plexity. Therefore, it is necessary to construct a bounding box detector to reduce the di-
mensionality of the 3D detection information of the lidar.
When constructing a bounding box detector, a plane fitting algorithm based on Ran-
dom Sample Consensus (RANSAC) [29] is used to preprocess the 3D data of the lidar to
remove redundant point cloud information such as the ground, and then obtain the target
point cloud, reducing the computational overhead. The specific process is as follows: first,
randomly select three points in the initial point cloud, and calculate the sampling plane
formed by the three points; secondly, calculate the distance from all point clouds to the
sampling plane, set the distance threshold, and divide many point clouds into inner points
(normal data) and outer points (abnormal data) according to the threshold; then, count
the number of interior points and exterior points, repeat the above process to the maxi-
mum number of iterations, and select the plane with the most interior points; and finally,
based on the plane with the most interior points, refit the plane with all its interior points
to obtain the final fitting plane equation, remove redundant point clouds such as ground
points according to the final fitting plane equation, and obtain the target point cloud.
After removing redundant point clouds, point clouds belonging to various main ob-
jects will be presented in a state of floating, separated in space, and cuboid bounding box
detectors will be used to detect objects. First, the Euclidean algorithm is used to cluster the
target point cloud in this state, and the state vector of the bounding box detector is ob-
tained according to the clustering result. The state vector of the bounding box detector is
expressed as:
, , , ,
[ ]
, , , , ,
T
x y v z z L W H
x, (5)
among them, xrepresents the state vector, compared with the state vector of the rectan-
gular model of the millimeter-wave radar; the state vector of the cuboid bounding box
detector has three more variables: z,
z
, and H. z represents the vertical coordinate of the
detection target,
z
represents the detection target’s vertical velocity, and H represents
the height of the detection target.
Since a single motion model is not effective for describing maneuvering targets, an
interactive multimodel unscented Kalman filter (IMM-UKF) was used in this paper to es-
timate and update the state. The IMM algorithm can synthesize multiple moving targets
[30] and analyze the weights of each model through Bayesian theory, which can more
accurately describe the law of vehicle motion. The basic idea is to use a finite number of
models to describe the motion state of the target as comprehensively as possible; the tran-
sition between the models is subject to the Markov state chain, and the interaction between
multiple models is considered. The JPDA tracker configured with IMM-UKF is a point
target tracker, as shown in Figure 2. The tracker consists of an input interaction module,
a UKF filter module, a probability update module, a JPDA data association module, and
an output fusion module; it is used to achieve local track generation for objects detected
by lidar.
Sensors 2023, 23, 6920 7 of 20
1
1
ˆ
j
j
UKF1
UKF2
UKFr
k
Z
1
1
j
k
1 2
r
j j j
ˆ1 1
1 1
x k k
P k k
1
j
v k
2
2
ˆ
j
j
x k k
P k k
ˆ
r
j
r
j
x k k
P k k
01
01
ˆ
j
j
x k k
P k k
02
02
ˆ
j
j
x k k
P k k
0
0
ˆ
r
j
r
j
x k k
P k k
ˆ1
1
r
j
r
j
x k k
P k k
2
2
ˆ
1
1
j
j
x k k
P k k
1
1
ˆ
1
1
j
j
x k k
P k k
2
1
j
k
1
r
j
k
1 2
r
j j j
1
i
j
v k
Input Interaction Module
Probability Update Module
JPDA
Module
Comprehensive
Output Module
Figure 2. JPDA tracker configured with IMM-UKF.
The UKF filter obtains the first state estimate
ˆi
j
k k
x based on the state vector x' of
the bounding box detector at time k.
The input interaction module calculates the second state estimate
0
ˆi
j
k k
x and the
second covariance matrix
0i
j
k k
P after multiple target interactions according to the first
state estimate
ˆi
j
k k
x and the first covariance matrix
i
j
k k
P of the UKF filter in the
UKF filter module at time k and outputs, where j represents the motion model, which is a
velocity model constructed according to the motion state of the target, i = 1, 2, ..., r; r is the
number of UKF filters.
0
1
ˆ ˆ
N
i i
j j j i
i
k k k k k
x x
, (6)
0 0 0
1
ˆˆˆˆ
NT
i i i i i
j i j j j j j i
i
k k k k k k k k k k k k k
P P x x x x
, (7)
in Formulas (6) and (7),
ˆi
k k
x is the state estimation of the target at time k under model
i,
i
k k
P is the estimated variance of the target at time k under model i, and
j i
k
is
the interactive input probability.
The UKF filter in the UKF filter module runs the UKF filter in parallel through the
time update process described by Formula (8) and the measurement update process de-
scribed by Formula (9) according to the output of the input interaction module and the
effective observation vector Zk at time k, then outputting the third state estimate
ˆ
1
i
j
k k
x, and the third covariance matrix
1
i
j
k k
P at time k + 1 respectively.
0
ˆ ˆ
1
i i
j j j
k k k k k
x F x , (8)
0
1T
i i
j j j j j
k k k k k k k
P F P F Q , (9)
among them, Fj(k) is the state transition matrix of model j at time k; Qj(k) is the process
noise of model j at time k.
Then, the information of the i-th measurement of the target under model j is as fol-
lows:
ˆ
1 1 1 1
i i i
j j j
k k k k k
v z H x , (10)
its covariance matrix Sj(k + 1) is as follows:
1 1 1 1 1
T
i
j j j j j
k k k k k k
S H P H R , (11)
Sensors 2023, 23, 6920 8 of 20
among them, z(i)(k + 1) is the i-th observation value associated with the target at time k + 1;
Hj(k + 1) is the observation matrix of model j at time k + 1; and Rj(k + 1) is the noise covar-
iance matrix.
The probability update module calculates the conditional probability
1
i
jk
of the
motion model j at time k + 1 according to the residual information of the UKF filter mod-
ule, where the residual information of the UKF filter is denoted by Λ
( + 1).
1
1 1
1
1 1
i
j j
jN
i i
j j
i
k k k
k
k k k
, (12)
the JPDA data association module calculates the fusion information under the model j at
the time k + 1 of the target according to Formula (13):
1
1 1
Mi i
j j j
i
k k
v v
, (13)
among them, M is the number of observations currently associated; β(i)
j is the probability
of association with the target k + 1 time and the i-th observation under model j.
The Kalman gain matrix can be expressed as:
1
1 1 1 1
T
i
j j j j
k k k k k
K P H S , (14)
updating the state variance, we can obtain:
1
1
1 1 1 1 1 1
1 1 1 1 1
MT
i
i
j j j j j j
i
MT
T T
i i i
j j j j j j
i
k k k k k k k
k k k k k
P P K S K
K v v v K
(15)
The output fusion module calculates the fused state estimation and covariance matrix ac-
cording to the conditional probability
1
jk
of motion model j at time k + 1, and ob-
tains the fused state estimation
ˆ
1 1
k k
x and covariance matrix P(k + 1|k + 1).
1
ˆ ˆ
1 1 1 1 1
N
j j
j
k k k k k
x x , (16)
1
ˆ
ˆ
1 1 1 1 1
N
T
j j j j
j
k k k k k
P P x x , (17)
among them,
ˆˆˆ
1 1 1 1
j j k k k k
x x x .
3.2. Sensor Spatiotemporal Registration
The measurement information of the millimeter-wave radar and the laser radar are
respectively processed by their respective local trackers to form two local tracks, that is,
the millimeter-wave radar track information of the detected object and the millimeter-
wave radar track information of the detected object. Due to the different transmission
rates and sampling periods of different sensors and there being sensor system deviations
and measurement errors, direct conversion will reduce the accuracy of data fusion. There-
fore, sensor spatiotemporal registration is required when processing millimeter-wave ra-
dar and lidar data.
Sensors 2023, 23, 6920 9 of 20
3.2.1. Time Registration
Time registration is to register the asynchronous measurement information of each
sensor to the same target at the same time [31]. Typical time registration methods include
the least squares method [32], curve fitting method [33], and Kalman filter method [30].
The least squares method is only suitable for uniform moving objects, and it will cause
model mismatch for nonuniform moving objects, resulting in unsatisfactory registration
results. The curve fitting and Kalman filtering methods are suitable for uniform and non-
uniform sampling scenarios. The Kalman filtering method can adjust the target motion
model. Compared with the curve fitting method, the registration accuracy is higher when
the moving target is complex. Therefore, in this paper, a Kalman filter-like method was
used for the temporal registration of sensor fusion. In the highway driving scenario, the
tracked target may maneuver, and a single model is not enough to describe its state. To
achieve temporal registration in a maneuvering target-tracking system, the IMM-TF
method [34] configured with a Kalman filter algorithm was adopted to solve the temporal
offset estimation problem.
For a multisensor system, the relative time offset between sensors i and n at time k is
defined as
,
ˆi n
k
t
. First, a two-stage relative time offset estimation algorithm is used to cal-
culate the relative time offset estimation between sensor 1 and other sensors, such as
1,2
ˆk
t
,
1,3
ˆ
k
t
,…,
1,
ˆr
N
k
t
. Secondly, the global accurate time stamp is obtained using Formula (18).
,
,
ˆ
1,if max , 2, 3, , 0
ˆ
arg max , 2,3, , , otherwise.
i n
k r
n
i n
k r
n
t n N
Nt n N
(18)
Therefore, the relative time offset between sensor i and the global precise timestamp is as
follows:
, 1, 1,
ˆˆ
ˆ
i N i N
k k k
t t t
, (19)
and the unbiased measurement coordinate system conversion algorithm is used to con-
vert the polar coordinate measurement into Cartesian coordinates. The conversion meas-
urement equation of sensor i at time stamp t-i
k is as follows:
i i i i
i k k k i k
t t t t
Z H x v , (20)
among them,
i
k
t
H is the linear measurement transfer matrix,
i
k
t
x is the passive dy-
namic equation, and
i
i k
t
v is the zero-mean white Gaussian noise vector. Due to the time
offset, the target dynamic equation that needs to be filtered is as follows:
1 1 1
i i i i i i
k k k k k k
t t t t t t
x F x w , (21)
and among them,
1
i i
k k
t t
F is the state transition matrix from time
i
k
t
to time
1
i
k
t
, and
1
i i
k k
t t
w is the zero-mean white Gaussian noise of the covariance matrix.
For different target dynamic models, such as CV, CA, or CT models, the correspond-
ing estimators are used to calculate constant or time-varying relative time offset estimates.
For the calculation methods, refer to [34].
Finally, the time offset estimates of different dynamic models are fused, and the ob-
tained time offset estimates and covariance are as follows:
1,2 1, 2
ˆˆ
j, j i
k k k
j
t t t
, (22)
Sensors 2023, 23, 6920 10 of 20
1,2 ,1,2 1,2 1,2 1,2 1,2
ˆ ˆ ˆ ˆ T
j
j, j, j i
k k k k k
k k
j
t t t t t
, (23)
and among them,
j i
k
t
is the conditional probability of motion model j.
3.2.2. Spatial Registration
Spatial registration is the process of using multisensors to detect the information of
common space targets and estimating and compensating the system deviation of sensors,
which can improve the accuracy of information fusion [35]. Lidar can obtain the complete
three-dimensional position information of the target (distance R, pitch angle θ, and azi-
muth angle φ), while millimeter-wave radar cannot obtain complete three-dimensional
information, and can only provide angle and distance information. Therefore, in order to
deal with the problem of incomplete measurement data registration model mismatch, this
paper uses the Residual Bias Estimation Registration (RBER) method [36] to estimate and
compensate for the detection information of space public targets.
All measurement values are divided into two parts: complete measurement and non-
complete measurement. First, a maximum likelihood estimation is performed on the com-
plete measurements in a common coordinate system to find the target position estimate
ˆc
k
x
:
,
,
1
,
ˆˆ
1
1
1
ˆ ˆ
1
ˆC
k i k
C
k i k
N
c
k i k
i
N
i
x x
x x
x x
, (24)
Then, based on sequential filtering technology, the incomplete measurement data are used
to sequentially update
ˆc
k
x
to obtain the updated target position estimate
ˆall
k
x
.
In
1
In
In
1
1In
In In
ˆˆˆ
k
k
all c c
k k k k k
T
all
k k x
T
all c
k k x x
z
z
x x K z h x
K P H
P P H H
, (25)
and among them,
ˆc
k
x
and
c
k
P
are the complete measurement target position estimation
and covariance matrix estimation, respectively; HIn
x is the connection Jacobian matrix
about L sensors; and In
k
z is the covariance matrix composed of incomplete measure-
ment data.
Then,
ˆall
k
x
is converted to the measurement coordinate system, a maximum likeli-
hood estimation is performed for all measurements in the measurement coordinate sys-
tem, the parameter ρ is obtained to be estimated through iteration, and the measurement
information of the salient target is used to eliminate the systematic deviation of the sensor.
The specific calculation method can be found in reference [36].
3.3. Sensor Track Association
In a distributed multisensor multitarget tracking system, the purpose of track associ-
ation is to study how to match the tracks reported by different sensors. Due to the asyn-
chrony of the track and the existence of systematic errors [37], the difficulty of track cor-
relation is increased, and the above problems must be solved in order to obtain reliable
correlation results. In order to solve this problem, a sequential m-best track association
algorithm based on the new target density (SMBTANTD) [36] is sampled. This method
introduces the history of track information to improve track correlation performance. In
Sensors 2023, 23, 6920 11 of 20
addition, this method can effectively solve the coupling problem between track associa-
tion and spatial registration by utilizing the repeated process of salient target selection,
multitarget filter tracking, and different sensor target association [36].
After each sensor obtains the target measurement value, the measurement value is
filtered to obtain the target state estimation value
ˆ
x
and its error covariance matrix P. If
Γi
j represents the tracking result of sensor i on target j, then the tuple,
1, 2, ,
i
j j
i M
, (26)
represents the hypothesis tracked from the same target.
In a multitarget multisensor, when creating a global association hypothesis H con-
sisting of tuples, the goal is to find the most probable hypothesis in the set of all global
hypotheses. The likelihood function of a tuple can be expressed as:
, , ,
i i M
j j j j j
L p
, (27)
and the most likely global hypothesis can be obtained by finding the tuple with the largest
likelihood function [38], namely:
ˆarg max
H
H L

, (28)
Assuming that multiple trajectories representing the same target obey a zero-mean nor-
mal distribution with variance P, there is no correlation between trajectories. The likeli-
hood function can be expressed as:
1
, ,
1/ 2
1
1 1 ˆ ˆ ˆ ˆ
exp 2
2
MT
i F M if i F M
iif
L
x x P x x
P
, (29)
and among them,
if
P
represents the cross-covariance matrix of random variables, and
ˆi
x
represents the state estimation of the i-th sensor.
,
ˆF M
x
represents the global state es-
timate.
1 1
, , , 1 ,
2
if ii F M F M ii F M F M
P P P P P P P
, (30)
and among them,
ii
P
is the cross-covariance matrix.
Let cji denote the j-th row and i-th column of the matrix; then, the target allocation
cost among different sensors can be expressed as:
ln
ji
c L
, (31)
and among them, ln L(ϒ) does not depend on the target indices i and j; it depends on
different combinations of different target trajectories from different sensors.
In order to solve the correlation problem when the number of targets measured by
multiple sensors is inconsistent, a new target density is introduced in the correlation ma-
trix. The target j acquired from the next sensor is defined as the new target, namely:
ln
j N
c
, (32)
where is the density of new objects, which has similarity with the correlation thresh-
old [39].
Assuming that object trajectories from the same sensor are not correlated, the alloca-
tion cost is set to infinity. Based on the Murty algorithm [40], the optimal solution to the
two-dimensional assignment problem is obtained:
Sensors 2023, 23, 6920 12 of 20
2 2 ,1 2,2 2 ,3
1 2 2,1
2,1 1 1 2, 2
1 2 2,1
2,2 2 2 2, 2
1
2,3 3
, ,
,
,
, (33)
and among them, Γ1
1, Γ1
2, and Γ1
3 represent the track information of sensor 1 for three tar-
gets; Γ2
1 and Γ2
2 represent the track information of sensor 2 for two targets; ϒ2,i is the i-th
tuple; and Γj
i represents the state estimation of the j-th sensor to i targets and the tracking
track information of the covariance. ϒ2,1 = 1
1, Γ2
1} means the matching information of the
first target received by sensor 1 and the first target received by sensor 2. ϒ2,2 = 1
2, Γ2
2} means
that the information of the second target received by sensor 1 matches the information of
the first target received by sensor 2. ϒ2,3 = 1
3} means that the third target information re-
ceived by sensor 1 does not match.
Finally, each time in an iterative manner, tracks from the next sensor are introduced
and correlated with the previous results.
3.4. Heterogeneous Track-to-Track Fusion
Millimeter-wave radar and lidar are sensors of different configurations. Heterogene-
ous sensor data fusion can use the incomplete measurement information in the measure-
ment coordinate system to update the complete target state estimation in the Cartesian
coordinate system [41]. In addition, the system is a nonlinear system; since the closed-
form solution of the posterior probability density function is very difficult, this paper
adopts the nonlinear system suboptimal filtering algorithm, that is, the IF heterogeneous
sensor fusion algorithm [42] completes the fusion of millimeter-wave radar and lidar
measurement data.
|
ˆi
k k
x
and
|
i
k k
P
denote the state estimation of the i-th tracker at the moment
i
k
t
and
the corresponding covariance of the Cartesian coordinate system, respectively, since two
different trackers are used for millimeter-wave radar and lidar, respectively (let them be
tracker 1 and tracker 2); so, here, i = 1, 2. For positive integers j and l, their information
matrix and information state are estimated as:
1
| |
:
j l j l
P
, (34)
1 1
| | | | |
ˆˆ
ˆ
:
j l j l j l j l j l
P x x
, (35)
The starting moment of fusion is
1 1
1 2
k j
t t
,
1 1 1 1
1 1
| |
ˆ,
k k k k
x P and
1 1 1 1
2 2
| |
ˆ,
j j j j
x P are obtained
from tracker 1 and tracker 2, respectively, and the state estimation and covariance in the
Cartesian coordinate system are transformed into a corresponding information state esti-
mation and information matrix.
1 1 1 1 1 1 1 1 1 1
1
1 1 1 1 1
| | | | |
ˆ
ˆ
,
k k k k k k k k k k
P x
, (36)
1 1 1 1 1 1 1 1 1 1
1
2 2 2 2 2
| | | | |
ˆ
ˆ
,
j j j j j j j j j j
P x
, (37)
At time
1
1
k
t
, tracker 1 and tracker 2 send
1 1 1 1
1 1
| |
ˆ
,
k k k k
and
1 1 1 1
2 2
| |
ˆ
,
j j j j
to the fusion
center, respectively. Make
1 1
1
k k
t t
,
1 1 1
1 1
|k k k
t , and
1 1 1
2 2
|j
k j
t . By analogy, the fusion
information matrix and the corresponding state matrix at time
1
k
t
of the calculation by
the fusion center are as follows:
1 1 1 1 1 1
1 2 1 2
ˆ ˆ ˆ
,
k k k k k k
t t t t t t
, (38)
Sensors 2023, 23, 6920 13 of 20
Let
cf
t
represent the current fusion moment and set
1
cf
k
t t
; then, the time points of the
data fusion center are
1
k
t
,
1
fk
t N
, …,
f
t
in turn. Among them, Δ is a constant data
measurement interval, Nf is a positive integer, and tf is the final time.
Let
cf f
k
t t N
; the tracker sends information to the fusion center at time
k
t
. For
local tracker 1, the prediction information matrix and the corresponding state estimation
from time
cf
t
to time
k
t
are as follows:
1
1 1
cf cf cf cf cf
, , , ,
k k k k
t t t t t t t t t
F F Q , (39)
1
1 1 1 1
cf cf cf cf cf
ˆ ˆ
, , , ,
k k k k
t t t t t t t t t
F , (40)
among them,
cf
,
k
t t
F and
cf
,
k
t t
Q are the state transition matrix and process noise
covariance matrix from time
cf
t
to time
k
t
, respectively. For local tracker 2, a similar
method can be used to obtain the prediction information matrix
2
cf
,
k
t t
from time
cf
t
to time
k
t
and the corresponding state estimation
2
cf
ˆ
,
k
t t
. Furthermore, the information
matrix
cf
,
k
t t
and state estimation
cf
ˆ
,
k
t t
of the fusion center can be obtained.
When new information is sent to the fusion center, the information state matrix and
information matrix of time
k
t
tracker 1 and tracker 2 need to be calculated.
1
1 1 1 1 1
ˆˆ
,
k k k k k
t t t t t
P x , (41)
1
2 2 2 2 2
ˆˆ
,
k k k k k
t t t t t
P x , (42)
The new information matrix and information state estimate sent by tracker 1 to the
fusion center at time
k
t
are as follows:
1 1 1
cf
,
k k k
t t t t
, (43)
1 1 1
cf
ˆˆˆ
,
k k k
t t t t
, (44)
For local tracker 2, a similar method can be used to obtain the new information matrix
2
k
t
and information state estimation
2
ˆ
k
t
sent to the fusion center from time
k
t
.
Based on the predicted information matrix
cf
,
k
t t
and information state estimation
cf
ˆ
,
k
t t
at the fusion center and the new information received by the tracker, the data
fusion information of the fusion center at time
k
t
can be calculated:
1 2
cf
,
k k k k
t t t t t
, (45)
1 2
cf
ˆ ˆ ˆ ˆ
,
k k k k
t t t t t
, (46)
Then, the fusion covariance
IF
k
t
P and state estimate
IF
ˆ
k
t
x can be computed.
IF 1 IF IF ˆ
ˆ
,
k k k k k
t t t t t
P x P , (47)
and if, at this time,
f
k
t t
, the fusion ends; otherwise, let cf
k
t t
and start to predict the
relevant information from Formula (39).
4. Simulation Verification and Analysis
In order to verify the effectiveness of the aforementioned autonomous-vehicle mul-
titarget-tracking method based on the fusion of millimeter-wave radar and lidar infor-
mation, it was necessary to conduct simulation experiments, and the performance of the
Sensors 2023, 23, 6920 14 of 20
multitarget-tracking method proposed in this paper was verified by using the Generalized
Optimal SubPattern Assignment Metric (GOSPA) [43]. The high-speed driving scene of
the autonomous vehicle and the target vehicle is shown in Figure 3, in which there is one
autonomous vehicle and four target vehicles.
The autonomous vehicle is located in the middle lane of the three lanes, and there are
two target vehicles in front of the autonomous vehicle, which are located in the middle
lane and the right lane, respectively; behind the ego vehicle, there is a target car in the
center lane and another target car in the left lane. The autonomous vehicle is equipped
with four millimeter-wave radars and one lidar, and the radar detection coverage over-
laps. At the beginning of the simulation, the autonomous vehicle and target vehicles 1, 3,
and 4 travel at a speed of 25 m/s, and target vehicle 2 travels at a speed of 35 m/s. It can be
seen that the simulation scene includes a car-following scene and an overtaking scene.
The parameters of the autonomous vehicle, target vehicle, and sensors are shown in
Tables 1 and 2.
100 50 0 -50 -100
X (m)
Y (m)
350
400
250
300
150
200
50
100
0
80
90
60
70
40
50
20
30
10
0
20 10 0 -10 -20
Y (m)
X (m)
2
1
3
4
Ego vehicle Target vehicle
Figure 3. Highspeed driving scene of autonomous vehicle and target vehicle.
Table 1. Motion parameter setting table.
Parameter
Ego Vehicle
Target Vehicle 1
Target Vehicle 2
Target Vehicle 3
Target Vehicle 4
Speed (m/s) 25 25 35 25 25
Length (m) 4.7 4.7 4.7 4.7 4.7
Width (m)
1.8
1.8
1.8
1.8
1.8
Height (m) 1.4 1.4 1.4 1.4 1.4
Road centers [0, 0; 50, 0; 100, 0; 250, 20; 400, 35]
Sample time (s)
0.1
Lane specifica-
tions 3.0
Sensors 2023, 23, 6920 15 of 20
Table 2. Sensor setting table.
Front Radar
Rear Radar
Left Radar
Right Radar
LiDAR
Mounting location (m) (3.7, 0.0, 0.6) (−1.0, 0.0, 0.6) (1.3, 0.9, 0.6) (1.3, −0.9, 0.6) (3.7, 0.0, 0.6)
Mounting angles (deg)
(0.0, 0.0, 0.0)
(180.0, 0.0, 0.0)
(90.0, 0.0, 0.0)
(−90.0, 0.0, 0.0)
(0.0, 0.0, 2.0)
Horizontal field of view (deg) (5.0, 30.0) (5.0, 90.0) (5.0, 160.0) (5.0, 160.0) (5.0, 360.0)
Vertical field of view (deg) - - - - (5.0 40.0)
Azimuth resolution (deg) 6.0 6.0 6.0 6.0 0.2
Range limits (m)
(0.0, 250.0)
(0.0, 100.0)
(0.0, 30.0)
(0.0, 30.0)
(0.0, 200.0)
Range resolution (m) 2.5 2.5 2.5 2.5 1.25
We carried out a performance analysis of a single-radar tracker (such as the millime-
ter-wave radar tracker and lidar tracker) and fusion estimation tracker from two aspects
of the simulation visualization interface and quantitative indicators, and judged the com-
prehensive performance of the fusion estimation tracker based on different events in the
scene.
4.1. Simulation Scene Visualization Analysis
Figure 4 is the visual interface of the simulation scene of the millimeter-wave radar,
lidar, and track fusion; the interface consists of five figures, namely the symbol feature
figure, the front view scene figure, the rear-view scene figure, the single-sensor figure,
and the close-up figure. The symbolic feature graph explains the target information of dif-
ferent types of sensors. The front view scene graph and the rear-view scene graph are the
forward scene and the backward scene at the same moment under the condition of the
third perspective, respectively. The single-sensor figure shows only target-tracking infor-
mation from millimeter-wave radar or lidar, while the close-up figure shows key infor-
mation about the target vehicle. The prefixes “R”, “L”, and “F” represent the target infor-
mation of the millimeter-wave radar tracker, lidar tracker, and fusion estimation tracker,
respectively, and the Arabic numerals represent the unique identifier.
The millimeter-wave radar can obtain multiple pieces of detection information of the
same target vehicle, which seriously affects the target detection effect. After being pro-
cessed by the extended target tracker, the detection effect is greatly improved. The scene
visualization interface with a simulation time step of 50 is shown in Figure 4a. It can be
seen from the figure that whether it is millimeter-wave radar, lidar, or fusion estimation
tracker, they can all effectively track four target vehicles and generate the respective target
tracks.
It can be seen from Section 3.1 that when using lidar for target tracking, the Euclidean
algorithm will be used to cluster the target point cloud in this state, and then the state
vector of the bounding box detector will be obtained according to the clustering results.
However, when the distance between different detection objects is very close, the effect of
the clustering algorithm will be affected. The scene visualization interface with a simula-
tion time step of 70 is shown in Figure 4b. It can be seen from the figure that the distance
between target vehicle 2 and target vehicle 3 in front of the autonomous vehicle is very
close; at this point, the bounding box detector aggregates the point cloud of each vehicle
into a larger bounding box, and as a result, the target point detected by the lidar deviates
from the center of the vehicle. Compared with the lidar tracker, the target points detected
by the fusion estimation tracker are highly consistent with the center of the vehicle.
Sensors 2023, 23, 6920 16 of 20
Front View(Time Step = 50) Radar
Rear View
Lidar
Close-up Display
Ground Truth Radar Tracks Lidar Tracks Fused Tracks P oint Cloud
Radar Detections Lidar Bounding Box Detections
Ground Truth Radar Tracks Lidar Tracks Fused Tracks Point Cloud
Radar Detections Lidar Bounding Box Detections
Front View(Time Step = 70) Radar
Rear View Close-up Display
F3
F2 F4
Lidar
(a) (b)
Figure 4. Radar, lidar, and the track-level fusion simulation scenarios. (a) Track mainte-
nance; (b) closely spaced targets.
4.2. Quantitative Index Analysis
We used quantitative indicators to evaluate the tracking performance of single-radar
trackers (such as millimeter-wave radar trackers and lidar trackers) and fusion estimation
trackers, and analyzed the performance of different trackers in simulation scenarios. First,
the tracking performance of the millimeter-wave radar, lidar, and fused estimation track-
ers was evaluated using estimation errors for position, velocity, size, and orientation.
The average estimation errors of millimeter-wave radar, LiDAR, and fusion estima-
tion tracker are shown in Figure 5. It can be seen from the figure that the tracking perfor-
mance of the fusion estimation tracker for the target vehicle was significantly better than
that of the single-radar trackers in terms of the tracking effects of different targets; the
estimation errors of position, velocity, size, and direction were reduced by 85.5%, 64.6%,
75.3%, and 9.5%, respectively. It should be noted that since the millimeter-wave radar had
no size and direction information, it is not shown in the figure.
Sensors 2023, 23, 6920 17 of 20
0
0.5
1
1.5
2
2.5
3
Velocity Error(m/s)
0
0.5
1
1.5
2
0
0.5
1
1.5
2
2.5
3
3.5
Yaw Error(deg)
0
0.2
0.4
0.6
0.8
1
1.2
1.4
Position Error(m)
1 2 3 4
Truth ID
1 2 3 4
Truth ID
1 2 3 4
Truth ID
1 2 3 4
Truth ID
Dimension Error(m)
Radar Lidar Fused
Figure 5. Average estimation error of different trackers.
The tracking performance or localization accuracy of different trackers can be quan-
titatively evaluated with the GOSPA index at each time step, with lower index values
indicating higher tracking accuracy. The tracking performance of millimeter-wave radar,
lidar, and track fusion trackers was quantitatively measured using the “missing target”
and “false tracking” components of the GOSPA metric. As shown in Figure 6, the “lost
target” dropped to 0 after the tracking system ran for 8 time steps. The reason for the high
value at the beginning of the period was due to the delay of the system. The “fault trace”
component was always 0, thus indicating that the trace system did not generate fault
traces. When the simulation time step was 70, the GOSPA index value of the lidar in-
creased significantly; the main reason is that the distance between different detection tar-
gets was too close. The average value of the GOSPA index obtained using the fusion esti-
mation method was 19.8% lower than that of the single-type sensor, which shows that the
overall tracking accuracy was effectively improved after the information fusion of differ-
ent sensors. The average value of the GOSPA index obtained using the fusion estimation
method was 19.8% lower than that of the single-type sensor, which shows that the overall
tracking accuracy was effectively improved after the information fusion of different sen-
sors.
Sensors 2023, 23, 6920 18 of 20
0
20
40
0 20 40 60 80 100 120
Time step
Lidar
Radar
Fused
Missed Target Metric
-1
-0.5
0
0.5
1
False Track Metric
0 20 40 60 80 100 120
Time step
Lidar
Radar
Fused
0
20
40
GOSPA Metric
0 20 40 60 80 100 120
Time step
Lidar
Radar
Fused
Figure 6. GOSPA tracking performance evaluation.
Through the performance analysis of the single-radar tracker (such as the millimeter-
wave radar tracker and the lidar tracker) and the fusion estimation tracker through the
two aspects of the simulation visual interface and the quantitative index, it can be con-
cluded that the multitarget-tracking method for autonomous vehicles based on millime-
ter-wave radar and lidar proposed in this paper can realize the track tracking of multiple
target vehicles in high-speed driving scenarios. Compared with the single-radar tracker,
the fusion estimation tracker can effectively avoid the detection target point deviation
problem caused by the too-close distance between different detection targets.
5. Conclusions
Most of the current autonomous-vehicle target-tracking methods based on the fusion
of millimeter-wave radar and lidar information struggle to guarantee accuracy and relia-
bility in the measured data, and cannot effectively solve the multitarget-tracking problem
in complex scenes. This paper takes the distributed multisensor multitarget tracking
(DMMT) system as the basic framework, based on millimeter-wave radar and lidar; a mul-
titarget-tracking method for autonomous vehicles is proposed that comprehensively con-
siders key technologies such as target tracking, sensor registration, track correlation, and
data fusion. In order to verify the effectiveness of the proposed method, a performance
analysis of the single-radar tracker (such as the millimeter-wave radar tracker and the
lidar tracker) and the fusion estimation tracker is carried out from two aspects of the sim-
ulation visualization interface and quantitative indicators. The multitarget-tracking
method for autonomous vehicles based on millimeter-wave radar and lidar proposed in
this paper can realize the track tracking of multiple target vehicles in high-speed driving
scenarios. Compared with the single-radar tracker, the position, velocity, size, and
Sensors 2023, 23, 6920 19 of 20
direction estimation errors of the track fusion tracker are reduced by 85.5%, 64.6%, 75.3%,
and 9.5%, respectively, and the average value of GOSPA indicators is reduced by 19.8%.
The fusion estimation tracker can effectively avoid the detection target point deviation
problem caused by the too-close distance between different detection targets, and can ob-
tain more accurate target state information than the single-radar tracker. In future re-
search, we will further consider scenarios such as rainy and snowy days, viaducts, tun-
nels, etc., and use different combinations of millimeter-wave radar, lidar, and cameras to
verify the effectiveness of the proposed method in the track tracking of multiple targets
by way of combining a simulation and a real vehicle experiment.
Author Contributions: Formal analysis, Y.T.; investigation, J.G.; methodology, J.S.; software, Y.T.
and J.G.; validation, C.P. and Z.W.; visualization, C.P. and Z.W.; writing—original draft, J.S. All
authors have read and agreed to the published version of the manuscript.
Funding: This research was funded by the National Key Research and Development Program of
China, grant number No. 2022YFE0101000; Chongqing Postdoctoral Research Special Funding Pro-
ject, grant number No. 2022CQBSHTB2010; and partially supported by school-level research pro-
jects, grant number 22XJZXZD05.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Data supporting this systematic review are available in the reference
section. In addition, the data presented in this study are available upon request from the correspond-
ing author.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Hou, W.; Li, W.; Li, P. Fault Diagnosis of the Autonomous Driving Perception System Based on Information Fusion. Sensors
2023, 23, 5110.
2. Shi, J.; Sun, D.; Qin, D.; Hu, M.; Kan, Y.; Ma, K.; Chen, R. Planning the trajectory of an autonomous wheel loader and tracking
its trajectory via adaptive model predictive control. Robot. Auton. Syst. 2020, 131, 103570.
3. Piao, C.; Gao, J.; Yang, Q.; Shi, J. Adaptive cruise control method based on hierarchical control and multi-objective optimization.
Trans. Inst. Meas. Control 2022, 45, 1298–1312.
4. Ramadani, A.; Bui, M.; Wendler, T.; Schunkert, H.; Ewert, P.; Navab, N. A survey of catheter tracking concepts and methodol-
ogies. Med. Image Anal. 2022, 82, 102584.
5. Beck, J.; Arvin, R.; Lee, S.; Khattak, A.; Chakraborty, S. Automated vehicle data pipeline for accident reconstruction: New in-
sights from LiDAR, camera, and radar data. Accid. Anal. Prev. 2023, 180, 106923.
6. Zhang, Z.; Jiang, J.; Wu, J.; Zhu, X. Efficient and optimal penetration path planning for stealth unmanned aerial vehicle using
minimal radar cross-section tactics and modified A-Star algorithm. ISA Trans. 2023, 134, 42–57.
7. Adurthi, N. Scan Matching-Based Particle Filter for LIDAR-Only Localization. Sensors 2023, 23, 4010.
8. Wang, Z.; Wu, Y.; Niu, Q. Multi-Sensor Fusion in Automated Driving: A Survey. IEEE Access 2020, 8, 2847–2868.
9. Nazari, M.; Pashazadeh, S.; Mohammad-Khanli, L. An Adaptive Density-Based Fuzzy Clustering Track Association for Distrib-
uted Tracking System. IEEE Access 2019, 7, 135972–135981.
10. Sharma, A.; Chauhan, S. Sensor Fusion for Distributed Detection of Mobile Intruders in Surveillance Wireless Sensor Networks.
IEEE Sens. J. 2020, 20, 15224–15231.
11. Dash, D.; Jayaraman, V. A Probabilistic Model for Sensor Fusion Using Range-Only Measurements in Multistatic Radar. IEEE
Sens. Lett. 2020, 4, 1–4.
12. Joo, J.E.; Choi, S.; Chon, Y.; Park, S.M. A Low-Cost Measurement Methodology for LiDAR Receiver Integrated Circuits. Sensors
2023, 23, 6002.
13. Xiao, F. Multi-sensor data fusion based on the belief divergence measure of evidences and the belief entropy. Inform. Fusion
2019, 46, 23–32.
14. Khaleghi, B.; Khamis, A.; Karray, F.O.; Razavi, S.N. Multisensor data fusion: A review of the state-of-the-art. Inform. Fusion 2013,
14, 28–44.
15. Kahraman, S.; Bacher, R. A comprehensive review of hyperspectral data fusion with lidar and sar data. Annu. Rev. Control 2021,
51, 236–253.
16. Lee, J.; Hong, J.; Park, K. Frequency Modulation Control of an FMCW LiDAR Using a Frequency-to-Voltage Converter. Sensors
2023, 23, 4981.
Sensors 2023, 23, 6920 20 of 20
17. Caron, F.; Davy, M.; Duflos, E.; Vanheeghe, P. Particle Filtering for Multisensor Data Fusion with Switching Observation Mod-
els: Application to Land Vehicle Positioning. IEEE Trans. Signal Process. 2007, 55, 2703–2719.
18. Gao, S.; Zhong, Y.; Li, W. Random Weighting Method for Multisensor Data Fusion. IEEE Sens. J. 2011, 11, 1955–1961.
19. Wang, X.; Fu, C.; Li, Z.; Lai, Y.; He, J. DeepFusionMOT: A 3D multi-object tracking framework based on camera-LiDAR fusion
with deep association. IEEE Robot. Autom. Lett. 2022, 7, 8260–8267.
20. Wang, X.; Fu, C.; He, J.; Wang, S.; Wang, J. StrongFusionMOT: A Multi-Object Tracking Method Based on LiDAR-Camera
Fusion. IEEE Sens. J. 2022, 23, 11241–11252.
21. Chang, K.; Bar-Shalom, Y. Joint probabilistic data association for multitarget tracking with possibly unresolved measurements
and maneuvers. IEEE Trans. Automat. Control 1984, 29, 585–594.
22. Blackman, S.S. Multiple hypothesis tracking for multiple target tracking. IEEE Aerosp. Electron. Syst. Mag. 2004, 19, 5–18.
23. Mahler, R. PHD filters of higher order in target number. IEEE Trans. Aerosp. Electron. Syst. 2007, 43, 1523–1543.
24. Xing, J.; Wu, W.; Cheng, Q.; Liu, R. Traffic state estimation of urban road networks by multi-source data fusion: Review and
new insights. Phys. A Stat. Mech. Appl. 2022, 595, 127079.
25. Pastor, A.; Sanjurjo-Rivo, M.; Escobar, D. Initial orbit determination methods for track-to-track association. Adv. Space Res. 2021,
68, 2677–2694.
26. You, L.; Zhao, F.; Cheah, L.; Jeong, K.; Zegras, P.C.; Ben-Akiva, M. A Generic Future Mobility Sensing System for Travel Data
Collection, Management, Fusion, and Visualization. IEEE Trans. Intell. Transp. 2020, 21, 4149–4160.
27. Yao, G.; Saltus, R.; Dani, A. Image Moment-Based Extended Object Tracking for Complex Motions. IEEE Sens. J. 2020, 20, 6560–
6572.
28. Liu, Z.; Shang, Y.; Li, T.; Chen, G.; Wang, Y.; Hu, Q.; Zhu, P. Robust Multi-Drone Multi-Target Tracking to Resolve Target
Occlusion: A Benchmark. IEEE Trans. Multimed. 2023, 25, 1462–1476.
29. Ghahremani, M.; Williams, K.; Corke, F.; Tiddeman, B.; Liu, Y.; Wang, X.; Doonan, J.H. Direct and accurate feature extraction
from 3D point clouds of plants using RANSAC. Comput. Electron. Agric. 2021, 187, 106240.
30. Akca, A.; Efe, M.Ö. Multiple Model Kalman and Particle Filters and Applications: A Survey. IFAC PapersOnLine 2019, 52, 73–
78.
31. Zhang, F.; Ge, R.; Zhao, Y.; Kang, Y.; Wang, Z.; Qu, X. Multi-sensor registration method based on a composite standard artefact.
Opt. Laser Eng. 2020, 134, 106205.
32. Blair, W.D.; Rice, T.R.; Alouani, A.T.; Xia, P. Asynchronous Data Fusion for Target Tracking with a Multitasking Radar and
Optical Sensor. In Proceedings of the Proceedings Volume 1482, Acquisition, Tracking, and Pointing V, Orlando, FL, USA, 1
August 1991; pp. 234–245.
33. Zhu, H.; Wang, C. Joint track-to-track association and sensor registration at the track level. Digit. Signal Process. 2015, 41, 48–59.
34. Li, S.; Cheng, Y.; Brown, D.; Tharmarasa, R.; Zhou, G.; Kirubarajan, T. Comprehensive Time-Offset Estimation for Multisensor
Target Tracking. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 2351–2373.
35. Taghavi, E.; Tharmarasa, R.; Kirubarajan, T.; Bar-Shalom, Y.; Mcdonald, M. A practical bias estimation algorithm for multisen-
sor-multitarget tracking. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 2–19.
36. Wang, J.; Zeng, Y.; Wei, S.; Wei, Z.; Wu, Q.; Savaria, Y. Multi-Sensor Track-to-Track Association and Spatial Registration Algo-
rithm Under Incomplete Measurements. IEEE Trans. Signal Process. 2021, 69, 3337–3350.
37. Rakai, L.; Song, H.; Sun, S.; Zhang, W.; Yang, Y. Data association in multiple object tracking: A survey of recent techniques.
Expert Syst. Appl. 2022, 192, 116300.
38. Kaplan, L.; Bar-Shalom, Y.; Blair, W. Assignment costs for multiple sensor track-to-track association. IEEE Trans. Aerosp. Electron.
Syst. 2008, 44, 655–677.
39. Danchick, R.; Newnam, G.E. Reformulating Reid’s MHT Method with Generalised Murty K-best Ranked Linear Assignment
Algorithm. IET Radar Sonar Navig. 2006, 153, 13–22.
40. He, X.; Tharmarasa, R.; Pelletier, M.; Kirubarajan, T. Accurate Murty's Algorithm for Multitarget Top Hypothesis Extraction; IEEE:
Piscataway, NJ, USA, 2011; pp. 1–8.
41. Mallick, M.; Chang, K.; Arulampalam, S.; Yan, Y. Heterogeneous Track-to-Track Fusion in 3-D Using IRST Sensor and Air MTI
Radar. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 3062–3079.
42. Tian, X.; Yuan, T.; Bar-Shalom, Y. Track-to-Track Fusion in Linear and Nonlinear Systems; Springer: Berlin/Heidelberg, Germany,
2015; pp. 21–41.
43. Rahmathullah, A.S.; García-Fernández, Á.F.; Svensson, L. Generalized Optimal Sub-Pattern Assignment Metric. In Proceedings
of the 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017; pp. 1–8.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual au-
thor(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
... Lidar is typically employed in automobiles and offers good comprehensive performance, but it is costly. Millimeter-wave radar provides longer development potential in the field of traffic because of its advantages, which include its all-weather operation, compact size, low cost, and increased detection range [10][11][12][13][14]. ...
Article
Full-text available
Kalman filtering is a common filtering method for millimeter-wave traffic radars. The proposal is for an Adaptive Strong Tracking Extended Kalman Filter (EKF) algorithm that aims to address the issues of classic EKF’s low accuracy and lengthy convergence time. This method, which incorporates time-varying fading effects into the covariance matrix of the traditional EKF, is based on the ST algorithm. It allows the recalibration of the covariance matrix and precise filtering and state estimation of the target vehicle. By altering the fading and attenuating factors of the ST algorithm and using orthogonality principles, many fine-tuned fading factors produced from least-squares optimization are introduced together with regionally optimum attenuation factors. The results of Monte Carlo experiments indicate that the average velocity inaccuracy is reduced by at least 38% in comparison to existing counterparts. The results validate the efficacy of this methodology in observing vehicular movements in metropolitan regions, satisfying the prerequisites of millimeter-wave radar technology for traffic monitoring.
... While some of these prior works do indeed achieve noise/sideband suppression comparable to that of the designed IC, all entail either or both significant processing time intervals and/or unrealistic degrees of computational complexity, both of which ultimately preclude real-time, edgebased operation. Among sensor fusion techniques utilizing alternative modalities such as LiDAR and camera-based vision, several existing works addressing tracking in the presence of unpredictable target motion do indeed operate at real-timecompatible intervals on the order of 100 ms [36], [37]. However, such techniques entail some combination of ML-based detection procedures, extended/unscented Kalman filters, and the interacting multiple model algorithm, all demonstrated in back-end software only, with computational complexities significantly greater than that of the proposed approach. ...
Article
Full-text available
High-resolution mmWave radar represents an attractive sensing modality for precise, real-time sensing on resource-limited edge devices. Inherent parasitic platform motion such as vibration, however, significantly degrades perception resolution. Although existing methods of motion compensation have addressed this issue, such algorithms entail compute-heavy post-processing and thus would not be feasible for short-time-scale operation on a resource-constrained edge device. We present the first known IC-enabled solution for vibratory motion compensation, in the form of a mixed-signal, real-time deconvolution kernel generator chip. The custom IC, implemented in the open-source SkyWater 130-nm technology, functions as a sensor-fusion-based filter generation engine, operating in processing windows of less than 95 ms and fusing synchronized data from an auxiliary accelerometer to synthesize a kernel capable of correcting the received radar signals in real time. Measurement results demonstrate accurate kernel generation, while yielding an average spur suppression of 35 dB across a range of vibration frequencies and amplitudes and 26 dB across measured target velocity. Such performance, in conjunction with the demonstrated operation on short time windows of radar data, validate the chip’s potential for realistic deployment in real-time applications requiring high-resolution perception.
Article
Full-text available
This paper presents a test methodology to facilitate the measuring processes of LiDAR receiver ICs by avoiding the inherent walk error issue. In a typical LiDAR system, a costly laser diode driver emits narrow light pulses with fast rising edges, and the reflected pulses from targets enter an optical detector followed by an analog front-end (AFE) circuit. Then, the received signals pass through the cascaded amplifiers down to the time-to-digital converter (TDC) that can estimate the detection range. However, this relatively long signal journey leads to the significant decline of rising-edge slopes and the output pulse spreading, thus producing inherent walk errors in LiDAR receiver ICs. Compensation methods requiring complex algorithms and extra chip area have frequently been exploited to lessen the walk errors. In this paper, however, a simpler and lower-cost methodology is proposed to test LiDAR receiver ICs by employing a high-speed buffer and variable delay cells right before the TDC. With these circuits, both START and STOP pulses show very similar pulse shapes, thus effectively avoiding the walk error issue. Additionally, the time interval between two pulses is easily determined by varying the number of the delay cells. Test chips of the proposed receiver IC implemented in a 180-nm CMOS process successfully demonstrate easier and more accurate measurement results.
Article
Full-text available
The reliability of autonomous driving sensing systems impacts the overall safety of the driving system. However, perception system fault diagnosis is currently a weak area of research, with limited attention and solutions. In this paper, we present an information-fusion-based fault-diagnosis method for autonomous driving perception systems. To begin, we built an autonomous driving simulation scenario using PreScan software, which collects information from a single millimeter wave (MMW) radar and a single camera sensor. The photos are then identified and labeled via the convolutional neural network (CNN). Then, we fused the sensory inputs from a single MMW radar sensor and a single camera sensor in space and time and mapped the MMW radar points onto the camera image to obtain the region of interest (ROI). Lastly, we developed a method to use information from a single MMW radar to aid in diagnosing defects in a single camera sensor. As the simulation results show, for missing row/column pixel failure, the deviation typically falls between 34.11% and 99.84%, with a response time of 0.02 s to 1.6 s; for pixel shift faults, the deviation range is between 0.32% and 9.92%, with a response time of 0 s to 0.16 s; for target color loss, faults have a deviation range of 0.26% to 2.88% and a response time of 0 s to 0.05 s. These results prove the technology is effective in detecting sensor faults and issuing real-time fault alerts, providing a basis for designing and developing simpler and more user-friendly autonomous driving systems. Furthermore, this method illustrates the principles and methods of information fusion between camera and MMW radar sensors, establishing the foundation for creating more complicated autonomous driving systems.
Article
Full-text available
An FMCW LiDAR (frequency-modulated continuous-wave light detection and ranging) is a sensor that can measure distance using optical interference frequency (fb). This sensor has recently attracted interest because it is robust to harsh environmental conditions and sunlight due to the wave properties of the laser. Theoretically, when the frequency of the reference beam is linearly modulated, a constant fb is obtained with respect to the distance. However, when the frequency of the reference beam fails to be linearly modulated, the distance measurement is not accurate. In this work, linear frequency modulation control using frequency detection is proposed to improve the distance accuracy. The FVC (frequency to voltage converting) method is used to measure fb for high-speed frequency modulation control. The experimental results show that linear frequency modulation control using an FVC improves FMCW LiDAR performance in terms of control speed and frequency accuracy.
Article
Full-text available
This paper deals with the development of a localization methodology for autonomous vehicles using only a 3D LIDAR sensor. In the context of this paper, localizing a vehicle in a known 3D global map of the environment is equivalent to finding the vehicle’s global 3D pose (position and orientation), in addition to other vehicle states, within this map. Once localized, the problem of tracking uses the sequential LIDAR scans to continuously estimate the states of the vehicle. While the proposed scan matching-based particle filters can be used for both localization and tracking, in this paper, we emphasize only the localization problem. Particle filters are a well-known solution for robot/vehicle localization, but they become computationally prohibitive as the states and the number of particles increases. Further, computing the likelihood of a LIDAR scan for each particle is in itself a computationally expensive task, thus limiting the number of particles that can be used for real-time performance. To this end, a hybrid approach is proposed that combines the advantages of a particle filter with a global-local scan matching method to better inform the resampling stage of the particle filter. We also use a pre-computed likelihood grid to speed up the computation of LIDAR scan likelihoods. Using simulation data of real-world LIDAR scans from the KITTI datasets, we show the efficacy of the proposed approach.
Article
Full-text available
Penetration path planning for stealth unmanned aerial vehicles (SUAVs) in the integrated air defense system (IADS) has been a hot research topic in recent years. The present study examines penetration path planning in different threat environments. Firstly, for the complex terrain and static radar threats, a modified A-Star algorithm containing the bidirectional sector expansion and variable step search strategy is proposed to elude static threats rapidly. Then, with regard to bandit threats, the minimal radar cross-section (RCS) tactics are presented to achieve path replanning. Futhermore, the combinatorial methodology of the minimum RCS tactics and the modified A-Star algorithm is applied to achieve the dynamic path planning for SUAV. The simulation results indicate that the modified A-Star algorithm and minimal RCS tactics can significantly reduce the probability of radar system, which has better superiority in calculation efficiency, path cost and safety. And the minimal RCS tactics have better real-time performance and are more convenient in dealing with dynamic threats, which enhances the survivability of SUAV and verifies the effectiveness of the proposed methodology.
Article
Multi-drone multi-target tracking aims at collaboratively detecting and tracking targets across multiple drones and associating the identities of objects from different drones, which can overcome the shortcomings of single-drone object tracking. To address the critical challenges of identity association and target occlusion in multi-drone multi-target tracking tasks, we collect an occlusion-aware multi-drone multi-target tracking dataset named MDMT. It contains 88 video sequences with 39,678 frames, including 11,454 different IDs of persons, bicycles, and cars. The MDMT dataset comprises 2,204,620 bounding boxes, of which 543,444 bounding boxes contain target occlusions. We also design a multi-device target association score (MDA) as the evaluation criteria for the ability of cross-view target association in multi-device tracking. Furthermore, we propose a Multi-matching Identity Authentication network (MIA-Net) for the multi-drone multi-target tracking task. The local-global matching algorithm in MIA-Net discovers the topological relationship of targets across drones, efficiently solves the problem of cross-drone association, and effectively complements occluded targets with the advantage of multiple drone view mapping. Extensive experiments on the MDMT dataset validate the effectiveness of our proposed MIA-Net for the task of identity association and multi-object tracking with occlusions. The dataset and code are available at https://github.com/VisDrone/Multi-Drone-Multi-Object-Detection-and-Tracking https://github.com/VisDrone/Multi-Drone-Multi-Object-Detection-and-Tracking .
Article
Considering the perception of a driver, driving safety, and riding comfort, an adaptive cruise control (ACC) method based on a multi-objective (MO) real-time optimization control algorithm is proposed. The ACC system can be divided into various working modes according to the driver’s perception, as well as the relative state of the host vehicle and the preceding vehicle. To distinguish the driver’s intention in the rapid acceleration, strong deceleration, steady, and larger clearance modes, a fuzzy tool is proposed to solve the problem. To maintain the inter-clearance, velocity, and jerking action in a reasonable region, an MO control algorithm based on model predictive control was used to obtain the optimal control vectors. Moreover, the control vector increment is restricted to suppressing the fluctuation caused by switching among various modes. Simulations were conducted for various scenarios to verify the effectiveness of the ACC method. The simulation results indicated that the driver’s comfort and fuel economy improved by 35.3% and 16.6%, respectively, compared with the linear quadratic algorithm in a complicated driving environment. Various simulation results demonstrated that the MO-ACC controller can reduce fuel consumption while barely sacrificing riding comfort or tracking performance, as compared to the linear quadratic controller.
Article
This paper proposes a multi-object tracking (MOT) method called StrongFusionMOT which fuses the information of LiDAR and camera sensors. The major contributions of the proposed StrongFusionMOT are in three aspects. Firstly, in the detection fusion stage, the depth information extracted by means of AD-census is supplemented to 2D detections to facilitate fusion of 2D and 3D detections. This detection fusion pattern enhances fusion robustness and provides accurate fusion performance. Secondly, a new cost function design named SDIoU is proposed by taking into account not only the intersection between the two bounding boxes, but also their shapes and relative distances. This cost function eliminates the shortcomings of the existing IoU designs and greatly enhances association precision. Thirdly, a multi-frame matching mechanism that involves tracks in the past n frames is proposed for reappeared tracks, which effectively suppresses cumulative errors resulting from consecutive frames of track predictions and greatly enhances association robustness. The effectiveness of the proposed StrongFusionMOT is evaluated by means of comparative experiments with state-of-the-art MOT solutions using the KITTI dataset and the nuScenes dataset. Both quantitative and qualitative results demonstrate the superiorities of the proposed method in terms of various performance metrics.
Article
As automated vehicles are deployed across the world, it has become critically important to understand how these vehicles interact with each other, as well as with other conventional vehicles on the road. One such method to achieve a deeper understanding of the safety implications for Automated Vehicles (AVs) is to analyze instances where AVs were involved in crashes. Unfortunately, this poses a steep challenge to crash-scene investigators. It is virtually impossible to fully understand the factors that contributed to an AV involved crash without taking into account the vehicle’s perception and decision making. Furthermore, there is a tremendous amount of data that could provide insight into these crashes that is currently unused, as it also requires a deep understanding of the sensors and data management of the vehicle. To alleviate these problems, we propose a data pipeline that takes raw data from all on-board AV sensors such as LiDAR, radar, cameras, IMU’s, and GPS’s. We process this data into visual results that can be analyzed by crash scene investigators with no underlying knowledge of the vehicle’s perception system. To demonstrate the utility of this pipeline, we first analyze the latest information on AV crashes that have occurred in California and then select two crash scenarios that are analyzed in-depth using high-fidelity synthetic data generated from the automated vehicle simulator CARLA. The data visualization procedure is demonstrated on the real-world Kitti dataset by using the YOLO object detector and a monocular depth estimator called AdaBins. Depth from LIDAR is used as ground truth to calibrate and assess the effect of noise and errors in depth estimation. The visualization and data analysis from these scenarios clearly demonstrate the vast improvement in crash investigations that can be obtained from utilizing state-of-the-art sensing and perception systems used on AVs.
Article
Catheter tracking has become an integral part of interventional radiology. Over the last decades, researchers have significantly contributed to theoretical and technical catheter tracking solutions. However, most of the published work thus far focuses on a single application or a single tracking technology. This paper provides an exhaustive review of the state-of-the-art for catheter tracking in general by analyzing significant contributions in this field. We first present a historical overview that led to catheter tracking and continue with a survey of leading tracking technologies. These include image-based tracking, active and passive tracking, electromagnetic tracking, fiber optic shape sensing, bioelectric navigation, robotic tracking solutions and hybrid tracking. As for imaging modalities, the focus is on x-ray based modalities, ultrasound, and magnetic resonance imaging. Finally, we review each tracking technology with respect to the imaging modality and establish the relation between the two and the underlying anatomy of interest.