ArticlePDF AvailableLiterature Review

Recent Advances in mmWave-Radar-Based Sensing, Its Applications, and Machine Learning Techniques: A Review

Authors:

Abstract

Human gesture detection, obstacle detection, collision avoidance, parking aids, automotive driving, medical, meteorological, industrial, agriculture, defense, space, and other relevant fields have all benefited from recent advancements in mmWave radar sensor technology. A mmWave radar has several advantages that set it apart from other types of sensors. A mmWave radar can operate in bright, dazzling, or no-light conditions. A mmWave radar has better antenna miniaturization than other traditional radars, and it has better range resolution. However, as more data sets have been made available, there has been a significant increase in the potential for incorporating radar data into different machine learning methods for various applications. This review focuses on key performance metrics in mmWave-radar-based sensing, detailed applications, and machine learning techniques used with mmWave radar for a variety of tasks. This article starts out with a discussion of the various working bands of mmWave radars, then moves on to various types of mmWave radars and their key specifications, mmWave radar data interpretation, vast applications in various domains, and, in the end, a discussion of machine learning algorithms applied with radar data for various applications. Our review serves as a practical reference for beginners developing mmWave-radar-based applications by utilizing machine learning techniques.
Citation: Soumya, A.; Krishna
Mohan, C.; Cenkeramaddi, L.R.
Recent Advances in mmWave-Radar-
Based Sensing, Its Applications,
and Machine Learning Techniques:
A Review. Sensors 2023,23, 8901.
https://doi.org/10.3390/s23218901
Academic Editor: Antonio Lázaro
Received: 25 August 2023
Revised: 6 October 2023
Accepted: 21 October 2023
Published: 1 November 2023
Copyright: © 2023 by the authors.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license (https://
creativecommons.org/licenses/by/
4.0/).
sensors
Review
Recent Advances in mmWave-Radar-Based Sensing,
Its Applications, and Machine Learning Techniques: A Review
A. Soumya 1, C. Krishna Mohan 1and Linga Reddy Cenkeramaddi 2,*
1Department of Computer Science Engineering, Indian Institute of Technology, Hyderabad 502285, India;
cs21resch15003@iith.ac.in (A.S.); ckm@cse.iith.ac.in (C.K.M.)
2Department of Information and Communication Technology, University of Agder, 4879 Grimstad, Norway
*Correspondence: linga.cenkeramaddi@uia.no
Abstract:
Human gesture detection, obstacle detection, collision avoidance, parking aids, automotive
driving, medical, meteorological, industrial, agriculture, defense, space, and other relevant fields
have all benefited from recent advancements in mmWave radar sensor technology. A mmWave radar
has several advantages that set it apart from other types of sensors. A mmWave radar can operate
in bright, dazzling, or no-light conditions. A mmWave radar has better antenna miniaturization
than other traditional radars, and it has better range resolution. However, as more data sets have
been made available, there has been a significant increase in the potential for incorporating radar
data into different machine learning methods for various applications. This review focuses on key
performance metrics in mmWave-radar-based sensing, detailed applications, and machine learning
techniques used with mmWave radar for a variety of tasks. This article starts out with a discussion of
the various working bands of mmWave radars, then moves on to various types of mmWave radars
and their key specifications, mmWave radar data interpretation, vast applications in various domains,
and, in the end, a discussion of machine learning algorithms applied with radar data for various
applications. Our review serves as a practical reference for beginners developing mmWave-radar-
based applications by utilizing machine learning techniques.
Keywords:
mmWave radar; mmWave radar applications; machine learning; industrial applications;
medical applications; automotive applications; military applications; computer vision
1. Introduction
The development of millimeter wave (mmWave) radar sensors during the past ten
years has been spurred on by numerous research applications, including civilian and
non-civilian applications [
1
,
2
]. With the latest improvements in chip technology and
lowered cost, the mmWave radar sensor has gained widespread popularity in a wide range
of applications. The mmWave radar system includes a transmitting antenna, a receiving
antenna, and a signal processing system to determine an object’s dynamic information, such
as range, velocity, and angle of arrival (AoA). The mmWave radar transmits a mmWave
signal into space by striking an object, and this signal gets reflected. The receiving antenna
captures the echo signal, which is then mixed with a transmitting signal to obtain an
intermediate-frequency (IF) signal. This IF signal is processed to obtain object information.
Various mmWave radars and working bands are shown in Table 1. mmWave radars operate
in the frequency range between 24 GHz and 300 GHz. Processing IF signals allow for the
measurement of an object’s range, velocity, and angle of arrival (AoA) [3].
Sensors 2023,23, 8901. https://doi.org/10.3390/s23218901 https://www.mdpi.com/journal/sensors
Sensors 2023,23, 8901 2 of 25
Table 1. Operational frequencies within the bandwidth spectrum.
Working Band (GHz) Bandwidth (GHz) Resolution
24 0.25 GHz 60 cm
77 4 GHz 3.75 cm
60 7 GHz 2.1 cm
94 2 GHz 7.5 cm
100 9 GHz 1.6 cm
300 40 GHz 0.375 cm
300 20 GHz 0.75 cm
The performance of signal processing is continually being improved along with hard-
ware components. The performance of mmWave-radar-based sensing is entirely based on
detecting mmWaves reflected by objects and subsequent signal processing. The perfor-
mance of the mmwave radar is independent of external lighting conditions and works well
even in low-light conditions and dazzling light. mmWave radar sensors are now widely
used in a variety of civilian applications, including obstacle detection, motion recognition,
localization, and tracking, owing to low-cost chip technology and improved reliability [
4
].
As a result, these improvements in radar technology and digital signal processing lead to
good accuracy in range and velocity estimation and better resolution in contrast with other
traditional radars. In addition to the benefits listed above, mmWave radar has superior
penetration capacity through various weather conditions like rain, fog, and snow.
Today’s radar-based sensing is more diverse. Applications range from civilian to
military and include a variety of automotive, industrial, and medical applications. Research
in the field is being refined as a result of technological advancements and more accurate
detection. To perceive the surrounding environment, mmWave radar sensors can easily
be integrated with other imaging sensors. Multi-sensor fusion effectively uses data from
multiple sensors to augment one another and improves the ability to extract detailed
information about targets under a variety of environmental and climatic conditions [
3
].
The mmWave radar is more often used in various applications when compared to other
sensors, such as RGB cameras, ultrasonic sensors, infrared sensors, and light detection and
ranging (LiDAR). Various applications demand the fusion of one or two sensors along with
mmWave radar [5].
The advantages mentioned above have increased the use and popularity of mmWave
radars. This article focuses on different state-of-the-art mmWave radars with key technical
specifications. The main focus of our study is on state-of-the-art mmWave radars, their
key performance metrics along with suitable applications, key measurements and their
interpretation, mmWave radar frequency selection criteria based on specifications and
application requirements, and machine learning techniques for mmWave-radar-based
sensing. Our review stands out as unique from other reviews due to its application focus
on the use of machine learning methods and provides a quick review of mmWave radar
models with specifications. The main significant aspects of this paper are highlighted in
Table 2, presenting how it differs from earlier surveys.
The remainder of this article is organized as follows: Section 2focuses on the per-
formance metrics of mmWave radar sensing. Section 3presents the measurements and
interpretation of the mmWave radar sensor. Various mmWave-radar-based applications
and machine learning techniques are focused on in Section 4. Finally, Section 5concludes
the article.
Sensors 2023,23, 8901 3 of 25
Table 2. Comparison of our work with earlier review articles.
Reference Year Focus Area SP MI Appl PM ML/DL
Tech. Comments
[6] 2011
Different mmWave technology specifications
7 7 3 7 7 Restricted to limited technologies
and data processing
[7] 2017
Focus on architecture, estimation techniques
3333 7 Restricted to advanced driver
assistance system
[4] 2019 Review of digital modulation and
interference mitigation methods 3 3 3 7 7 Restricted to automated
driving technology
[5] 2020
Working principles of vision sensors and
performance parameters for autonomous
systems
3 7 7 3 3 Restricted to autonomous systems
[8] 2021 Uses deep learning and fusion models 3 7 3 3 3 Deep fusion operations and
datasets need to be improved
[3] 2022 To determine working principles, data
representation methods, and challenges 3 7 3 3 7 Restricted to detection applications
[1] 2022
On various on-board sensors, hardware
components, software environments, and
machine learning algorithms
7373 3 Restricted to unmanned aerial
vehicle applications
Our Work 2022
Performance metrics of mmWave radar,
state-of-the-art mmWave radar models
available on the market, radar data
interpretation, applications of mmWave
radars, machine learning techniques for
mmWave radar
3333 3
Considers sensing applications
using mmWave radar sensor in
broad areas of science
and engineering
SP
—signal processing;
MI
—measurement interpretation;
Appl
—applications;
PM
—performance metrics;
ML/DL
Tech.—machine learning or deep learning techniques.
2. Performance Metrics in mmWave-Radar-Based Sensing
Currently, mmWave radars have become ubiquitous, with a wide variety of applica-
tions. The types of components in mmWave radars are shown in Figure 1. The synthesizer
generates the chirp, the transmitter transmits this chirp signal, the transmitted chirp is
reflected off the objects in front of the radar, and the reflected signal is received via the
receiver antenna. The RX signal and TX signal are mixed to generate the resultant IF signal.
By processing the IF signal, all the key parameters can be estimated. When multiple objects
are in front of the radar, one Tx chirp generates multiple reflected chirps from different
objects, and the IF signal will have multiple tones corresponding to each of the reflections.
The purpose of the analog-to-digital converter (ADC) is to digitize the IF signal. Fast
Fourier transform (FFT) is performed on this digitized IF signal to obtain the range profile.
Early radar systems were primarily utilized for navigation and object detection at
short ranges. They were used in maritime navigation, for example, to detect other vessels or
impediments in foggy circumstances. However, advancements in semiconductor technol-
ogy and processing led to advancements in the development of advanced radar systems [
7
].
The mmWave radar provides better antenna miniaturization than other traditional radars.
Broadband radio frequency (RF) signals allow for the better resolution capability of the
radar, which is the key to many high-performance automotive, medical, and industrial
applications. Additionally, their enhanced portability and affordable chip have made them
suitable for usage in various day-to-day applications.
A mmWave frequency-modulated continuous wave (FMCW) radar accurately esti-
mates the range and velocity of multiple targets from the radar sensor without the need
for more transceivers [
9
]. For the accurate estimation of the range and velocity of multiple
targets, one TX antenna and one RX antenna are sufficient. However, the angle of arrival
(AoA) estimation of multiple targets demands more hardware resources, such as more
TX antennas and more RX antennas. The greater the number of transceiver antennas, the
better the AoA estimation performance will be. However, while detecting objects using
one mmWave radar, signals from other nearby radars cause interference, which eventually
degrades the performance of the mmWave radar. However, there are some mitigation
techniques, though it is still challenging, and this is still an ongoing research topic.
Sensors 2023,23, 8901 4 of 25
To select the suitable mmWave radar based on the application, the following perfor-
mance metrics play a crucial role [10,11]:
Range —The range is estimated by analyzing the frequency content in the IF signal. As
shown in Figure 2a, transmitted and received chirps as a function of time for a single
object detected. It can be observed that the received chirp is a time-delay (
τ
) version
of the transmitted chirp.
The time delay can be measured as:
(τ) = 2R/C
where Ris the distance to the detected object and Cis the speed of light.
The initial phase of the IF signal is φ0
(φ0) = 4πR/λ
The range is computed as:
R=C×fIF /2S
The maximum range is decided via the sampling frequency of this IF signal. The larger
the sampling frequency, the better the maximum range capacity of the radar will be.
The other deciding factor for the maximum range is the transmission power. There is
no hardware constraint on the accurate range estimation of multiple targets. One TX
antenna and one RX antenna are sufficient.
Rmax =FsC/2S
Another fundamental limitation for maximum range comes from the transmitting
power:
Rmax = ((σPtGTX GRX λ2Tmeas)/(4π3SNRminkTF) ) 1
4
where
Pt
is output power of device;
GTX
is the TX antenna gain,
GRX
is the RX
antenna fain;
σ
is the radar cross-section of the target (RCS);
Tmeas
is total measurement
time; SNR is the signal-to-noise ratio;
k
is the Boltzmann constant; and
T
is antenna
temperature.
Velocity—mmWave radar estimates the velocity of multiple targets using the phase
difference between IF signals, as illustrated in Figure 2b. There is no hardware
constraint for the accurate velocity estimation of the multiple targets.
One TX
antenna
and one RX antenna are sufficient.
The phase difference is derived as
∆Φ
= 4
πVTc
/
λ
, where
Tc
is the chirp duration
time. The velocity is computed as:
V=λ∆Φ/4πTc
The measurement is unambiguous only if |
∆Φ
| <
π
. We can derive that vs. <
λ
/4
Tc
.
The maximum possible velocity estimation depends on how fast chirps can be trans-
mitted.
Vmax =λ/4Tc
Angle of arrival—Estimating the angle of arrival of one object requires at least one
TX antenna and two RX antennas as shown in Figure 3. The greater the number of
RX antennas, the better the AoA estimation performance will be. By using MIMO,
more transceiver elements can be made with a limited number of TX and RX antennas.
However, accurate AoA estimation of multiple targets demands more hardware
resources, such as more TX and RX antennas.
The phase difference between the IF signals of the two receivers is derived as:
(∆Φ) = 2πR/λ
Sensors 2023,23, 8901 5 of 25
Under the assumption of a planar wavefront basic geometry
R = d
sin(θ
), where d
is the distance between the receiving antennas. The angle of arrival
θ
is computed
from φ
(θ) = sin((λ Φ)/(2πd))
The unambiguous measurement of the angle of arrival requires that (∆Φ) < π.
This leads to:
(θmax ) = sin((λ)/(2d))
Range resolution—The shortest distance at which two objects can become close while
still being detected as two distinct objects via radar. The smaller this distance, the
better the resolving capability of the radar will be. A radar with greater RF bandwidth
gives better range resolution.
The smallest frequency differences of an IF signal are related to the chirp duration,
Tc
.
(f)>1
Tc
where Tcis the observation interval or chirp duration.
(since(f) = S2R
c)
(R)>c/2STc=c/2B(sinceB =STc)
Range resolution depends only on the RF bandwidth swept by the chirp.
Range resolution =C/(2B)
where C is the speed of light, and B is the radar bandwidth. For example, a bandwidth
of 4 GHz gives a range resolution of 3.75 cm.
Velocity resolution—The smallest velocity difference the two targets can have while still
being detected via radar as two distinct targets with two distinct velocities. Velocity
resolution can be improved by increasing the frame time. A frame consists of a number
of series of chirps.
(∆Φ) = 4πV Tc/λ
One can mathematically derive the velocity resolution (Vres) if the frame period
Tf= NTc.
V>Vres = λ/(2 Tf)
Vres =λ/(2Tf)
where Tfis the frame time.
Angle of arrival resolution—The angle of arrival resolution is the smallest angle that can
be formed between two targets and radar while still being detected via the radar as
two distinct targets. The smaller the angle of arrival resolution, the better the radar’s
resolving capability. Extending the number of antennas (both TX and RX) improves
the AoA resolution.
(θ(res)) = 2/NRX
All the performance parameters, such as range, velocity, angle of arrival, range res-
olution, velocity resolution, AoA resolution, maximum range, maximum velocity, and
maximum AoA, are analytically tabulated in Table 3.
Sensors 2023,23, 8901 6 of 25
Figure 1. Architecture of the mmWave radar sensor.
Figure 2. (a) IF signal and (b) chirp frame [12].
Figure 3.
(
a
) Two antennas are required to estimate AoA and (
b
) maximum angular field of view [
12
].
Table 3. Performance metrics of mmWave radar.
Range Velocity Angle of arrival
R=(C×fIF )/2SV= λ∆Φ/4πTcθ= sin((λ∆Φ )/(2πd))
Range resolution Velocity resolution Angle of arrival resolution
dres = C/2BVres = λ/(2 Tf)θres =2/NRX
Max range Max velocity Max angle of arrival
dmax = FsC/2SVmax= λ/4 Tcθmax= sin((λ/2 d))
R
—target’s range from the radar;
V
—target’s Velocity;
θ
—target’s angle of arrival;
C
—speed of light;
S
—slope of
chirp, B/
Tc
;
φ
—phase difference between IF signals;
Tf
—frame time,
NTC
;
B
—chirp RF bandwidth;
fIF
—IF
signal frequency;
Fs
—sampling rate of the IF signal;
Tc
—chirp duration time;
λ
—wavelength of the chirp signal;
N—number of chirps; NRX —number of receiving antennas; d—distance between receiving antennas.
Sensors 2023,23, 8901 7 of 25
The following are important factors to take into account when choosing a mmWave
radar [10,11]:
Sensor type—The RF bandwidth, IF bandwidth, ADC sampling rate, range resolution,
velocity resolution, and AoA resolution are important key factors in deciding the
performance. mmWave radar sensors are broadly categorized into three sensor types:
long-range, medium-range, and short-range radars, as listed in Table 4. The appropriate
sensor should be chosen depending on the application demands and needs.
Frequency and bandwidth—It is advantageous to have a high-frequency sensor; it uses
a low antenna size and gives a better angular resolution. High bandwidth offers a
high-range resolution. A 77 GHz frequency radar with a 4 GHz bandwidth gives a
range resolution of 3.75 cm. Popular mmWave radar models with specifications are
listed in Table 5.
Table 4. Suitable mmWave radar types with respect to the applications.
Type Range (m) Bandwidth
(MHz)
Azimuth View Angle
(deg.)
Elevation View Angle
(deg.) Applications
Long-range radar 10–250 600 ±15 ±5 Autonomous cruise control
Medium-range radar 1–100 600 ±40 ±5Lane change assistance
system, collision mitigation
Short-range radar 1–30 4000 ±80 ±10 Blind spot detection, parking
assistance
Table 5. Popular mmWave radar models and their specifications.
Reference Model Name Range
(m)
Working Band
(GHz)
Azimuth Field
of View (deg)
Elevation
Field of
View (deg)
Chip
Memory
(MB)
User Interface and Connectivity
[13] TI-IWR6843AOPEVM 180 60 GHz–64GHz ±120 ±120 1.75 TMMWAVEICBOOST, DCA1000, I2C,
LVDS, QSPI, SPI, UART
[14] TI-IWR1843 180 76 GHz–81 GHz 100 ±15 2 CAN, LVDS, QSPI, I2C, SPI, UART,
CSI-2
[15] Delphi ESR 174 76 GHz ±10 ±45 - CAN
[16] NavTech CIR204-h 200 76 GHz–77 GHz 360 - - 1 Gbps Ethernet
[17] TI-AWRL6432 250 57 GHz–64 GHz ±18 - 1 QSPI, PWM, I2C
[18] BoschLRR3 250 77 GHz 30 - - CAN
[19] Continental ARS408-21 250 77 GHz ±9 14 - CAN 500 kbps
[20] Continental SRR600 >180 76 GHz–81 GHz ±90 ±40 - Ethernet, CAN-FD
[21] TI-AWR1843AOPEVM 150 76 GHz–81 GHz 140 140 2 DCA1000EVM, CAN, CAN-FD
[22] TI-AWR1642 100 76 GHz–81 GHz ±60 ±10 1.5 CAN, CAN-FD, SPI, I2C, UART
[23] TI-IWR1642 150 76 GHz–77 GHz ±100 ±15 1.5 CAN, CAN-FD, QSPI
[24] NXP-TEF810X 250 76 GHz–81 GHz ±18 - 1.5 LVDS, CSI2
[25] NXP-SAF85xx 250 76 GHz–81 GHz ±18 - 5.5 SGMII Ethernet, dual CAN FD
[26] NXP-TEF82xx 250 76 GHz–81 GHz ±18 - 0.576 CSI-2, LVDS
[27] Continental-ARS540 300 76 GHz–81 GHz ±60 ±60 1 CAN, Ethernet
[23] TI-IWR1443 60 77 GHz–79 GHz ±65 ±15 1.5 LVDS, DCAN, QSPI, CSI2
[28] TI-AWR1443 60 76 GHz–81 GHz ±65 ±15 1.5 LVDS, DCAN, QSPI
[29] TI-IWRL6432 60 57 GHz–64 GHz ±65 ±15 1 I2C, SPI, UART, QSPI
[30] Continental-ARS4-A 250 77 GHz ±75 ±20 1 CAN, Ethernet
[31] TI-AWR2243BOOST 150 76 GHz–81 GHz ±90 ±90 1.5 SPI, UART, I2C, CAN-FD
[32] TI-AWR1243 250 76 GHz–81 GHz ±60 ±15 1 SPI, MIPI-CSI2, UART
[33] NXP4D-S32R45 300 77 GHz ±56 ±56 1 PCIe, Ethernet, CAN-FD
[34] RDK-S32R274 200 79–81 GHz ±30 ±30 1 Ethernet, CAN-FD
Sensors 2023,23, 8901 8 of 25
Accessories—As the needs of mmWave radars change, upgrading the firmware and
software becomes necessary. Hence, choosing a manufacturer that can offer stable
software updates is important. mmWave radars have been integrated with Lidars
in [
35
,
36
] to provide better results; additionally, radars fused with cameras or infrared
sensors are studied in [
37
39
]. Texas Instruments (TI) introduced a commercial radar,
TDA3x, board for radar camera fusion to provide effective tracking and detection
applied in [
40
]. There is a discussion of fusion techniques in [
41
,
42
], such as low-level
and feature-level fusion.
3. Measurements from mmWave Radar Sensor and Interpretation of the Data
3.1. Range Profile
A sample range profile measured using the millimeter wave radar sensor is illustrated
in Figure 4a. Taking the fast Fourier transform (FFT) of an intermediate-frequency signal
yields the range profile. This range profile depicts the relative reflected power from the
targets as a function of range. The target is indicated via the peaks in the range profile.
Targets with a large cross-section reflect more power, producing a stronger peak. Targets
located near the radar produce a strong peak in the range profile.
Figure 4. (a) Range profile and (b) range–Doppler heatmap.
A range–Doppler heatmap displays multiple targets with their speeds as a function
of range, as shown in Figure 4b. Stationary targets have zero Doppler while moving
targets have a range–Doppler map with non-zero Doppler values. This essentially provides
dynamic information about the targets, such as their velocity and range.
3.2. Range–Azimuth Heatmap
The range–azimuth heatmap in Figure 5a is intended to display radar cube information
corresponding to a zero Doppler bin for every combination of range and angle bin. It
essentially gives the location information about the targets with respect to radar in a
Cartesian coordinate system. The colored signal power representation within the plot
shows the range and angle coordinate points where the targets are located.
3.3. Three-dimensional Scatter Plot
The detected targets are displayed in 3D space by selecting a non-zero elevation
resolution as the antenna configuration, as shown in Figure 5b.
Sensors 2023,23, 8901 9 of 25
Figure 5. (a) Range–azimuth heatmap and (b) 3D scatter plot.
4. Applications of mmWave Radars and Machine Learning Techniques
mmWave radars are used in a variety of applications, as illustrated in Figure 6. These
include automotive applications, industrial applications, military applications, medical
applications, robotics and automation applications, civilian applications, and security and
surveillance applications.
Figure 6. Applications of mmWave radar sensing.
Sensors 2023,23, 8901 10 of 25
4.1. Automotive Applications
Automotive applications use mmWave radar sensors to accurately localize and mea-
sure the radial range, velocity, and AoA of moving objects. The applications include
adaptive cruise control, autonomous emergency braking, blind spot detection, lane change
assistance, front cross-traffic alert, rear cross-traffic alert, automated parking, body/chassis
applications, and in-cabin applications. The relative positioning of vehicles has been
estimated using mmWave radar in [
43
]. Vehicle detection in advanced driving assistant sys-
tems using automotive radar with range–azimuth–Doppler dimensions is studied in [
44
].
A 2D car detection system for autonomous driving applications is studied in [
45
]. The
ability of an autonomous vehicle to perceive and comprehend its surroundings is studied
in [
46
]. The use of mmWave radars for vehicle detection in self-driving applications is
studied in [
47
]. The use of mmWave radar sensor and vision sensor fusion for obsta-
cle detection in autonomous driving is studied in [
48
]. To obtain high accuracy in new
advanced driver-assistance systems (ADASs), mmWave radars have been used in vehi-
cles [
49
], as shown in Figure 7. A 60 GHz mmWave radar has been used to reduce driver
distractions with real-time hand gesture recognition instead of touchscreens and wearable
components [
50
]. Machine-learning-based hand gesture recognition is studied in [
51
] using
CNN and LSTM. People occupancy detection in a vehicle has been implemented utilizing
mmWave radar [
52
]. Furthermore, contactless non-intrusive vehicle occupant detection
is studied in [
53
], and autonomous navigation by predicting vehicle location utilizing
mmWave radar is studied in [
54
]. Radar sensors placed at the front and rear corner of the
car that form beams for front and rear blind spot detection, as well as for cross-traffic alert,
are demonstrated in [
55
]. mmWave radar became a solution for ground-based traffic moni-
toring, and the management of both terrestrial and aerial vehicles using angle estimations
in [
56
]. The detailed automotive applications and associated radar details are tabulated in
Table 6.
Figure 7. Advanced driver-assistance system [57].
Sensors 2023,23, 8901 11 of 25
Table 6. Automotive applications using popular radars.
Reference Year Application Radar Used
[58] 1997 Intelligent cruise control with collision warning FMCW (76 GHz–77 GHz)
[59] 2017 Blind spot detection and warning system AWR1843 (76 GHz–77 GHz)
[60] 2017 Automated emergency breaking TI-AWR1243 (76 GHz–78 GHz)
[60] 2017 In-car occupant detection TI-AWR1642 (76 GHz–81 GHz)
[61] 2017 Driver vital sign monitoring TI-AWR1642 (77 GHz)
[62] 2018
Automotive body and chassis sensing applications
TI-AWR1642 (77 GHz)
[50] 2018 In-car controlling with gestures FMCW-mmWave (60 GHz)
[63] 2019 Automated parking system TI-AWR1843 (77 GHz)
[64] 2020 Lane change assistance with obstacle detection TI-AWR1843AOPEVM (77 GHz)
[65] 2020 Parking assistance with obstacle detection TI-AWR1642BOOST (77 GHz–81 GHz)
[66] 2020 Debris detection for automotive radar mmWave (76 GHz–81 GHz)
[67] 2021 Automotive vehicle detection in parking lot
TI AWR2243BOOST-MIMO (76 GHz–81 GHz)
[65] 2022 Motor cycle safety and Blind spot detection TI-AWR1843AOP (76 GHz–81 GHz)
[55] 2022 Automotive corner radar for cross traffic alert TI-AWR1843EVM (76 GHz–81 GHz)
4.2. Industrial Applications
mmWave radars have become popular as they provide precise measurements that
are useful in industrial applications. Industrial applications include level sensing of fluids,
volume identification for solids, infrastructure systems, surface quality assessment in
production industries, and vibration monitoring. Utilizing mmWave radar for automatic
crack detection to distinguish between cracked and uncracked ceramic tiles and for quality
control of packaged ceramic tiles is presented in [
68
]. In industrial processes, to have
control over the usage of liquid and identify leakages, mmWave radars have been utilized
for accurately measuring fluid levels in tanks, as presented in [
69
]. In [
70
], a framework
for accurate material identification for six different materials using mmWave radar is
presented. In addition, the volume of the materials has been determined. Using low-power
transmission signals and reflections with a non-line-of-sight (NLOS) method for detecting
moving objects has been studied in [
71
]. This study includes a model for the echo signal
of the NLOS target by considering the multipath effect and the weak target echo signal
issues. The detection and classification of gases and aerosols have been implemented in [
72
].
Detecting the vibrational target objects by modifying shaking frequencies and assessing
the performance is studied in [
10
]. The monitoring of the mass flow of pneumatically
transported bulk materials using mmWave measuring is presented in [
73
]. mmWave radars
are useful in metal production industries, where they require the precise measurement of
slabs of copper, steel, and aluminum in production. In rolling mills, mmWave radar sensor
technology is useful to provide accurate measurements, even in smoky, hot, steamy, and
dusty conditions, as shown in Figure 8. The detailed industrial applications along with the
radar utilized are tabulated in Table 7.
Sensors 2023,23, 8901 12 of 25
Figure 8. Width measurement in a cold and hot rolling mill [74].
Table 7. Industrial applications using popular radars.
Reference Year Application Radar Used
[75] 2006 Surface sensing mmWave sensor (29.72 GHz–37.7 GHz)
[76] 2013
Measuring the liquid level and interface sensing
mmWave Doppler sensor (77 GHz)
[68] 2015 Crack detection in ceramic tiles V-Band Imaging Radar (60 GHz)
[69] 2017 Fluid level sensing TI-IWR1443 (77 GHz)
[77] 2018 Material classification
FMCW radr with Infineon’s DEMO-BGT60TR24
sensor (60 GHz)
[78] 2018 Motion detection and intersection monitoring IWR6843 60 GHz radar
[79] 2019 Foam detection in chemical applications IC with mmWave ssensor (80 GHz)
[10] 2020 Obtaining the performance on detecting
vibrational targets FMCW 80 GHz sensor integrated on SiGe chip
[80] 2022 Eavesdropping and spying on phone calls TI-AWR1843BOOST (77 GHz)
[70] 2022 Material identification TI-IWR1443 FMCW (77 GHz–81 GHz)
4.3. Medical Applications
Medical applications for mmWave radars have also gained importance due to their
sensitive detection capability and the penetration of mmWave signals in biological tissues.
Using mmWave radar, various glucose concentration levels in blood samples to distinguish
the healthy or diabetic have been studied in [
81
]. In [
82
], contactless breathing rate and
heart rate monitoring of patients using mmWave were implemented, as shown in Figure 9.
mmWave radar sensors have been used for monitoring vital signs via non-contact means in
a robust way [
83
]. The use of mmWave radar in real-time human motion behavior detection
is presented in [
84
]. The recognition of multiple patient behaviors has been simultaneously
studied utilizing mmWave radars in [
85
]. The real-time detection and tracking of human
skeletal positions for patient monitoring is studied in [
86
]. Considering body movements as
micro-motion parameters for real-time fitness tracking via non-contact means has been stud-
ied in [
87
]; additionally, fitness tracking by classifying and counting exercises is presented
in [
88
]. The detection of sleeping pose identification utilizing mmWave radar has been
implemented in [
89
]. For skin diagnosis applications, mmWave radar has been utilized
in [
90
]. The utilization of a wearable radar sensor for continuous blood pressure monitoring
is presented in [
91
]. Detailed medical applications utilizing mmWave radars are tabulated
in Table 8.
Sensors 2023,23, 8901 13 of 25
Figure 9. Contactless patient monitoring application [82].
Table 8. Medical applications with popular radars.
Reference Year Application Radar Used
[81] 2018 Blood glucose level detection FMCW-XENSIV (60 GHz)
[85] 2019 Multiple patients behavior detection TI-AWR1642BOOST (77 GHz)
[90] 2020 Skin cancer detection Designed sensor (77 GHz)
[87] 2021 Contactless fitness tracking TI-IWR1642 (77 GHz–81 GHz)
[82] 2022
Contactless monitoring of patients and elderly people alone
IWR6843AOPEVM (60 GHz–64 GHz)
[92] 2022 Measuring systolic blood pressure TI-IWR6843AOP (60 GHz–64 GHz)
[93] 2022 Vital sign measuring TI-IWR1443 (77 GHz–81 GHz)
[94] 2022 Health monitoring with posture estimation TI-IWR6843 (60 GHz–64 GHz)
[95] 2022 Blood pressure monitoring TI-AWR1843 (77 GHz–81 GHz)
[96] 2022 Cardiorespiratory rate monitoring Commercial FMCW (122 GHz)
[97] 2022 Galvanic skin test to assess mental acuity and stress levels TI-AWR1843 (77 GHz)
[98] 2022 Automated heart rate and breathing rate monitoring TI-AWR1443BOOST (77 GHz)
4.4. Robotics and Automation Applications
Robotics applications include both indoor and outdoor environments. Detecting trans-
parent objects such as glass walls is very important in autonomous navigation. mmWave
radars reliably detect glass walls [
99
]. They are also quite reliable as ground-speed radars in
agricultural and warehouse robots [
99
]. Using the same ground-speed radar, it is possible
to sense the surface edges if the radar is mounted in front of it, facing toward the ground.
Safeguards around robotic arms are another important field wherein mmWave radar plays
a vital role. Mapping and navigation is another important application where mmWave
radars are used in indoor environments. Robotic applications for human path tracking
and collision avoidance are explored utilizing mmWave radars [
100
]. The detection of
obstacles and avoiding collision in a 360-degree path in robotics using mmWave radars
is demonstrated in [
101
], as shown in Figure 10. Incorporating an antenna on package
sensors with a wider field of view in both azimuth and elevation helps in the intelligent
sensing of transparent objects and dark objects, which is studied in [
102
]. Glass walls and
the materials behind them can be detected using mmWave radar sensors [
99
]. Detailed
robotics and automation applications and associated radar details are tabulated in Table 9.
Table 9. Robotics and automation applications with popular radars.
Reference Year Application Radar Used
[102] 2019 Intelligent robot for transparent object sensing IWR6843 (60 GHz)
[103] 2020
Robot-mounted mmWave radar for tracking heart rate
IWR6843 (62 GHz)
[54] 2020 Predicting autonomous robot navigation FMCW (77 GHz)
[101] 2020 Collision detection and avoidance IWR6843 (60 GHz)
[104] 2020 mmWave radars as safe guard robots IWR6843 (60 GHz)
[100] 2021 Automated indoor navigation and path tracking AWR6843 (77 GHz)
[99] 2021 Glass wall and partition detection IWR1443BOOSTEVM (77 GHz)
Sensors 2023,23, 8901 14 of 25
Figure 10. Robotics application [100].
4.5. Security and Surveillance Applications
mmWave radars are useful in security and surveillance applications because they
can detect moving objects and obstacles in low-light conditions. In particular, personal
screening and maintaining security aspects are discussed in [
105
]. mmWave radars are
being used in air traffic control systems and low-altitude space surveillance applications
to detect and display the position of aerial vehicles. Aerial vehicle activity monitoring
with radar range and angle measurements is studied in [
106
]. Airborne radars for obstacle
avoidance, landing aids, automotive radars for collision avoidance, and driving safety
support are studied in [
107
]. In [
108
], airborne surveillance with navigational aid on the
ground using mmWave synthetic aperture radar (SAR) is implemented, which tracks the
actual flight route and records it. Unmanned aerial vehicle (UAV) detection with respect
to a range of up to 40 m using low-cost mmWave sensors is reported in [
2
]. In airports
and other sensitive places, mmWave radars are utilized in the identification of intrusions
in [
109
]. Vehicle detection and tracking in traffic monitoring applications with a range
greater than 100 m are shown in Figure 11. A richer radar point cloud representation for a
traffic monitoring scenario is shown in [110]. People tracking using radar applications for
consumers in indoor and outdoor environments is presented in [
111
]. Furthermore, channel
tracking for a vehicular communication system is studied in [
112
]. Detailed security and
surveillance and civilian applications and associated radar details are tabulated
in Table 10.
Sensors 2023,23, 8901 15 of 25
Figure 11. Traffic monitoring application [113].
Table 10. Security and surveillance and civilian applications with popular radars.
Reference Year Application Radar Used
[114] 2006 Power line prediction in helicopter rescue mmWave radar (94 GHz)
[115] 2008 mmWave radars for safe helicopter landing Radar module with 94 GHz
[116] 2010 Providing indoor security of short range mmWave SAR (77 GHz)
[117] 2010 Debris detection on airport runways mmWave radar (73 GHz–80 GHz)
[118] 2013 Concealed threat detection W-band (75 GHz–110 GHz)
[108] 2015 Surveillance imaging applications MIRANDA radar (35 GHz and 94 GHz)
[113] 2018 Traffic monitoring IWR1642EVM 77 GHz radar
[119] 2019 Human target detection, classification,
tracking ISM band (24 GHz MIMIC)
[120] 2020 Tracking of malicious and hidden drones mmWave (77 GHz)
[121] 2021 Ego-motion estimating in indoor
environments TI-AWR1843BOOST (76 GHz–81 GHz)
[2] 2021 Unmanned aircraft system detection and
localization AWR1843 Boost (76 GHz–81 GHz)
[106] 2021 Aerial vehicle locating and air traffic
management AWR1843 (76 GHz–79 GHz)
[122] 2021 3D human skeletal pose estimation TI-AWR1843 (77 GHz)
[123] 2023 Indoor positioning system IWR6843ISK (60 GHz–64 GHz)
4.6. Civilian Applications
mmWave radars have grown in popularity because they are robust to adverse weather
conditions and find many uses in the civilian sector. The creation of a drone setup by
integrating a mmWave sensor to detect power lines up to a 40-m range with improved
performance and fast detection is shown in Figure 12. Debris detection on airport runways,
early risk warnings for helicopters, power line detection in flying paths, and malicious
drone detection are some of the applications of mmWave radars. Detecting small foreign
Sensors 2023,23, 8901 16 of 25
objects on airport runways for a safe landing is studied in [
117
], as shown in Figure 13.
mmWave radars reliably detect high-voltage invisible power lines in snowy and inclement
weather. Risk avoidance with early warnings from mmWave sensors to rescue helicopters
is discussed in [114].
Figure 12. Power line communication [124].
Figure 13. Debris detection on airport runways [125].
In [
120
], a cooperative radar sensing network is implemented for tracking small,
hidden, unauthorized unmanned aerial vehicles (UAVs). Micro-UAV detection for defense
applications using 24 GHz mmWave radars is presented in [
126
]. The classification of birds
and drones using radar micro-Doppler signatures is presented in [
127
]. The detection and
mitigation of GPS spoofing for drones is explored in [
128
]. Aircraft runway extraction in
low-visibility conditions for a safe landing is investigated in [
129
]. Furthermore, in [
115
], the
use of mmWave radars for the safe landing of helicopters in inclement weather conditions
based on height drift data is studied. Another interesting application is implemented
in [
130
], which utilizes a CNN architecture to generate radar maps for recognizing and
classifying various real road images captured from gravel, mud, and river surfaces.
Sensors 2023,23, 8901 17 of 25
4.7. Other Applications
Object detection is one of the earliest applications of mmWave radar, which extends
to human fall detection, as presented in [
131
]. Hand gesture recognition for user interac-
tions with computers is studied in [
132
]. A model for long-range gesture recognition is
investigated in [
133
]. Furthermore, the real-time recognition of macro-gestures is presented
in [
134
]. Human pose estimation through occlusions and walls is explored in [
135
]. The
implementation of mmWave harmonic sensors to track small insects is conducted in [
136
].
The use of 61 GHz mmWave radar for human face classification is investigated in [
137
].
Non-contact skin sensing for analyzing human emotional arousal and stress status using
mmWave radars is implemented in [
97
]. The use of mmWave radar combined with GNN
and LSTM for human activity recognition is explored in [138].
In addition to the aforementioned areas, mmWave radars are also used in underground
mining with range measurements [
139
], spying on phone calls [
80
], efficient soil moisture
sensing [
140
], micro-action recognition systems [
141
], mmWave radar and audio signal
fusion for speech recognition [
142
]. Recent advances in sensor technology, combined
with machine learning techniques, have also enabled new applications for mmWave radar
sensors to be developed.
The use of mmWave radar combined with machine learning algorithms has grown
in popularity in recent years. As shown in Table 11, we provide an overview of machine
learning algorithms that are widely used in computer vision and related fields and have
been applied to radar signal processing. The applications of machine learning include
object detection, classification, clustering, and tracking, utilizing radar data. IF signals from
radars contain a predefined set of target features, and these IF signals are used in machine
learning models to make subsequent predictions. However, deep learning algorithms are
based on multiple layers of neural networks to learn high-level feature representations
from input radar IF data, which are then used to make intelligent decisions.
Furthermore, many research works are in progress that input radar signals into var-
ious deep learning techniques for object detection, such as those in [
45
,
143
145
]; object
classification, such as those in [
146
,
147
]; object segmentation, such as those in [
46
,
148
]; and
multi-class target classification, such as those in [
149
], using mmWave radar range–angle
images. Furthermore, target classification using the range FFT of a mmWave radar’s sta-
tistical features is studied in [
150
]. The utilization of various deep learning techniques
and micro-Doppler patterns from radar data for object classification is explored in [
151
].
Multi-person identification with distinct micro-Doppler signatures is studied in [152].
Sensors 2023,23, 8901 18 of 25
Table 11. Machine learning techniques for mmWave radar sensing.
References Year Method Application Comments
[88] 2016 CNN, data transformation
techniques Fitness tracking
1. Can classify different exercises with 95.53 accuracy and is capable
of counting repetitive exercises.
2. Counting repetitive exercises improves accuracy.
[132] 2016 Random forest algorithm Hand gesture recognition
1. RF offers 86% per-gesture accuracy with raw data.
2. RF with Bayesian Filter offers 92% per-gesture accuracy with
raw data.
[84] 2019 Convolution neural
network (CNN) Human behavior detection
1. Point cloud data are processed using the CFAR algorithm.
2. The usage of micro-Doppler information on human activities with
CNN produces an accuracy above 99%.
[131] 2019 NN is compared with SVM,
DT Fall detection
1. Attains 98% accuracy with NN backpropagation.
2. Evaluated on only three possible human positions with coordina-
tion points.
[153] 2019 CNN, ConvLSTM, RF Received power prediction
1. Power prediction from image works effectively with rotated 3D
CNN, and spatiotemporal features are predicted with a Random
forest algorithm.
2. Received power of 500ms with high accuracy and RMS errors less
than 1.0 is achieved.
[112] 2019 LSTM Channel tracking in
vehicular system
1. Accurate user channel prediction, and less overhead rate.
2. Usage of LSTMs is to predict the user channel based on past channel-
state information.
[45] 2019 PointNets 2D car detection 1. Using PointNets for classification with segmentation.
2. Mutli-class object detection needs to be investigated.
[46] 2020 CNN, RNN Scene understanding via
classification
1. CNN with grid maps as input for classifying static objects.
2. RNN with point clouds as input for classifying dynamic instances.
[47] 2020 DBSCAN, Faster R-CNN Vehicle detection
1. The proposed method performs better as it is a DBSCAN method
based on elevation resolution and also removes noise points using
filters.
2. Using Faster R-CNN achieves 96% accuracy by representing the tar-
get with the density of the point cloud.
[110] 2020 Point clouds, GMM Multimodal traffic
monitoring
1. GMM performs point cloud segmentation from sensor-collected
point clouds.
2. Can extend with DBSCAN for classifying more transportation
modes.
[48] 2020 SAF-FOC framework Obstacle detection
1. Feature-level fusion performs well compared with data-level and
decision-level fusion.
2. To cover 360coverage, the framework can be extended with multi-
ple sensors.
[86] 2020 CNN Detecting human
skeletal pose
1. Radar data to image representation with the help of depth, az-
imuth, and elevation information of reflection points to identify
skeletal position.
2. Proposed an architecture with significantly reduced computational
complexity with reused weights, and it also provided lower local-
ization error, such as 3.2 cm depth and 2.7 cm elevation.
[138] 2021 Graph neural network with
LSTM
Human activity recognition
and gesture recognition
1. Iteratively extracts the point cloud features and updates the graphs.
2. Excellent action recognition performance compared to
other methods.
[154] 2021 SVM Shape classification and
object detection
1. Uses SVM with RFB kernel to achieve an accuracy of 96%.
2. Comparatively less accuracy is obtained to classify multiple
target objects.
[98] 2022 CNN
Automatic monitoring of
heart rate and
breathing rate
1. Obtained 87% classification accuracy by forming low, average, high,
and a combination of six classes using CNN.
2. Removes the noise caused via vibrations and gives clear rate.
[70] 2022 CNN, K-nearest neighbor Material identification
1. K-NN uses two feature sets, while CNN uses the material’s distinc-
tive features for identifying the materials.
2. Enhanced classification accuracy of 98% in identifying the six mate-
rials at three different volume levels.
5. Conclusions
mmWave radar sensors have significant advantages compared to other sensors, mak-
ing them an ideal solution for a vast number of applications. This article discussed the key
performance parameters as well as the interpretation of radar measurements for mmWave
radar sensors. The most recent mmWave radar advances and cutting-edge mmWave radars
were thoroughly reviewed. The use of mmWave radar sensors was discussed in a variety
of applications, such as automotive, industrial, robotics and automation, medical, security,
and surveillance fields, as well as others. Finally, machine learning techniques applied to
mmWave radar sensor data were investigated.
Sensors 2023,23, 8901 19 of 25
The future of mmWave radar technology seems promising, with plenty of room for
growth and expansion. Here, we present some major trends and developments to keep an
eye on in future years. mmWave radars are projected to play a critical role in advanced
driver-assistance systems (ADASs) and driverless vehicles in automotive applications. In
the future, the increased integration of mmWave radar sensors in automobiles is likely
to improve safety, enable autonomous driving, and improve situational awareness in a
variety of weather conditions. mmWave radars have the potential to transform industrial
applications, such as non-destructive testing, quality control, and process automation.
Future advancements may result in smaller, more adaptable, and cost-effective industrial
mmWave radar systems. There is significant interest in employing mmWave radar for
medical applications, such as the remote monitoring of vital signs, fall detection for the
elderly, and early illness identification in healthcare and medical imaging. Healthcare-
related mmWave radar equipment may advance in the future. mmWave radar systems are
useful in security and surveillance applications, such as perimeter monitoring, intrusion
detection, and surveillance in complex settings. Future advancements could result in more
complex and integrated security solutions. In the field of IoT and smart cities, mmWave
radar sensors could find uses in smart cities for traffic control, environmental monitoring,
and public safety. Space exploration uses mmWave radar technology for remote sensing,
landing, and planetary exploration. As space exploration advances, mmWave radar devices
may play an important part in future space missions. Continued advances in signal
processing methods and the application of machine learning techniques will enhance the
capabilities of mmWave radar systems, allowing for better object detection, tracking, and
imaging. The miniaturization of mmWave radar components and their integration into
smaller and more diversified devices may be future trends, making them more accessible
for a larger range of applications. Challenges include mitigating interference from other
mmWave sources, dealing with atmospheric effects, and ensuring regulatory compliance.
Author Contributions:
Conceptualization, L.R.C.; methodology, L.R.C. and A.S.; software, L.R.C.
and A.S.; validation, L.R.C. and A.S.; formal analysis, L.R.C. and A.S.; investigation, L.R.C. and A.S.;
resources, L.R.C., A.S. and C.K.M.; data curation, L.R.C. and A.S.; writing—original draft preparation,
L.R.C. and A.S.; writing—review and editing, L.R.C., A.S. and C.K.M.; visualization, L.R.C. and
A.S.; supervision, L.R.C. and C.K.M.; project administration, L.R.C.; funding acquisition, L.R.C. All
authors have read and agreed to the published version of the manuscript.
Funding:
This work was supported by the INCAPS project: 287918 of the International Partnerships
for Excellent Education, Research and Innovation (INTPART) program from the Research Council of
Norway and the Low-Altitude UAV Communication and Tracking (LUCAT) project: 280835 of the
IKTPLUSS program from the Research Council of Norway.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Data sharing not applicable.
Conflicts of Interest: The authors declare no conflict of interest.
References
1.
Wilson, A.; Kumar, A.; Jha, A.; Cenkeramaddi, L.R. Embedded Sensors, Communication Technologies, Computing Platforms and
Machine Learning for UAVs: A Review. IEEE Sens. J. 2021,22, 1807–1826. [CrossRef]
2.
Morris, P.J.B.; Hari, K. Detection and localization of unmanned aircraft systems using millimeter-wave automotive radar sensors.
IEEE Sens. Lett. 2021,5, 1–4. [CrossRef]
3.
Venon, A.; Dupuis, Y.; Vasseur, P.; Merriaux, P. Millimeter Wave FMCW RADARs for perception, recognition, and localization in
automotive applications: A survey. IEEE Trans. Intell. Veh. 2022,7, 533–555. [CrossRef]
4.
Hakobyan, G.; Yang, B. High-performance automotive radar: A review of signal processing algorithms and modulation schemes.
IEEE Signal Process. Mag. 2019,36, 32–44. [CrossRef]
5.
Cenkeramaddi, L.R.; Bhatia, J.; Jha, A.; Vishkarma, S.K.; Soumya, J. A survey on sensors for autonomous systems. In Proceedings
of the 2020 15th IEEE Conference on Industrial Electronics and Applications (ICIEA), Kristiansand, Norway, 9–13 November
2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1182–1187.
Sensors 2023,23, 8901 20 of 25
6.
Roodaki, P.M.; Taghian, F.; Bashirzadeh, S.; Jalaali, M. A survey of millimeter-wave technologies. In Proceedings of the 2011
International Conference on Electrical and Control Engineering, Yichang, China, 16–18 September 2011; IEEE: Piscataway, NJ,
USA, 2011; pp. 5726–5728.
7.
Patole, S.M.; Torlak, M.; Wang, D.; Ali, M. Automotive radars: A review of signal processing techniques. IEEE Signal Process.
Mag. 2017,34, 22–35. [CrossRef]
8.
Abdu, F.J.; Zhang, Y.; Fu, M.; Li, Y.; Deng, Z. Application of deep learning on millimeter-wave radar signals: A review. Sensors
2021,21, 1951. [CrossRef]
9.
Ikram, M.Z.; Ahmad, A.; Wang, D. High-accuracy distance measurement using millimeter-wave radar. In Proceedings of the
2018 IEEE Radar Conference (RadarConf18), Oklahoma City, OK, USA, 23–27 April 2018; IEEE: Piscataway, NJ, USA, 2018;
pp. 1296–1300.
10.
Ciattaglia, G.; De Santis, A.; Disha, D.; Spinsante, S.; Castellini, P.; Gambi, E. Performance evaluation of vibrational measure-
ments through mmWave radars. In Proceedings of the 2020 IEEE 7th International Workshop on Metrology for AeroSpace
(MetroAeroSpace), Pisa, Italy, 22–24 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 160–165.
11.
Darlis, A.R.; Ibrahim, N.; Kusumoputro, B. Performance Analysis of 77 GHz mmWave Radar Based Object Behavior. J. Commun.
2021,16, 576–582. [CrossRef]
12.
Lovescu, C.; Rao, S. The Fundamentals of Millimeter Wave Radar Sensors. Available online: https://www.ti.com/lit/pdf/spyy0
05 (accessed on 10 February 2023) .
13. IWR6843aop. Available online: https://www.ti.com/tool/IWR6843AOPEVM (accessed on 3 January 2023).
14.
IWR1843. Available online: https://www.ti.com/lit/ds/swrs228a/swrs228a.pdf?ts=1673330497709&ref_url=https%253A%25
2F%252Fwww.ti.com%252Ftool%252FIWR1843BOOST (accessed on 3 January 2023).
15.
Stanislas, L.; Peynot, T. Characterisation of the Delphi Electronically Scanning Radar for robotics applications. In Proceedings of
the Australasian Conference on Robotics and Automation 2015, Canberra, Australia, 2–4 December 2015; Australian Robotics and
Automation Association: Sydney, NSW, Australia, 2015; pp. 1–10.
16.
NavTech CIR204-h. Available online: https://levelfivesupplies.com/wp- content/uploads/2019/01/CIR-datasheet.pdf (accessed
on 30 December 2022).
17.
AWR1843. Available online: https://www.ti.com/lit/ds/symlink/awr1843aop.pdf?ts=1673203974876&ref_url=https%253A%25
2F%252Fwww.ti.com%252Fsensors%252Fmmwave-radar%252Fautomotive%252Fproducts.html (accessed on 3 January 2023).
18.
BOSCH. Available online: https://www.bosch-engineering.jp/media/jp/pdfs_3/einsatzgebiete_4/produktdatenblaetter_2/12
0903_LRR3_EN_V05_final.pdf (accessed on 23 December 2022).
19.
Continental Engineering Services. ARS 408-21. Available online: https://conti-engineering.com/wp-content/uploads/2020/02/
ARS-408-21_EN_HS-1.pdf (accessed on 31 December 2022).
20.
SRR600. Available online: https://www.continental-automotive.com/en-gl/Passenger-Cars/Autonomous- Mobility/Enablers/
Radars/SRR600 (accessed on 22 December 2022).
21.
AWR1843AOP. Available online: https://www.ti.com/product/AWR1843AOP#product-details (accessed on 22 December 2022).
22.
AWR1642. Available online: https://www.ti.com/lit/ds/symlink/awr1642.pdf?ts=1673244878752&ref_url=https%253A%252F%
252Fwww.ti.com%252Fproduct%252FAWR1642 (accessed on 31 December 2022).
23.
IWR1642. TI Robots. Available online: https://www.ti.com/lit/ds/symlink/iwr1642.pdf?ts=1673279643829&ref_url=https%25
3A%252F%252Fwww.ti.com%252Fproduct%252FIWR1642 (accessed on 31 December 2022).
24.
TEF810X. Nxp Radar. Available online: https://www.nxp.com/docs/en/data-sheet/TEF810XDS.pdf (accessed on
31 December 2022).
25.
SAF85xx. Nxp Model. Available online: https://www.nxp.com/products/radio-frequency/radar-transceivers-and- socs/high-
performance-77ghz-rfcmos-automotive-radar-one-chip-soc:SAF85XX (accessed on 31 December 2022).
26.
TEF82xx. Nxp Radar Model. Available online: https://www.nxp.com/products/radio-frequency/radar-transceivers/fully-
integrated-77-ghz-rfcmos-automotive-radar-transceiver:TEF82xx (accessed on 31 December 2022).
27.
ARS540. Available online: https://www.continental-automotive.com/en-gl/Passenger-Cars/Autonomous-Mobility/Enablers/
Radars/Long-Range-Radar/ARS540 (accessed on 31 December 2022).
28.
AWRL1443. Ti Automative Radar Model. Available online: https://www.ti.com/lit/ds/symlink/awr1443.pdf?ts=167352176862
8&ref_url=https%253A%252F%252Fwww.ti.com%252Fproduct%252FAWR1443 (accessed on 31 December 2022).
29.
IWRL6432. Industrial Radar Model. Available online: https://www.ti.com/lit/ds/symlink/iwrl6432.pdf?ts=1673520655112&
ref_url=https%253A%252F%252Fwww.ti.com%252Fproduct%252FIWRL6432 (accessed on 31 December 2022).
30. ARS4-A. Available online: https://apps.fcc.gov/els/GetAtt.html?id=144885&x= (accessed on 31 December 2022).
31.
AWR2243. Available online: https://www.ti.com/lit/ug/spruit8d/spruit8d.pdf?ts=1673514555771 (accessed on
31 December 2022).
32. AWR1243. Available online: https://www.ti.com/lit/wp/spyy003/spyy003.pdf (accessed on 22 December 2022).
33.
NXP4D-S32R45. Available online: https://www.electronicproducts.com/nxp-unveils-4d-imaging-radar-processor-for-l2
-autonomy/ (accessed on 10 January 2022).
34.
RDK-S32R274. Available online: https://www.nxp.com/docs/en/fact-sheet/RDK-S32R274_FS.pdf (accessed on
22 December 2022).
Sensors 2023,23, 8901 21 of 25
35.
Hajri, H.; Rahal, M.C. Real time lidar and radar high-level fusion for obstacle detection and tracking with evaluation on a ground
truth. arXiv 2018, arXiv:1807.11264.
36.
Kwon, S.K.; Hyun, E.; Lee, J.H.; Lee, J.; Son, S.H. A low-complexity scheme for partially occluded pedestrian detection using
LiDAR-radar sensor fusion. In Proceedings of the 2016 IEEE 22nd International Conference on Embedded and Real-Time
Computing Systems and Applications (RTCSA), Daegu, Republic of Korea, 17–19 August 2016; IEEE: Piscataway, NJ, USA, 2016.
37.
Sengupta, A.; Cheng, L.; Cao, S. Robust multiobject tracking using mmwave radar-camera sensor fusion. IEEE Sens. Lett.
2022
,
6, 1–4. [CrossRef]
38.
Ulrich, M.; Maile, F.; Löcklin, A.; Yang, B.; Kleiner, B.; Ziegenspeck, N. A model for improved association of radar and camera
objects in an indoor environment. In Proceedings of the 2017 Sensor Data Fusion: Trends, Solutions, Applications (SDF), Bonn,
Germany, 10–12 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6.
39.
Sikdar, A.; Cao, S.; Zheng, Y.F.; Ewing, R.L. Radar depth association with vision detected vehicles on a highway. In Proceedings
of the 2014 IEEE Radar Conference, Cincinnati, OH, USA, 19–23 May 2014; IEEE: 2014; pp. 1159–1164.
40.
Zhong, Z.; Liu, S.; Mathew, M.; Dubey, A. Camera radar fusion for increased reliability in ADAS applications. Electron. Imaging
2018,2018, 258-1. [CrossRef]
41.
Steux, B.; Laurgeau, C.; Salesse, L.; Wautier, D. Fade: A vehicle detection and tracking system featuring monocular color vision
and radar data fusion. In Proceedings of the Intelligent Vehicle Symposium, Versailles, France, 17–21 June 2002; IEEE: Piscataway,
NJ, USA, 2002; Volume 2, pp. 632–639.
42.
Mahlisch, M.; Hering, R.; Ritter, W.; Dietmayer, K. Heterogeneous fusion of Video, LIDAR and ESP data for automotive ACC
vehicle tracking. In Proceedings of the 2006 IEEE International Conference on Multisensor Fusion and Integration for Intelligent
Systems, Heidelberg, Germany, 3–6 September 2006; IEEE: Piscataway, NJ, USA, 2006; pp. 139–144.
43.
de Ponte Müller, F. Survey on ranging sensors and cooperative techniques for relative positioning of vehicles. Sensors
2017
,
17, 271. [CrossRef] [PubMed]
44.
Major, B.; Fontijne, D.; Ansari, A.; Teja Sukhavasi, R.; Gowaikar, R.; Hamilton, M.; Lee, S.; Grzechnik, S.; Subramanian, S. Vehicle
detection with automotive radar using deep learning on range-azimuth-doppler tensors. In Proceedings of the IEEE/CVF
International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27 October–2 November 2019.
45.
Danzer, A.; Griebel, T.; Bach, M.; Dietmayer, K. 2D car detection in radar data with pointnets. In Proceedings of the 2019 IEEE
Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; IEEE: Piscataway, NJ, USA,
2019; pp. 61–66.
46.
Schumann, O.; Lombacher, J.; Hahn, M.; Wöhler, C.; Dickmann, J. Scene understanding with automotive radar. IEEE Trans. Intell.
Veh. 2019,5, 188–203. [CrossRef]
47.
Huang, Y.; Zhang, H.; Guo, K.; Li, J.; Xu, G.; Chen, Z. Density-based vehicle detection approach for automotive millimeter-wave
radar. In Proceedings of the 2020 IEEE 3rd International Conference on Electronic Information and Communication Technology
(ICEICT), Shenzhen, China, 13–15 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 534–537.
48.
Chang, S.; Zhang, Y.; Zhang, F.; Zhao, X.; Huang, S.; Feng, Z.; Wei, Z. Spatial attention fusion for obstacle detection using
mmwave radar and vision sensor. Sensors 2020,20, 956. [CrossRef] [PubMed]
49.
Gao, X.; Xing, G.; Roy, S.; Liu, H. Experiments with mmwave automotive radar test-bed. In Proceedings of the 2019 53rd Asilomar
Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 3–6 November 2019; IEEE: Piscataway, NJ, USA, 2019;
pp. 1–6.
50.
Smith, K.A.; Csech, C.; Murdoch, D.; Shaker, G. Gesture recognition using mm-wave sensor for human-car interface. IEEE Sens.
Lett. 2018,2, 1–4. [CrossRef]
51.
Yu, J.T.; Yen, L.; Tseng, P.H. mmWave radar-based hand gesture recognition using range-angle image. In Proceedings of the 2020
IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; IEEE: Piscataway, NJ, USA,
2020; pp. 1–5.
52.
Munte, N.; Lazaro, A.; Villarino, R.; Girbau, D. Vehicle occupancy detector based on FMCW mm-wave radar at 77 GHz. IEEE
Sens. J. 2022,22, 24504–24515. [CrossRef]
53.
Texas Instruments. Vehicle Occupant Detection using TI mmWave Sensors. Available online: https://training.ti.com/vehicle-
occupant-detection-using-ti-mmwave-sensors (accessed on 1 May 2022).
54.
Jose, E.; Adams, M.; Mullane, J.S.; Patrikalakis, N.M. Predicting millimeter wave radar spectra for autonomous navigation. IEEE
Sens. J. 2010,10, 960–971. [CrossRef]
55.
Texas Instruments. Crosstraffic Alert Radar Sensor. Available online: https://training.ti.com/automotive-corner-radar-using-ti-
mmwave-sensors. (accessed on 24 January 2023).
56.
Cenkeramaddi, L.R.; Rai, P.K.; Dayal, A.; Bhatia, J.; Pandya, A.; Soumya, J.; Kumar, A.; Jha, A. A Novel Angle Estimation for
mmWave FMCW Radars Using Machine Learning. IEEE Sens. J. 2021,21, 9833–9843. [CrossRef]
57.
Texas Instruments. Advanced Driver Assistance Systems (ADAS). Available online: https://www.ti.com/applications/
automotive/adas/overview.html#aem-application-Browse (accessed on 1 May 2022).
58.
Russell, M.; Crain, A.; Curran, A.; Campbell, R.; Drubin, C.; Miccioli, W. Millimeter-wave radar sensor for automotive intelligent
cruise control (ICC). IEEE Trans. Microw. Theory Tech. 1997,45, 2444–2453. [CrossRef]
59.
Liu, G.; Zhou, M.; Wang, L.; Wang, H.; Guo, X. A blind spot detection and warning system based on millimeter wave radar for
driver assistance. Optik 2017,135, 353–365. [CrossRef]
Sensors 2023,23, 8901 22 of 25
60. AWR1243 Radar. Available online: https://www.ti.com/lit/wp/spyy009/spyy009.pdf (accessed on 22 December 2022).
61. AWR1642 Radar. Available online: https://www.ti.com/video/5703076705001 (accessed on 22 December 2022).
62.
77GHz Single Chip Radar Sensor Enables Automotive Body and Chassis Applications. Available online: https://www.ti.
com/lit/wp/spry315/spry315.pdf?ts=1674543928796&ref_url=https%253A%252F%252Fwww.google.com%252F (accessed on
22 December 2022).
63.
Texas Instruments. Webinar—Automotive Parking System Using TI mmWave Sensors. Available online: https://training.ti.com/
webinar-automotive-parking-system-using-ti-mmwave-sensors?context=1128486-1139157-1137700 (accessed on 1 May 2022).
64.
AWR1843AOPEVM Radar. Available online: https://e2e.ti.com/blogs_/b/behind_the_wheel/posts/how-aop-technology-
expands-radar-sensor-placement-for-automotive-applications (accessed on 22 December 2022).
65.
BlindspotAWR1843AOPEVM Radar. Available online: https://training.ti.com/zh-tw/blind-spot-detection-motorcycles-using-
ti-mmwave-radar?context=1149404-1149401 (accessed on 22 December 2022).
66.
Yurduseven, O.; Fromenteze, T.; Decroze, C.; Fusco, V.F. Frequency-diverse computational automotive radar technique for debris
detection. IEEE Sens. J. 2020,20, 13167–13177. [CrossRef]
67.
Li, X.; Wang, X.; Yang, Q.; Fu, S. Signal Processing for TDM MIMO FMCW Millimeter-Wave Radar Sensors. IEEE Access
2021
,
9, 167959–167971. [CrossRef]
68.
Agarwal, S.; Singh, D. An Adaptive Statistical Approach for Non-Destructive Underline Crack Detection of Ceramic Tiles Using
Millimeter Wave Imaging Radar for Industrial Application. IEEE Sens. J. 2015,15, 7036–7044. [CrossRef]
69. Fluidlevel Radar. Available online: https://www.ti.com/lit/wp/spyy004/spyy004.pdf (accessed on 20 January 2022).
70.
Skaria, S.; Hendy, N.; Al-Hourani, A. Machine Learning Methods for Material Identification Using mmWave Radar Sensor. IEEE
Sens. J. 2022,23, 1471–1478. [CrossRef]
71.
Wei, Y.; Sun, B.; Zhou, Y.; Wang, H. Non-Line-of-Sight Moving Target Detection Method Based on Noise Suppression. Remote
Sens. 2022,14, 1614. [CrossRef]
72.
Hattenhorst, B.; Piotrowsky, L.; Pohl, N.; Musch, T. An mmWave sensor for real-time monitoring of gases based on real refractive
index. IEEE Trans. Microw. Theory Tech. 2021,69, 5033–5044. [CrossRef]
73.
Baer, C.; Jaeschke, T.; Mertmann, P.; Pohl, N.; Musch, T. A mmWave measuring procedure for mass flow monitoring of pneumatic
conveyed bulk materials. IEEE Sens. J. 2014,14, 3201–3209. [CrossRef]
74.
OndoSense. Metal Production, Shortrange. Available online: https://ondosense.com/en/applications/width-measurement-of-
slabs-in-hot-rolling-mills/ (accessed on 1 May 2022).
75. Park, J.; Nguyen, C. Development of a new millimeter-wave integrated-circuit sensor for surface and subsurface sensing. IEEE
Sens. J. 2006,6, 650–655. [CrossRef]
76.
Nakagawa, T.; Hyodo, A.; Kogo, K.; Kurata, H.; Osada, K.; Oho, S. Contactless Liquid-Level Measurement With Frequency-
Modulated Millimeter Wave through Opaque Container. IEEE Sens. J. 2013,13, 926–933. [CrossRef]
77.
Weiß, J.; Santra, A. One-Shot Learning for Robust Material Classification Using Millimeter-Wave Radar System. IEEE Sens. Lett.
2018,2, 1–4. [CrossRef]
78.
Texas Instruments, mmWave. Available online: https://www.ti.com/lit/wp/spry328/spry328.pdf (accessed on
23 January 2022
).
79.
Delden, M.v.; Westerdick, S.; Musch, T. Investigations on Foam Detection Utilizing Ultra-Broadband Millimeter Wave FMCW
Radar. In Proceedings of the 2019 IEEE MTT-S International Microwave Workshop Series on Advanced Materials and Processes
for RF and THz Applications (IMWS-AMP), Bochum, Germany, 16–18 July 2019; pp. 103–105. [CrossRef]
80.
Basak, S.; Gowda, M. mmspy: Spying phone calls using mmwave radars. In Proceedings of the 2022 IEEE Symposium on
Security and Privacy (SP), San Francisco, CA, USA, 22–26 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1211–1228.
81.
Omer, A.E.; Shaker, G.; Safavi-Naeini, S.; Murray, K.; Hughson, R. Glucose levels detection using mm-wave radar. IEEE Sens.
Lett. 2018,2, 1–4. [CrossRef]
82.
Texas Instruments. Mmwave, Sensor. Available online: https://training.ti.com/contactless- patient-and-elderly- care-using-
mmwave-sensors (accessed on 1 May 2023).
83.
Ahmad, A.; Roh, J.C.; Wang, D.; Dubey, A. Vital signs monitoring of multiple people using a FMCW millimeter-wave sensor. In
Proceedings of the 2018 IEEE Radar Conference (RadarConf18), Oklahoma City, OK, USA, 23–27 April 2018; IEEE: Piscataway, NJ,
USA, 2018; pp. 1450–1455.
84.
Zhang, R.; Cao, S. Real-time human motion behavior detection via CNN using mmWave radar. IEEE Sens. Lett.
2018
,3, 1–4.
[CrossRef]
85.
Jin, F.; Zhang, R.; Sengupta, A.; Cao, S.; Hariri, S.; Agarwal, N.K.; Agarwal, S.K. Multiple patients behavior detection in real-time
using mmWave radar and deep CNNs. In Proceedings of the 2019 IEEE Radar Conference (RadarConf), Boston, MA, USA,
22–26 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6.
86.
Sengupta, A.; Jin, F.; Zhang, R.; Cao, S. mm-Pose: Real-time human skeletal posture estimation using mmWave radars and CNNs.
IEEE Sens. J. 2020,20, 10032–10044. [CrossRef]
87.
Tiwari, G.; Gupta, S. An mmWave radar based real-time contactless fitness tracker using deep CNNs. IEEE Sens. J.
2021
,
21, 17262–17270. [CrossRef]
88.
Tiwari, G.; Bajaj, P.; Gupta, S. mmFiT: Contactless Fitness Tracker Using mmWave Radar and Edge Computing Enabled Deep
Learning. IEEE Internet Things J. 2021,XX, 1–8.
Sensors 2023,23, 8901 23 of 25
89.
Yang, Z.; Pathak, P.H.; Zeng, Y.; Liran, X.; Mohapatra, P. Vital sign and sleep monitoring using millimeter wave. ACM Trans. Sens.
Netw. TOSN 2017,13, 1–32. [CrossRef]
90.
Arab, H.; Chioukh, L.; Ardakani, M.D.; Dufour, S.; Tatu, S.O. Early-stage detection of melanoma skin cancer using contactless
millimeter-wave sensors. IEEE Sensors J. 2020,20, 7310–7317. [CrossRef]
91.
Wen, L.; Dong, S.; Zhang, Z.; Gu, C.; Mao, J. Noninvasive Continuous Blood Pressure Monitoring Based on Wearable Radar
Sensor with Preliminary Clinical Validation. In Proceedings of the 2022 IEEE/MTT-S International Microwave Symposium–IMS
2022, Denver, CO, USA, 19–24 June 2022; pp. 707–710. [CrossRef]
92.
Shi, J.; Lee, K. Systolic blood pressure measurement algorithm with mmWave radar sensor. KSII Trans. Internet Inf. Syst. TIIS
2022,16, 1209–1223.
93.
Hu, Y.; Toda, T. Remote Vital Signs Measurement of Indoor Walking Persons Using mm-Wave FMCW Radar. IEEE Access
2022
,
10, 78219–78230. [CrossRef]
94.
Wu, J.; Dahnoun, N. A health monitoring system with posture estimation and heart rate detection based on millimeter-wave
radar. Microprocess. Microsyst. 2022,94, 104670. [CrossRef]
95.
Ran, Y.; Zhang, D.; Chen, J.; Hu, Y.; Chen, Y. Contactless Blood Pressure Monitoring with mmWave Radar. In Proceedings of
the GLOBECOM 2022–2022 IEEE Global Communications Conference, Rio de Janeiro, Brazil, 4–8 December 2022; pp. 541–546.
[CrossRef]
96.
Antolinos, E.; García-Rial, F.; Hernández, C.; Montesano, D.; Godino-Llorente, J.I.; Grajal, J. Cardiopulmonary activity monitoring
using millimeter wave radars. Remote Sens. 2020,12, 2265. [CrossRef]
97.
Xu, X.; Zhang, D.; Chen, J.; Wu, Z.; Sun, Q.; Chen, Y. Contactless GSR Sensing Using mmWave Radar. IEEE Sens. J.
2022
,
22, 24264–24275. [CrossRef]
98.
Gupta, K.; Srinivas, M.; Soumya, J.; Pandey, O.J.; Cenkeramaddi, L.R. Automatic Contact-less Monitoring of Breathing Rate and
Heart Rate utilizing the Fusion of mmWave Radar and Camera Steering System. IEEE Sens. J.
2022
,22, 22179–22191. [CrossRef]
99.
Texas Instruments. Robotic Applications. Available online: https://www.ti.com/lit/wp/spry311a/spry311a.pdf?ts=1674551319
819 (accessed on 23 January 2022).
100.
TIsafe, M. Safe Robots. Available online: https://training.ti.com/autonomous-robotics-using-ti-mmwave-sensors?context=1128
486-1139156-1147844 (accessed on 23 December 2022).
101.
Texas Instruments. 360mmwave Radar Sensor. Available online: https://training.ti.com/360-degree-safety-bubble-robotics-
using-ti-mmwave-sensors?context=1128486-1139156-1147582 (accessed on 1 May 2023).
102.
Texas Instruments. Intelligent, Aop. Available online: https://training.ti.com/intelligent-robotics-ti-mmwave-aop-sensors?
context=1128486-1139156-1138090 (accessed on 1 May 2022).
103.
Zhao, P.; Lu, C.X.; Wang, B.; Chen, C.; Xie, L.; Wang, M.; Trigoni, N.; Markham, A. Heart Rate Sensing with a Robot Mounted
mmWave Radar. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France,
31 May 2020–31 August 2020; pp. 2812–2818. [CrossRef]
104.
Texas Instruments. Safe Guard Robots. Available online: https://training.ti.com/safety-guards-industrial-robots?context=1128
486-1139156-1146381 (accessed on 31 December 2022).
105.
Luukanen, A.; Appleby, R.; Kemp, M.; Salmon, N. Millimeter-wave and terahertz imaging in security applications. In Terahertz
Spectroscopy and Imaging; Springer: Berlin/Heidelberg, Germany, 2012; pp. 491–520.
106.
Rai, P.K.; Idsøe, H.; Yakkati, R.R.; Kumar, A.; Khan, M.Z.A.; Yalavarthy, P.K.; Cenkeramaddi, L.R. Localization and activity
classification of unmanned aerial vehicle using mmWave FMCW radars. IEEE Sens. J. 2021,21, 16043–16053. [CrossRef]
107.
Wenger, J. Automotive radar-status and perspectives. In Proceedings of the IEEE Compound Semiconductor Integrated Circuit
Symposium, CSIC’05, Palm Springs, CA, USA, 30 October 2005–2 November 2005; IEEE:Piscataway, NJ, USA, 2005; p. 4.
108.
Caris, M.; Stanko, S.; Palm, S.; Sommer, R.; Pohl, N. Synthetic aperture radar at millimeter wavelength for UAV surveillance
applications. In Proceedings of the 2015 IEEE 1st International Forum on Research and Technologies for Society and Industry
Leveraging a better tomorrow (RTSI), Turin, Italy, 16–18 September 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 349–352.
109.
Sheen, D.; McMakin, D.; Hall, T. Three-dimensional millimeter-wave imaging for concealed weapon detection. IEEE Trans.
Microw. Theory Tech. 2001,49, 1581–1592. [CrossRef]
110.
Jin, F.; Sengupta, A.; Cao, S.; Wu, Y.J. Mmwave radar point cloud segmentation using gmm in multimodal traffic monitoring.
In Proceedings of the 2020 IEEE International Radar Conference (RADAR), Washington, DC, USA, 28–30 April 2020; IEEE:
Piscataway, NJ, USA, 2020; pp. 732–737.
111.
Ninos, A.; Hasch, J.; Heizmann, M.; Zwick, T. Radar-Based Robust People Tracking and Consumer Applications. IEEE Sens. J.
2022,22, 3726–3735. [CrossRef]
112.
Guo, Y.; Wang, Z.; Li, M.; Liu, Q. Machine learning based mmWave channel tracking in vehicular scenario. In Proceedings of the
2019 IEEE International Conference on Communications Workshops (ICC Workshops), Shanghai, China, 20–24 May 2019; IEEE:
Piscataway, NJ, USA, 2019; pp. 1–6.
113.
Texas Instruments. Traffic, Mmwave. Available online: https://www.ti.com/lit/wp/spyy002b/spyy002b.pdf?ts=167439454023
0&ref_url=https%253A%252F%252Fwww.ti.com%252Fsensors%252Fmmwave-radar%252Findustrial%252Foverview.html (ac-
cessed on 1 May 2022).
Sensors 2023,23, 8901 24 of 25
114.
Migliaccio, C.; Nguyen, B.; Pichot, C.; Yonemoto, N.; Yamamoto, K.; Yamada, K.; Nasui, H.; Mayer, W.; Gronau, A.; Menzel, W.
Millimeter-wave radar for rescue helicopters. In Proceedings of the 2006 9th International Conference on Control, Automation,
Robotics and Vision, Singapore, 5–8 December 2006; IEEE: Piscataway, NJ, USA, 2006; pp. 1–6.
115.
Hagelen, M.; Briese, G.; Essen, H.; Bertuch, T.; Knott, P.; Tessmann, A. A millimetrewave landing aid approach for helicopters
under brown-out conditions. In Proceedings of the 2008 IEEE Radar Conference, Rome, Italy, 26–30 May 2008; pp. 1–4. [CrossRef]
116.
Feil, P.; Kraus, T.; Menzel, W. Short range mm-wave SAR for surveillance and security applications. In Proceedings of the 8th
European Conference on Synthetic Aperture Radar, Aachen, Germany, 7–10 June 2010; VDE: Berlin, Germany, 2010; pp. 1–4.
117.
Menzel, W. Millimeter-wave radar for civil applications. In Proceedings of the The 7th European Radar Conference, Paris, France,
30 September–1 October 2010; IEEE: Piscataway, NJ, USA, 2020; pp. 89–92.
118.
Andrews, D.A.; Harmer, S.W.; Bowring, N.J.; Rezgui, N.D.; Southgate, M.J. Active millimeter wave sensor for standoff concealed
threat detection. IEEE Sens. J. 2013,13, 4948–4954. [CrossRef]
119.
Will, C.; Vaishnav, P.; Chakraborty, A.; Santra, A. Human target detection, tracking, and classification using 24-GHz FMCW radar.
IEEE Sens. J. 2019,19, 7283–7299. [CrossRef]
120.
Guerra, A.; Dardari, D.; Djuri´c, P.M. Dynamic radar network of UAVs: A joint navigation and tracking approach. IEEE Access
2020,8, 116454–116469. [CrossRef]
121.
Almalioglu, Y.; Turan, M.; Lu, C.X.; Trigoni, N.; Markham, A. Milli-RIO: Ego-motion estimation with low-cost millimetre-wave
radar. IEEE Sens. J. 2020,21, 3314–3323. [CrossRef]
122.
Ding, W.; Cao, Z.; Zhang, J.; Chen, R.; Guo, X.; Wang, G. Radar-based 3D human skeleton estimation by kinematic constrained
learning. IEEE Sens. J. 2021,21, 23174–23184. [CrossRef]
123.
Li, W.; Chen, R.; Wu, Y.; Zhou, H. Indoor Positioning System Using a Single-Chip Millimeter Wave Radar. IEEE Sens. J.
2023
,23,
5232–5242. [CrossRef]
124.
Texas Instruments. Power Line Communication Using TI mmWave Sensors. Available online: https://www.ti.com/lit/wp/
slyy038/slyy038.pdf (accessed on 26 October 2023).
125.
Navtech Debris Detection Using Navtech Radar Sensor. Available online: https://navtechradar.com/explore/debris-detection/
(accessed on 26 October 2023).
126.
Ezuma, M.; Ozdemir, O.; Anjinappa, C.K.; Gulzar, W.A.; Guvenc, I. Micro-UAV detection with a low-grazing angle millimeter
wave radar. In Proceedings of the 2019 IEEE Radio and Wireless Symposium (RWS), Orlando, FL, USA, 20–23 January 2019; IEEE:
Piscataway, NJ, USA, 2019; pp. 1–4.
127.
Rahman, S.; Robertson, D.A. Radar micro-Doppler signatures of drones and birds at K-band and W-band. Sci. Rep.
2018
,8, 1–11.
[CrossRef] [PubMed]
128.
Pardhasaradhi, B.; Cenkeramaddi, L.R. GPS spoofing detection and mitigation for drones using distributed radar tracking and
fusion. IEEE Sens. J. 2022,22, 11122–11134. [CrossRef]
129.
Liu, C.; Zhao, Q.; Zhang, Y.; Tan, K. Runway extraction in low visibility conditions based on sensor fusion method. IEEE Sens. J.
2014,14, 1980–1987.
130.
Sabery, S.M.; Bystrov, A.; Gardner, P.; Stroescu, A.; Gashinova, M. Road Surface Classification Based on Radar Imaging Using
Convolutional Neural Network. IEEE Sens. J. 2021,21, 18725–18732. [CrossRef]
131.
Alanazi, M.A.; Alhazmi, A.K.; Yakopcic, C.; Chodavarapu, V.P. Machine learning models for human fall detection using millimeter
wave sensor. In Proceedings of the 2021 55th Annual Conference on Information Sciences and Systems (CISS), Baltimore, MD,
USA, 24–26 March 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–5.
132.
Lien, J.; Gillian, N.; Karagozler, M.E.; Amihood, P.; Schwesig, C.; Olson, E.; Raja, H.; Poupyrev, I. Soli: Ubiquitous gesture sensing
with millimeter wave radar. ACM Trans. Graph. TOG 2016,35, 1–19. [CrossRef]
133.
Liu, Y.; Wang, Y.; Liu, H.; Zhou, A.; Liu, J.; Yang, N. Long-range gesture recognition using millimeter wave radar. In Proceedings
of the Green, Pervasive, and Cloud Computing: 15th International Conference, GPC 2020, Xi’an, China, 13–15 November 2020;
Springer: Berlin/Heidelberg, Germany, 2020; pp. 30–44.
134.
Ninos, A.; Hasch, J.; Zwick, T. Real-time macro gesture recognition using efficient empirical feature extraction with millimeter-
wave technology. IEEE Sens. J. 2021,21, 15161–15170. [CrossRef]
135.
Zhao, M.; Li, T.; Abu Alsheikh, M.; Tian, Y.; Zhao, H.; Torralba, A.; Katabi, D. Through-wall human pose estimation using radio
signals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June
2018; pp. 7356–7365.
136.
Tahir, N.; Brooker, G. Toward the development of millimeter wave harmonic sensors for tracking small insects. IEEE Sens. J.
2015
,
15, 5669–5676. [CrossRef]
137.
Lim, H.S.; Jung, J.; Lee, J.E.; Park, H.M.; Lee, S. DNN-based human face classification using 61 GHz FMCW radar sensor. IEEE
Sens. J. 2020,20, 12217–12224. [CrossRef]
138.
Gong, P.; Wang, C.; Zhang, L. Mmpoint-GNN: Graph neural network with dynamic edges for human activity recognition through
a millimeter-wave radar. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen,
China, 18–22 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–7.
139.
Brooker, G.M.; Scheding, S.; Bishop, M.V.; Hennessy, R.C. Development and application of millimeter wave radar sensors for
underground mining. IEEE Sens. J. 2005,5, 1270–1280. [CrossRef]
Sensors 2023,23, 8901 25 of 25
140.
Chen, W.; Feng, Y.; Cardamis, M.; Jiang, C.; Song, W.; Ghannoum, O.; Hu, W. Soil moisture sensing with mmWave radar. In
Proceedings of the Proceedings of the 6th ACM Workshop on Millimeter-Wave and Terahertz Networks and Sensing Systems,
Sydney, NSW, Australia, 17 October 2022; pp. 19–24.
141.
Zhang, Z.; Meng, W.; Song, M.; Liu, Y.; Zhao, Y.; Feng, X.; Li, F. Application of multi-angle millimeter-wave radar detection in
human motion behavior and micro-action recognition. Meas. Sci. Technol. 2022,33, 105107. [CrossRef]
142.
Liu, T.; Gao, M.; Lin, F.; Wang, C.; Ba, Z.; Han, J.; Xu, W.; Ren, K. Wavoice: A noise-resistant multi-modal speech recognition
system fusing mmwave and audio signals. In Proceedings of the 19th ACM Conference on Embedded Networked Sensor
Systems, Coimbra, Portugal, 15–17 November 2021; pp. 97–110.
143.
Wang, L.; Tang, J.; Liao, Q. A study on radar target detection based on deep neural networks. IEEE Sens. Lett.
2019
,3, 1–4.
[CrossRef]
144.
Nabati, R.; Qi, H. Rrpn: Radar region proposal network for object detection in autonomous vehicles. In Proceedings of the 2019
IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; IEEE: Piscataway, NJ, USA,
2019; pp. 3093–3097.
145.
Gao, X.; Xing, G.; Roy, S.; Liu, H. Ramp-cnn: A novel neural network for enhanced automotive radar object recognition. IEEE
Sens. J. 2020,21, 5119–5132. [CrossRef]
146.
Heuel, S.; Rohling, H. Two-stage pedestrian classification in automotive radar systems. In Proceedings of the 2011 12th
International Radar Symposium (IRS), Leipzig, Germany, 7–9 September 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 477–484.
147.
Patel, K.; Rambach, K.; Visentin, T.; Rusev, D.; Pfeiffer, M.; Yang, B. Deep learning-based object classification on automotive radar
spectra. In Proceedings of the 2019 IEEE Radar Conference (RadarConf), Boston, MA, USA, 22–26 April 2019; IEEE: Piscataway,
NJ, USA, 2019; pp. 1–6.
148.
Schumann, O.; Hahn, M.; Dickmann, J.; Wöhler, C. Semantic segmentation on radar point clouds. In Proceedings of the 2018 21st
International Conference on Information Fusion (FUSION), Cambridge, UK, 10–13 July 2018; IEEE: Piscataway, NJ, USA, 2018;
pp. 2179–2186.
149.
Gupta, S.; Rai, P.K.; Kumar, A.; Yalavarthy, P.K.; Cenkeramaddi, L.R. Target classification by mmWave FMCW radars using
machine learning on range-angle images. IEEE Sens. J. 2021,21, 19993–20001. [CrossRef]
150.
Bhatia, J.; Dayal, A.; Jha, A.; Vishvakarma, S.K.; Joshi, S.; Srinivas, M.; Yalavarthy, P.K.; Kumar, A.; Lalitha, V.; Koorapati, S.; et al.
Classification of targets using statistical features from range fft of mmwave fmcw radars. Electronics 2021,10, 1965. [CrossRef]
151.
Angelov, A.; Robertson, A.; Murray-Smith, R.; Fioranelli, F. Practical classification of different moving targets using automotive
radar and deep neural networks. IET Radar Sonar Navig. 2018,12, 1082–1089. [CrossRef]
152.
Huang, X.; Ding, J.; Liang, D.; Wen, L. Multi-person recognition using separated micro-Doppler signatures. IEEE Sens. J.
2020
,
20, 6605–6611. [CrossRef]
153.
Nishio, T.; Okamoto, H.; Nakashima, K.; Koda, Y.; Yamamoto, K.; Morikura, M.; Asai, Y.; Miyatake, R. Proactive received power
prediction using machine learning and depth images for mmWave networks. IEEE J. Sel. Areas Commun.
2019
,37, 2413–2427.
[CrossRef]
154.
Arab, H.; Ghaffari, I.; Chioukh, L.; Tatu, S.; Dufour, S. Machine learning based object classification and identification scheme
using an embedded millimeter-wave radar sensor. Sensors 2021,21, 4291. [CrossRef]
Disclaimer/Publisher’s Note:
The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
... This breakthrough has opened massive opportunities in terms of highly flexible waveform design leading to a high performance in a miniature and cost-effective package. Initially designed for advanced driver assistance systems (ADASs) and limited only to high-end luxury cars, they now find widespread applications beyond the automotive domain [8]. Given the cost-effectiveness and wide availability, it is intrinsically appealing for application engineers and academicians to devise novel methods and techniques to stretch performance beyond hardware constraints. ...
... It is crucial to note that, at the time this article is being published, the dynamic nature of technological advancements may bring new sensors on the horizon. A review of recent advances in mmWave radar sensors offered a comparison of complete radar systems and single-chipset transceivers in a common table [8]. This served as a good repository for a comparison of the specifications but offered little insight apropos the efficacy for a specific use case. ...
... Plugging (5) into (8) and expressing (7) in terms of wavelength, λ, gives us the following: ...
Article
Full-text available
Precise altitude data are indispensable for flight navigation, particularly during the autonomous landing of unmanned aerial systems (UASs). Conventional light and barometric sensors employed for altitude estimation are limited by poor visibility and temperature conditions, respectively, whilst global positioning system (GPS) receivers provide the altitude from the mean sea level (MSL) marred with a slow update rate. To cater to the landing safety requirements, UASs necessitate precise altitude information above ground level (AGL) impervious to environmental conditions. Radar altimeters, a mainstay in commercial aviation for at least half a century, realize these requirements through minimum operational performance standards (MOPSs). More recently, the proliferation of 5G technology and interference with the universally allocated band for radar altimeters from 4.2 to 4.4 GHz underscores the necessity to explore novel avenues. Notably, there is no dedicated MOPS tailored for radar altimeters of UASs. To gauge the performance of a radar altimeter offering for UASs, existing MOPSs are the de facto choice. Historically, frequency-modulated continuous wave (FMCW) radars have been extensively used in a broad spectrum of ranging applications including radar altimeters. Modern monolithic millimeter wave (mmWave) automotive radars, albeit designed for automotive applications, also employ FMCW for precise ranging with a cost-effective and compact footprint. Given the technology maturation with excellent size, weight, and power (SWaP) metrics, there is a growing trend in industry and academia to explore their efficacy beyond the realm of the automotive industry. To this end, their feasibility for UAS altimetry remains largely untapped. While the literature on theoretical discourse is prevalent, a specific focus on mmWave radar altimetry is lacking. Moreover, clutter estimation with hardware specifications of a pure look-down mmWave radar is unreported. This article argues the applicability of MOPSs for commercial aviation for adaptation to a UAS use case. The theme of the work is a tutorial based on a simplified mathematical and theoretical discussion on the understanding of performance metrics and inherent intricacies. A systems engineering approach for deriving waveform specifications from operational requirements of a UAS is offered. Lastly, proposed future research directions and insights are included.
... From a different perspective, mmWave radar sensors provide accurate target range and velocity information, surpassing the other alternatives under extreme weather conditions [293]. One example is IWR6843, which provides a range of detection of 180 m and an ±120 elevation field of view [294]. At the final destination, and although LiDAR mainly aims at generating 3D maps of the environment (via analysis of the reflect light from emitted laser pulses), it can as well be utilized for UAV positioning (e.g., providing precise measurements of a UAV's altitude above ground level [295]). ...
Article
Full-text available
The evolving technologies regarding Unmanned Aerial Vehicles (UAVs) have led to their extended applicability in diverse domains, including surveillance, commerce, military, and smart electric grid monitoring. Modern UAV avionics enable precise aircraft operations through autonomous navigation, obstacle identification, and collision prevention. The structures of avionics are generally complex, and thorough hierarchies and intricate connections exist in between. For a comprehensive understanding of a UAV design, this paper aims to assess and critically review the purpose-classified electronics hardware inside UAVs, each with the corresponding performance metrics thoroughly analyzed. This review includes an exploration of different algorithms used for data processing, flight control, surveillance, navigation, protection, and communication. Consequently, this paper enriches the knowledge base of UAVs, offering an informative background on various UAV design processes, particularly those related to electric smart grid applications. As a future work recommendation, an actual relevant project is openly discussed.
... These losses can result in signal attenuation and reduced coverage, particularly in urban environments with dense infrastructure, meaning they are more susceptible to obstacles and even atmospheric conditions. This limits their effective range [4], making them more suitable for localized communication rather than wide-area coverage [5,6]. Incorporating the multiple input multiple output (MIMO) technology, which is a potential solution where multiple antennas are employed at both ends of the communication channel, can enhance data transmission and spectral efficiency without requiring extra bandwidth or transmission power. ...
Article
Full-text available
In this paper, a 3D conformal meta-lens designed for manipulating electromagnetic beams via height-to-phase control is proposed. The structure consists of a 40 × 20 array of tunable unit cells fabricated using 3D printing, enabling full 360° phase compensation. A novel automatic synthesizing method (ASM) with an integrated optimization process based on genetic algorithm (GA) is adopted here to create the meta-lens. Simulation using CST Microwave Studio and MATLAB reveals the antenna’s beam deflection capability by adjusting phase compensations for each unit cell. Various beam scanning techniques are demonstrated, including single-beam, dual-beam generation, and orbital angular momentum (OAM) beam deflection at different angles of 0°, 10°, 15°, 25°, 30°, and 45°. A 3D-printed prototype of the dual-beam feature has been fabricated and measured for validation purposes, with good agreement between both simulation and measurement results, with small discrepancies due to 3D printing’s low resolution and fabrication errors. This meta-lens shows promise for low-cost, high-gain beam deflection in mm-wave wireless communication systems, especially for sensing applications, with potential for wider 2D beam scanning and independent beam deflection enhancements.
... While the traditional parabolic antenna is a high-performance solution, its large size makes it unsuitable for most of those applications. Therefore, there is a growing need for flat, compact, lightweight, and mechanically robust antennas that align with the specific requirements of these systems [7][8][9][10][11]. Additionally, the antennas' gain and beam coverage range play a significant role in overcoming path loss in high-frequency bands and enhancing resolution [12][13][14][15][16][17][18][19]. ...
Article
Full-text available
Wide-angle mechanical beam steering for on-the-move satellite communications is presented in this paper based on a closed-form pillbox antenna system. It includes three main parts: a fixed-feed part, which is a substrate-integrated waveguide (SIW) horn with an extended aperture attached to a parabolic reflector; a novel quasi-optical system, which is a single coupling slot alongside and without spacing from the parabolic reflector; and a radiating disc, which is a leaky-wave metallic pattern. To make the antenna compact, pillbox-based feeding is implemented underneath the metallic patterns. The antenna is designed based on a substrate-guided grounded concept using leaky-wave metallic patterns operating at 20 GHz. Beam scanning is achieved using mechanical rotation of the leaky-wave metallic patterns. The proposed antenna has an overall size of 340 × 335 × 2 mm3, a gain of 23.2 dBi, wide beam scanning range of 120°, from −60° to +60° in the azimuthal plane, and a low side lobe level of −17.8 dB at a maximum scan angle of 60°. The proposed antenna terminal is suitable for next-generation ubiquitous connectivity for households and small businesses in remote areas, ships, unmanned aerial vehicles, and disaster management.
Chapter
The content of this chapter provides a thoughtful analysis of case studies that highlight the detection capabilities of FMCW radar systems operating in mmWave configurations. The case studies demonstrate how mmWave FMCW radar technology may be used to detect objects, motions, and changes in both line of sight and non-line of sight settings with accuracy, and efficiency. Each case study explores the unique difficulties presented by the application environment, which can include anything from identifying impediments in automotive safety systems to detecting minute movements for vital sign monitoring in healthcare. The steps for detecting different gesture recognition using IWR 1843 BOOST FMCW radar system and its processing are focussed upon. The document highlights the technology's excellent resolution, motion sensitivity, and adaptability to a variety of challenging environments in LoS and NLoS scenarios and with the technical details of operating mmWave radars, including signal processing methods, machine learning algorithms, and mitigating interference from surrounding objects.
Article
This paper presents the design and experimental assessment of a novel sensing system for controlling the orientation of a robotic hand in grasping non-metallic objects. This goal is pursued by using a radiofrequency (RF) probe placed on the robotic hand and a printed resonator located on the tagged item. The real and imaginary part of the probe input impedance are exploited to infer the angular orientation and distance between the object and the hand to help in finding the most suitable position for grasping. To assess the proposed design, a thorough experimental investigation has been carried out by using an ad-hoc testing platform mimicking the robotic hand. Two different types of probes have been considered to obtain the proper coupling with the adopted U-shaped ( hair pin ) compact resonator. The set of performed measurements prove that the individuated RF system configuration allows the identification of the relative angular position of the probe and the distance between the probe and the resonator offering advantages compared to others approaches in terms of range and accuracy.
Article
Full-text available
In recent years, radar sensors are gaining a paramount role in non-invasive inspection of different objects and materials. In this paper, we present a framework for using machine learning in material identification based on their reflected radar signature. We employ multiple receiving channels of the radar module to capture the signatures of the reflected signal from different target materials. Within the proposed framework, we present three approaches suitable for material classification, namely: (i) Convolutional Neural Networks (CNN), (ii) k-Nearest Neighbor, and (iii) Dynamic Time Warping (DTW). The proposed framework is tested using extensive experimentation and found to provide near-ideal classification accuracy in classifying six distinct material types. Furthermore, we explore the possibility of utilizing the framework to detect the volume of the identified material, where the obtained classification accuracy is above 98% in distinguishing three different volume levels.
Article
Estimation of indoor location is a key capability for indoor location-based services. Among various solutions, single-chip low-cost millimeter-wave radar as an emerging technology are getting more and more attention due to their inherent advantages such as penetration ability, low-power consumption. Currently, the vast majority of work is focused on hybrid positioning based on millimeter wave radar with other sensors, and there is a lack of research on single radar indoor positioning. In this paper, we propose a novel indoor positioning system based on single-radar sensor, which determines the trajectory by estimating the velocity and heading of moving radar from sparse point clouds. Doppler velocity of detected object can be able to feedback the radar velocity, and for another hand, the proposed Cluster Center Association method combined with Normal Distributions Transform (NDT) can be used for heading estimation. In addition, we proposed the motion consistency constraints to enhance the robustness of the system. The experimental results show that the traditional NDT algorithm will fail due to the sparse point cloud and severe indoor multipath, and the proposed algorithm can provide the positioning accuracy within 1 meter and the heading estimation error of 3 degrees.
Article
The galvanic skin response (GSR), which is a manifestation of the activity in eccrine sweat glands, has always been an important psychological test to evaluate emotional arousal and mental stress. Although, researches have focused on the radio reflections of skin and revealed that sweat gland activity can change the reflection signal of skin, the complicated experimental settings including vector network analyzer (VNA) and even optical path design limit the further daily life application. In this paper, we demonstrate the potential of using commodity millimeter-wave (mmWave) Multiple-Input Multiple-Output (MIMO) radar for contactless GSR sensing. To simulate the reflected signals of Frequency-Modulated Continuous-Wave (FMCW) radar to the skin of different individuals, a four layered skin model focused on the sweat duct is utilized and the change of conductivity in sweat ducts is regarded as mapping of sweat gland activity. Extensive experiments are conducted, consisting of 3 different physiological status: relaxing status, mental and physical stress status. We design a beamformer to transform raw signal to space domain so that the reflection signals from the palm are separated. The principal components of processed signal are extracted to eliminate the interference caused by breathing and random noise. We evaluate the performance in terms of the correlation between the radar signal and GSR before and after stress-induced events. The experimental results preliminarily verify the feasibility of contactless GSR sensing.
Article
Millimeter-wave (mmWave) radars are a solution in automotive applications such as Adaptive Driver Assisted Systems (ADAS). This work explores the use of commercially-available mmWave radar technology for in-cab vehicle applications. A framework for vehicle occupancy detector based on mmWave radar at 77 GHz is presented. A multiple-input multiple-output (MIMO) frequency-modulated continuous wave (FMCW) radar from Texas Instruments (AWR1642) is used. The system can detect the presence of people occupying the seats by measuring small movements of the body. To this end, the static clutter is removed, therefore the point cloud returned by the radar contains information about the body. The radar is installed on the roof of the vehicle to have the maximum field of view of the seats. An algorithm that uses both the density and the dispersion of points around each seat is proposed to determine its degree of occupation. Experimental results have been presented for 4 and 5- seater vehicles with high accuracy.
Article
With the recent hike in the autonomous and automotive industries, sensor-fusion-based perception has garnered significant attention for multiobject classification and tracking applications. Furthering our previous work on sensor-fusion-based multiobject classification, this letter presents a robust tracking framework using a high-level monocular-camera and millimeter wave radar sensor-fusion. The proposed method aims to improve the localization accuracy by leveraging the radar's depth and the camera's cross-range resolutions using decision-level sensor fusion and make the system robust by continuously tracking objects despite single sensor failures using a tri-Kalman filter setup. The camera's intrinsic calibration parameters and the height of the sensor placement are used to estimate a birds-eye view of the scene, which in turn aids in estimating 2-D position of the targets from the camera. The radar and camera measurements in a given frame is associated using the Hungarian algorithm. Finally, a tri-Kalman filter-based framework is used as the tracking approach. The proposed approach offers promising MOTA and MOTP metrics including significantly low missed detection rates that could aid large-scale and small-scale autonomous or robotics applications with safe perception.
Article
The demand for non-contact breathing and heart rate measurement is increasing. In addition, because of the high demand for medical services and the scarcity of on-site personnel, the measurement process must be automated in unsupervised conditions with high reliability and accuracy. In this paper, we propose a novel automated process for measuring breathing rate and heart rate with mmWave radar and classifying these two vital signs with machine learning. A frequency modulated continuous wave (FMCW) mmWave Radar is integrated with a pan, tilt and zoom (PTZ) camera to automate camera steering and direct the radar toward the person facing the camera. The obtained signals are then fed into a deep convolutional neural network to classify them into breathing and heart signals that are individually low, normal, and high in combination, yielding 6 classes. This classification can be used in medical diagnostics by medical personnel. The average classification accuracy obtained is 87% with a precision, recall and F1 score of 0.93.
Article
Millimeter-wave radar is a device gaining popularity in many applications such as occupancy detection and people counting, vital sign detection and robotics. It has recently been used for human detection and tracking within home health care as a potential application area. Furthermore, increasing number of researchers utilize mmWave radar for heart rate detection. In this paper, we introduce a new home health monitoring system based on millimeter-wave radar, which is composed of posture estimation and heart rate detection. We focus on categorisation of postures (standing, sitting and lying) for a single person in indoor environments. When the person is estimated as lying down, our system can further detect the heart rate. We establish a new point cloud dataset and show that, with simple algorithms employed, the system has an accuracy of around 99% for posture estimation and approximately 91% accuracy for detecting heart rate with only a single millimeter-wave radar.