ChapterPDF Available

Authentication Framework for Healthcare Devices Through Internet of Things and Machine Learning

Authors:

Abstract and Figures

The rise in the application and use of IoT has made tech gadgets and gizmos smarter and interconnected than ever before. IoT has expanded in a number of fields including smart cities and homes, healthcare, finance, etc. A typical IoT device is likely to integrate computation, networking, and physical processes from embedded computational gadgets. To monitor and extract valuable existential patterns from the large volume of data that is generated, Machine Learning (ML) helps a lot by the different number of algorithms that can be developed. Typically, ML is a discipline that covers the two major fields of constructing intelligent computer systems and governing these systems. ML has progressed dramatically over the past two decades and has emerged as one of the major methods for developing computer vision systems and other pedagogies. Rapid advancements and enhancements in these fields are leading to extensive interactions among the devices in the heterogeneous pool of gadgets. However, with advancements, there always will be a few bottlenecks that hinder the security and safety of the device or the gadget. Making use of such methodologies and techniques for user access control exposes this to endless vulnerabilities including numerous attacks and other complications. It is extremely important to protect the authenticity and privacy of the users and the data that is stored in these smart devices. This paper discusses the variety of ways in which a smart device can validate the user with proper authentication and verification.KeywordsSecure machine learningCloud computingBiometric authenticationInternet of Things
Content may be subject to copyright.
Authentication Framework for Healthcare Devices
through Internet of Things and Machine Learning
Shruti Kute1, Shreyas Madhav A V2, Amit Kumar Tyagi 2,3[0000-0003-2657-8700 ],
Atharva Deshmukh4
1,2School of Computing Science and Engineering, Vellore Institute of Technology,
Chennai, 600127, Tamilnadu, India.
3Centre for Advanced Data Science, Vellore Institute of Technology, Chennai,
600127, Tamilnadu, India
Terna Engineering College, Navi Mumbai
shrutikute99@gmail.com, shreyas.madhav@gmail.com
amitkrtyagi025@gmail.com atharva1525@gmail.com
Abstract. The rise in the application and use of IoT has made tech gadgets and
gizmos smarter and interconnected than ever before. IoT has expanded in a
number of fields including smart cities and homes, healthcare, finance, etc. A
typical IoT device is likely to integrate computation, networking, and physical
processes from embedded computational gadgets. To monitor and extract
valuable existential patterns from the large volume of data that is generated,
Machine Learning (ML) helps a lot by the different number of algorithms that
can be developed. Typically, ML is a discipline that covers the two major fields
of constructing intelligent computer systems and governing these systems. ML
has progressed dramatically over the past two decades and has emerged as one
of the major methods for developing computer vision systems and other
pedagogies. Rapid advancements and enhancements in these fields are leading to
extensive interactions among the devices in the heterogeneous pool of gadgets.
However, with advancements, there always will be a few bottlenecks that hinder
the security and safety of the device or the gadget. Making use of such
methodologies and techniques for user access control exposes this to endless
vulnerabilities including numerous attacks and other complications. It‟s
extremely important to protect the authenticity and privacy of the users and the
data that‟s stored in these smart devices. This paper discusses the variety of
ways in which a smart device can validate the user with proper authentication
and verification.
Keywords: Secure Machine learning, Cloud computing, biometric
authentication, Internet of Things.
1 Introduction
The actual items or 'things, around us (for example, actuators, sensors, and many
more) are progressively becoming interconnected, by and large, by means of the Web.
These interconnected things gather (or sense), offer, but also cycle data in a serious
information-driven Internet of Things (IoT) setting. Cloud computing is been used in
2
IoT organisations, across both regular citizen and military applications, to manage the
dynamic prerequisites like capacity, computation (e.g., information examination and
representation, etc.).
Like most worker customer or organized frameworks, there seems to be a danger
that IoT application workers could be focused on by assailants, along with pernicious
insiders, mostly with points of gaining unapproved admittance to information. In other
words, the customer hubs are powerless against dangers, for example, disconnected
secret key speculating assaults, client pantomime assaults, insider assaults, and client
explicit key robbery assaults. Biometric-put together validation systems (for example,
once based with respect to the person's unique mark, face, voice, fingerprint,
keystroke, and iris) are being demonstrated for being genuinely strong towards
ordinary dangers than with traditional secret key-based verification plans. In
particular, in such a framework, the highlights of the enlisted clients are put away in a
concentrated information base in the cloud to improve the cycle of validation. During
validation, a client is resolved to be either real or not founded on the coordinating
likelihood with the test layout that is put away in the information base.
IoT devices give a variety of new options for healthcare providers to monitor
patients as well as for patients to monitor themselves, such as Ingestible sensors,
which gather information from the digestive system as well as provide information
about stomach PH levels. Automated insulin delivery (AID) systems, also known as
closed-loop insulin delivery systems or Artificial Pancreas systems, are
groundbreaking for persons with diabetes, a disorder that affects around 8.5 percent of
adults globally, according to World Health Organization data.
2 Literature Survey
The blast of IoT has prompted the expanded utilization of biometrics
authentication in different cloud computing focused administrations. Peer et al.[1] had
imagined necessities for developing cloud based biometric arrangements. They
likewise proposed a contextual analysis in a cloud climate utilizing a unique mark-
based verification administration.
Yuan et al.[2] proposed a novel biometric recognizable proof plan with
security safeguarding. The greater parts of the tasks were safely moved to cloud
workers. The information base proprietors produced accreditations. These
qualifications were utilized in the recognizable proof help actualized by cloud workers
in an encoded information base. Such a technique would guarantee the secrecy of
personal biometric information in a cloud. Haghigat et al.[3] proposed a cloud based
cross-endeavor biometric recognizable proof plan known as "CloudId" where
encryption methods are utilized to give a dependable biometric validation framework.
Be that as it may, the framework experienced weighty computational multifaceted
nature. Das and Goswami [4] proposed a productive, secure, safe, and protection
saving biometric-based far off confirmation administration for E-healthcare
frameworks. By and by, the framework experienced different weaknesses.
The cancellable biometrics framework is a stride in front of the ordinary
biometric framework. The arrangement of a cancellable biometric framework in the
cloud climate is as yet a difficult issue and an open exploration territory. Amin et al.
[5] have proposed a common verification and meeting key arrangement in a distant
climate utilizing the cancellable bio-decimal measuring standard. During client
3
verification, they used a cancellable biometric framework known as bio hashing. In a
cloud computing climate, Bommagani et al. [6] proposed a cancellable face
acknowledgment structure where acknowledgment tasks had been moved to the cloud
to misuse the equal execution ability of the cloud. At the same time, a cancellable face
layout age calculation dependent on an irregular projection framework[7] was likewise
proposed. Wu et al. [8] also proposed a method for creating multiple keys from a
client's finger vein for validating various cloud-based administrations utilising a high-
dimensional space self-adjustment calculation. When compared to cancellable
biometrics, the method for producing various keys with a fluffy vault plot is
computationally perplexing.
Hu et al. [9] proposed a cloud-based face confirmation plot in which they re-
evaluated every one of the processes, from image pre-processing to coordinating cloud
workers.[10] They also additionally abused the cloud's equal cycle execution limit and
used facial confirmation for cloud based IoT applications. Notwithstanding, there is a
gap that requires lightweight change methods to be conveyed in the cloud climate. A
few scientists have received cloud computing innovation to satisfy the needs regarding
force and capacity limits.[11] As a result, there is indeed a prerequisite for cloud-based
administrations to serve as a viable cross-stage for different IoT applications.
Karimian et al. [12] used electrocardiogram (ECG) signals to authenticate an
Internet of things in, observing that Electrocardiography biometrics are efficient,
secure, and simple to deploy. Kantarci et al. [13] proposed a cloud-centric model in
which biometric authentication architecture combines a biometric method with
context-awareness method for preventing unwanted access to mobile apps. Dhillon
and Kalra et al. [14] devised a lightweight multi-factor identity - based cryptographic
method using less expensive hash functions. Ratha et al. [15] was the first one to
propose cancelable biometrics utilising 3 separate transformation functions where first
is Cartesian transformation, second would be Polar transformation, and last one
is Functional transformation.
3 Internet of Things and Machine Learning based
Authorization Model
Biometric distinguishing proof empowers end-clients to utilize actual
characteristics rather than passwords or PINs as a safe strategy for getting to a
framework or an information base. Biometric innovation depends on the idea of
supplanting "one thing you have with you" with "what your identity is," which has
been viewed as a more secure innovation to protect individual data. The conceivable
outcomes of applying biometric ID are truly colossal.[16] Biometric ID is applied
these days in areas where security is a first concern, similar to air terminals, and could
be utilized as a way to control outskirt crossing adrift, land, and air wilderness.
Particularly for the air traffic region, where the quantity of flights will be expanded by
40% before 2013, the confirmation of portable IoT gadgets will be accomplished when
the bio-features models become adequately full grown, productive, and impervious to
IoT assaults.
Another region where biometric ID strategies are beginning to be received is
electronic IDs. Biometric ID cards, for example, the Estonian and Belgian public ID
cards were utilized to distinguish and confirm qualified citizens during races. Moving
above and beyond, Estonia has presented the Versatile ID framework that permits
4
residents to lead Web casting a ballot and consolidates biometric distinguishing proof
and cell phones. This framework that was very creative when it was at first acquainted
has a few dangers with the constituent technique and was reprimanded for being
uncertain.
As per an overview by Spear System and Exploration, in 2014, $16 billion
was taken by 12.7 million individuals who were survivors of wholesale fraud in the
US as it were. This sum is determined without considering the financial issues and
mental mistreatment that survivors of this extortion endure. From the financial area
and organizations to admittance to homes, vehicles, PCs, and cell phones, biometric
innovation offers the most elevated level of security as far as protection and security
insurance and secure access.
Cell phones are these days a fundamental piece of our regular daily existence,
as they are utilized for an assortment of portable applications. Performing biometric
validation through cell phones can give a more grounded system to personality check
as the two verification factors, "something you have" and "something you are," are
joined. A few arrangements that incorporate multibiometric and conduct validation
stages for telecom transporters, banks, and different businesses were as of late
presented.
3.1 Biometric based user authentication for smart IoT devices
There are diverse related reviews that manage client verification. Albeit some of them
covered distinctive validation strategies, we just consider those that were completely
devoted for biometric confirmation. We characterize the studies as per the
accompanying rules:
a) Deployment scope: it shows whether the confirmation plot is sent on cell
phones or not.
b) Focus biometric region: it shows whether the review zeroed in on
all/particular biometric highlights.
c) Threat models: it demonstrates whether the study considered the dangers
against the verification plans.
d) Countermeasures: it shows whether the study zeroed in on and considered the
countermeasures to shield the verification plans.
e) Machine learning (ML) and data mining (DM) calculations: they show
whether the study makes reference to for every arrangement the pre-owned
AI or data mining technique.
A few reviews depicted the validation conspires that just consider explicit bio
features. For example, the studies just centered around the keystroke elements. Then
again, Gafurov introduced biometric stride acknowledgment frameworks. Revett et al.
overviewed biometric verification frameworks that depend on mouse developments.
Yampolskiy and Govindaraju introduced an extensive report on conduct biometrics.
Mahadi et al. studied social based biometric client verification and decided the
arrangement of best classifiers for conduct based biometric confirmation.[17],[18]
Sundararajan and Woodard surveyed 100 unique methodologies that utilized deep
learning and different biometric modalities to distinguish clients. Teh et al.presented
distinctive confirmation arrangements that depend on touch elements in cell phones.
Rattani and Derakhshani gave the best in the class identified with face biometric
confirmation plots that are intended for cell phones. They additionally examined the
5
satire assaults that target portable face biometrics just as the anti-spoofing techniques.
Mahfouz et al. reviewed the social biometric confirmation conspire that are applied on
cell phones. Meng et al. studied the validation systems utilizing biometric clients on
cell phones. They recognized eight possible assaults against these confirmation
frameworks alongside promising countermeasures. Our study and both spotlight on
verification plots that are intended for cell phones and consider all the biometric
highlights and manage danger models and countermeasures. Nonetheless, doesn't give
data identified with the pre-owned AI or data mining strategy for all the reviewed
arrangements. Likewise, just conceals papers to 2014, while the inclusion of our study
is up to 2018. Supposedly, this work is the primary that altogether covers dangers,
models, countermeasures, and the AI calculations of the biometric confirmation plans.
3.2 Biometric based user authentication cancellation
The idea of cancellable biometrics is that the first layout information is
changed into an alternate adaptation by utilizing a non-invertible change work in the
enlistment stage. Question information in the confirmation stage is applied a similar
non-invertible change. Coordinating is directed in the changed space utilizing the
changed layout and question information.
Ratha et al. started three distinctive change capacities, known as Cartesian,
polar, and practical changes. The proposed change works deliberately misshape the
first highlights, with the goal that it is infeasible or computationally hard to recover
crude format information. Notwithstanding, one disadvantage is that the proposed
strategy is enlistment based, and subsequently, the exact discovery of solitary focuses
is required. Typically, exact enrolment is hard as a result of biometric vulnerability
(e.g., picture relocation, non-direct bending, and obtaining condition). Jin et al.
proposed a two-factor validation strategy called bio-hashing. Bio-hashing joins token-
based information with unique mark highlights by the iterative inward item to make
another list of capabilities. At that point, each incentive in the list of capabilities is
changed over to a paired number dependent on a predefined edge. Lee et al. produced
cancellable unique mark formats by separating a revolution and interpretation
invariant element for each detail, which is considered to be the principal arrangement
free cancellable unique mark layout plan. Ahn et al. utilized trios of particulars as a list
of capabilities, and change is performed on mathematical properties got from the trios.
Yang et al. made cancellable layouts by utilizing both neighbourhood and worldwide
highlights. Neighbourhood highlights incorporate distances and relative points
between particulars sets, while worldwide highlights incorporate direction and edge
recurrence. In this exploration, the distance of a couple of particulars is changed
utilizing an opposite projection, to determine the non-invertible change.
Ahmad and Hu proposed an arrangement free structure dependent on a couple
of polar organize. In this structure, the general situation of every minutia to all other
particulars among a polar facilitate range is used. From any two particulars, three
neighbourhood highlights are extricated and changed by a useful change to produce
the cancellable layout. In view of the minutia structure, Wang et al. further improved
framework security and precision by proposing some new change capacities, for
example, endless to-one planning, abridged roundabout convolution, and incomplete
Hadamard change. Zhang et al. planned a combo plate and a utilitarian change to
deliver cancellable layouts dependent on the Minutia Cylinder Code (MCC). MCC is a
notable neighbourhood minutia descriptor, which depends on 3D nearby structures
related to every minutia. The creators of the MCC later proposed a format insurance
6
strategy named P-MCC, which plays out a KL change on the MCC highlight portrayal.
Nonetheless, P-MCC doesn't have the property of revocability. At that point, 2P-MCC
was proposed to add cancelability to P-MCC utilizing a halfway change-based plan.
Afterward, Arjona et al. introduced a protected unique mark coordinating
methodology, named P-MCC-PUFs, which contains two components dependent on P-
MCC and PUFs (Genuinely Unclonable Capacities). The proposed conspire
accomplishes the best presentation when the length of the element vector is set to 1024
pieces and gives solid information protection and security. Yang et al. planned a
cancellable unique mark layout dependent on arbitrary projection. The planned format
can shield assaults through record variety (ARM) inferable from the component
decorrelation calculation. Meanwhile, a Delaunay triangulation-based nearby structure
proposed in the plan can diminish the negative impact of nonlinear twisting on
coordinating execution. Sandhya and Prasad intertwined two structures, nearby
structure, and inaccessible structure, at the elementary level to create twofold esteemed
highlights, which are then ensured by an arbitrary projection-based cancellable
insurance strategy.
To additional upgrade security and acknowledgment execution, a few analysts
proposed the utilization of multimodal cancellable biometrics. For instance, Yang et
al. proposed a multimodal cancellable biometric framework that wires unique mark
highlights and finger-vein highlights to accomplish better acknowledgment precision
and higher security. In the proposed framework, an improved incomplete discrete
Fourier change is used to give non-invertibility and revocability. Additionally,
Dwivedi and Dey proposed a mixture combination (score level and choice level
combination) plan to coordinate cancellable unique mark and iris modalities to
decrease impediments in every individual methodology. Test consequences of
multimodal cancellable biometric frameworks show execution improvement over their
unimodal partner.
In this segment, the development of cancellable biometrics, from the
presentation of the possibility of cancellable biometrics and some early change work
plans, to the ongoing different cancellable biometrics, is introduced. There are two
classifications in the plan of cancellable biometrics. One class revolves around the
extraction and portrayal of stable biometric includes in order to accomplish better
acknowledgment precision, and the other classification centers around planning secure
change capacities, which are relied upon to be numerically non-invertible. It is
foreseen that future exploration work in cancellable biometrics will endeavour to
accomplish both better acknowledgment precision and more grounded security by
utilizing numerous cancellable biometrics.
4 Proposed framework for Authorization and authentication of
Smart IoT devices
An adaptable security system is required to address the vastly different IoT climate
and the associated security challenges. Given below Figure - 1 shows a structure to
make sure about the IoT climate and is contained four segments:
Authentication
Authorization
Network Enforced Policy
Secure Analytics: Visibility and Control
7
Fig. 1. Internet of Things Framework
Authentication: The validation layer is at the core of this system, and it is
used to provide and confirm the distinguish data of an IoT substance. When
connected IoT/M2M gadgets (example, inserted sensors and actuators or
endpoints) require access to IoT framework, a relationship of trust is started
dependent with character of the gadget. The best approach of storing as well
as manifesting personality data may be exceedingly exceptional for IoT
gadgets. It should be noted that in average venture organisations, its
endpoints might be distinguished by only a human accreditation (example,
username and secret key, biometrics, or token). IoT/M2M endpoints should
always be fingerprinted by implies that they do not require any human
collaboration. All these identifiers incorporate radio-recurrence distinguishing
proof (RFID), shared mystery, X.509 testaments, the endpoint's Macintosh
address, or other kind of unchanging equipment-based foundation of trust.
Setting up character by means of X.509 testaments give a solid confirmation
framework. In any case, in the IoT area, numerous gadgets might not have
enough memory to store an endorsement or may not have the necessary
computer processor capacity to execute the cryptographic activities of
approving the X.509 authentications (or any kind of open key activity).
Existing personality impressions, e.g., 802.1AR and verification norms as
defined by IEEE 802.1X, could be used for devices that can manage both the
central processor load and memory to store solid qualifications. Nonetheless,
the problems of a new structure factors, as well as new modalities, open the
door for additional examination in describing more modest impression
qualification types as well as less process focused cryptographic builds and
verification norms.
Authorization: This is the second layer of this system is authorization, which
controls a device's entry throughout the organization's texture. This layer
extends the central validation layer by utilising an element's character data. A
trust relationship is established between IoT devices to trade appropriate data
8
using confirmation and approval segments. For example, a vehicle might well
set up a trust alliance with some other vehicle from the same seller.
Regardless of the trust relationship, vehicles may only be able to trade one„s
well-being capacities. When a trustworthy partnership is established between
a similar vehicle and their vendor's organisation, the vehicle may be
permitted to start sharing additional data, e.g., the odometer reading, last
upkeep track, and so on. Fortunately, current arrangement instruments for
both overseeing and controlling access to shopper and undertaking networks
map exceptionally well to the IoT/M2M requirements. The major challenge
will be to create a technology that can scale to deal with billions of IoT/M2M
gadgets with varying trust connections in the texture. To fragment
information traffic and build up start to finish correspondence, traffic
approaches and appropriate controls will be used throughout the organisation.
Network Enforced Policy: This layer includes all components that safely
route and transport endpoint traffic over the foundation, whether control,
board, or real information traffic. There are currently established conventions
and instruments to ensure about the organisation foundation and influence
strategy that are appropriate to the IoT/M2M use cases, similar to the
Approval layer.
Secure Analytics / Visibility and Control: This safe examination layer
characterizes the administrations by which all components (endpoints and
organization framework, inclusive of server farms) may partake in giving
telemetry to the reason for picking up perceivability and ultimately
controlling the IoT/M2M biological system. We can convey a huge equal
information base (MPP) stage that can cycle huge volumes of information in
close to continuous with the development of massive information
frameworks. When we combine this innovation with examination, we can
conduct a genuine factual investigation on the security data to identify
anomalies. Furthermore, it includes all components that total and correlate
data, including telemetry, to provide observation and danger identification.
Danger alleviation can range from preventing the assailant from accessing
additional assets to running specific content to begin legitimate remediation.
The data generated by IoT devices is only useful if the privilege investigation
calculations or other security knowledge measures are labelled to identify the
threat. We can enhance scientific results by collecting information from
multiple sources and applying security profiles and measurable models based
on various layers of security calculations.
We are all aware that network foundations are becoming increasingly
complex. Consider geographies with both public and private mists; the threat
knowledge and protection capacities should also be cloud-based. To achieve
precise insight, coordination of perceivability, setting, and control is required.
This layer's components include the following:
The genuine IoT/M2M foundation from which telemetry and observation
information is obtained and accumulated. The center arrangement of
capacities to blend, examine the information for the purposes of giving
perceivability, and offer relevant mindfulness and control. The conveyance
stage for the genuine analysis, operated from the first two segments,
discussed above.
9
While real-world IoT/M2M implementations may be unique, their structure
could be applied to almost any engineering. The framework is simple and
adaptable enough to support controlled gadgets (e.g., PCs, mobile scanners,
etc) if they live in the IoT base.
5 Classification of various analytic technique for Smart IoT
devices
The quantity of gadgets associating with the Web is ballooning, introducing
the time of the "Internet of Things" (IoT). IoT alludes to the huge number of minimal
effort gadgets that speak with one another and with far off workers on the Web
independently. It includes ordinary articles, for example, lights, cameras, movement
sensors, entryway locks, thermostats, power switches, and family machines, with
shipments extended to arrive at almost 20 billion by 2020. A great many IoT gadgets
are required to discover their way in homes, ventures, grounds, and urban communities
of the not-so-distant future inciting "shrewd" conditions profiting our general public
and our lives.
The multiplication of IoT, in any case, makes a significant issue.
Administrators of shrewd conditions can think that it‟s hard to figure out what IoT
gadgets are associated with their organization and further to determine whether every
gadget is working ordinarily. This is for the most part ascribed to the undertaking of
overseeing resources in an association, which is commonly circulated across various
offices. For instance, in a neighbourhood board, lighting sensors might be introduced
by the office's group, sewage and trash sensors by the disinfection office, and
observation cameras by the nearby police division.
Organizing across different divisions to get a stock of IoT resources is tedious,
burdensome and blunder inclined, making it almost difficult to know absolutely what
IoT gadgets are working on the organization anytime. Acquiring "perceivability" into
IoT gadgets in an opportune way is of fundamental significance to the administrator,
who is entrusted with guaranteeing that gadgets are in suitable organization security
fragments, are provisioned for imperative nature of administration, and can be isolated
quickly when penetrated. The significance of perceivability is accentuated in Cisco's
latest IoT security report, and further featured by two late occasions: sensors of a fish
tank that undermined a gambling club in Jul 2017, and assaults on a college grounds
network from its own candy machines in Feb 2017. In the two cases, network division
might have possibly forestalled the assault and better perceivability would have
permitted quick isolating to restrict the harm of the digital assault on the endeavour
organization.
One would expect that gadgets can be distinguished by their Macintosh
address and DHCP exchange. Notwithstanding, this faces a few difficulties: (a) IoT
gadget makers ordinarily use NICs provided by outsider merchants, and consequently
the Authoritatively Extraordinary Identifier (OUI) prefix of the Macintosh address
may not pass on any data about the IoT gadget; (b) Macintosh locations can be mock
by vindictive de-indecencies; (c) numerous IoT gadgets don't set the Host Name
choice in their DHCP demands; in fact, we found that about a large portion of the IoT
gadgets we considered don't uncover their hostnames, as appeared in Table 1; (d) in
any event, when the IoT gadget uncovered its hostname it may not generally be
important (for example WBP-EE4C for Withing‟s infant screen in Table 1); and
10
finally (e) these hostnames can be changed by the client (for example the HP printer
can be given a self-assertive hostname). Hence, depending on the DHCP foundation is
anything but a feasible answer for accurately recognize gadgets at scale.
5.1 Descriptive analysis for IoT
The illustrative investigation is the most fundamental type of scientific knowledge that
permits clients to portray and total approaching IoT information. Illustrative
investigation - even computations as straightforward as a mean and standard deviation
- can be utilized to rapidly sort out gathered information. In an associated plant use
case, portrayal examination may be utilized to respond to the inquiry, "What are the
normal siphon temperature, stream rate, and RPM throughout a 30-minute time span?"
When distinguishing top tier clear investigation abilities on an IoT stage, undertakings
ought to assess:
On-stage clear examination capacities: The capacity of a stage to perform
expressive insightful requests, for example, amassing or figuring essential
measurements of ingested information focuses across sensors, gadgets, or
gatherings of gadgets just as outwardly introducing the outcomes.
On-stage information lake/huge information stockpiling capacities: The
capacity of the stage to both store and inquiry against huge amounts of
ingested IoT information including table-based information stores with more
noteworthy than 10 million lines or unstructured information stores with
more prominent than 50 million records.
5.2 Predictive analysis for IoT
Prescient examination looks to demonstrate future information and practices by
breaking down recorded information.[19] Relapse examination, for example, direct
relapse is an illustration of prescient investigation. In a similar use case, prescient
examination may be utilized to respond to the inquiry, "What is the assessed time-to-
disappointment for a siphon that is showing a 20% expansion in estimated
temperature?" When recognizing top tier prescient examination abilities on an IoT
stage, endeavours ought to assess:
On-stage prescient scientific model structure: The capacity of the stage to
consequently or through automatic interfaces create a prescient model of the
fundamental stage ingested IoT information. Models, for example, direct or
polynomial relapses are average, albeit more intricate displaying decisions
are accessible in refined stages.
On-stage prescient logical model activity: The capacity of the stage to use
either a stage created or stage coordinated information model, (for example,
R or Python) to group information or recognize exceptions through
abnormality identification. Clients should put accentuation on the capacity to
oversee models, for example, model forming and refreshing just as the
capacity to incorporate a prescient model inside an unpredictable occasion
preparing (CEP) structure[20].
5.3 Prescriptive analysis for IoT
Prescriptive examinations are investigations to assist ventures with advancing a future
course to be taken. Picture handling, AI, and characteristic language preparing are a
11
portion of the strategies used to finish prescriptive investigation. A prescriptive
investigation may be utilized to respond to the inquiry, "To boost siphon uptime and
limit administration spans, what is the most extreme permitted temperature increment
for a siphon before a protection siphon adjusting must be booked?" When recognizing
top tier prescriptive investigation capacities on an IoT stage, endeavours ought to
assess
On-stage prescriptive insightful model capacities: The capacity of the stage to
use either a stage produced or stage coordinated information model, for
example, R or Python, to upgrade a business result or applicable KPI. A
prescriptive model ought to augment or limit a business-important KPI, for
example, an ideal opportunity to-conveyance in course arranging or hardware
uptime for prescient upkeep.
6 Proposed System: IoT-based Lightweight Cancellable
Biometric System
Fig. 2. Proposed Biometric System
The Figure 2 shows a unique mark-based verification plot in a cloud climate. In such a
setting, the concentrated information base is a prominent objective because of the
capacity of biometric formats. Since the quantity of each biometric methodology is
restricted, the outcomes of any trade off will be extensive (e.g., one's biometrics, for
example, iris can't be supplanted). Consequently, premium in cancellable biometrics is
created. The Cancellable Biometric Framework (CBS) by and large includes a sign or
12
highlight level change of the first biometric highlight dependent on some client
explicit key.[21] As such, instead of the first/crude biometric features, the changed
layout is put away in a concentrated information base. Consequently, if the
incorporated information base is compromised, the CBS can essentially drop the
undermined format and conjure another changed layout from the first bio-metric
highlights with the end goal that the undermined format and the new layout are
completely extraordinary. The client can likewise try out various administrations (e.g.,
internet banking, web-based business, and e-government) utilizing diverse cancellable
formats created out of a solitary biometric methodology. Key advantages of the CBS
over a traditional (or 'plain vanilla') biometric framework are as per the following:
Cancellable biometric formats hold a single direction property since the
cancellable changes are equal to cryptographic salting.
A single biometric source may be used to generate a plethora of different
formats. Each layout is unrelated to the others.
The client could enlist a variety of applications in a variety of various
formats. The first/crude biometric example won't be uncovered.
Revocation is useful in the sense that a newly created model replaces the
old/bargained interface (similar to changing a password).
The focal point of this examination is on the advancement of a lightweight, cloud-
based cancellable biometric system, which is intended to give secure client verification
to various IoT applications [22, 23]. Figure 2 shows our proposed cancellable
biometric-based client validation conspire in a cloud climate com-prising both
customer hubs (e.g., portable processing or IoT gadgets) and cloud server(s). Every
client (in charge of at least one customer hub) has various cancellable unique mark
layouts produced from a solitary finger impression picture. Clients can be confirmed
utilizing the cancellable unique mark format, and each cancellable unique mark layout
is one of a kind and doesn't correspond with different formats despite the fact that they
are created from the same biometric methodology. Likewise, we re-appropriated
exercises, for example, picture pre-processing, include extraction, highlight change
and layout coordinating to the cloud [19]. Such exercises can be gotten to through a
UI, which can be a program or a portable (application). Both the cancellable biometric
information base and applicable programming modules are likewise moved to the
cloud. Such an arrangement permits us to give constant and equal preparation and
understand the accompanying advantages:
Negligible framework arrangement time.
Arrangement of momentary and on-request administration alongside the
chance of the expansion and additionally erasure of parts.
Reasonable innovation, in any event, for little and medium-sized
organizations, because of insignificant arrangement expenses and time.
Exceptionally adaptable, since the cancellable biometric information base can
be scaled to fit any sort of utilization, for example, 1: 1 or 1: N check
situations, with accessible asset pooling.
The following are the components of the proposed cloud-based, lightweight
cancellable biometric framework:
Lightweight cancellable layout age in the cloud climate, and
Lightweight cancellable layout coordinating in the cloud climate.
The vital commitments of this paper are as per the following:
13
In a cloud climate, the proposed cloud-based lightweight cancellable
biometric confirmation framework can successfully distinguish a person;
The proposed cloud-based lightweight cancellable biometric confirmation
framework may also likewise adequately understand the cancellable
biometric framework-based verification administrations in different cloud-
based IoT applications; and
When conveyed in confirmation administrations, the proposed framework
consumes less time; henceforth, it is known as the lightweight cancellable
biometric framework.
Many academics have been working on creating biometric-based techniques for
authentication process in IoT networks because of the advantages like uniqueness of
biometrics over password- and token-based traditional authentication. A framework
may be built utilising biometrics and wireless device radio fingerprinting for user
authentication, which not only verifies that the observed healthy data is from the
proper patient, but also guarantees the data's security.
Fig. 3. Cancellable biometric-based user authentication
7 Conclusion
Given the growing trend as well as variety of IoT devices in our general public, the
need to validate such devices will become more articulated. We introduced a cloud-
based, lightweight cancellable biometric authentication system for IoT applications in
this paper. The use of cancellable biometric layouts addresses the security concerns
14
that underpin the use of biometrics for confirmation. We exhibited our proposed
framework's lightweight trademark, precision, and security.
8 Future works and limitations
In the process of developing a biometric authentication system for IoT applications,
use of lightweight adaption has some other privacy-related flaws that can make it
sometimes unsuitable for use in biometrics. These flaws can be solved in the future by
employing one-way functions in cryptography with biometrics to bring clarity in the
above lightweight method and its shortcomings. In the future, we will examine other
sorts of transformation functions and investigate how to effectively hide the secret key
in different circumstances.
References
1. P. Peer, J. Bule, J. Z. Gros, and V. Struc. Building cloud-based biometric services.
Informatica, 37(2):115, 2013
2. J. Yuan and S. Yu. Efficient privacy-preserving biometric identification in cloud
computing. In INFOCOM, 2013 Proceedings IEEE, pages26522660, 2013.
3. M. Haghighat, S. Zonouz, and M. Abdel-Mottaleb. Cloudid: Trustworthy cloud-
based and cross-enterprise biometric identification. ExpertSystems with
Applications, 42(21):79057916, 2015.
4. A. K. Das and A. Goswami. A secure and efficient uniqueness-and-anonymity-
preserving remote user authentication scheme for connectedhealth care. Journal of
medical systems, 37(3):9948, 2013.
5. R. Amin, S. H. Islam, G. P. Biswas, M. K. Khan, and X. Li. Cryptanalysis and
enhancement of anonymity preserving remote user mutualauthentication and
session key agreement scheme for e-health care systems. Journal of medical
systems, 39(11):140, 2015.
6. A. S. Bommagani, M. C. Valenti, and A. Ross. A framework for secure cloud-
empowered mobile biometrics. In 2014 IEEE MilitaryCommunications
Conference (MILCOM‟14), pages 255–261, 2014.
7. E. Bingham and H. Mannila. Random projection in dimensionality reduction:
applications to image and text data. In Proceedings of theseventh ACM SIGKDD
international conference on Knowledge discovery and data mining, pages 245
250, 2001
8. Z. Wu, L. Tian, P. Li, T. Wu, M. Jiang, and C. Wu. Generating stable biometric
keys for flexible cloud computing authentication using fingervein. Information
Sciences, 433-434.
9. P. Hu, H. Ning, T. Qiu, Y. Xu, X. Luo, and A. K. Sangaiah. A unified face
identification and resolution scheme using cloud computing ininternet of things.
Future Generation Computer Systems, 81:582592, 2018.
10. Adam, Edriss Eisa Babikir. “Survey on Medical Imaging of Electrical Impedance
Tomography (EIT) by Variable Current Pattern Methods.” Journal of ISMAC 3,
no. 02(2021): 82-95.
11. Amit Kumar Tyagi, Poonam Chahal, “Artificial Intelligence and Machine
Learning Algorithms”, Book: Challenges and Applications for Implementing
15
Machine Learning in Computer Vision, IGI Global, 2020.DOI: 10.4018/978-1-
7998-0182-5.ch008
12. Karimian, N.; Wortman, P.A.; Tehranipoor, F. Evolving authentication design
considerations for the internet of biometric things (IoBT). In Proceedings of the
Eleventh IEEE/ACM/IFIP International Conference on Hardware/Software
Codesign and System Synthesis, Pittsburgh, PA, USA, 17 October 2016; p. 10.
13. Kantarci, B.; Erol-Kantarci, M.; Schuckers, S. Towards secure cloud-centric
internet of biometric things. In Proceedings of the 2015 IEEE 4th International
Conference on Cloud Networking (CloudNet), Niagara Falls, ON, Canada, 57
October 2015; pp. 8183.
14. Dhillon, P.K.; Kalra, S. A lightweight biometrics based remote user authentication
scheme for IoT services. J. Inf. Secur. Appl. 2017, 34, 255270.
15. Ratha, N.K.; Connell, J.H.; Bolle, R.M. Enhancing security and privacy in
biometrics-based authentication systems. IBM Syst. J. 2001, 40, 614634.
16. Tyagi, Amit Kumar; Nair, Meghna Manoj; Niladhuri, Sreenath; Abraham, Ajith,
"Security, Privacy Research issues in Various Computing Platforms: A Survey
and the Road Ahead", Journal of Information Assurance & Security . 2020, Vol.
15 Issue 1, p1-16. 16p.
17. Akshara Pramod, Harsh Sankar Naicker, Amit Kumar Tyagi, “Machine Learning
and Deep Learning: Open Issues and Future Research Directions for Next Ten
Years”, Book: Computational Analysis and Understanding of Deep Learning for
Medical Care: Principles, Methods, and Applications, 2020, Wiley Scrivener,
2020.
18. Amit Kumar Tyagi, G. Rekha, “Challenges of Applying Deep Learning in Real-
World Applications”, Book: Challenges and Applications for Implementing
Machine Learning in Computer Vision, IGI Global 2020, p. 92-118. DOI:
10.4018/978-1-7998-0182-5.ch004
19. B. Gudeti, S. Mishra, S. Malik, T. F. Fernandez, A. K. Tyagi and S. Kumari, "A
Novel Approach to Predict Chronic Kidney Disease using Machine Learning
Algorithms," 2020 4th International Conference on Electronics, Communication
and Aerospace Technology (ICECA), Coimbatore, 2020, pp. 1630-1635, doi:
10.1109/ICECA49313.2020.9297392.
20. Smys, S., and Wang Haoxiang. “Data Elimination on Repetition using a Block-
chain based Cyber Threat Intelligence.” IRO Journal on Sustainable Wireless Sys-
tems 2, no. 4(2021): 149-154.
21. Amit Kumar Tyagi, Aswathy S U "AARIN: Affordable, Accurate, Reliable and
INnovative Mechanism to Protect a Medical Cyber-Physical System using
Blockchain Technology" IJIN, KaiS, Decemeber 2021.
22. Shamila, M, Vinuthna, K. and Tyagi, Amit. (2019). A Review on Several Critical
Issues and Challenges in IoT based e-Healthcare System. 1036-1043.
10.1109/ICCS45141.2019.9065831.
23. Pandian, Dr A. Pasumpon. “Development of Secure Cloud Based Storage Using
the Elgamal Hyper Elliptic Curve Cryptography with Fuzzy Logic Based Integer
Selection.” Journal of Soft Computing Paradigm 2, no. 1 (2020): 24-35.
Chapter
This chapter examines the current landscape and future prospects of emerging technologies in drug discovery, focusing on their potential to provide efficient and innovative services to patients. The integration of technologies such as artificial intelligence (AI), machine learning, high-throughput screening, and advanced analytics is reshaping the drug discovery process, promising accelerated development timelines and improved therapeutic outcomes. AI and machine learning algorithms play a pivotal role in data analysis, aiding in the identification of potential drug candidates, target validation, and predictive modeling. High-throughput screening technologies enable the rapid testing of large compound libraries, expediting the identification of lead compounds and optimizing drug development pipelines. Additionally, advanced analytics facilitate the interpretation of complex biological data, enhancing our understanding of disease mechanisms and drug interactions. The convergence of these technologies holds immense promise for personalized medicine.
Chapter
Full-text available
Digital health communication (DHC) has become an increasingly popular domain of the healthcare industry, enabling effective and efficient communication between healthcare providers, patients, and other stakeholders. However, the growing importance of digital platforms and the exchange of sensitive health information also present cybersecurity challenges. This chapter explains the utilization of AI-based cybersecurity in DHC to enhance security and protect patient privacy. Artificial intelligence (AI) plays an important role in cybersecurity by enabling advanced threat detection, rapid response, and intelligent risk management. AI algorithms can analyze large amounts of data, identify patterns, and detect potential security breaches or malicious activities in real-time. By adding AI-based cybersecurity solutions, DHC platforms can enhance their security measures and protect sensitive patient data. In DHC, AI-based cybersecurity can be utilized to ensure secure data transmission and storage.
Article
Full-text available
The enhancement (in the past decade) in electronics devices/technology, a rise in most of the accidents concerning security and surveillance intruding the private lives of the users question the existing systems being used to combat this challenge wherein, the third-parties gather and handle large amounts of individual details. As part of the constant evolution of the cyber physical system architecture, one of the goals of our system is to reduce the latency time for enrollment of new information. Generally, the efficiency and benefit of a Cyber Physical Systems (CPS) depends heavily on interconnection of individual devices or nodes. Exchange of data and information relevant to an overall task or functionality is the key to many applications such as smart grids, smart cities, and many others. Trustworthiness of data is needed to make such systems (especially in MCPS systems) successful. To be able to fulfill policies to guarantee the safety of all entities within a Medical Cyber Physical Systems (MCPS) and to provide security measures to enforce these cryptographic solutions have to be embedded. This paper describes a decentralized e-healthcare application framework for personal data management that ensures that users own and access their data. This work proposes a novel mechanism to secure Medical Cyber Physical Systems (MCPS), i.e., as an automated access-control manager (including building trust in a third party). This work also integrates some features in security building blocks in ultra-small devices to provide essential properties to secure embedded systems.
Chapter
Full-text available
With the development in technology, many other technologies like machine learning (ML), deep learning, blockchain technology, Internet of Things, and quantum computing have taken place in this current era. These technologies are helping human being to live their life comfortably and without any hurdle. Today, technology is helping human and protecting nature with minimum waste of available/limited resources. Among these inventions, ML and deep learning are two unique inventions which have attract many researchers or computer science researchers (or many research communities) to solve complex problems through ML. Today, ML use has been moved in many sectors to increase productivity of businesses; for example, for retail/marketing purpose, churn prediction of customers, for e-healthcare, and detecting disease in early stages. These are the few examples where ML is used in this current smart era. Together, this deep learning also has increased its importance over ML in many applications like bio-informatics, health informatics, identification of images or handwritten languages, and audio recognition. Many researchers get problematic scenario when they are not sure about particular use of machine and deep learning. This work fulfil such conditions/requirements and provide a complete details about ML and deep learning, i.e., with its evolution to forefront use, to use in many applications, to benefiting to the society, and to challenges and potential limitation in the respective learning techniques.
Article
Full-text available
Recently, the image reconstruction study on EIT plays a vital role in the medical application field for validation and calibration purpose. This research article analyzes the different types of reconstruction algorithms of EIT in medical imaging applications. Besides, it reviews many methods involved in constructing the electrical impedance tomography. The spatial distribution and resolution with different sensitivity has been discussed here. The electrode arrangement of various methods involved in the EIT system is discussed here. This research article comprises of adjacent drive method, cross method, and alternative opposite current direction method based on the voltage driven pattern. The assessment process of biomedical EIT has been discussed and investigated through the impedance imaging of the existent substances. The locality of the electrodes can be calculated and fixed for appropriate methods. More specifically, this research article discusses about the EIT image reconstruction methods and the significance of the alternative opposite current direction approach in the biomedical system. The change in conductivity test is further investigated based on the injection of current flow in the system. It has been established by the use of Electrical Impedance Tomography and Diffuse Optical Tomography Reconstruction Software (EDITORS) software, which is open-source software.
Chapter
Full-text available
Due to development in technology, millions of devices (internet of things: IoTs) are generating a large amount of data (which is called as big data). This data is required for analysis processes or analytics tools or techniques. In the past several decades, a lot of research has been using data mining, machine learning, and deep learning techniques. Here, machine learning is a subset of artificial intelligence and deep learning is a subset of machine leaning. Deep learning is more efficient than machine learning technique (in terms of providing result accurate) because in this, it uses perceptron and neuron or back propagation method (i.e., in these techniques, solve a problem by learning by itself [with being programmed by a human being]). In several applications like healthcare, retails, etc. (or any real-world problems), deep learning is used. But, using deep learning techniques in such applications creates several problems and raises several critical issues and challenges, which are need to be overcome to determine accurate results.
Chapter
Full-text available
With the recent development in technologies and integration of millions of internet of things devices, a lot of data is being generated every day (known as Big Data). This is required to improve the growth of several organizations or in applications like e-healthcare, etc. Also, we are entering into an era of smart world, where robotics is going to take place in most of the applications (to solve the world's problems). Implementing robotics in applications like medical, automobile, etc. is an aim/goal of computer vision. Computer vision (CV) is fulfilled by several components like artificial intelligence (AI), machine learning (ML), and deep learning (DL). Here, machine learning and deep learning techniques/algorithms are used to analyze Big Data. Today's various organizations like Google, Facebook, etc. are using ML techniques to search particular data or recommend any post. Hence, the requirement of a computer vision is fulfilled through these three terms: AI, ML, and DL.
Article
Cyber threat is a major issue that has been terrorizing the computing work. A typical cyber-physical system is crucial in ensuring a safe and secure architecture of a sustainable computing ecosystem. Cyber Threat Intelligence (CTI) is a new methodology that is used to address some of the existing cyber threats and ensure a more secure environment for communication. Data credibility and reliability plays a vital role in increasing the potential of a typical CTI and the data collected for this purpose is said to be highly reliable. In this paper, we have introduced a CTI system using blockchain to tackle the issues of sustainability, scalability, privacy and reliability. This novel approach is capable of measuring organizations contributions, reducing network load, creating a reliable dataset and collecting CTI data with multiple feeds. We have testing various parameters to determine the efficiency of the proposed methodology. Experimental results show that when compared to other methodologies, we can save upto 20% of storage space using the proposed methodology.
Article
The technological advancements in the field of the information and communication technology led to the development of more promising cloud paradigm that allows online provision of services such as platform, software and infrastructure. The infrastructure services provided by the cloud allows the user data to be stored and accessed ubiquitously unlike the arrays that are available on the premises. However the security of the data that are being stored in the internet are still under research. This remains as major inhibitor for the adoption of cloud service in spite of its reliability, elasticity, high computational capabilities and the pay as you go possibilities. So the paper puts forth the construction of a secure storage as service in the cloud computing by utilizing the cryptography system based on Elgamal. The proposed method encrypts the data to be stored using the Elgamal that incorporates the hyper elliptic curve cryptography for twofold encryption and further utilizes the fuzzy logic to perform the integer selection that serves as the significant attribute in defining the cloud storage security. The point addition and the double based ECC is used for generating the keys for the hyper elliptic curve cryptography. The twofold security model put forth is validated using the MATLAB on the terms of efficiency observed on securing the data stored in cloud as well as the cost and the execution time endured by the proposed twofold security. The results shows that the protection offered by the twofold security is much better compared to the prevailing.