Conference PaperPDF Available

Cloud semantic-based dynamic multimodal platform for building mhealth context-aware services

Authors:

Figures

Content may be subject to copyright.
Cloud Semantic-based Dynamic Multimodal Platform
for Building mHealth Context-aware Services
Adel ALTI
LRSD, University of Setif-1,
19000, Setif – Algeria
altiadel2002@yahoo.fr
Sébastien LABORIE
LIUPPA/IUT Bayonne,
64600, AngletFrance
Sebastien.Laborie@iutbayonne.univ-
pau.fr
Philippe ROOSE
LIUPPA, University of Pau,
64600, AngletFrance
Philippe.Roose@iutbayonne.univ-pau.fr
Abstract—Currently, everybody wish to access to applications
from a wide variety of devices (PC, Tablet, Smartphone, Set-top-
box, etc.) in situations including various interactions and
modalities (mouse, tactile screen, voice, gesture detection, etc.). At
home, users interact with many devices and get access to many
multimedia oriented documents (hosted on local drives, on cloud
storage, online streaming, etc.) in various situations with multiple
(and sometimes at the same time) devices. The diversity and
heterogeneity of users profiles and service sources can be a
barrier to discover the available services sources that can come
from anywhere from the home or the city. The objective of this
paper is to suggest a meta-level architecture for increasing the
high level of context concepts abstracting for heterogeneous
profiles and service sources via a top-level ontology. We
particularly focus on context-aware mHealth applications and
propose an ontologies-based architecture, OntoSmart (a top-
ONTOlogy SMART), which provides adapted services that help
users to broadcast of multimedia documents and their use with
interactive services in order to help in maintaining old people at
home and achieving their preferences. In order to validate our
proposal, we have used Semantic Web, Cloud and Middlewares
by specifying and matching OWL profiles and experiment their
usage on several platforms.
Keywords—Cloud; Profil; Modality; Health; Context-aware;
Location-Based Smart Service.
I. INTRODUCTION
Nowadays, everyone can access to a variety of services at
anytime, anywhere (from the home or the city) and anyhow
(using laptop, Smartphone, Set-top-box, etc.). These mobile
devices include variety of sensors and are becoming smarter.
Meanwhile, hosted services are becoming richer with
contextual information and more complex, especially in the
field of smart-* environments (Home, City, etc.). The diversity
of context information is generated by platforms heterogeneity,
user’s preferences and context variations. For instance; at
home, the variety of situations and contexts can rely on the set
of home sensors and mostly manipulated media contents of
home services from heterogeneous mobile terminals
(Smartphones, tablets, etc.) and/or remotely on the cloud.
In order to efficiently manage the explosion of users
profiles and service sources in smart-* environments (Home,
Health, etc.) and relations among them, in this article we define
a top-ontology called OntoSmart (a top-ONTOlogy SMART).
Our goal is describing heteregnous service sources and user
profiles of smart-* environments (Home, Home, City, etc.)
from semantic-independent level, and to present a new cloud
management tool that provides a great flexibility and enables
automatic semantic adaptation and customization of mobile
client services. All resources (smart and interactive services,
user context information and multimedia documents) are
managed and classified in a dynamic way. We also discuss the
usefulness and the importance of ontologies to specify
semantics of smart health clearly.
Our proposed cloud ontology captures a shared conceptual
schema common in the smart-* environments and maintains
semantic smart and interactive service information in
heterogeneous service sources for service model. It includes
reasoning properties in higher level of system, providing a set
of generic and common concepts of smart-* domains, by the
identification of general concepts and their relations. This will
cope with the context of ubiquitous environments, where top-
ontology is presented as “a convenient way for isolating
concerns of a system”. [4].
Our approach based on three layers architectures consisting
of a top-level dedicated to manage heterogeneous user profiles
and context constraints (media and modality preferences)
across multiple smart domains. The middle layer consists in
the smart domain specific ontologies (City, Home, Vehicule,
etc.) including digital health preoccupations. The basic layer
consists in Kalimucho [1] [4] as Middleware platform for local
or cloud (re-)deployment of services, the discovery services
and the adaptation services components. The separation of
smart domains reduces the scale of context knowledge, and the
separation of adaptation from contexts, makes these concerns
reusable. Thus, the user can access services from anywhere
from the home or the city. For instance, when the doctor leaves
his office to city, our platform will redeploy dynamically
services with a high-level set of rules using upper ontology and
a new set of context-aware services in the city now can be
accessed. Thanks to ontologies, to exploit heterogeneous
profiles and situations and create an interesting knowledge for
context-aware services reasoning.
In the next section, we will summarize related works. In the
section 3, we detail a top-level ontology OntoSmart and high
level inference rules in context-aware mHealth applications.
We give the general architecture of the platform in section 4
and evaluation results in section 5. Finally, section 6 concludes
our article.
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
978-1-4673-7701-0/15/$31.00 ©2015 IEEE 357
II. OBJECTIVES AND RELATED WORKS
The first related area of research are some works involving
adaptations of component-based applications refer to the
evolving needs of the users and the execution context by
exploiting a descriptive view of context information (e.g.
CC/PP, SGP, CSCP). CC/PP (Composite Capability/
Preference Profiles) [2] is a W3C recommendation for
specifying device capabilities and user preferences. This profile
language is based on RDF and was maintained by the W3C
Ubiquitous Web Applications descriptive since it lists sets of
values which correspond to the screen size, the browser
version, the memory capacity, etc. However, the CC/PP
structure lacks functionality, for instance it limits complex
structure description by forcing a strict hierarchy with two
levels. Furthermore, it does not consider the description of
relationships and constraints between some context
information. CSCP (Comprehensive Structured Context
Profiles) [3] uses RDF and is also based on CC/PP. In contrast
to CC/PP, CSCP has a multi-level structure and models
alternative values according to predefined situations. Even if
CSCP provides a description of the context, which is not
limited to two hierarchical levels, this proposal does not allow
the specification of complex user constraints (e.g., avoiding
playing videos while the mobile phone battery level is lower
than 15%) Moreover, authors stated that this proposal was
developed as a proprietary model for specific domains. SGP
(Semantic Generic Profile) [5] organizes profile information
into tree kind of facets; the device characteristics, the user
context information and the document composition. The SGP
links profile information with the specification of high-level
explicit constraints. These constraints enable to model different
types of actions under rich conditions. Complex explicit
constraints can be specified and context information about the
device and the user are provided by services, thus guiding the
adaptation process in order to provide an adapted document
that complies with rich user constraints. Hence, profiles may
migrate from different platforms without modifying several
values. However, SGP uses manual constraints (if condition
then action). Furthermore, SGP is specified in RDF/XML
which lucks the semantic relationships expressiveness.
Mohamed BAHAJ et al [6] [7] applied a semantic approach to
multimedia documents and their adaptation through an
ontology that allows users to control how to receive their
multimedia documents in a given condition. They extended
UPOS (User-Profile Ontology with Situation-Dependent
Preferences Support) for multimedia document adaptation
UPOMA. They defined a sub-profile, as a subset of the profile.
In their case, such a sub-profile contains a set of multimedia
documents preferences. This work didn’t talk exactly about
cloud model adaptation of the documents to have full control
of how documents should be received in different devices
(Laptop, Tablet, Smartphone…) in order to be played and nor
multi-parts profiles.
The second related areas of research are smart health
projects that have been proposed for monitoring of the old
people at home [8] [10] [11]. However, these works only have
kept the focus under the old person’s welfare. Roberto et al. [8]
proposed a framework for assisting old people in
heterogeneous and dynamic contexts related to the user’s
location. The framework describes the parameters for
customizing the service information; target the use of
ontologies and semantic models to share knowledge among
mobile devices. This work provides mechanisms to allow end-
users to play a central role in the selection service of their
contexts. This work does not consider the use of automatic
knowledge sharing techniques and explicit user multimodal
preferences in the semantic representation and composition
selection process. Primal Pappachan et al. [10] developed a
framework called Rafiki for guiding health community in order
to provide an adapted diagnostic process that complies with
users requirements (health professionals, patients, elderly,
dependent persons). The framework provides facilities for
collaboration among patients and health-care providers; target
the use of patient contexts (e.g., age, gender, location,
profession, etc.) to identify possible diseases. This work
exploits correlations about context parameters between
diseases, symptoms, and patient. Internet-based and P2P-based
approachs are two intercations mechanisms used in the
community health-care. The first approach represents
dissemination of knowledge between health-care providers and
Community Health Workers (CHW). However, in remote
areas, Internet access is not guaranteed and can be unreliable.
Therefore, Raki leverages P2P ad hoc networks using other
wireless communication mechanisms (e.g., Bluetooth and
WiFi) to update their knowledge base and share interesting
facts. The authors have presented an interesting approach for
describing service with semantic data. However, neither
various multimodal services, not heterogeneous profiles are
discussed. Lemlouma and al. [12] proposed a framework for
automatic dependency evaluation of the elderly people and the
management of the heterogeneity of elderly profiles and
service sources that can come from anywhere: from the home
or the city. The proposed home profile service is extended with
context service taken into account. This architecture is context-
aware but suffered from a semantic-based multimodal
evaluation for achieving an efficient and flexible framework.
To improve performance trade-offs of smart mobile healthcare
applications for intelligent environments, Joyce Chai et al. [11]
uses mobile cloud computing technologies to access remote
running health applications especially that require contextual
intelligence. The idea has exploited the cloud environment
hosts a Service Adaptation Module to optimize the runtime
deployment of the required components and acts as an entry
point for the adapter on the mobile client. The authors have
presented an interesting approach for annotating service
description with QoS data. However, neither various
multimodal interactions, not cloud profiles are exploited.
The methods and contributions of this paper are different
from above related works. Firstly, the paper exploits cloud
resources (smart and interactive services, hardware resources
and multimedia documents) for maximizing user QoS under
user constraint in mobile health applications. Secondly, the
paper proposes an intelligent service management system,
which is based on the on upper-level ontology model, as a way
to inference users’ preferences according to different sensitive
situations. The success of this approach provides its members
remote access (e.g. on the cloud) to external interactive
services (e.g. modalities changes) and sensors according to
user context (e.g. preferences, activity and resources changes).
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
358
Due to heterogeneous context profiles as well as its various
service sources (home, city, bus, etc.) can be specified more
easily using a semantic representation to better support
context-awareness. The main objectives are: (1) - develop a
top-level ontology model in order to describe heterogeneous
user profiles, (2) - identify critical situations, inferring user
constraints and determining the necessary adaptation so that
the user has the full exploitation of digital health documents,
(3) - maximize sharing and reuse of resources in different
contexts, (4) - group some specific profiles into generic and
optimized profiles by activity, preferences, location and time.
Our framework will define the semantic profile matcher
which compares the user profile constraints (implicit or
explicit) with other profiles constraints. It produces an
adaptation guide that contains some adaptation directives for
solving the detected matches. Our approach motivates mobile
users to become actively involved in groups that have yet
shared experiences. We prefer the hybrid approach that gives
more exibly to select freely the adaptation emplacement.
Consequently, we follow to execute services locally at the
home (where data can be easily updated) and/or remotely on
the cloud, depending on the capabilities of the client and the
context application.
III. ONTOSMART: A TOP-ONTOLOGY SMART
A top-level ontology OntoSmart conceptualizes knowledge
that addresses several smart-* domains with the discipline of
context-aware service modeling. The motivation for creating
the ontology is the ability to reason, to re-use existing profile
models and to share previous user experiences and preferences.
The OntoSmart define common concepts as well as
relationships among those concepts to represent context-aware
services that support all activities related to a user and his
current context. It can be used to publish metadata about their
services within their context constraints in the UDDI registry.
An overview of a top-level ontology structure, where the most
important concepts are shown, is given in Fig.1. The most
important concepts in the ontology are the Context class and its
specialization, user profile, sub-context profile and user
preferences. We divided context sub-ontology into seven main
sub-contexts:
The RessourceContext describes current usage of
processing power, memory, etc, is a prerequisite to
guarantee a minimum quality of service.
The HardContext are mobile devices like tablets or
Smartphone, are constrained in their resources
(memory size, CPU speed, battery energy, etc) and act
as execution environment for smart and interactive
services.
The UserContext contains personnel, location and
multimedia information. User profile describes user’s
personnel information such as name, role, ID, phone,
address, email, multimedia_content. The user
preference includes some preferred languages and
preferred modalities (voice, gesture, pen click, mouse
click, etc.). This is very interesting for a user to explicit
its constraints between some contextual information.
For example: For instance, a user can specify the
following constraint in his profile: “If it is lunch time,
remind me to take some medicines using audio
contents”. We can find also a description of the user’s
health situation, as the user can be healthy or handicap.
The PlaceContext describes related informations about
user’s location {longitude, altitude and attitude} and
the available resources sharing.
The ActivityContext: according to a schedule, a user
can engage in scheduled activity.
The DocumentContext describes the nature of the
documents (text, video, audio).The document context
specifies a set of properties related to a specific media
type: Text, Image, Video, Sound
The BiomedicalContext gathers all contextual
information related to smart-health. The
BiomedicalContext is divided into two sub-contexts:
BioContext and MedicalContext.
Fig. 1. An overview of the structure of the common Smart Ontology.
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
359
User wishes new media/modality preferences, have new
requirements. Media and modality preferences and context
variations require documents adaptation according to execution
constraints, e.g., audio contents and voice modality may not be
played while a user is participating at a meeting.
User interactions are now potentially spread over multiple
devices with different interacting modalities and should be
interpreted according to the user’s situation and context
variations. Modality preferences in mobile and ubiquitous
environments, evolved dynamically as it is one of the
important tasks that can be provide a high felxibity to mobile
clients. An explicit modality constraint is a set of conditions
and a set of actions:
9 Include modality: according to the user situation, this
action includes a modality for him. For example, if the
user has a visual handicap, voice modality is included.
9 Exclude modality: according to the user situation, this
action excludes a modality for him. For example, if the
user has a visual handicap, touch and gesture
modalities are excluded.
9 Equivalent modality: equivalent modality is required
under the same circumstances as the previous used
modality.
9 Complement modality: an extra modality is added as a
need to fulfill a specific task.
A. Cloud-based Service Ontology
Nowadays environment are getting smarter in order to reply
for user requests anytime and anywhere according to his
location, interaction between users seeks to get quatliy services
from providers. Service could be either smart or interactive.
Any service can be used in a local way or can be used on the
cloud (see Fig.2). The smart service handles the data storage
that users need to run their applications. An interactive service
allows unimodal and multimodal interactions.
Fig. 2. Cloud-based Service Ontology.
B. Health-Domain Ontology
User profiles in the mHealth domain are poor semantically,
not generic and not enough dynamic to take into account the
evolving situation of a patient. To model mHealth domain,
specialized classes and their corresponding attributes and
relations have been specialized in the subclasses, which are:
mHealth users’ profiles: The core of OntoSmart is the
conceptualization of the User Context Profile concept.
It should be noted that the goal is not only to provide a
useful categorization of the user profiles, but also to
provide each category with its own specific attributes.
There are many subclasses (Doctor, Nurse, Pharmacist
and Patient). The Patient has symptoms and a
document and attaches it with sensors.
mHealth services: mHealth domain includes a variety
of services that collect data about patient activity and
analyze these data to extract clinical knowledge.
mHealth documents: mHealth documents consist on
three subclasses (Report, Monitoring and Treatment).
Our goal is to provide the most appropriate treats in
some critical situations (e.g. hypoglycemic diabetic
coma).
mHealth hardware and sensors: mHealth situations are
accessible on a wide variety of embedded sensors. The
heterogeneity of such sensors and the diversity of
user’s needs require management, quality of service
and adapt to critical situations.
mHealth symptoms: A symptom is used during
diagnostics to detect possible diseases. Symptoms may
be multiple for a given disease. Our study includes all
symptoms (fever, weakness, nausea, hypoglycemic
diabetic coma).
Signs of disease : Signes are indicators of the patient's
health, glucose level, weight temperature, etc.
mHealth institutions: defined as hospitals, clinics,
health professional polyclinic and academic health
center.
mHealth preferences: In our health-domain ontology,
we specify explicit semantic constraints with
qualitative and quantitative information. For instance,
a use can specify the following constraint in his
profile: “If it is lunch time, remind me to take some
medicines using audio contents”.
C. Context-aware services and ontology reasoning
Developer rules: The developer defines a set of rules in
order to assure the continuity of services and to provide
context-aware service to execution constraints and user’
preferences.
Rule 1: Looking for modality preferences in the Cloud, the
user modality type should be equal to other users’ modality
types, modality operator is “Include” and target interactive
services are transport services. This rule can be described in
SWRL [15] as follows:
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
360
Fig. 3. Health-domain ontology extended from Top-level OntoSmart
Profil(?p) ^ ComposedOf(?p, ?sp) ^
DependsOn(?sp, ?act) ^ HasPreferences(?sp,? pref) ^
ModalityPreferences (?pref) ^
HasModalitypreferences(?pref, ?mdpref)
HasModalityOperatortype (?pref, ?mop) ^
swrlb:equal(?mop, "Include") ^
InteractiveService(?s) ^ ServiceCategory(?s, "Health") ^
IsDeployed(?s, ?host) ^ swrlb:equal(?host, "Cloud") ^
DependsOn(?sp3, ?s) ^ ServiceModality(?s, ?mp) ^
ModalityType (?mp, ?t) ^ swrlb:equal(?mdpref, ?t)
->sqwrl:select (?s)
Rule 2: The service is migrated on the cloud when the
battery level is low, so data could be stored separately and that
could help minimizing use of the energy. This rule can be
described as follows:
Profile(?p) ^ ComposedOf (?p, ?sp) ^ DependsOn (?sp, ?d) ^
Device(?d) ^ Battery_Level(?d, ?betterylevel) ^
swrlb:equal(?betterylevel, "Low") ^ HealthService(?s) ^
DeployedOn(?s, ?hot) ^ DefinedAs(?d, ?hot) ^ Cloud(?cloud)
->Migrated(?s, ?cloud)
Health-domain rules: The expert defines a set of rules in
order delivers relevant information to the clinician, to
include/exclude different interacting modalities and to detect
maladies.
Rule 3: Exclude undesirable preferences for visual
handicap users. This rule can be described in SWRL as:
Profil(?p) ^ ComposedOf(?p, ?sp) ^
DependsOn(?sp, ?user) ^User(?user) ^
HealthStatus(?user, "VisualHandicap") ^
HasPreferences (?sp, ?mp) ^ ModalityPreferences (?mp) ^
HasPreferences (?sp, ?medp) ^ MediaPreferences (?medp) ^
DependsOn(?sp, ?service) ^ InteractiveService (?service) ^
ServiceModality(?service, ?mt) ^
DependsOn (?sp, ?doc) ^Document(?doc) ^
-> HasMediapreferences(?medp, "Audio") ^
documentType (?doc, "Audio") ^ ModalityType(?mt, "Voice") ^
HasModalitypreferences(?mp, "Voice")
Rule 4 : The detection of possible diseases based on the
patient context (age, vital signs and symptom). This rule can be
described as follows:
Profile(?p) ^ ComposedOf(?p, ?sp) ^ DependsOn(?sp, ?d) ^
Device(?d) ^ DefinedAs(?d, ?hot) ^ HealthService(?s) ^
DeployedOn(?s, ?hote) ^ HasCategory(?s, "Digestive") ^
Inputs (?s, ?i1) ^ Fever (?i1) ^ isExistence(?i1, "YES") ^
Inputs (?s, ?i2) ^ Diar (?i2) ^ isExistence (?i2, "YES") ^
Inputs(?s, ?i3) ^ Abdominal_pain(?i3)^isExistence (?i3, "YES") ^
Inputs(?s, ?i4) ^ Nausee(?i4) ^ isExistence (?i4, "YES") ^
Inputs(?s, ?i6) ^ Vomissement(?i6) ^ isExistence (?i6, "YES") ^
Inputs(?s, ?i5) ^ Bio_Sensor_Temperature(?i5) ^
Qualitative_value(?i5, “High”) ^ Disease (?m)
-> Flu(?m)
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
361
IV. CLOUD SEMANTIC-BASED DYNAMIC ARCHITETURE
The proposed platform is built according to a layered
architecture as shown in Fig. 4. The architecture consists in
three layers: a top-level layer dedicated to the semantic query
layer. The middle layer consists in the generic cloud-based
semantic multimodal adaptation core layer. The basic layer is
the Kalimucho [1] layer offers service-level functions: (re)
deployment and reconfiguration strategy, according to its
system (Android, Laptop with QoS requirements, dynamic
supervision of adaptation components and communication
protocols between mobiles) nodes.
A. General Architecture
The cornerstone of our framework for dynamic interactive
services selection is the generic cloud-based semantic
multimodal adaptation core layer. It provides abstractions to
hide the complexity and the heterogeneity of the underlying
service-based P2P principle and implements a cloud-based
semantic social-based service strategy for each user according
to its context, which is inferred automatically from user
profiles and inferences rules based on OntoSmart. All
resources (smart services, user information and multimedia
documents) are managed and classified in dynamic platform,
which relies on the following components: (1) - Cloud
Management (CM) component; (2) - Profiles and Resource
Management (PRM) component and (3) - Dynamic
Multimodal Adaptation Management (DMAM) component;
The CM component allows setting and modeling a cloud-
based profile by specifying the characteristics of its members.
It allows modeling and managing adaptation process. It allows:
(1) - to enrich user profiles from different multimodal
techniques (gesture, voice, pen click and mouse click) with
various QoS (media quality, execution time, security, etc.), (2)-
to share health services in the online cloud registry and (2)- to
dynamic manage shared context aware adaptation services,
shared interactive health services with various modalities and
shared multimedia documents.
The PRM component allows the user of requests
multimedia documents about on-line relevant adaptation
services into shared cloud registry. Furthermore, this
component supports the association of shared adaptation
services by categories (health, transport, multimedia, etc.)
classified by QoS, and publish of shared services as well as
user experiments in the cloud registry as OWL files.
Semantic profile matcher module compares the user
profile constraints (implicit and explicit) with other
profiles properties (context and modality reasoning). It
produces an adaptation guide that contains some
adaptation directives for solving the detected
constraints. For instance :
ª User A wants to get health services only with text
and sound
ª User B does have required health text services and
other extra health documents.
ª User C does have required health sound services
and other extra health documents.
Fig. 4. Layers and component of Cloud Dynamic Semantic-based Platform.
It starts with comparing the user A profile constraints with
user “B” profile, in our case user B have only one required
request which is (text), so user A will look for required the
sound services at the user C profile, in our case he get required
one which is sound, so research is done and accomplished .
In case user A get all required services from user B. So
research is accomplished without looking for
documents in user C profile.
In case User A didn’t get any required services from
users near him (B and C in our case) so he would look
for them in the cloud.
If an action corresponds to a modality inclusion or
exclusion, then the adaptation directive contains at
least one complement operator.
If an action corresponds to modality duplication then
the adaptation directive contains at least one
redundancy operator.
If an action corresponds to a modality update then the
adaptation directive contains at least one substitution
operator.
If an action corresponds to migrate component from a
device to another, then the adaptation directive
contains at least one substitution operator.
Add other adaptation directives related to implicit profile
constraints, such as the supported modalities, the supported
media types, the battery level, etc. For example, from a profile
and a current situation, if we confirm that we must exclude
voice modality (because the current noisy environment), the
adaptation process should start by executing a substitution and
propose voice with button OK.
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
362
Semantic interactive services discovery module.
This module discovers some interactive services that
will match the semantic constraints of the user: action
type (complement operator, the substitution operator,
equivalent operator), inputs/outputs, modality types
media type (image, video, sound and text), a specific
user context information (user health context, user
location, user language, user age, screen resolution,
battery level, memory size, etc.). After the discovery
module finds the matched interactive services, it sends
them to the automatic adaptation decision component.
Decision module. This module is responsible of
selecting the best interaction service after calculating
the score of each discovered services. The comparison
of score interactive services allows us to select the best
service. In this case, values of quality formula [13] are
used for classifying the relevant interactive services
that have potential benefit.
Reconfiguration and redeployment module: This
module is responsible for the total or partial (re)
deployment of selected services.
B. Search Algorithm
Our strategy of semantic profile matching method tackles
different aspects (location, time, service category, media and
modality preferences). It would provide users with an
improved services research, as it gives the user the opportunity
to look for missing preferences in other near or cloud profiles.
It takes user’s current context (battery level, CPU load,
bandwidth, user location) and all his around resources as input
(local, remote), groups services on clusters based on (location,
time, activity) and filtering available services in cloud.
V. POSSIBLES SCENARIOS AND VALIDATION
A. Real-life Scenarios of Health and Social Care
Within the scope of our work let’s suggest this scenario:
Ahmed is an old man and has a visual handicap. Today,
when his health schedules activities start time begin, his Smart
TV remembers him to take medicines using a voice modality.
Then, when his serious health detected by camera sensor, an
urgent event is triggered, we identify categories of people who
are able to intervene (neighbor, family member, doctor) with an
order of preference. If the contacted was the Doctor, he takes
ambulance to go to Ahmed’s Home and uses his Smartphone to
follow health status of Ahmed using gesture modality. When it
comes a low battery level is detected, the Doctor wants to
continue flow the health status of Ahmed. Data flow could be
stored in the Cloud (See Fig.6).
We have utilized our approach for semantic user profiles
and interactive health services exchanges that is widespread in
the local or cloud environment. An identical semantic model is
created for every user profile and it is connected to other
profiles using the same context (location, time, service
category, user preferences). . These scenarios are done sending
continuous sensors data (GPS location, speed, video stream,
Wifi and 3G, etc.).
Input : User Profile Constraints
Output : Interactive services matching results
Found_1=false; /* required media preference is found */
Found_2=false; /* required modality preference is found */
IF (media preference is matched with local user profile)
Found_1= true;
IF (modality preference is matched with local user profile)
Found_2=true;
IF (Found_1==false || Found_2==false) {
/* Search interactive services at users nearby */
IF (Found_1==false){
IF (required media preference is found)
Found_1=true;}
IF (Found_2==false){
IF (required modality preference is found)
Found_2=true;}
}
IF (Found_1==false || Found_2==false) {
/* Cloud Search interactive and multimedia services */
IF (Found_1==false){
IF (required media preference is found)
Found_1=true;}
IF (Found_2==false){
IF (required modality preference is found)
Found_2=true;}
}
IF (Found_1 && Found_2) both preferences found
IF (Found_1 || Found_2) Only media preferences was found
IF (!Found_1 && !Found_2) None of preferences was found
Fig. 5. Alignment Algorithm.
First scenario (Health status): This scenario occurs when
a user has a visual handicap, service modalities and media
contents should change automatically according to his health
status, so we suggest that only audio media is authorized, video
text and images are excluded. Service modalities are excluded
except voice modality.A user can command his mobile device
using only his voice. In case a user has a good health status and
press the button to infer an exception occurs telling him that he
is not authorized to do it.
Second scenario (Schedule interactive health services): A
user has a daily schedule, so if any of his schedule activities
start time begins, the required one should be inferred and get
started in user’s device. A user is able to look for services at
users near him; they should be in the same area where is he
located and have the service he is up to get. The result service
should be the best one for user choice according to his media
and modality preferences. The cloud can handle a big data
storage so users are able to put their services in the cloud and
other users can look for their services.
Third scenario (Low battery level): This scenario occurs
when a user has a low battery level and wants to get his service
in the cloud in order to reduce the use of his device data,
exceptions were made so devices with medium or high battery
levels could not put their services in the cloud according to the
given scenario.
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
363
Fig. 6. Possible scenarios
B. Performance Comparison
Originally, the context searching service used SWRL rules
(available on standard OWL Java API) to access mHealth
services from the cloud ontology. We have performed these
experiments on a Laptop running Windows 7 (x64) with 6GB
of RAM and i7-2630QM quadruple coreprocessor (2 GHz).
After some performance tests, we realized that SWRL query
execution gets alarmingly slow when the ontology grows in
user profiles instances as can be seen in Table 1. Once this
problem was detected, we decided to use our JAVA API
algorithm, which made the query time considerably faster by
applying each rule separately. Moreover, the execution time
remains constant at any specific domain ontology size.
TABLE I. PERFORMANCE EVALUATIONS
Case 1: Looking all over and finding preferences in local devices
SWRL rules Elapsed time : 9975 ms
Java rules Elapsed time : 9487 ms
Case 2: Looking all over and finding preferences in nearby users
SWRL rules Elapsed time : 10289 ms
Java rules Elapsed time : 427 ms
Case 3 : Looking all over and finding preferences in the Cloud
SWRL rules Elapsed time : 12613 ms
Java rules Elapsed time : 12613 ms
VI. CONCLUSION
This paper presents a cloud semantic-
based dynamic multimodal adaptation platform. Our platform
proposes to identify situations, infer constraints and determine
the necessary adaptations. Novelty and originality of our
approach can be viewed in the combination of the following
points: (1) - minimize the execution time of SWRL queries
which facilitate alignment and inference in the ontology, (2) -
experimentations of ontology with more than hundred user
profiles. In a future work we would like to improve our Java
algorithms efficiency. Moreover, we suggest to test alignment
using more than one ontology.
REFERENCES
[1] C. Cassagnes, P. Roose, M. Dalmau, “Kalimucho: software architecture
for limited mobile devices”, ACM SIGBED Review, v.6 n.3, October
2009 DOI= http://doi.acm.org/10.1145/1851340.1851354.
[2] G. Klyne, F. Reynolds, C.Woodrow, O. Hidetaka, J. Hjelm, M. H.
Butler, L. Tran, "Composite Capability/Preference Profiles (CC/PP):
Structure and Vocabularies 1.0," 2004.
[3] S. Buchholz, T. Hamann, G. HUbsch, “Comprehensive Structured
Context Profiles (CSCP): Design and experiences”, 2nd IEEE Annual
Conference on Pervasive Computing and Communications Workshop
(PERCOMW'04). Washington, USA: IEEE Computer Society, pp. 43-48
[4] D. Keling, P. Roose, M. Dalmau and J. Novado, “Kali2Much: An
autonomic middleware for adaptation-driven platform”. In Proceedings
of the International Workshop on Middleware for Context-Aware
Applications in the IoT (M4IOT 2014), pp. 25-30
[5] C. Dromzee, S. Laborie, P. Roose, “A Semantic Generic Profile for
Multimedia Documents Adaptation”, Intelligent Multimedia
Technologies for Networking Applications: Techniques and Tools. IGI
Global, 225-246, 2013.
[6] H. Khallouki, M. Bahaj, P. Roose, S. Laborie, “SMPMA: Semantic
multimodal Profile for Multimedia documents adaptation”, 5th IEEE
Inter. Workshop on Codes, Cryptography and Communication Systems
(IWCCCS'14), 27-28 Nov 2014, 142-147, El Jadida, Morocco.
[7] M. BAHAJ, H. Khallouki, “UPOMA: User Profile Ontology for
Multimedia documents Adaptation”, The Journal of Technology. Vol. 06
(October. 2014), 448-457, 2014.
[8] R.Yus, E. Mena, S. Ilarri, A. Illarramendi, “SHERLOCK: Semantic
management of Location-Based Services in wireless environments”,
Pervasive and Mobile Computing, 5 (Dec. 2014), 87-99.
[9] N.Z. Naqvi, D. Preuveneers, Y. Berbers, “A Quality-aware Federated
Framework for Smart Mobile Applications in the Cloud”, Proceedings
of the 5th Int. Conf. on Ambient Systems, Networks and Technologies
(ANT 2014), Procedia Computer Science, pp. 253–260, 2014.
[10] A. Alti, S. Laborie, P. Roose, “USSAP: Universal Smart Social
Adaptation Platform”, In Proceedings of the 6th International Conference
on Ambient Systems, Networks and Technologies (ANT 2015),
Procedia Computer Science, pp. 670–674, 2015.
[11] J. Chai, S. Pan, , M.X. Zhou, “MIND: A Semantics-based Multimodal
Interpretation Framework for Conversational Systems”, In Proceedings
of the International Workshop on Natural, Intelligent and Effective
Interaction in Multimodal Dialogue Systems. Springer-Verlag, 37-46.
[12] T. Lemlouma, S.Laborie, P. Roose, A., Rachedi, K., Abdelaziz,
“mHealth Contents and Services Delivery and Adaptation Challenges
for Smart Environments”,mHealth Multidisciplinary Verticals, CRC
Press/Taylor & Francis, pp.295-314, 2014.
[13] A. Alti, S. Laborie, P. Roose, “USSAP: Universal Smart Social
Adaptation Platform”, Proceedings of the 6th International Conference
on Ambient Systems, Networks and Technologies (ANT 2015),
Procedia Computer Science, pp. 670–674, 2015.
[14] Q.P. Hai, S. Laborie, P. Roose, “On-the-fly Multimedia Document
Adaptation Architecture”, In Proceedings of International Workshop on
Service Discovery and Composition in Ubiquitous and Pervasive
Environments (SUPE), 10 :1188-1193, 2012.
[15] Protégé. 2007. Why Protégé. http://protege.stanford.edu/.
e-Health Pervasive Wireless Applications and Services (eHPWAS'15)
364
... Many studies works Abdelmoneem et al. (2019), Alti et al. (2015Alti et al. ( , 2016, Angsuchotmetee et al. (2020), Bayer et al. (2019), Bhunia et al. (2014), Cassagnes et al. (2009), Chai et al. (2002), Dey (2001), Devitt (2016), Esposito et al. (2018), Garg et al. (2020), Hameur Laine et al. (2017, He et al. (2015), Keling et al. (2014), Name et al. (2017) and Naqvi et al. (2014;Rhayem et al. 2020) has been proposed for reconfigurable system for smart health. The existing works has proposed the conceptual approach capable of reconfiguring distributed health applications based on services hosted on user devices, context monitoring, and diseases detecting using semantic approaches. ...
... In this paper, we extend the previous framework presented in Alti et al. (2015) and Alti et al. (2016) with a new layer called AGENT-based autonomic semantic service adaptation controller and reconfiguration (Agent-ASSACR) in order to facilitate service search in the mobile cloud and optimize the deployment of health services. We also provide independent smart home management allowing: (1) analyze, evaluate and improve the redundancy rate for equivalent services providing access to remote service providers, (2) parallel discovery of equivalent services with different QoS and alternative paths by analyzing the behavior of all routers in the same local network, based on simple network management protocol (SNMP) communication protocol that is the means of controlling management information describing the state of gates (routers). ...
... The execution time is the average response time needed to accomplish the detection of potential health disease and triggering corresponding action services through fifty different patients with different biomedical sensors. The execution time is tested with Kali-smart framework (Alti et al. 2015) in similar test conditions, which are shown in Fig. 14. The evaluations are performed for the biomedical ...
Article
Full-text available
Nowadays, mobile applications have been widely used in various health and social domains. These applications must be flexible enough to support context changes with auto-adaptation. In spite of the diversity and heterogeneity of users’ profiles and smart objects, most existing smart-health domains works adopted rigid strategies. To address this problem, this paper comes up with a new autonomic and agent-based semantic context-aware platform for smart health monitoring and disease detection. The classification of diseases reveals new insight on using efficiently interactive services in order to help in maintaining elderly individuals at home and accomplishing their preferences. It discovers semantic health services according to users’ preferences with consistent mobility of users and limited resources (low battery, processing capabilities, memory, and others). In order to validate our proposal, we have used Semantic Web, Cloud and Kalimucho middleware by specifying some profiles and experiment their usage on several platforms.
... The availability of a large number of heterogeneous services with different QoS is provided by several medical providers; developers make of significant profit from them to provide customized health application that meets specific user's needs. Several platforms and approaches for self-reconfiguring, context monitoring and diseases detecting using semantic approaches have been proposed [1,[3][4][5][6][7][8][9][10]. However, there is a glaring lack of techniques that are allowing users to identify and access to services automatically and optimally. ...
... A dynamic reconfiguration (adding, updating or migration of services) of these systems will increase the maintenance costs, often related to the high (re)connectivity of wide smart objects that continuously evolve faster to achieve the sole purpose of satisfying the human needs. Other approaches [1,[6][7][8][9][10] have a focus on deploying quality services on the mobile device and Cloud and not examine in detail their impact on long-life Quality of Service (QoS) of medical systems. In the best of our knowledge, none of the proposed approaches has been considered due to the complexity of reacting rapidly to urgent situations under several moving locations nor deduced the QoS using parallel collaborative smart agents. ...
... In this paper, we extend the previous framework presented in [7,17] with a new layer called AGENT-based Autonomic Semantic Service Adaptation Controller and Reconfiguration (Agent-ASSACR) in order to meet the quality of user's constraints service by guaranteeing a minimum execution time for context aware of distributed mobile applications using agentbased cooperative strategy. Instead of launching one agent for optimal service discovery, layer further launches several agents for parallel optimal mobile and Cloud services discovery and selection. ...
Article
Nowadays, mobile applications have been widely used in various health and social domains. While there are large number of smart devices that connect and transmit their context data in free manner, these mobile applications have limitations regarding the increasing number of heterogeneous context data sent by devices, conducting the problem of identifying user-centric situations and providing services in real-time. Many smart mobile health systems were proposed to ensure context-aware personal health monitoring and diseases detection, while most of them are failed to ensure a right level of dynamicity and enough flexibility for assisting users everywhere, anytime, through widespread sensors and mobile devices. In fact, it becomes necessary to rethink a new way to minimize the response time combining competitive agents with semantic-based situation reasoning strategy. In this paper, we introduce a novel agent-based platform with three-layered ontology for the semantic description and parallel management of services selection approach dedicated to context-aware health mobile applications. In addition, we propose an innovative parallel services discovery and optimal selection process, which involved a set of filtered and classified semantic health multipath according to user’s context and preferences with consistent mobility of users and limited resources (low battery, processing capabilities, memory and others). Experimental results show the effectiveness of the proposed approach as it includes the semantic service information on cooperative agents. In addition, the proposed approach ensures fast response time and prolongs the continuity of the mobile application.
... Considering existing production lines, the approach enables to make equipment smart to integrate it into the smart service world. A similar approach resulted in a platform as basis for smart mobile health services (Alti et al. 2015). ...
... Real-time data such as sensor data, were emphasized as being interesting for satisfying individual customer needs (Lê Tuán et al. 2012). Specific applications were discussed that are suitable for a specific industry, such as the health sector (e.g., Alti et al. 2015). Prototypes were developed for a specific application and predefined types of customers, such as private end users (e.g., Mukudu et al. 2016 Fig. 4 Summary of research intensity structured analysis of the role of data in the design phase is missing. ...
... To sum up, this article presented and discussed concrete ideas for further research that will help to draw a clearer and more comprehensive picture of smart services in the academic literature as well as for practical applications. Nuaimi et al. (2015) C i t y B 2 C Allmendinger and Lombreglia (2005) Manufacturing B2B Alti et al. (2015) H e a l t h B 2 C Anthopoulos and Fitsilis (2014) C i t y B 2 C Anthopoulos et al. General Both Kamp et al. (2016) Manufacturing B2B Kennedy and Keskin (2016) Manufacturing B2C Keskin and Kennedy (2015) Manufacturing Both Kim et al. (2015) Health, Home B2C Knote and Blohm (2016) General B2C Kölmel and Bulander (2015) Manufacturing Both Korzun et al. (2015) H e a l t h B 2 C Kryvinska et al. (2008) Telecommunications B2C Kuebel and Zarnekow (2015) H o m e B o t h Kwak et al. (2014) General B2C Kynsilehto and Olsson (2011) General B2C Kynsilehto and Olsson (2012) General, home B2C Lê Tuán et al. (2012) C i t y B 2 C Lebedev et al. (2016) H e a l t h B 2 C Lee et al. (2010) General B2C Lee et al. (2012) General B2C Lee et al. (2015) H o m e B 2 C Lee et al. (2016) General B2C Lesjak et al. (2014) Manufacturing B2B Lesjak et al. (2015) Manufacturing B2B Li et al. (2015) H e a l t h B 2 B Ma et al. (2010) Health, home, robotic Kamp et al. (2016) Keskin and Kennedy (2015) Knote and Blohm (2016) Keskin and Kennedy (2015) Wünderlich et al. (2015) Wünderlich et al. ...
Article
Full-text available
Smart services serve customers and their individual, continuously changing needs; information and communications technology enables such services. The interactions between customers and service providers form the basis for co-created value. A growing interest in smart services has been reported in the literature in recent years. However, a categorization of the literature and relevant research fields is still missing. This article presents a structured literature search in which 109 relevant publications were identified. The publications are clustered in 13 topics and across five phases of the lifecycle of a smart service. The status quo is analyzed, and a heat map is created that graphically shows the research intensity in various dimensions. The results show that there is diverse knowledge related to the various topics associated with smart services. One finding suggests that economic aspects such as new business models or pricing strategies are rarely considered in the literature. Additionally, the customer plays a minor role in IS publications. Machine learning and knowledge management are identified as promising fields that should be the focus of further research and practical applications. Concrete ideas for future research are presented and discussed and will contribute to academic knowledge. Addressing the identified research gaps can help practitioners successfully provide smart services.
... A KMS must support both ways to integrate new knowledge; general services and smart services. Therefore, it must be considered that knowledge is not static but dynamic (R3) (Alti et al., 2015), just like smart e-Service Journal Volume 13 Issue 2 services (Batubara, 2015). Knowledge management only works reliably when this is ensured. ...
... For example, classifications of knowledge can increase efficiency (Chatterjee & Armentano, 2015). Another aspect that must be considered when looking at efficiency is that the KMS must be able to identify redundant information (R10) (Alti et al., 2015). This provides several advantages. ...
Article
Full-text available
Smart services have become increasingly important in the last few years. The provision of smart services combined with product portfolios is advantageous because individual contexts of customers who use the products are addressed. While various and frequently changing data are collected and analyzed, specific knowledge is necessary to turn data into valuable information. A knowledge management system (KMS) adapted to requirements of provided smart services is necessary to aggregate, maintain, and provide knowledge. With a comprehensive literature review over 157 papers, we extracted requirements for KMS for smart services (Smart Service KMS). Smart services are highly complex, so requirements are existing especially for Smart Service KMS in addition to those of KMS. We present ten design principles derived from the requirements focusing on how tailored Smart Service KMS can be designed. We contribute to the theory by providing an overview about crucial requirements of Smart Service KMS. Practitioners can use our design principles to develop efficient Smart Service KMS.
... For instance, their seven main sub-contexts are the "UserContext", "PlaceContext, and "ActivityContext", which similarly define user details, their location, and activity, respectively. The other four sub-contexts are the "ResourceContext" describing system resources (e.g., processing power), the "HardContext" defining the devices that the smart services execute on, the "DocumentContext" specifying the properties of relevant media files, and the "BiomedicalContext" that gathers context data pertaining to the user's health [46]. ...
Article
Full-text available
Context awareness is a field in pervasive computing, which has begun to impact medical systems via an increasing number of healthcare applications that are starting to use context awareness. The present work seeks to determine which contexts are important for medical applications and which domains of context-aware applications exist in healthcare. A systematic scoping review of context-aware medical systems currently used by patients or healthcare providers (inclusion criteria) was conducted between April 2021 and June 2023. A search strategy was designed and applied to Pub Med, EBSCO, IEEE Explore, Wiley, Science Direct, Springer Link, and ACM, articles from the databases were then filtered based on their abstract, and relevant articles were screened using a questionnaire applied to their full texts prior to data extraction. Applications were grouped into context-aware healthcare application domains based on past reviews and screening results. A total of 25 articles passed all screening levels and underwent data extraction. The most common contexts used were user location (8 out of 25 studies), demographic information (6 out of 25 studies), movement status/activity level (7 out of 25 studies), time of day (5 out of 25 studies), phone usage patterns (5 out of 25 studies), lab/vitals (7 out of 25 studies), and patient history data (8 out of 23 studies). Through a systematic review process, the current study determined the key contexts within context-aware healthcare applications that have reached healthcare providers and patients. The present work has illuminated many of the early successful context-aware healthcare applications. Additionally, the primary contexts leveraged by these systems have been identified, allowing future systems to focus on prioritizing the integration of these key contexts.
... openu rc. org Alti et al. (2015) Kalimucho - Huang et al. (2018) WuKong https:// tinyu rl. com/ unacl wu Mshali et al. (2018) Survey - Dragone et al. (2015) Peis https:// tinyu rl. ...
Article
Full-text available
Human activity recognition (HAR) essentially uses (past) sensor data or complex context information for inferring the activities a user performs in his daily tasks. HAR has been extensively studied using different paradigms, such as different reasoning mechanisms, including probabilistic, rule-based, statistical, logical reasoning, or the machine learning (ML) paradigm, to construct inference models to recognize or predict user activities. ML for HAR allows that activities can be recognized and even anticipated through the analysis of collected data from different sensors, with greater accuracy than the other paradigms. On the other hand, context-aware middlewares (CAMs) can efficiently integrate a large number of different devices and sensors. Moreover, they provide a programmable and auto-configurable infrastructure for streamline the design and construction of software solutions in scenarios where lots of sensors and data are their bases, such as ambient intelligence, smart cities, and e-health domains. In this way, the full integration of ML capabilities as services in CAMs can advance the development of software solutions in these domains when ML is necessary, specially for HAR, which is the basis for many scenarios in these domains. In this work, we present a survey for identifying the state-of-the-art in using ML for HAR in CAMs through a systematic literature review (SLR). In our SLR, we worked to answer four research questions: (i) what are the different types of context reasoners available in CAMs; (ii) what are the ML algorithms and methods used for generating models for context reasoning; (iii) which CAMs support data processing in real time; and (iv) what are the HAR scenarios usually tackled by the research works. In our analysis, we observed that, although ML offers viable approaches to construct inference models for HAR using different ML approaches, including batch learning, adaptive learning and data stream learning, there are yet some gaps and research challenges to be tackled, specially on the use of data stream learning considering concept drift on data, mechanisms for adapting the inference models, and further considering all of this as services in CAMs, specially for HAR.
... Specifically, Santana-Perez et al. (2017) proposed a novel approach that describes scientific workflows' execution in the cloud using semantic vocabularies. In addition, the work by Alti et al. (2015) proposed a multi-level ontology-based architecture referred to as OntoSmart which can be used to enhance ''the high level of context concepts abstraction for heterogeneous service sources and profiles using a top-level ontology''. This architecture is beneficial to overcoming the barrier resulting from the heterogeneity and diversity of cloud-based service sources and profiles. ...
Article
Full-text available
Semantic and cloud computing technologies have become vital elements for developing and deploying solutions across diverse fields in computing. While they are independent of each other, they can be integrated in diverse ways for developing solutions and this has been significantly explored in recent times. With the migration of web-based data and applications to cloud platforms and the evolution of the web itself from a social, web 2.0 to a semantic, web 3.0 comes as the convergence of both technologies. While several concepts and implementations have been provided regarding interactions between the two technologies from existing research, without an explicit classification of the modes of interaction, it can be quite challenging to articulate the interaction modes; hence, building upon them can be a very daunting task. Hence, this research identifies and describes the modes of interaction between them. Furthermore, a “cloud-driven” interaction mode which focuses on fully maximising cloud computing characteristics and benefits for driving the semantic web is described, providing an approach for evolving the semantic web and delivering automated semantic annotation on a large scale to web applications.
Article
Full-text available
Currently, the field of smart (home, city, health, tourism, etc.) is naturally heterogeneous and multimedia oriented. In such a domain, there is an increasing usage of heterogeneous mobile devices, as well as captors transmitting data (IoT). They are highly connected and can be used for many different services, such as to monitor, to analyze and to display information to users. In this context, data management and adaptation in real time are becoming a challenging task. More precisely, at one time, it is necessary to handle in a dynamic, intelligent and transparent framework various data provided by multiple devices with several modalities. This paper presents a Kali-Smart platform, which is an autonomic semantic-based context-aware platform. It is based on semantic web technologies and a middleware providing autonomy and reasoning facilities. Moreover, Kali-Smart is generic and, as a consequence, offers to users a flexible infrastructure where they can easily control various interaction modalities of their own situations. An experimental study has been made to evaluate the performance and feasibility of the proposed platform.
Article
Full-text available
Multimedia documents can be accessed at anytime and anywhere on a wide variety of devices, such as laptops, tablets and smartphones. The heterogeneity of devices and user preferences has raised the problem of these documents adaptation according to user context condition and device characteristics. In this paper, we propose a semantic multimodal profile for multimedia documents adaptation. Thus, we propose a mechanism allowing dynamic generation of user conditions by including multimodal technologies. The proposed profile has the great advantage to offer a flexible way to users to satisfy their constraints and easily choose how to display their multimedia documents.
Article
Full-text available
This paper presents an approach to enhance users experience through the use of recommendations and social networks for on-the-fly (at runtime) adaptation of multimedia documents. The originality of the dedicated social and context-aware of quality service composition paths is that relies on contextual information collection with history-based service cloud selection and social media analysis technics, for providing the right service to the right user on the right time and on the right place and to deploy customizable services inside one application. We show that our approach successfully and efficiently captures online social experiences in order to improve assembly of potential adaptation services and the effectiveness of our approach
Article
Full-text available
Currently, multimedia contents can be accessed at anytime and anywhere with a wide variety of devices, such as laptops, smartphones and tablets. In this paper, we present a service-based multimedia adaptation architecture where adaptation services can be available locally and remotely on multiple platforms. When conflicts are detected between some multimedia document properties and profile constraints, we propose a procedure that selects relevant adaptation services in order to compute an adapted document. This procedure may be executed locally or remotely, moreover for unsovled conflicts this one indicates which services may have to be installed. (C) 2011 Published by Elsevier Ltd. Selection and/or peer-review under responsibility of [name organizer]
Article
Full-text available
Mobile Cloud Computing is an active research domain to investigate the ways of achieving a federation between mobile computing and cloud computing. Whereas, Ambient Intelligence aims to teach our mobile devices how to behave smart in our environment, narrowing the gap between people, their devices and the environment. These two domains behave as backbone for ubiquitous environment around us where everything is connected by sharing information and resources. Mobile devices use cloud resources to cope with ever-growing computing and storage demands for smart applications. Context provisioning in the cloud and mobile connectivity will give rise to new interactions with cloud infrastructure to effectively manage context information. However, as a lot of contextual information is being sensed and captured by mobile devices, it is not always clear whether it makes sense to upload all the raw sensed data to the cloud and have it processed there. Many resource and performance trade-offs exist for federated deployment of smart mobile applications. In this paper, we present a quality-aware federated framework to overcome the challenges for smart applications and to support the development and deployment of these applications, while opportunistically using cloud computing. The framework highlights the aspects of Quality-of-Context (QoC) provisioning and dynamic adaptability under uncertain context.
Article
Full-text available
Location-Based Services (LBSs) are attracting great interest nowadays, mainly due to the economic value they can provide. So, different applications are being developed for tracking, navigation, advertising, etc., but most of those applications are designed for specific scenarios and goals with implicit knowledge about the application context. However, currently it is a challenge to provide a common framework that allows us to manage knowledge obtained from data sent by heterogeneous moving objects (textual data, multimedia data, sensor data, etc.). Moreover, the challenge is even greater considering situations where the system must adapt itself to contexts where the knowledge changes dynamically and in which moving objects can use different underlying wireless technologies and positioning systems. In this paper we present the system SHERLOCK, that offers a common framework with new functionalities for LBSs. Our system processes user requests continuously to provide up-to-date answers in heterogeneous and dynamic contexts. Ontologies and semantic techniques are used to share knowledge among devices, which enables the system to guide the user selecting the service that best fits his/her needs in the given context. Moreover, the system uses mobile agent technology to carry out the processing tasks wherever necessary in the dynamic underlying networks at any time.
Article
Full-text available
This paper presents a software platform (called Kalimucho) for mobile and embedded hardware. The architecture of such application is based on two component models: Osagaia for software containers and Korrontea for connector containers. Kalimucho uses service factory permitting the creation of instances of software components. This factory allows making the most suitable configuration according to services decision managing reconfiguration. The reconfiguration service decisions depend on data sent by control interfaces and the users defined rules (via the platform and control interfaces).
Article
Currently, multimedia documents can be accessed at anytime and anywhere with a wide variety of mobile devices, such as laptops, smartphones, and tablets. Obviously, platform heterogeneity, users' preferences, and context variations require document adaptation according to execution constraints. For example, audio contents may not be played while a user is participating in a meeting. Current context modeling languages do not handle such real life user constraints. These languages generally list multiple information values that are interpreted by adaptation processes in order to deduce implicitly such high-level constraints. In this chapter, the authors overcome this limitation by proposing a novel context modeling approach based on services, where context information is linked according to explicit high-level constraints. In order to validate the proposal, the authors have used semantic Web technologies by specifying RDF profiles and experimenting on their usage on several platforms.
Conference Paper
To support context-based multimodal interpretation in conversational systems, we have developed a semantics-based representation to capture salient information from user inputs and the overall conversation. In particular, we present three unique characteristics: fine-grained semantic models, flexible composition of feature structures, and consistent representation at multiple levels. This representation allows our system to use rich contexts to resolve ambiguities, infer unspecified information, and improve multimodal alignment. As a result, our system is able to enhance understanding of multimodal inputs including those abbreviated, imprecise, or complex ones.
Conference Paper
In dynamic heterogeneous environments, such as pervasive computing, context-aware adaptation is a key concept to meet the varying requirements of different clients. To enable such functionality, context information must be gathered and eventually presented to the application performing the adaptation. Therefore, a common representation format for the context information is required. We introduce a novel representation language for context information: comprehensive structured context profiles (CSCP). CSCP is based on the resource description framework (RDF) and is designed to be comprehensive and thoroughly structured to describe the entire context of mobile sessions. Besides the design of CSCP, we describe our experiences with CSCP in a running system, a mobility portal for context-aware adaptive access to email and Web information services.