ArticlePDF Available

Eye Segmentation Method for Telehealth: Application to the Myasthenia Gravis Physical Examination

Authors:

Abstract

Due to the precautions put in place during the COVID-19 pandemic, utilization of telemedicine has increased quickly for patient care and clinical trials. Unfortunately, teleconsultation is closer to a video conference than a medical consultation, with the current solutions setting the patient and doctor into an evaluation that relies entirely on a two-dimensional view of each other. We are developing a patented telehealth platform that assists with diagnostic testing of ocular manifestations of myasthenia gravis. We present a hybrid algorithm combining deep learning with computer vision to give quantitative metrics of ptosis and ocular muscle fatigue leading to eyelid droop and diplopia. The method works both on a fixed image and frame by frame of the video in real-time, allowing capture of dynamic muscular weakness during the examination. We then use signal processing and filtering to derive robust metrics of ptosis and l ocular misalignment. In our construction, we have prioritized the robustness of the method versus accuracy obtained in controlled conditions in order to provide a method that can operate in standard telehealth conditions. The approach is general and can be applied to many disorders of ocular motility and ptosis.
Citation: Lesport, Q.; Joerger, G.;
Kaminski, H.J.; Girma, H.; McNett, S.;
Abu-Rub, M.; Garbey, M. Eye
Segmentation Method for Telehealth:
Application to the Myasthenia Gravis
Physical Examination. Sensors 2023,
23, 7744. https://doi.org/10.3390/
s23187744
Academic Editors: Md Zia Uddin
and Pei-Ju Chiang
Received: 19 June 2023
Revised: 28 August 2023
Accepted: 4 September 2023
Published: 7 September 2023
Copyright: © 2023 by the authors.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license (https://
creativecommons.org/licenses/by/
4.0/).
sensors
Article
Eye Segmentation Method for Telehealth: Application to the
Myasthenia Gravis Physical Examination
Quentin Lesport 1, Guillaume Joerger 2, Henry J. Kaminski 3, Helen Girma 3, Sienna McNett 3,
Mohammad Abu-Rub 3and Marc Garbey 1,2,4,*
1Department of Surgery, School of Medicine and Health Sciences, George Washington University,
Washington, DC 20037, USA; lesport@gwu.edu
2Care Constitution Corp., Newark, DE 19702, USA; joerger@orintelligence.com
3Department of Neurology & Rehabilitation Medicine, School of Medicine and Health Sciences,
George Washington University, Washington, DC 20037, USA; hkaminski@mfa.gwu.edu (H.J.K.);
hgirma@mfa.gwu.edu (H.G.); smcnett@gwmail.gwu.edu (S.M.); maburub@mfa.gwu.edu (M.A.-R.)
4LaSIE, UMR CNRS 7356, Universitéde la Rochelle, 17000 La Rochelle, France
*Correspondence: garbeymarc@gwu.edu
Abstract:
Due to the precautions put in place during the COVID-19 pandemic, utilization of
telemedicine has increased quickly for patient care and clinical trials. Unfortunately, teleconsul-
tation is closer to a video conference than a medical consultation, with the current solutions setting
the patient and doctor into an evaluation that relies entirely on a two-dimensional view of each
other. We are developing a patented telehealth platform that assists with diagnostic testing of ocular
manifestations of myasthenia gravis. We present a hybrid algorithm combining deep learning with
computer vision to give quantitative metrics of ptosis and ocular muscle fatigue leading to eyelid
droop and diplopia. The method works both on a fixed image and frame by frame of the video in
real-time, allowing capture of dynamic muscular weakness during the examination. We then use
signal processing and filtering to derive robust metrics of ptosis and l ocular misalignment. In our
construction, we have prioritized the robustness of the method versus accuracy obtained in controlled
conditions in order to provide a method that can operate in standard telehealth conditions. The
approach is general and can be applied to many disorders of ocular motility and ptosis.
Keywords:
telehealth; telemedicine; myasthenia gravis; ptosis; diplopia; deep learning; computer
vision; eyes tracking; neurological disease
1. Introduction
Telemedicine (TM) is an emerging tool for monitoring patients with neuromuscular
disorders and has significant potential to improve clinical care [
1
,
2
], with patients having
favorable impressions of telehealth during the COVID-19 pandemic [
3
,
4
]. There is great
promise in taking advantage of the video environment to provide remote alternatives
to physiological testing and disability assessment [
2
]. Telehealth is particularly well-
suited for the management of patients with myasthenia gravis (MG) due to its fluctuating
severity and the potential for early detection of significant exacerbations. MG is a chronic,
autoimmune neuromuscular disorder that manifests with generalized fatiguing weakness
with a propensity to involve the ocular muscles. For this purpose, the Myasthenia Gravis
Core Exam (MG-CE) [
5
] was designed to be conducted via telemedicine. The validated
patient reported outcome measures typically used in clinical trials may also be added to
the standard TM visit to enhance the rigor of the virtual examination [
6
]. In this study, we
address the first two components of the MG-CE [
5
], evaluation of ptosis (exercise 1) and
diplopia (exercise 2), thus focusing on the tracking eye and eyelid movement.
Today’s standard medical examination of MG relies entirely on the expertise of the
medical doctor, who grades each exercise of the protocol by watching the patient. The
Sensors 2023,23, 7744. https://doi.org/10.3390/s23187744 https://www.mdpi.com/journal/sensors
Sensors 2023,23, 7744 2 of 16
examiner rates the severity of ptosis by qualitatively judging the position of the eyelid in
relationship to the pupil and monitoring for ptosis becoming more severe over the course
of the one-minute assessment [
7
]. The determination of diplopia is entirely dependent on
the patient’s report. Further, the exam is dependent entirely on the patient’s interpretation
of what is meant by double vision (versus blurred vision) and further complicated by the
potential suppression of the false image by central adaptation, and in some situations,
monocular blindness, which eliminates the complaint of double vision. The objective of
our method is to provide unbiased, automatic quantitative metrics of ptosis and ocular
misalignment during the telemedicine visit.
Our long-term goal is to complement the neurological exam with computer algorithms
that quantitatively and reliably report information directly to the examiner, along with
an error estimate on the metric output. The algorithm should be fast enough to provide
feedback in real-time and automatically enter the medical record. A similar approach was
used by Liu and colleagues [
8
], monitoring patients during ocular exercises to bring a
computer-aided diagnosis, but with highly controlled data and environment. We aim to
use a more versatile approach by extracting data from more generic telehealth footage and
requiring as little additional effort from the patient and clinician as possible.
By use of videos, the algorithm should capture the dynamic of ptosis and ocular
misalignment over time that is inherent to the neuromuscular fatigue of MG. This feature
may not be readily detected by the human examiner, who simply watches the patient
perform tasks: the doctor estimates the eye motion at the end of the ptosis exercise versus
the initial eye position with a simple scale from 0 to 3. For the diplopia exercise, the doctor
enters in the report if the patient experienced double vision or not because of the great
difficulty in judging ocular eye misalignment during a telemedicine visit.
It is understood that the medical doctor is the final judge of the diagnostic. Our
method is a supporting tool, as are AI generated image annotations in radiography [
9
], and
it is not intended to replace the medical doctor’s diagnostic skill. Further, our approach
does not supplement the sophisticated technology used to study ocular motility for over
five decades [10].
Symptoms of double vision and ptosis are expected in essentially all patients with
MG, and the evaluation of lid position and ocular motility is a key aspect of the diagnostic
examination and ongoing assessment of patients, with ocular myasthenia being the most
common first form of the disease, before progressing to the generalized form [
7
]. In
many neurological diseases, including dementias, multiple sclerosis, stroke, and cranial
nerve palsies, eye movement examination is important in diagnosis. We expect that
our algorithm will also be useful in the diagnosis and monitoring of many neurological
disorders via telehealth [
11
13
]. The technology may also be adapted for assessment in the
in-person setting as a means to objectively quantitate the ocular motility examination in a
simple fashion.
The manuscript is organized as follows: First, we provide background on the develop-
ment of our study. Second, we describe in detail our method and algorithms and place into
context of previous studies to illustrate the novelty of our approach. Then we report the
results followed by discussion. Finally, we summarize our finding in the conclusion and
provide some recommendations on follow-up investigations.
2. Background
The NIH Rare Disease Clinical Research Network dedicated to myasthenia gravis
(MGNet) initiated an evaluation of examinations performed by telemedicine. The study
recorded the evaluations, including the MG Core Exam, to assess reproducibility and exam
performance. We took advantage of the Zoom recordings performed at George Washington
University to evaluate our technology. We used two videos of each subject performed on
different days for quantitative assessment of the severity of ptosis and diplopia for patients
with a confirmed diagnosis of MG. The patients were provided instructions regarding their
position in relationship to their cameras and levels of illumination, as well as to follow the
Sensors 2023,23, 7744 3 of 16
examining neurologist’s instructions during the examinations. In exercise 1, the patient is
asked to hold their gaze upwards for 61 s (see Figure 1). The goal is to assess the severity of
ptosis (uncontrolled closing of eyelid), if any, during the exercise [14] and rating severity:
Sensors 2023, 23, 7744 3 of 17
instructions regarding their position in relationship to their cameras and levels of illumi-
nation, as well as to follow the examining neurologists instructions during the examina-
tions. In exercise 1, the patient is asked to hold their gaze upwards for 61 s (see Figure 1).
The goal is to assess the severity of ptosis (uncontrolled closing of eyelid), if any, during
the exercise [14] and rating severity:
Figure 1. Subject looking up for evaluation of ptosis with identication of key landmarks. The red
dots are the landmarks obtained with the machine learning method. The green lines correspond to
the detected position of upper and lower eyelids and the iris boundary.
0: No visible ptosis within 45 s
1: Visible ptosis within 1145 s
2: visible ptosis within 10 s
3: Immediate ptosis
However, another grading system was proposed for the MG-CE using the following
ratings:
0: No ptosis
1: Mild, eyelid above pupil
2: Moderate, eyelid at pupil
3: Severe, eyelid below pupil
In exercises 2, the patient must hold their gaze right or left for 61 ssee Figure 2. The
goal is to assess for double vision, and when it appears. Severity rating ranges from 0 to 3:
0: No diplopia with 61 s sustained gaze
1: Diplopia with 1160 s sustained gaze
2: Diplopia within 110 s but not immediately
3: Immediate diplopia with primary or lateral gaze
Figure 2. Subject looking eccentrically in exercise 2 to evaluate for development of diplopia.
Our goal was to take accurate and robust measurements of the eye anatomy in real-
time during the exercises, and to automatically grade ptosis and ocular misalignment. The
algorithm should reconstruct the eye geometry of the patient from the video and the posi-
tion of the pupil inside that geometric domain. The difficulty is to precisely recover these
geometric elements from a video of the patient where the eye dimension in pixels is at best
about 1/10 of the overall image dimension. Most of the studies of oculometry assume that
the image is centered on the eye and that it occupies most of the image. Alternatively, eye
Figure 1.
Subject looking up for evaluation of ptosis with identification of key landmarks. The red
dots are the landmarks obtained with the machine learning method. The green lines correspond to
the detected position of upper and lower eyelids and the iris boundary.
0: No visible ptosis within 45 s
1: Visible ptosis within 11–45 s
2: visible ptosis within 10 s
3: Immediate ptosis
However, another grading system was proposed for the MG-CE using the
following ratings:
0: No ptosis
1: Mild, eyelid above pupil
2: Moderate, eyelid at pupil
3: Severe, eyelid below pupil
In exercises 2, the patient must hold their gaze right or left for 61 s—see Figure 2. The
goal is to assess for double vision, and when it appears. Severity rating ranges from 0 to 3:
Sensors 2023, 23, 7744 3 of 17
instructions regarding their position in relationship to their cameras and levels of illumi-
nation, as well as to follow the examining neurologists instructions during the examina-
tions. In exercise 1, the patient is asked to hold their gaze upwards for 61 s (see Figure 1).
The goal is to assess the severity of ptosis (uncontrolled closing of eyelid), if any, during
the exercise [14] and rating severity:
Figure 1. Subject looking up for evaluation of ptosis with identication of key landmarks. The red
dots are the landmarks obtained with the machine learning method. The green lines correspond to
the detected position of upper and lower eyelids and the iris boundary.
0: No visible ptosis within 45 s
1: Visible ptosis within 1145 s
2: visible ptosis within 10 s
3: Immediate ptosis
However, another grading system was proposed for the MG-CE using the following
ratings:
0: No ptosis
1: Mild, eyelid above pupil
2: Moderate, eyelid at pupil
3: Severe, eyelid below pupil
In exercises 2, the patient must hold their gaze right or left for 61 ssee Figure 2. The
goal is to assess for double vision, and when it appears. Severity rating ranges from 0 to 3:
0: No diplopia with 61 s sustained gaze
1: Diplopia with 1160 s sustained gaze
2: Diplopia within 110 s but not immediately
3: Immediate diplopia with primary or lateral gaze
Figure 2. Subject looking eccentrically in exercise 2 to evaluate for development of diplopia.
Our goal was to take accurate and robust measurements of the eye anatomy in real-
time during the exercises, and to automatically grade ptosis and ocular misalignment. The
algorithm should reconstruct the eye geometry of the patient from the video and the posi-
tion of the pupil inside that geometric domain. The difficulty is to precisely recover these
geometric elements from a video of the patient where the eye dimension in pixels is at best
about 1/10 of the overall image dimension. Most of the studies of oculometry assume that
the image is centered on the eye and that it occupies most of the image. Alternatively, eye
Figure 2. Subject looking eccentrically in exercise 2 to evaluate for development of diplopia.
0: No diplopia with 61 s sustained gaze
1: Diplopia with 11–60 s sustained gaze
2: Diplopia within 1–10 s but not immediately
3: Immediate diplopia with primary or lateral gaze
Our goal was to take accurate and robust measurements of the eye anatomy in real-
time during the exercises, and to automatically grade ptosis and ocular misalignment.
The algorithm should reconstruct the eye geometry of the patient from the video and the
position of the pupil inside that geometric domain. The difficulty is to precisely recover
these geometric elements from a video of the patient where the eye dimension in pixels is at
best about 1/10 of the overall image dimension. Most of the studies of oculometry assume
that the image is centered on the eye and that it occupies most of the image. Alternatively,
eye trackers do not rely on a standard camera using the visual spectrum but rather use
infrared in order to isolate clearly the pupil as a feature in a corneal reflection image [
15
17
].
Localization of eye position can take advantage of deep learning methods but requires
large, annotated data sets for training [
18
,
19
]. As described later, we use an existing open-
Sensors 2023,23, 7744 4 of 16
source library to take advantage of such information. From a model of eye detection, we
can focus the search for pupil and iris location in the region of interest [
20
]. Among the
popular techniques to detect the iris location [
21
] are the circular Hough transform [
22
,
23
]
and the Daughman’s algorithm method [24].
We found in our telehealth application, using a standard camera operating in the
visual spectrum, that these methods have a robustness issue due to their insensitivity to
low resolution of the eyes’ Region Of Interest (ROI), poor control on illumination of the
subject, and specific eye geometry consequent to ptosis. Next, we will present our hybrid
method, combining an existing deep learning library for face tracking and a local computer
vision method to build ptosis and diplopia metrics.
3. Method
3.1. Dataset
We used twelve videos acquired by Zoom during the ADAPT study telehealth sessions
of six patients with MG. Each subject had telemedicine evaluations within 48 h of each
other and participated in a set of standardized outcome measures including the MGNet
Core Exam [
5
]. Telehealth sessions were organized as Zoom meetings by a board-certified
neurologist with subspecialty training in neuromuscular disease in the clinic, providing the
assessments of all patients at their homes. In practice, these Zoom sessions were limited in
video quality to a relatively low resolution in order to accommodate the available internet
bandwidth and because they were recorded on the doctor side during streaming. We
extracted fixed images at various steps of the exercise to test our algorithm, as well as on
video clips of about 60 s each for each exercise 1 and 2, described above. The number of
pixels per frame was as low as 450
×
800 at a rate of 30 Frames Per Second (FPS). We also in-
cluded half a dozen videos of healthy subjects acquired in similar conditions to the ADAPT
patients in order to have a base line on eye motion during both eye tracking exercises.
We aimed to achieve 2-pixel accuracy of anatomic markers, allowing us to compute
ptosis and eye alignment. This was chosen for the following reasons: (1) When one looks
for an interface location, two pixels is about the best accuracy one may achieve since it is
projected on a discrete grid of pixels; (2) this metric is independent of the resolution of the
image. As a matter of fact, there is no control of the dimension of eyes expressed in pixels
in a video frame during the telemedicine session, since the assessment is posteriori on
existing videos; (3) in addition, high definition camera footage with a patient far away from
the camera will have lower resolution of ocular anatomy than a standard Zoom recording
with a patient close to the camera, as the eye dimension is expressed in pixels.
The distance of the patient to the camera and illumination of the subject leads to
variability of the evaluations. These conditions are inherent limitations of the telehealth
standard to accommodate patients’ equipment and home environment.
We used our new telehealth platform, which includes a Lumens B30U PTZ camera
(Lumens Digital Optics Inc., Hsinchu, Taiwan) with a resolution of 1080
×
1920 at
30 FPS
and is connected to a Dell Optiplex 3080 small form factor computer (Intel processor i5-
10500 t, 2.3 GHz, 8 Gb Ram), for processing video analysis. This system, tested initially on
healthy subjects, was used on one patient, following the MG-CE protocol. Through this
process, we have acquired a data set that is large enough to evaluate the robustness and
quality of the algorithms. Error rates depending on resolution and other human factors are
compared in the Section 4. Next, we describe the algorithm construction.
Full session videos were divided into short clips of each exercise using an editing
software. The clips were then processed one at a time. For every applicable exercise, the
video went through the Machine Learning algorithm, saving landmark positions of the face
using [
15
] in Python. Once the landmarks were saved, they were exported to a Matlab script
where the rest of the processing using conventional computer vision methods coded in
Matlab was performed. All processes were executed automatically on a laptop. We can also
use the Matlab compiler to accelerate processing and protect the software
from alterations.
Sensors 2023,23, 7744 5 of 16
3.2. Face and Eyes Detection
The first step was to detect the face in the image. Previous investigations have devel-
oped face tracking algorithms and compared methods for facial detection [
25
,
26
]. Among
the most widely used algorithms and the one we chose to use was OpenCV’s implementa-
tion of the Haar Cascade algorithm [
27
], based on the detector of R. Lienhart [
28
], which is
a fast method and, overall, the most reliable for real-time detection.
Once the bounding box of the face is detected, key facial landmarks were required to
monitor the patient’s facial features, which would be the foundation of our segmentation
and analysis to evaluate ptosis and ocular misalignment. For facial alignment, many
methods exist. Some of these image-based techniques were reviewed by Johnston and
Chazal [
29
]. One of the most time-efficient for real-time application is based on the shape
regression approach [
30
]. We used DLib’s implementation of the regression tree technique
from V. Kazemi and J. Sullivan [
31
], which was trained on the 300 W dataset [
32
] fitting
a 68 points landmark to the face (Figures 3and 4). The ROI for each eye is the polygon
formed by points 37 to 42 for the right eye, and 43 to 48 for the left eye, in reference to the
model in Figure 3.
Sensors 2023, 23, 7744 5 of 17
went through the Machine Learning algorithm, saving landmark positions of the face using
[15] in Python. Once the landmarks were saved, they were exported to a Matlab script where
the rest of the processing using conventional computer vision methods coded in Matlab was
performed. All processes were executed automatically on a laptop. We can also use the
Matlab compiler to accelerate processing and protect the software from alterations.
3.2. Face and Eyes Detection
The rst step was to detect the face in the image. Previous investigations have devel-
oped face tracking algorithms and compared methods for facial detection [25,26]. Among
the most widely used algorithms and the one we chose to use was OpenCVs implemen-
tation of the Haar Cascade algorithm [27], based on the detector of R. Lienhart [28], which
is a fast method and, overall, the most reliable for real-time detection.
Once the bounding box of the face is detected, key facial landmarks were required to
monitor the patients facial features, which would be the foundation of our segmentation
and analysis to evaluate ptosis and ocular misalignment. For facial alignment, many meth-
ods exist. Some of these image-based techniques were reviewed by Johnston and Chazal
[29]. One of the most time-ecient for real-time application is based on the shape regres-
sion approach [30]. We used DLibs implementation of the regression tree technique from
V. Kazemi and J. Sullivan [31], which was trained on the 300 W dataset [32] ing a 68
points landmark to the face (Figures 3 and 4). The ROI for each eye is the polygon formed
by points 37 to 42 for the right eye, and 43 to 48 for the left eye, in reference to the model
in Figure 3.
Figure 3. Dlib Facial Landmarks [15].
Figure 4. Eye Opening distance (blue arrow) and eye area (in green), obtained from the machine
learning landmarks (red dots).
We present below the postprocessing of each video frame, concentrating on the ROI
specic to each anatomic landmark (upper lid, lowerlid, and part of the iris boundary).
While in principle, processing of the video instead of serial images one by one would be
advantageous, blinking and rapid eye movements interferes with the process and make a
robust solution more dicult to build.
Figure 3. Dlib Facial Landmarks [15].
Sensors 2023, 23, 7744 5 of 17
went through the Machine Learning algorithm, saving landmark positions of the face using
[15] in Python. Once the landmarks were saved, they were exported to a Matlab script where
the rest of the processing using conventional computer vision methods coded in Matlab was
performed. All processes were executed automatically on a laptop. We can also use the
Matlab compiler to accelerate processing and protect the software from alterations.
3.2. Face and Eyes Detection
The rst step was to detect the face in the image. Previous investigations have devel-
oped face tracking algorithms and compared methods for facial detection [25,26]. Among
the most widely used algorithms and the one we chose to use was OpenCVs implemen-
tation of the Haar Cascade algorithm [27], based on the detector of R. Lienhart [28], which
is a fast method and, overall, the most reliable for real-time detection.
Once the bounding box of the face is detected, key facial landmarks were required to
monitor the patients facial features, which would be the foundation of our segmentation
and analysis to evaluate ptosis and ocular misalignment. For facial alignment, many meth-
ods exist. Some of these image-based techniques were reviewed by Johnston and Chazal
[29]. One of the most time-ecient for real-time application is based on the shape regres-
sion approach [30]. We used DLibs implementation of the regression tree technique from
V. Kazemi and J. Sullivan [31], which was trained on the 300 W dataset [32] ing a 68
points landmark to the face (Figures 3 and 4). The ROI for each eye is the polygon formed
by points 37 to 42 for the right eye, and 43 to 48 for the left eye, in reference to the model
in Figure 3.
Figure 3. Dlib Facial Landmarks [15].
Figure 4. Eye Opening distance (blue arrow) and eye area (in green), obtained from the machine
learning landmarks (red dots).
We present below the postprocessing of each video frame, concentrating on the ROI
specic to each anatomic landmark (upper lid, lowerlid, and part of the iris boundary).
While in principle, processing of the video instead of serial images one by one would be
advantageous, blinking and rapid eye movements interferes with the process and make a
robust solution more dicult to build.
Figure 4.
Eye Opening distance (blue arrow) and eye area (in green), obtained from the machine
learning landmarks (red dots).
We present below the postprocessing of each video frame, concentrating on the ROI
specific to each anatomic landmark (upper lid, lowerlid, and part of the iris boundary).
While in principle, processing of the video instead of serial images one by one would be
advantageous, blinking and rapid eye movements interferes with the process and make a
robust solution more difficult to build.
3.2.1. Computing the Ptosis Metric
First, we processed the time window of the video clip when the patient is focusing eye
gaze upwards.
The ROI construction for each eye may give a first approximation of ptosis for exercise
1 of the MG-CE as follows: We used eyelid distance and eye area, as shown in Figure 4.
The eyelid distance (ED) approximation is the average distance between points of the
Sensors 2023,23, 7744 6 of 16
upper eyelid (segment 38–39 for the right eye and 44–45 for left eye) and lower eyelids
(segment 42–41 for the right eye and 48–47 for left eye), as determined by the landmarks
in Figure 3. Eye area is the area contained in the outline of the eye determined by the
landmark (Figure 4). We normalize these measurements by the eye length (EL), as the
horizontal distance between the corners of each eye (Figure 5).
Sensors 2023, 23, 7744 6 of 17
3.2.1. Computing the Ptosis Metric
First, we processed the time window of the video clip when the patient is focusing
eye gaze upwards.
The ROI construction for each eye may give a rst approximation of ptosis for exer-
cise 1 of the MG-CE as follows: We used eyelid distance and eye area, as shown in Figure
4. The eyelid distance (ED) approximation is the average distance between points of the
upper eyelid (segment 3839 for the right eye and 4445 for left eye) and lower eyelids
(segment 4241 for the right eye and 4847 for left eye), as determined by the landmarks
in Figure 3. Eye area is the area contained in the outline of the eye determined by the
landmark (Figure 4). We normalize these measurements by the eye length (EL), as the hor-
izontal distance between the corners of each eye (Figure 5).
Figure 5. Eye length measurement.
As a byproduct, we also identify the blink rate, as shown in Figure 6.
Figure 6. Blinking Identication: The downward spikes of the graph are perfectly synchronized and
correspond to blinks.
The eye lid location provided by the deep learning algorithm may not be accurate, as
shown in Figure 7. For example, here the lower landmarks (41) and (42) are quite far o
the contour of the eye, and landmarks (37) and (40) are not precisely localized at the cor-
ners of the eye. The accuracy of this library varies depending on the characteristic of the
subject, such as iris color, contrast with sclera, skin color, etc. The accuracy is also depend-
ent on the frame of the video clip and eect of lightning or variations of head position. We
also found that the hexagon of the model identied by the deep learning algorithm may
degenerate to a pentagon when a corner point overlaps another point of the hexagon. In
some extreme cases, we found the ROIs at the wrong location; for example, the algorithm
confused the nares with the eye location. Such an error was relatively easy to detect but
improving the accuracy of the deep learning library for a patient performing an eccentric
gaze position, such as in exercise 1 and 2, would require re-training the algorithm with a
model having a larger number of landmarks concentrating on the ROI. This is feasible but
would require a large traing set. MG is a rare disease, and no large data sets suitable for
training a deep learning techniques exist at present. Further, we have no way to predict
the size of data set necessary to eectively train the algorithm.
Many eye detection methods have been developed in the eld of ocular motility re-
search, but they rely on images taken in controlled environments with specic infrared
Figure 5. Eye length measurement.
As a byproduct, we also identify the blink rate, as shown in Figure 6.
Sensors 2023, 23, 7744 6 of 17
3.2.1. Computing the Ptosis Metric
First, we processed the time window of the video clip when the patient is focusing
eye gaze upwards.
The ROI construction for each eye may give a rst approximation of ptosis for exer-
cise 1 of the MG-CE as follows: We used eyelid distance and eye area, as shown in Figure
4. The eyelid distance (ED) approximation is the average distance between points of the
upper eyelid (segment 3839 for the right eye and 4445 for left eye) and lower eyelids
(segment 4241 for the right eye and 4847 for left eye), as determined by the landmarks
in Figure 3. Eye area is the area contained in the outline of the eye determined by the
landmark (Figure 4). We normalize these measurements by the eye length (EL), as the hor-
izontal distance between the corners of each eye (Figure 5).
Figure 5. Eye length measurement.
As a byproduct, we also identify the blink rate, as shown in Figure 6.
Figure 6. Blinking Identication: The downward spikes of the graph are perfectly synchronized and
correspond to blinks.
The eye lid location provided by the deep learning algorithm may not be accurate, as
shown in Figure 7. For example, here the lower landmarks (41) and (42) are quite far o
the contour of the eye, and landmarks (37) and (40) are not precisely localized at the cor-
ners of the eye. The accuracy of this library varies depending on the characteristic of the
subject, such as iris color, contrast with sclera, skin color, etc. The accuracy is also depend-
ent on the frame of the video clip and eect of lightning or variations of head position. We
also found that the hexagon of the model identied by the deep learning algorithm may
degenerate to a pentagon when a corner point overlaps another point of the hexagon. In
some extreme cases, we found the ROIs at the wrong location; for example, the algorithm
confused the nares with the eye location. Such an error was relatively easy to detect but
improving the accuracy of the deep learning library for a patient performing an eccentric
gaze position, such as in exercise 1 and 2, would require re-training the algorithm with a
model having a larger number of landmarks concentrating on the ROI. This is feasible but
would require a large traing set. MG is a rare disease, and no large data sets suitable for
training a deep learning techniques exist at present. Further, we have no way to predict
the size of data set necessary to eectively train the algorithm.
Many eye detection methods have been developed in the eld of ocular motility re-
search, but they rely on images taken in controlled environments with specic infrared
Figure 6.
Blinking Identification: The downward spikes of the graph are perfectly synchronized and
correspond to blinks.
The eye lid location provided by the deep learning algorithm may not be accurate, as
shown in Figure 7. For example, here the lower landmarks (41) and (42) are quite far off the
contour of the eye, and landmarks (37) and (40) are not precisely localized at the corners of
the eye. The accuracy of this library varies depending on the characteristic of the subject,
such as iris color, contrast with sclera, skin color, etc. The accuracy is also dependent on the
frame of the video clip and effect of lightning or variations of head position. We also found
that the hexagon of the model identified by the deep learning algorithm may degenerate to
a pentagon when a corner point overlaps another point of the hexagon. In some extreme
cases, we found the ROIs at the wrong location; for example, the algorithm confused the
nares with the eye location. Such an error was relatively easy to detect but improving the
accuracy of the deep learning library for a patient performing an eccentric gaze position,
such as in exercise 1 and 2, would require re-training the algorithm with a model having a
larger number of landmarks concentrating on the ROI. This is feasible but would require a
large traing set. MG is a rare disease, and no large data sets suitable for training a deep
learning techniques exist at present. Further, we have no way to predict the size of data set
necessary to effectively train the algorithm.
Many eye detection methods have been developed in the field of ocular motility
research, but they rely on images taken in controlled environments with specific infrared
lights allowing for maximal contrast to define the eye and orbital anatomy, and are focused
exclusively on the eye.
The essential feature of our approach is to start from the ROI, i.e the polygons provided
by deep learning that delineate roughly the eye “contour”. This result is relatively robust
with a standard video but is not very accurate overall. Therefore, we zoomed in on
spatial subdomains of the ROI in a divide and conqueer manner to segment each interface
Sensors 2023,23, 7744 7 of 16
individually using standard computer vision algorithms. Special features are the upper
lid/lower lids and bottom of the iris boundary for ptosis and the visible side of the iris
boundary for diplopia. We have separated the upper lid, lower lid, and the iris boundaries
as we are looking for a single interface in each rectangular box, as in Figure 7. In principle,
these interface boundaries should cross the rectangle horizontally for lid position and
vertically for ocular misalignment. One must check with an algorithm that the interface
partitions the rectangle into two connex sub domains. The segmentation algorithm may
shrink the rectangle to a smaller dimension as much as necessary to separate each anatomic
feature.
Sensors 2023, 23, 7744 7 of 17
lights allowing for maximal contrast to dene the eye and orbital anatomy, and are fo-
cused exclusively on the eye.
Figure 7. Local Rectangle on left to search for the correct position of the lower lid and on right to
draw the interface between the iris and sclera below. The red dots are obtained with the machine
learning method. The blue squares are the regions used to search for the green landmarks: the iris
boundary and the upper and lower eyelids.
The essential feature of our approach is to start from the ROI, i.e the polygons pro-
vided by deep learning that delineate roughly the eye contour”. This result is relatively
robust with a standard video but is not very accurate overall. Therefore, we zoomed in on
spatial subdomains of the ROI in a divide and conqueer manner to segment each interface
individually using standard computer vision algorithms. Special features are the upper
lid/lower lids and boom of the iris boundary for ptosis and the visible side of the iris
boundary for diplopia. We have separated the upper lid, lower lid, and the iris boundaries
as we are looking for a single interface in each rectangular box, as in Figure 7. In principle,
these interface boundaries should cross the rectangle horizontally for lid position and ver-
tically for ocular misalignment. One must check with an algorithm that the interface par-
titions the rectangle into two connex sub domains. The segmentation algorithm may
shrink the rectangle to a smaller dimension as much as necessary to separate each ana-
tomic feature.
To be more specic, we rst describe the local search method to position the lower
lid and the lower boundary of the iris during ptosis exercise 1. The description below is
set for the right eye, with the processing of left eye being analagous. As shown in Figure
7, to improve the lower lid positioning, we draw a small rectangle including the landmark
points (42) (41) and look for the interface between the sclera and the skin of the lower lid.
Similarly, we draw a rectangle that contains (38), (39), (40), and (41) and identify the in-
terface of the iris and sclera.
The interface found by the computer vision algorithm would be acceptable only if it
is a smooth curve (H1) that crosses the rectangle horizontally (H2). For the iris boom, we
expect the curve to be convex (H3).
As the computer vision is concentrated in a rectangle of interest that contains the
interface we are looking for, the problem is simpler to solve, and the solution can be found
accurately provided that the two subdomains separated by the interface have clear dis-
tinct features. To that purpose, we rst enhance the contrast of the image in that rectangle
before further processing. Second, we used several simple techniques, such as kmeans,
restricting ourselves to two clusters, or open snake, which maximizes the gradient of the
image along a curve. Those numerical techniques come with numerical indicators to show
how well two regions are clearly separated in a rectangular box. For example, with the
kmean algorithm, we center the two clusters clearly separated, and each cluster should be
a connex set (H4). For the open snake method, we check on the smoothness of the curve
and the gradient value across that curve.
If the computer vision algorithm fails to nd an interface that satises all our hypoth-
eses (H1 to H4), we either reran the k-means algorithm, changing the seed, or shrink the
size of the rectangle until convergence to an acceptable solution. If the computer vision
algorithm fails to nd a smaller rectangle that ts the search, we cannot conclude on the
Figure 7.
Local Rectangle on left to search for the correct position of the lower lid and on right to
draw the interface between the iris and sclera below. The red dots are obtained with the machine
learning method. The blue squares are the regions used to search for the green landmarks: the iris
boundary and the upper and lower eyelids.
To be more specific, we first describe the local search method to position the lower lid
and the lower boundary of the iris during ptosis exercise 1. The description below is set
for the right eye, with the processing of left eye being analagous. As shown in Figure 7,
to improve the lower lid positioning, we draw a small rectangle including the landmark
points (42) (41) and look for the interface between the sclera and the skin of the lower
lid. Similarly, we draw a rectangle that contains (38), (39), (40), and (41) and identify the
interface of the iris and sclera.
The interface found by the computer vision algorithm would be acceptable only if it is
a smooth curve (H1) that crosses the rectangle horizontally (H2). For the iris bottom, we
expect the curve to be convex (H3).
As the computer vision is concentrated in a rectangle of interest that contains the
interface we are looking for, the problem is simpler to solve, and the solution can be found
accurately provided that the two subdomains separated by the interface have clear distinct
features. To that purpose, we first enhance the contrast of the image in that rectangle before
further processing. Second, we used several simple techniques, such as kmeans, restricting
ourselves to two clusters, or open snake, which maximizes the gradient of the image along
a curve. Those numerical techniques come with numerical indicators to show how well two
regions are clearly separated in a rectangular box. For example, with the kmean algorithm,
we center the two clusters clearly separated, and each cluster should be a connex set (H4).
For the open snake method, we check on the smoothness of the curve and the gradient
value across that curve.
If the computer vision algorithm fails to find an interface that satisfies all our hypothe-
ses (H1 to H4), we either reran the k-means algorithm, changing the seed, or shrink the
size of the rectangle until convergence to an acceptable solution. If the computer vision
algorithm fails to find a smaller rectangle that fits the search, we cannot conclude on the
lower lid and upper lid position and must skip that image frame from our reconstrution of
the ptosis/diplopia metrics we are building. This may happen if there is a lack of contrast
due to poor illumination or a diffuse image due to motion.
The kmean algorithm and snake are some of the simplest algorithms used to separate
the two subdomains of the rectangle and draw an interface that satisfies the criteria (H1) to
(H4). We used both algorithms simultaneously to determine concurrence of each algorithm.
As shown in the Section 4, this solution is robust and has satisfactory accuracy. There is
Sensors 2023,23, 7744 8 of 16
no difficulty in using a more sophisticated level set or graph cut technique, other than the
additional cpu requirement.
In the example seen in Figure 7, the model provides the correct location of the upper
lid, and the contrast between the iris and the skin above is clear.
Overall, our hybrid algorithm combines deep learning with local computer vision
output metrics, such as the distance between the lower lid and the bottom of the iris as well
as the lower lid and the upper lid. The first distance is useful to determine that the patient
is performing the exercise correctly, while the second distance provides an assessment
of ptosis. It is straightforward to assess the diameter of the iris as the patient is looking
straight and the pupil should be at the center of the iris circle.
3.2.2. Computing the Diplopia Metric
As illustrated in Figure 2, we use a similar method to identify the upper lid and lower
lid position. The only novelty here is to identify the correct side boundary of the iris as the
patient is looking left or right, using a computer vision algorithm in a small horizontal box
that starts from the corner of the eye landmark (37) or (40) and goes to the landmarks of the
upper lid and lower lid on the opposite side, i.e., (39) and (41) or (38) and (42). The same
algorithm is applied to the right eye, as described above, and the left eye, except that the
inteface is now crossing the rectangle of interest vertically. We compute the barycentric
coordinate denoted
α
and the horizontal distance of the point P that is the visible lateral
point of the iris boundary to the corner of the eye, relative to eye length, as shown in
Figures 2and 8. We denote these as
α
left and
α
right the measurement
α
for the left and
right eyes, respectively. The distance from the face of the patient to the camera is much
larger than the dimension of the eye and makes the barycentric coordinate quasi-invariant
to the small motion of the patient head during the exercise.
Figure 8. Barycentric Coordinate (α) used in Diplopia Assessment.
In principle, Pleft and Pright should be of the same order, as the subject is looking
straight at the camera;
α
left and
α
right should also be strongly corelated, as the subjects
direct their gaze horizontally. As fatigue occurs, the difference between
α
left and
α
right
should change with time and correspond to the misalignment of both eyes. We assume that
diplopia occurs when the difference between
α
left–
α
right deviates significantly from its
initial value at the beginning of the exercise.
3.2.3. Eye Gaze and Reconstruction of Ptosis and Diplopia Metrics in Time
Thus far, we have described a hybrid algorithm that we used for each frame of the
video clips during Exercise 1 and 2. As before, we used a simple clustering algorithm for
the ROI for each eye to reconstruct the scleral area and detect the time window for each
exercise: the sclera should be one side left or right of the iris in Exercise 2 and one side
Sensors 2023,23, 7744 9 of 16
below the iris in Exercise 1. Since we know, a priori, that each exercise lasts one minute, we
do not need a very accurate method to reconstruct when the exercise starts or ends. For
verification purposes, the result of left eye gaze and right eye gaze should be consistent.
We cannot guarantee that the computer vision algorithm converges for each frame.
We are required to check for:
Stability: the patient should keep their head in approximately the same position
Lightning effects: the k-means algorithm shows non-convex clusters in the rectangle
of interest when reflecting light affects the iris detection, for example.
Instability of the deep learning algorithm output: when the landmarks of the ROI
change in time independently of the head position.
Exception of quick eye movements due to blinking or reflex that should not enter the
ptosis or diplopia assessment.
As our method eliminates all the frames that do not pass these tests, we generate a
time series of measures for ptosis and diplopia during each one-minute exercise that is
not continuous in time. Rather, we use linear interpolation in time to fill the holes and
provide that the time gaps are small enough, i.e., a fraction of a second. All time gaps that
are greater than a second are identified in the time series and may correspond to marker
fatigue during the evaluation.
To obtain the dynamic of the ptosis and diplopia measure that is not part of the
standard core exam, and to present some interest for neuromuscular fatigue, we further
postprocess the signal with a special high order filter, as in [
13
], that can take advantage of
Fourier technique for nonperiodic time series.
4. Results
To construct the validation of our method, we visually compare the result of our
hybrid segmentation algorithm to a ground true result obtained on fixed images. In order
to obtain a representative data set, we extract an image every two seconds from the videos
of patients. We used six videos of the ADAPT series with the first visit of six patients. The
subjects were three women, three men—one African American/Black, one Asian, and three
White, with one person who identified as Hispanic.
We extracted one image every 2 s of the video clip for Exercise 1, assessing ptosis,
and the two video clips corresponding to Exercise 2, assessing ocular misalignment. We
conducted the same analysis with the patient video that was recorded with the Inteleclinic
system equipped with a high-definition camera.
Each exercise lasts about one minute, and we obtained about 540 images from the
ADAPT series and 90 from the Inteleclinic. The validation of the image segmentation was
performed for each eye, which doubles the amount of time.
For Exercise 1, we check three landmarks positions: the points on the upper lid, iris
bottom, and lower lid situated on the vertical line that crosses the center of the ROI. For
exercise 2, we identify the position of the iris boundary that is opposite to the direction of
gaze. If the patient looks to their left, we determine the position of the iris boundary that is
furthest to the right.
To facilitate the verification, our code automatically generates these images with an
overlay of the grid of spatial steps at two pixels. This rule is used vertically for exercise 1
and horizontally for exercise 2.
We consider that the segmentation is correct to assess ptosis and ocular misalignment
when the localization of the landmarks is correct within two pixels. It is often difficult to
judge the results based on our own inspection, as demonstrated in the image of Figure 9.
We used independent visual verifications by two reviewers to make our determinations.
Sensors 2023,23, 7744 10 of 16
Sensors 2023, 23, 7744 10 of 17
judge the results based on our own inspection, as demonstrated in the image of Figure 9.
We used independent visual verications by two reviewers to make our determinations.
Figure 9. Visual verication on zoomed image of eyes using a 2-pixel rule for Exercise 1 on left,
Exercise 2 on right. Green lines are the detected positions of upper and lower eyelids and iris bound-
ary. The red dots correspond to the landmarks detected with the machine learning method.
Not all images are resolved by our hybrid algorithm. It is critical to have an adequate
time frame in the video to reconstruct the dynamic of ptosis and ocular misalignment. First,
we eliminated all the images from the data set in which the Deep Learning library failed to
correctly localize the eyes. This can be easily detected in a video, since the library operates
on each frame individually and may jump from one position to a completely different one
while the patient stays still. For example, for one of the patients, the deep learning algo-
rithms confused each of the nostrils with the eyes.
The MG-CE videos had low resolution, especially when the displays of the patient
and the medical doctor were side by side, or have poor contrast, focus, or lighting. There-
fore, it is impressive that, on average, we were able to use 74% of the data set for further
processing with our hybrid algorithm.
Our algorithm cannot precisely nd the landmark of interest when the deep learning
library gives an ROI that is signicantly o the target. The bias of the deep learning algo-
rithm was particularly signicant during exercise 1, in which the eyes are wide open, and
the scleral area is decentered below the iris. The lower points of the polygon that mark the
ROI are often far inside the white scleral area above the lower lid. The end points of the
hexagon in the horizontal direction may become misaligned with the iris too far o the
rectangular area of local search we wish to identify.
We automatically eliminated 44% of the images of the video clips of the ADAPT se-
ries and 10% of the Inteleclinic series for exercise 1. The Inteleclinic result was acquired in
beer lightning conditions and with a higher resolution than the ADAPT videos.
We consider that the segmentation is correct if the three landmarks of the ptosis ex-
ercises, as dened earlier, are within two pixels of the ground true result obtained manu-
ally.
For exercise 1 with the ADAPT series, we obtained a success rate for identication of
73% for the lower lid, 89% for the boom of the iris, and 78% for the upper lid. For exercise
1 and for the Inteleclinic series of images, we obtained a success rate of 77%, 100%, and
77%, respectively.
For exercise 2, the quality of the acquisition was somewhat beer. We eliminated 18%
of the image ROIs for the ADAPT series and 13% for the Inteleclinic series.
The localization of the iris boundary, used to check ocular misalignment, was beer,
with a success rate of 95%. The eyelids are less open than in exercise 1 and are held in the
primary position. The upper and lower lid landmarks were obtained with a success rate
of 73% and 86%, respectively.
Overall, the success rates between the ADAPT series, provided by standard Zoom
call, and the Inteleclinic system result, which uses an HD camera, were similar. Under the
same conditions of distance between the patient and the camera, as well as the 2-pixel
accuracy, Inteleclinic may provide a submillimeter accuracy of eye motion that is
Figure 9.
Visual verification on zoomed image of eyes using a 2-pixel rule for Exercise 1 on left,
Exercise 2 on right. Green lines are the detected positions of upper and lower eyelids and iris
boundary. The red dots correspond to the landmarks detected with the machine learning method.
Not all images are resolved by our hybrid algorithm. It is critical to have an adequate
time frame in the video to reconstruct the dynamic of ptosis and ocular misalignment. First,
we eliminated all the images from the data set in which the Deep Learning library failed to
correctly localize the eyes. This can be easily detected in a video, since the library operates
on each frame individually and may jump from one position to a completely different
one while the patient stays still. For example, for one of the patients, the deep learning
algorithms confused each of the nostrils with the eyes.
The MG-CE videos had low resolution, especially when the displays of the patient and
the medical doctor were side by side, or have poor contrast, focus, or lighting. Therefore, it
is impressive that, on average, we were able to use 74% of the data set for further processing
with our hybrid algorithm.
Our algorithm cannot precisely find the landmark of interest when the deep learning
library gives an ROI that is significantly off the target. The bias of the deep learning
algorithm was particularly significant during exercise 1, in which the eyes are wide open,
and the scleral area is decentered below the iris. The lower points of the polygon that mark
the ROI are often far inside the white scleral area above the lower lid. The end points of
the hexagon in the horizontal direction may become misaligned with the iris too far off the
rectangular area of local search we wish to identify.
We automatically eliminated 44% of the images of the video clips of the ADAPT series
and 10% of the Inteleclinic series for exercise 1. The Inteleclinic result was acquired in better
lightning conditions and with a higher resolution than the ADAPT videos.
We consider that the segmentation is correct if the three landmarks of the ptosis exer-
cises, as defined earlier, are within two pixels of the ground true result
obtained manually.
For exercise 1 with the ADAPT series, we obtained a success rate for identification of
73% for the lower lid, 89% for the bottom of the iris, and 78% for the upper lid. For exercise
1 and for the Inteleclinic series of images, we obtained a success rate of 77%, 100%, and
77%, respectively.
For exercise 2, the quality of the acquisition was somewhat better. We eliminated 18%
of the image ROIs for the ADAPT series and 13% for the Inteleclinic series.
The localization of the iris boundary, used to check ocular misalignment, was better,
with a success rate of 95%. The eyelids are less open than in exercise 1 and are held in the
primary position. The upper and lower lid landmarks were obtained with a success rate of
73% and 86%, respectively.
Overall, the success rates between the ADAPT series, provided by standard Zoom call,
and the Inteleclinic system result, which uses an HD camera, were similar. Under the same
conditions of distance between the patient and the camera, as well as the 2-pixel accuracy,
Inteleclinic may provide a submillimeter accuracy of eye motion that is remarkable and
much better than the ADAPT series result. However, the Inteleclinic is more sensitive to
the head motion of the subject.
We now present summary results of the ptosis and diplopia assesments of MG subjects,
which were the goal of the eye and lid motion algorithm development.
Sensors 2023,23, 7744 11 of 16
As illustrated in Figure 3, we can determine, from the polygon obtained by the deep
learning algorithm, a first approximation of ptosis severity by computing the area of the
eye that is exposed to the view, as well as the vertical dimension of the eye. As a byproduct
of this metric, we can identify blinking (see Figure 6). We appreciated that left and right eye
blinking occur simultaneously, as is expected. Not every subject with MG blinks during
the exercise.
The time dependent measure of diplopia or ptosis obtained by our algorithm contains
noise. We can improve the accuracy of the measures by ignoring the eyes with identified
detection outliers, provided that the time gaps corresponding to these outliers are small. To
recover the signal without losing accuracy, we use the same high order filtering technique
that we used in a previous paper to analyze thermal imagery signal [
13
]. The Inteleclinic
data set is working well, as shown in Figure 10.
Sensors 2023, 23, 7744 11 of 17
remarkable and much beer than the ADAPT series result. However, the Inteleclinic is
more sensitive to the head motion of the subject.
We now present summary results of the ptosis and diplopia assesments of MG sub-
jects, which were the goal of the eye and lid motion algorithm development.
As illustrated in Figure 3, we can determine, from the polygon obtained by the deep
learning algorithm, a rst approximation of ptosis severity by computing the area of the
eye that is exposed to the view, as well as the vertical dimension of the eye. As a byproduct
of this metric, we can identify blinking (see Figure 6). We appreciated that left and right
eye blinking occur simultaneously, as is expected. Not every subject with MG blinks dur-
ing the exercise.
The time dependent measure of diplopia or ptosis obtained by our algorithm con-
tains noise. We can improve the accuracy of the measures by ignoring the eyes with iden-
tied detection outliers, provided that the time gaps corresponding to these outliers are
small. To recover the signal without losing accuracy, we use the same high order ltering
technique that we used in a previous paper to analyze thermal imagery signal [13]. The
Inteleclinic data set is working well, as shown in Figure 10.
Figure 10. Linear least square ing coecient of the upper lid drop from right and left eye, respec-
tively = [0.15, −0.17]. The green line shows a least square approximation of the distance between
the lower lid and upper lids of the patient. The red curve shows the distance between the lower
point of the iris and the lower lid below. This second curve is used to check that the patient performs
the exercise correctly.
We observe a 15% decay in lid opening that is very dicult to appreciate by human
inspection of video clip or during the in person medical doctor examination. As a maer
of fact, this low shift of the upper lid is slow and almost unnoticeable during a 60 s obser-
vation. This patient was considered as asymptomatic during the neurologist examination,
but the lid droop has been identied by our method. Control subject patients, even the
elderly, did not exhibit lid drop of a similar degree during the ptosis exercise.
During exercise 2 with the patient of the inteleclinic data set, we obtained no eye
misalignment (see Figure 11), but the eye opening was about half of its value during the
rst ptosis exercise, and the eye opening does not stay perfectly constant. We observe, on
Figure 10.
Linear least square fitting coefficient of the upper lid drop from right and left eye,
respectively = [
0.15,
0.17]. The green line shows a least square approximation of the distance
between the lower lid and upper lids of the patient. The red curve shows the distance between the
lower point of the iris and the lower lid below. This second curve is used to check that the patient
performs the exercise correctly.
We observe a 15% decay in lid opening that is very difficult to appreciate by human
inspection of video clip or during the in person medical doctor examination. As a matter of
fact, this low shift of the upper lid is slow and almost unnoticeable during a 60 s observation.
This patient was considered as asymptomatic during the neurologist examination, but the
lid droop has been identified by our method. Control subject patients, even the elderly, did
not exhibit lid drop of a similar degree during the ptosis exercise.
During exercise 2 with the patient of the inteleclinic data set, we obtained no eye
misalignment (see Figure 11), but the eye opening was about half of its value during the
first ptosis exercise, and the eye opening does not stay perfectly constant. We observe, on
the Inteleclinic video, that the eye gaze direction to the left and to the right is so extreme
that one of the pupils might be covered, in part, by the skin at the corner of the eyes.
Sensors 2023,23, 7744 12 of 16
Sensors 2023, 23, 7744 12 of 17
the Inteleclinic video, that the eye gaze direction to the left and to the right is so extreme
that one of the pupils might be covered, in part, by the skin at the corner of the eyes.
Figure 11. Evolution of the barycentric coordinates for each eye during exercise 2. The green line is
the least square ing of the barycentric coordinates evolution. It is perfectly horizontal which
means no weakness in eye motion was shown during the exercise.
The results of ptosis and diplopia assessments from the ADAPT video were less sat-
isfactory but still allowed an assessment of lid droop and ocular misalignment, though
with less accuracy. Figure 12 shows a representative example of the limits of our approach,
when the gap of information between two time points cannot be recovered. It should be
appreciated that the eye opening was of the order of 10 pixels as opposed to about 45
pixels in the Inteleclinic data set. In this situation, the subject was not close enough to the
camera, which compromised the resolution signicantly. However, a posteriori review
identied the gap identied by our algorithm and it corresponded to a short period of
time when the patient stopped looking up to rest from looking straight. It should be ap-
preciated that holding an upward gaze for one-minute is a strenuous atypical activity for
anyone.
Figure 12. Example of the assessment of one the ADAPT patient series. Note the patient closes his
eyes at 35 s.
Figure 11.
Evolution of the barycentric coordinates for each eye during exercise 2. The green line
is the least square fitting of the barycentric coordinates evolution. It is perfectly horizontal which
means no weakness in eye motion was shown during the exercise.
The results of ptosis and diplopia assessments from the ADAPT video were less
satisfactory but still allowed an assessment of lid droop and ocular misalignment, though
with less accuracy. Figure 12 shows a representative example of the limits of our approach,
when the gap of information between two time points cannot be recovered. It should be
appreciated that the eye opening was of the order of 10 pixels as opposed to about 45 pixels
in the Inteleclinic data set. In this situation, the subject was not close enough to the camera,
which compromised the resolution significantly. However, a posteriori review identified
the gap identified by our algorithm and it corresponded to a short period of time when
the patient stopped looking up to rest from looking straight. It should be appreciated that
holding an upward gaze for one-minute is a strenuous atypical activity for anyone.
Sensors 2023, 23, 7744 12 of 17
the Inteleclinic video, that the eye gaze direction to the left and to the right is so extreme
that one of the pupils might be covered, in part, by the skin at the corner of the eyes.
Figure 11. Evolution of the barycentric coordinates for each eye during exercise 2. The green line is
the least square ing of the barycentric coordinates evolution. It is perfectly horizontal which
means no weakness in eye motion was shown during the exercise.
The results of ptosis and diplopia assessments from the ADAPT video were less sat-
isfactory but still allowed an assessment of lid droop and ocular misalignment, though
with less accuracy. Figure 12 shows a representative example of the limits of our approach,
when the gap of information between two time points cannot be recovered. It should be
appreciated that the eye opening was of the order of 10 pixels as opposed to about 45
pixels in the Inteleclinic data set. In this situation, the subject was not close enough to the
camera, which compromised the resolution signicantly. However, a posteriori review
identied the gap identied by our algorithm and it corresponded to a short period of
time when the patient stopped looking up to rest from looking straight. It should be ap-
preciated that holding an upward gaze for one-minute is a strenuous atypical activity for
anyone.
Figure 12. Example of the assessment of one the ADAPT patient series. Note the patient closes his
eyes at 35 s.
Figure 12.
Example of the assessment of one the ADAPT patient series. Note the patient closes his
eyes at 35 s.
5. Discussion
Due to the precautions caused by the COVID-19 pandemic, there has been a rapid
increase in the utilization of telemedicine in patient care and clinical trials. The move to
video evaluations offers the opportunity to objectify and quantify physical examination,
which presently relies on the subjective assessment of an examiner with varied levels
of experience and often limited time to perform a thorough examination. Physicians
Sensors 2023,23, 7744 13 of 16
still remain reticent to incorporate telemedicine into their clinical habits, in particular in
areas that require a physical examination (neuromuscular diseases, movement disorders)
compared to areas that are primarily symptom-focused (headache). Telemedicine, on the
other hand, has numerous features that can provide an enhanced assessment of muscle
weaknesses, deeper patient monitoring and education, reduced burden and cost of in-
person clinic visits, and increased patient access to care. The potential for clinical trials
to establish rigorous, reproducible examinations at home provides similar benefits for
research subjects.
MG is an autoimmune, neuromuscular disease. Outcome measures are established for
MG trials, but these are considered suboptimal [
33
]. The MG CE, in particular for ocular
muscle weakness, has been standardized and is well defined [
5
]. Because of the benefit
for frequent monitoring for MG patients, reliable and quantitative teleconsultation would
be highly valuable for patient care. However, the grading of ptosis and diplopia relies
on repetitive and tedious examinations that the examiner must perform. The dynamic
component of upper eyelid dropping can be overlooked during the typical examination by
non-experts. Diagnosis of diplopia in these telehealth sessions relies on subjective patient
feedback. Overall, the physical examination relies heavily on qualitative, experienced
judgment, rather than on unbiased, rigorous, quantitative metrics.
Our goal is to move from 2D teleconsultation and its limitations to a multi-dimension
consultation. The system presented in this paper addresses that need by introducing
modern image processing techniques that are quick and robust in the process of recovering
quantitative metrics that should be independent of the examiner. The diagnosis and
treatment decisions remain the responsibility of the medical doctor, who has the knowledge,
and not our algorithm output.
One of the difficulties of standard telehealth sessions is the poor quality of video. The
resolution may be severely limited by the bandwidth of the network at the patient location.
Zoom video footage is also limited by the poor compression/decompression methods used,
which could cause artifacts with respect to the original raw video recording acquired by the
patient’s camera. Eventually the patient evaluation would benefit from software used on
their computers or tablets in order to directly compute all exam metrics with the resolution
of the native image before any compression occurs. These metrics, which are just a few
numbers, can be sent on a network directly to the physician without bandwidth limitations
In our study, the quality of the video was adequate to allow the medical doctor assess ptosis
and diplopia as specified by the protocol, but not ideal for image processing, especially
because the videos were recorded on the doctor side rather than recording the raw video
footages on the patient side. Light conditions and positioning of the patient in front of the
camera was often poorly controlled when patients are at home with their personal computer
or tablet. These factors are all of crucial importance to optimize numerical algorithm and
image processing that are robust and transparent on the level of accuracy they provide.
Eye tracking is highly sensitive to patient motion, poor resolution, and eyelid droop with
gaze directed eccentrically. However, use of standard telemedicine video will be necessary,
despite limitations, to support scalability and adoption of our approach for broad use by
physicians and patients.
As we digitalize the exercise output for assessing ptosis, we must rigorously define the
metric. We assessed instantaneous measurements as well as time dependent ones. From the
dynamic perspective, we could identify patients who show steady upper eyelid position
from those who start well and develop progressive eyelid weakness. We also separate
global measurement related to the overall eye opening from measurement that computes
the distance from the pupil to the upper lid. This last metric is clinically significant for
the patient when the droop covers the pupil and impairs vision. A decision on how these
metrics should classify ptosis severity remains the physician’s decision.
Similarly, we assumed that diplopia could be measured by the “misalignment” of the
left and right pupil during exercise 2. Vision involves not only the alignment of the ocular
Sensors 2023,23, 7744 14 of 16
axes but also the brain’s ability to compensate for such the misalignment and eliminate the
complaint of double vision.
Both measurement of ptosis and diplopia were quite sensitive to the resolution of
the video. In Zoom recorded telehealth sessions, the distance from the pupil to the upper
lid is of the order of ten pixels. A two pixels error on the landmark positions may still
provide a relative error of about 20% on the ptosis metric. The deep learning algorithm
introduces even larger errors on the landmark points of the ROI polygon. However, with
the HD camera we tested, and the processing being conducted on raw footage rather than
on streamed recorded footage, this relative error is divided by two. Our analytical approach
has also been able to provide recommendations on how to improve the MG ocular exam.
For example, to ensure the reproducibility and quality of the result, our algorithm can
provide feedback in real-time to the medical doctor on how many pixels are available
to track the eyes and therefore give direction to the patient to position closer and better
with respect to the camera on their end. Similarly, exercise 2 may benefit from reduced
extreme eccentric gaze, which compromises definition of iris boundary, when covered by
the overlying eyelid. This would allow for a more appropriate situation to assess double
vision properly.
The general concept of our hybrid approach, starting from deep learning to obtain
the ROI, and zooming into specific regions of the eye using computer vision, might be
applied to assess eye movements in other neurological diseases [
34
]. However, this would
require significant additional analysis to adapt the method to a higher requirement on data
acquisition, for example in application of assessment of saccades or smooth pursuit.
6. Conclusions
We have presented a hybrid algorithm to estimate the eye and lid movements of
patients with myasthenia gravis using the established MG core examination. Our hybrid
algorithm starts from a standard machine learning algorithm in order to obtain a coarse
estimate of the eye contour with a hexagon. This off the shelf deep learning library is not
trained for ptosis and diplopia evaluation that changes dramatically with eye position,
but it provides a coarse estimate on eye “contour” in a robust way. We then use a divide
and conquer technique to define smaller rectangles that separate each anatomic marker
of interest, such as upper lid, lower lid, and part of iris boundary, needed to estimate the
lid and eye position. Because the problem of segmentation becomes much simpler in each
local rectangle, we can use a standard image segmentation technique and check that the
segmentation passes several tests to eliminate artifacts due to head motion, poor lightning
condition, and lack of image resolution in a frame-by-frame manner.
We assessed a large, representative data set of image frames that our algorithm deliv-
ered with a two-pixel accuracy in about 80% of cases, which is adequate to filter outliers
and compute ptosis and eye alignment metrics. As opposed to the standard MG core
examination, our method provides the dynamic evolution of ptosis and ocular alignment
during the standard 60 s exercise of the standard MG core examination. This information is
usually not accessible quantitatively to the neurologist during a standard telemedicine visit,
and has proven to be more sensitive than the neurologist assessment on several occasion.
It should be noted that our method highlights a few open clinical questions: how do
we make use of the time dynamic evolution of ptosis and diplopia during the exercises? Is it
meaningful to capture the fatigue effect or is overall observation by the examiner adequate?
Our methods have limitations, and there is still much room for improvement. Devel-
opment of a model of the eye geometry with iris and pupil geometric markers that extend
the model of Figure 3in greater detail, including upper lid droop, is necessary. Applying
deep learning technology to this model would be quite feasible. This is certainly a worthy
effort, but it would require hundreds of patient videos with correct annotation to train
the algorithm [
19
]. However, given the expected expansion of telemedicine with ongoing
limitation of a patient’s technology and internet access, this is an important endeavor.
Sensors 2023,23, 7744 15 of 16
Further, as we have seen, deep learning technology may have spectacular robustness that
were shown in annotated videos but may not ensure accuracy.
We currently are developing a high-performance telehealth platform [
35
] that can be
conveniently distributed at multiple medical facilities in order to build the large, annotated
quality data set that may help advance our understating of MG.
7. Patents
A smart Cyber Infrastructure to enhance usability and quality of telehealth consul-
tation, M. Garbey, G. Joerger, provisional 63305420 filed by GWU, January 2022 and as a
PCT/US23/61783, January 2023.
Author Contributions:
This project is highly interdisciplinary and required all co-authors’ contri-
butions to establish the concept and reach the goal of the paper. All authors contributed to the
writing of the paper. M.G. led the project, designed and tested the different algorithms, including
the sensitivity analysis. G.J. participated in the validation of the method, creation of the prototype
and in the computer vision analysis. Q.L. implemented the algorithms on the platform and ran the
tests. Q.L., H.J.K., H.G., M.A.-R. and S.M. participated in the gathering and analysis of the data. All
authors have read and agreed to the published version of the manuscript.
Funding:
The work was partially supported by the MGNet a member of the Rare Disease Clinical
Research Network Consortium (RDCRN) NIH U54 NS115054. Funding support for the DMCC is
provided by the National Center for Advancing Translational Sciences (NCATS) and the National
Institute of Neurological Disorders and Stroke (NINDS).
Institutional Review Board Statement:
The George Washington University IRB provided review
and approved these investigations (IRB# NCR224008).
Informed Consent Statement:
All subjects were provided written informed consent via REDCap,
secure web-based system.
Data Availability Statement:
Study videos are recorded on Zoom software. The video recordings
are downloaded and stored in secure-HIPAA compliant dropbox and GW MFA servers. Study data
is entered into REDCap, a secure web-based data collection system. The REDCap system complies
with all applicable guidelines to ensure patient confidentiality, data integrity, and reliability.
Conflicts of Interest:
Henry J. Kaminski is a consultant for Roche, Cabeletta Bio, Lincoln Therapeutics,
Takeda and UCB Pharmaceuticals; and is CEO and CMO of ARC Biotechnology, LLC based on US
Patent 8,961,98. He is principal investigator of the Rare Disease Network for Myasthenia Gravis
(MGNet) National Institute of Neurological Disorders and Stroke, U54 NS115054 and Targeted
Therapy for Myasthenia Gravis. R41NS110331-01 to ARC Biotechnology.
References
1.
Giannotta, M.; Petrelli, C.; Pini, A. Telemedicine applied to neuromuscular disorders: Focus on the COVID-19 pandemic era. Acta
Myol. 2022,41, 30. [PubMed]
2.
Spina, E.; Trojsi, F.; Tozza, S.; Iovino, A.; Iodice, R.; Passaniti, C.; Abbadessa, G.; Bonavita, S.; Leocani, L.; Tedeschi, G.; et al. How
to manage with telemedicine people with neuromuscular diseases? Neurol. Sci. 2021,42, 3553–3559. [CrossRef] [PubMed]
3.
Hooshmand, S.; Cho, J.; Singh, S.; Govindarajan, R. Satisfaction of Telehealth in Patients with Established Neuromuscular
Disorders. Front. Neurol. 2021,12, 667813. [CrossRef] [PubMed]
4.
Ricciardi, D.; Casagrande, S.; Iodice, F.; Orlando, B.; Trojsi, F.; Cirillo, G.; Clerico, M.; Bozzali, M.; Leocani, L.; Abbadessa, G.; et al.
Myasthenia gravis and telemedicine: A lesson from COVID-19 pandemic. Neurol. Sci.
2021
,42, 4889–4892. [CrossRef] [PubMed]
5.
Guidon, A.; Muppidi, S.; Nowak, R.J.; Guptill, J.T.; Hehir, M.K.; Ruzhansky, K.; Burton, L.B.; Post, D.; Cutter, G.; Conwit, R.; et al.
Telemedicine visits in myasthenia gravis: Expert guidance and the myasthenia gravis core exam (MG-CE). Muscle Nerve
2021
,64,
270–276. [CrossRef] [PubMed]
6.
Thomsen, J.L.S.; Andersen, H. Outcome Measures in Clinical Trials of Patients with Myasthenia Gravis. Front. Neurol.
2020
,11,
596382. [CrossRef]
7. Al-Haida, M.; Benatar, M.; Kaminski, H.J. Ocular Myasthenia. Neurol. Clin. 2018,36, 241–251. [CrossRef]
8.
Liu, G.; Wei, Y.; Xie, Y.; Li, J.; Qiao, L.; Yang, J.-J. A computer-aided system for ocular myasthenia gravis diagnosis. Tsinghua Sci.
Technol. 2021,26, 749–758. [CrossRef]
Sensors 2023,23, 7744 16 of 16
9.
Tang, A.; Tam, R.; Cadrin-Chênevert, A.; Guest, W.; Chong, J.; Baefett, J.; Chepelev, L.; Cairns, R.; Mitchell, J.R.; Cicero, M.D.; et al.
Health Policy and Practice/Sante: Politique et pratique medicale, Canadian Association of Radiologists White Paper on Artificial
Intelligence in Radiology. Can. Assoc. Radiol. J. 2018,69, 120–135. [CrossRef]
10.
Leigh, R.J.; Zee, D.S. Contemporary Neurology Series. In The Neurology of Eye Movements, 5th ed.; Oxford University Press: New
York, NY, USA, 2015. [CrossRef]
11.
Crutcher, M.L.D.; Calhoun-Haney, R.; Manzanares, C.M.; Lah, J.J.; Levey, A.I.; Zola, P.S.M. Eye Tracking During a Visual Paired
Comparison Task as a Predictor of Early Dementia. Am. J. Alzheimer’s Dis. Other Dement. 2009,24, 258–266. [CrossRef]
12.
Hutton, J.T.; Nagel, J.A.; Loewenson, R.B. Eye tracking dysfunction in Alzheimer-type dementia. Neurology
1984
,34, 99. [CrossRef]
13.
Garbey, M.; Sun, N.; Merla, A.; Pavlidis, I. Contact-free measurement of cardiac pulse based on the analysis of thermal imagery.
IEEE Trans. Biomed. Eng. 2007,54, 1418–1426. [CrossRef]
14.
Burns, T.M.; Conaway, M.; Sanders, D.B. The MG Composite: A valid and reliable outcome measure for myasthenia gravis.
Neurology 2010,74, 1434–1440. [CrossRef] [PubMed]
15. Rynkiewicz, F.; Daszuta, M.; Napieralski, P. Pupil Detection Methods for Eye Tracking. J. Appl. Comput. Sci. 2018,26, 201–221.
16.
Hansen, D.W.; Ji, Q. In the eye of the beholder: A survey of models for eyes and gaze. IEEE Trans. Pattern Anal. Mach. Intell.
2010
,
32, 478–500. [PubMed]
17. Singh, H.; Singh, J. Human Eye Tracking and Related Issues: A Review. Int. J. Sci. Res. Publ. 2012,2, 1–9.
18.
Khan, W.; Hussain, A.; Kuru, K.; Al-Askar, H. Pupil Localisation and Eye Centre Estimation Using Machine Learning and
Computer Vision. Sensors 2020,20, 3785. [CrossRef] [PubMed]
19.
Zhao, L.; Wang, Z.; Zhang, G.; Qi, Y.; Wang, X. Eye state recognition based on deep integrated neural network and transfer
learning. Multimed. Tools Appl. 2017,77, 19415–19438. [CrossRef]
20.
Kunka, B.; Kostek, B. Non-Intrusive Infrared-Free Eye Tracking Method. In Proceedings of the Signal Processing Algorithms,
Architectures, Arrangements, and Applications Conference Proceedings (SPA), Poznan, Poland, 24–26 September 2009.
21.
Ghali, A.A.; Jamel, S.; Mohamad, K.M.; Yakub, N.A.; Deris, M.M. A Review of Iris Recogntion Algorithms. JOIV Int. J. Inform. Vis.
2017,1, 175–178.
22.
Toennies, K.; Behrens, F.; Aurnhammer, M. Feasibility of Hough-Transform-Based Iris Localization for Real-Time Application. In
Proceedings of the 16th International Conference on Pattern Recognition, Quebec City, QC, Canada, 11–15 August 2002; 2, pp.
1053–1056.
23.
Liang, D.B.B.; Houi, L.K. Non-Intrusive Eye Gaze Direction Tracking Using Color Segmentation and Hough Transform. In
Proceedings of the International Symposium on Communications and Information Technologies, Sydney, NSW, Australia, 17–19
October 2007; pp. 602–607.
24.
Verma, P.; Maheedhar Dubey, M.; Verma, P.; Basu, S. Daughman’s Algorithm Method for Iris Recognition—A Biometric Approach.
Int. J. Emerg. Technol. Adv. Eng. 2012,2, 177–185. Available online: www.ijetae.com (accessed on 15 April 2023).
25. Jain, V.; Learned-Miller, E. FDDB: A Benchmark for Face Detection in Unconstrained Settings; Tech. Rep. UM-CS-2010-009; UMASS:
Amherst, MA, USA, 2010; p. 11.
26.
Kabakus, A.T. An Experimental Performance Comparison of Widely Used Face Detection Tools. ADCAIJ Adv. Distrib. Comput.
Artif. Intell. J. 2019,8, 5–12. [CrossRef]
27.
OpenCV Haar Cascade Eye Detector [En ligne]. Available online: https://github.com/opencv/opencv/blob/master/data/
haarcascades/haarcascade_eye.xml (accessed on 30 March 2022).
28.
An, M.H.; You, S.C.; Park, R.W.; Lee, S. Using an Extended Technology Acceptance Model to Understand the Factors Influencing
Telehealth Utilization After Flattening the COVID-19 Curve in South Korea: Cross-sectional Survey Study. JMIR Med. Inform.
2021,9, e25435. [CrossRef] [PubMed]
29.
Johnston, B.; de Chazal, P. A review of image-based automatic facial landmark identification techniques. EURASIP J. Image Video
Process. 2018,2018, 86. [CrossRef]
30. Cao, X.; Wei, Y.; Wen, F.; Sun, J. Face Alignment by Explicit Shape Regression. Int. J. Comput. Vis. 2014,107, 177–190. [CrossRef]
31.
Kazemi, V.; Sullivan, J. One Millisecond Face Alignment with an Ensemble of Regression Trees. In Proceedings of the 2014 IEEE
Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1867–1874. [CrossRef]
32.
Sagonas, C.; Antonakos, E.; Tzimiropoulos, G.; Zafeiriou, S.; Pantic, M. 300 Faces In-The-Wild Challenge: Database and results.
Image Vis. Comput. 2016,47, 3–18. [CrossRef]
33.
Reports and Data. Myasthenia Gravis Market Size, Share, Industry Analysis by Treatment, By End-Use and Forecast to 2028.
Retrieved from BioSpace. Available online: https://www.biospace.com/article/myasthenia-gravis-market-size-share-industry-
analysis-by-treatment-by-end-use- and-forecast-to-2028/ (accessed on 3 January 2022).
34.
Mao, Y.; He, Y.; Liu, L.; Chen, X. Disease Classification Based on Eye Movement Features with Decision Tree and Random Forest.
Front. Neurosci.
2020
,14, 798. Available online: https://www.frontiersin.org/articles/10.3389/fnins.2020.00798 (accessed on 10
March 2023).
35. Care Constitution. Available online: https://careconstitution.com (accessed on 25 June 2021).
Disclaimer/Publisher’s Note:
The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
Article
Full-text available
Introduction COVID-19 pandemic radically transformed our daily clinical practice, raising the need not to lose close contact with patients without being able to see them face-to-face. These issues are even more felt and evident in fragile patients, as those affected by neuromuscular disease. An important help came from new digital technologies that allow clinicians to remotely monitor health status and any deterioration of chronically ill patients. Methods In this mini-review, an initiative of the “Digital Technologies, Web and Social Media Study Group” of the Italian Society of Neurology, we propose to analyze the approach to neuromuscular patients by looking over raising evidence on the main cornerstones of Telemedicine (TM): clinician-patient interaction, remote clinical assessment, remote monitoring, and digital therapeutics. In particular, we explored the strategies developed by researchers and their impact on the physical and emotional status of the patients, with particular focusing on their adherence to the program of virtual monitoring. Results TM plays an important role in each of four stages of approach to neuromuscular disease, having demonstrated validity in keep close clinical patient interaction, clinical assessment, remote monitoring, and telerehabilitation. Nevertheless, there is no remote alternative to electrophysiological testing neither validate tools to assess disability. Conclusion The role of TM in neuromuscular care is yet underestimated but is crucial, beyond the pandemic era. Further development of TM is advisable, through making specific apps, remotely controlled by clinicians, and making more engaging clinicians-patients interaction. Last, it is necessary to ensure adequate internet access to everyone.
Article
Full-text available
Introduction/aims: Determine established neuromuscular disease patients' satisfaction with telehealth during the COVID-19 pandemic. Methods: We received 50 completed Utah telehealth satisfaction surveys from a cohort of 90 from April 2020 to June 2020. Returning neuromuscular disease patients rated seven aspects from 1 (strongly disagree) to 5 (strongly agree): Communication, timeliness of physician, picture quality, sound quality, protection of privacy, the comfort of the physical exam, the ease of healthcare, and whether patients would prefer “in-person” visits despite safety precaution. A favorable response was defined as a response of “Agree” or “Strongly Agree” to the survey questions. An independent t-test, Fisher's or chi-square test were used to compare demographic factors on outcomes for each survey question. Results: The average age was 47.54 ± 20.63, 54% were female, 70% from rural areas, 60% had family present “webside,” and 14% had family present remotely. The majority of patients reported “Agree” or “Strongly Agree” to each survey question assessing their telehealth satisfaction, except for whether patients preferred in-person appointments. Demographic factors, including location and clinical diagnosis, did not influence survey responses. Discussion: The vast majority of established neuromuscular disease patients responded favorably to their telehealth experience during the COVID-19 pandemic.
Article
Full-text available
Background Although telehealth is considered a key component in combating the worldwide crisis caused by COVID-19, the factors that influence its acceptance by the general population after the flattening of the COVID-19 curve remain unclear. Objective We aimed to identify factors affecting telehealth acceptance, including anxiety related to COVID-19, after the initial rapid spread of the disease in South Korea. Methods We proposed an extended technology acceptance model (TAM) and performed a cross-sectional survey of individuals aged ≥30 years. In total, 471 usable responses were collected. Confirmatory factor analysis was used to examine the validity of measurements, and the partial least squares (PLS) method was used to investigate factors influencing telehealth acceptance and the impacts of COVID-19. Results PLS analysis showed that increased accessibility, enhanced care, and ease of telehealth use had positive effects on its perceived usefulness (P=.002, P<.001, and P<.001, respectively). Furthermore, perceived usefulness, ease, and privacy/discomfort significantly impacted the acceptance of telehealth (P<.001, P<.001, and P<.001, respectively). However, anxiety toward COVID-19 was not associated with telehealth acceptance (P=.112), and this insignificant relationship was consistent in the cluster (n=216, 46%) of respondents with chronic diseases (P=.185). Conclusions Increased accessibility, enhanced care, usefulness, ease of use, and privacy/discomfort are decisive variables affecting telehealth acceptance in the Korean general population, whereas anxiety about COVID-19 is not. This study may lead to a tailored promotion of telehealth after the pandemic subsides.
Article
Full-text available
Myasthenia gravis (MG) is a heterogeneous disorder whose clinical presentation ranges from mild ocular deficits to severe widespread weakness. This variance poses a challenge when quantifying clinical deficits. Deficits and symptoms are quantified using standardized clinical scales and questionnaires which are often used as outcome measures. The past decades have seen the development of several validated outcome measures in MG, which are used in clinical trials to obtain regulatory approval. In recent years, emphasis has moved from objective assessments to patient-reported outcomes. Despite a growing body of literature on the validity of the MG-specific outcome measures, several unresolved factors remain. As several novel therapeutics are currently in clinical development, knowledge about capabilities and limitations of outcome measures is needed. In the present paper, we describe the most widely used clinical classifications and scales in MG. We highlight the choice of outcome measures in published and ongoing trials, and we denote whether trial efficacy was reached on these outcomes. We discuss advantages and limitations of the individual scales, and discuss some of the unresolved factors relating to outcome assessments in MG.
Book
This new edition comprises a modern synthesis of the anatomical, physiological, and pharmacological substrate for eye movements, including current views on the reflexive and voluntary control of gaze. This synthesis is based on electrophysiological and inactivation studies in macaque, and behavioural studies in humans that incorporate functional imaging and transcranial magnetic stimulation (TMS) in normals, and clinicopathological studies in patients with neurological, visual, or vestibular disorders. Sophisticated experimental paradigms have been applied to both species to explore aspects of cognition, memory, volition, and reward. This large body of research has demonstrated the power of eye movements as experimental tools. The second part of this online resource applies this synthesis to the clinical and laboratory evaluation of patients with abnormal eye movements due to a broad range of disorders - from muscular dystrophy, and genetic disorders, to dementia, including visual and vestibular conditions.
Article
Neuromuscular diseases are rare and usually chronic progressive disorders that require a multidisciplinary clinical evaluation and functional monitoring. The patient-physician relationship and therapies are also key elements to be provided. The COVID-19 pandemic dramatically changed the way patients' health was managed and national health care services underwent a radical reorganization. Telemedicine, with the use of Information and Communication Technology (ICT) by health professionals, became the main strategy to ensure the continuation of care. However, the experience regarding the use of Telemedicine in neuromuscular disorders is very limited and the scientific literature is extremely scarce. From the first experiences in the '50s, the development of Telemedicine has been supplemented and supported by the implementation of ICT to guarantee the secure and effective transmission of medical data. Italian national guidelines (2010-2020) describe the technical and professional guarantees necessary to provide Telemedicine services. Nevertheless, at the time the pandemic appeared, no guidelines for clinical evaluation or for the administration of functional scales remotely were available for neuromuscular diseases. This has been a critical point when clinical evaluations were mandatory also for the renewal of drug prescriptions. However, the common opinion that telemedicine basic services were important to overcome the change in medical practice due to COVID-19 in neuromuscular diseases, even in pediatric age, emerged. Moreover, alternative digital modalities to evaluate patients at home in a kind of virtual clinic were considered as a field of future development.
Article
The current mode of clinical aided diagnosis of Ocular Myasthenia Gravis (OMG) is time-consuming and laborious, and it lacks quantitative standards. An aided diagnostic system for OMG is proposed to solve this problem. The values calculated by the system include three clinical indicators: eyelid distance, sclera distance, and palpebra superior fatigability test time. For the first two indicators, the semantic segmentation method was used to extract the pathological features of the patient's eye image and a semantic segmentation model was constructed. The patient eye image was divided into three regions: iris, sclera, and background. The indicators were calculated based on the position of the pixels in the segmentation mask. For the last indicator, a calculation method based on the Eyelid Aspect Ratio (EAR) is proposed; this method can better reflect the change of eyelid distance overtime. The system was evaluated based on the collected patient data. The results show that the segmentation model achieves a mean Intersection-Over-Union (mIoU) value of 86.05%. The paired-sample T-test was used to compare the results obtained by the system and doctors, and the p values were all greater than 0.05. Thus, the system can reduce the cost of clinical diagnosis and has high application value.
Article
COVID-19 pandemic has induced an urgent reorganization of the healthcare system to ensure continuity of care for patients affected by chronic neurological diseases including myasthenia gravis (MG). Due to the fluctuating nature of the disease, early detection of disease worsening, adverse events, and possibly life-threatening complications is mandatory. This work analyzes the main unresolved issues in the management of the myasthenic patient, the possibilities offered so far by digital technologies, and proposes an online evaluation protocol based on 4 simple tests to improve MG management. Telemedicine and Digital Technology might help neurologists in the clinical decision-making process of MG management, avoiding unnecessary in presence consultations and allowing a rational use of the time and space reduced by the pandemic.
Article
Introduction/aims: Telemedicine may be particularly well-suited for myasthenia gravis (MG) due to the disorder's need for specialized care, its hallmark fluctuating muscle weakness, and the potential for increased risk of virus exposure among patients with MG during the coronavirus disease 2019 (COVID-19) pandemic during in-person clinical visits. A disease-specific telemedicine physical examination to reflect myasthenic weakness does not currently exist. Methods: This paper outlines step-by-step guidance on the fundamentals of a telemedicine assessment for MG. The Myasthenia Gravis Core Exam (MG-CE) is introduced as a MG-specific, telemedicine, physical examination, which contains eight components (ptosis, diplopia, facial strength, bulbar strength, dysarthria, single breath count, arm strength, and sit to stand) and takes approximately 10 minutes to complete. Results: Pre-visit preparation, remote ascertainment of patient-reported outcome scales and visit documentation are also addressed. Discussion: Additional knowledge gaps in telemedicine specific to MG care are identified for future investigation.