Fig 3 - uploaded by Stan Z Li
Content may be subject to copyright.
Color images taken by a color camera versus NIR images taken by the present NIR imaging system. While unfavorable lighting is obvious in the color face images, it is almost unseen in the NIR face images. 

Color images taken by a color camera versus NIR images taken by the present NIR imaging system. While unfavorable lighting is obvious in the color face images, it is almost unseen in the NIR face images. 

Source publication
Article
Full-text available
Most current face recognition systems are designed for indoor, cooperative-user applications. However, even in thus-constrained applications, most existing systems, academic and commercial, are compromised in accuracy by changes in environmental illumination. In this paper, we present a novel solution for illumination invariant face recognition for...

Context in source publication

Context 1
... spectral measurements of facial skin sampled at some facial points are used for face recognition; they are shown to differ significantly from person to person. Further investigations of using NIR images for face localization and recognition are found in [10], [44], [45]. All of those works use the “bright pupil” effect, namely, specular reflection of active NIR lights on pupils to detect eyes in NIR images. In addition, Zhao and Grigat’s system [44] uses DCT coefficients as features and an SVM as the classifier. Zou et al.’s work [45] derives their matching methods based on an LDA transform and shows that the NIR illuminated faces are better separable than faces under varying ambient illumination. The method of using the “bright pupil” effect to detect eyes has a serious drawback that limits its applications. It assumes that the “bright pupils” are present in the eyes and can be detected using an algorithm. However, the assumption can be invalidated, for example, when there are specular reflections of NIR lights on the eyeglasses, with “narrow eyes” where eyelids may occlude “bright pupils,” when an eye is closed, or when the eyes are looking aside (see examples in Fig. 7). These happen for most of the face images with eyeglasses and also for a significant percentage of face images without eyeglasses. Therefore, we have considered eye detection a main problem to be solved in face recognition using active NIR images. The goal of making this special-purpose hardware is to overcome the problem arising from uncontrolled environmental lights so as to produce face images of a good illumination condition for face recognition. By “a good illumination condition,” we mean that the lighting is from the frontal direction and the image has suitable pixel intensities, i.e., having good contrast and not saturated. We propose two strategies to control the light direction: 1) Mount active lights on the camera to provide frontal lighting and 2) minimize environmental lighting. We set two requirements on the active lighting: 1) The lights should be strong enough to produce a clear frontal-lighted face image without causing disturbance to human eyes and 2) the minimization of the environmental lighting should have minimum reduction of the intended active lighting. Radiation spectrum ranges are shown in Fig. 1. While far (thermal) infrared imaging reflects heat radiation, NIR imaging is more like normal visible light imaging, though NIR is invisible to the naked eyes. Ultraviolet radiation is harmful to the human body and cannot be used for face recognition applications. Our solution for requirement 1, is to choose the active lights in the near infrared (NIR) spectrum between 780-1,100 nm and mount them on the camera. We use NIR light- emitting diodes (LEDs) as active radiation sources, which are strong enough for indoor use and are power-effective. A convenient wavelength is 850 nm. Such NIR lights are almost invisible to the human eye, yet most CCD and CMOS sensors have sufficient response at this spectrum point. When mounted on the camera, the LEDs are approximately coaxial to the camera direction and, thus, provide the best possible straight frontal lighting, better than mounting anywhere else; moreover, when the LEDs and camera are together, control of the lights can be easier using a circuit in the box. The geometric layout of the LEDs on the camera panel may be carefully designed such that the illumination on the face is as homogeneous as possible. The strength of the total LED lighting should be such that it results in the NIR face images with good S/N ratio when the camera-face distance is between 50-100 cm, a convenient range for the user. A guideline is that it should be as strong as possible, at least stronger than expected environmental illumination, yet not cause sensor saturation. A concern is the safety of human eyes. When the sensor working in the normal mode is not saturated, the safety is guaranteed. Our solution for requirement 2, above is to use a long pass optical filter to cut off visible light while allowing NIR light to pass. We choose a filter such that ray passing rates are 0, 50, 88, and 99 percent at the wavelength points of 720, 800, 850, and 880 nm, respectively. The filter cuts off visible environmental lights (< 700 nm) while allowing most of the 850nm NIR light to pass. Fig. 2 illustrates a design of the hardware device and its relationship with the face. The device consists of 18 NIR LEDs, an NIR camera, a color camera, and the box. The NIR LEDs and camera are for NIR face image acquisition. The color camera capture color face images may be used for fusion with the NIR images or for other purposes. The hardware and the face are relatively positioned in such a way that the lighting is frontal and NIR rays provide nearly homogenous illumination on face. The imaging hardware works at a rate of 30 frames per second with the USB 2.0 protocol for 640 Â 480 images and costs less than 20 US dollars. Fig. 3 shows example images of a face illuminated by NIR LED lights from the front, a lamp aside and environmental lights. We can see the following: 1) The lighting conditions are likely to cause problems for face recognition with the color images. 2) The NIR images, with the visible light composition cut off by the filter, are mostly frontal-lighted by the NIR lights, with minimum influence from the side lighting, and provide a good basis for face recognition. In visible light images, an intrapersonal change due to different lighting directions could be larger than an extrapersonal change under similar lighting conditions. This can be illustrated by an analysis on correlation coefficients and matching scores, shown in Fig. 4. There are seven pairs of face images; each pair is taken of the same person’s face but illuminated by a visible light lamp from left and right, respectively. The correlation table shows intrapersonal and extrapersonal correlation coefficients. There, the diagonal entries (in bold font) are for the intrapersonal pairs (e.g., entry ð i; i Þ is the correlation between the two images in column i ); the upper triangles are for the extrapersonal right-right pairs (e.g., entry ð i; j Þ is the correlation between the two images in column i and j in the first row); the lower triangle entries are for the extrapersonal left-left pairs (e.g., entry ð i; j Þ is the correlation between the two images in column i and j in the second row). We see that the correlation coefficients between two images of faces illuminated from left and right are all negative numbers regardless of the face identity; those between two images under similar lighting directions can be either positive or negative. The mean and variance are -0.4738 and 0.1015 for the intrapersonal pairs and 0.0327 and 0.5932 for the extrapersonal pairs. Therefore, it is not a surprise that a PCA matching engine has no chance of making correct matches in this case. The score table is produced by an advanced matching engine trained using AdaBoost with LBP features on a visible light face image training set (the AdaBoost-trained matching engine produces better results than trained using LDA). The mean and variance of the scores are 0.4808 and 0.0238 for intrapersonal pairs and 0.5694 and 0.0359 for extrapersonal pairs. We see that the scores for intrapersonal pairs under different lighting directions are generally lower than those for extrapersonal pairs under similar lighting directions. This means that reliable recognition cannot be achieved with visible light images, even using the advanced matching engine. The impact of environmental lighting is much reduced by the present NIR imaging system, as shown by the correlations and scores in Fig. 5. There, the corresponding NIR face images are taken in the same visible light conditions as in Fig. 4 and the explanations of the two tables are similar to those for Fig. 4. The correlation coefficients between NIR images are all positive, regardless of the visible light conditions and person identity. They have mean and variance of 0.8785 and 0.1502 for intrapersonal pairs and 0.6806 and 0.0830 for extrapersonal pairs. However, the intrapersonal correlation coefficients may not necessarily be higher than the extrapersonal ones, meaning possible recognition errors, even with the NIR images. Therefore, a better matching engine than correlation or PCA is still needed for highly accurate face recognition, even with NIR face images. The score table shows the matching score produced by an LBP+AdaBoost classifier trained on active NIR images. The mean and variance of the scores are 0.6750 and 0.0296 for intrapersonal pairs and 0.3113 and 0.0358 for extrapersonal pairs. By examining all of the entries, we can see that the intrapersonal scores are consistently much higher than the extrapersonal ones. The above case studies suggest that the proposed active NIR imaging system with an advanced LBP+AdaBoost recognition engine can yield the highest recognition performance of all of the schemes. In this section, we first provide an analysis from the Lambertian imaging model to show that such images contain the most relevant, intrinsic information about a face, subject only to a multiplying constant or a monotonic transform due to lighting intensity changes. We then present an LBP-based representation to amend the degree of freedom of the monotonic transform to achieve an illumination invariant representation of faces for indoor face recognition applications. According to the Lambertian model, an image I ð x; y Þ under a point light source is formed according to the following: where ð x; y Þ is the albedo of the facial surface material at point ð x; y Þ , n 1⁄4 ð n x ; n y ; n z Þ is the surface normal (a unit row vector) in the 3D space, and s 1⁄4 ð s x ; s y ; s z Þ is the lighting direction (a column vector, with magnitude). Here, albedo ð x; y Þ reflects the photometric properties of facial skin and hairs; n ð ...

Citations

... CBSR NIR. It was a near-IR face dataset collected by the Chinese Academy of Sciences [136]. It contained 3,940 near-IR face images of 197 people. ...
Article
Full-text available
Recent years have witnessed the increasing popularity and dramatic progress of contactless sensing technologies, which are able to conduct remote signal acquisition without body contact. Both the physical signs and the physiological parameters can be acquired with contactless sensing. This paper introduces popular contactless sensing technologies, explores their application scenarios, and delves into the underlying theoretical principles. It comprehensively reviews the open datasets released in this field, encompassing collection scenarios, sample counts, data formats, and volunteer information. The performance baseline, typical work, and accessible links are also furnished. In addition, it includes discussions on the primary challenges and potential solutions in the context of contactless sensing with open datasets. Finally, suggestions for establishing a high-quality dataset are also given to the community.
... Most conventional image-based algorithms have an excellent performance in terms of accuracy when the face image is recorded under controlled conditions. However, these methods fail when presented with images captured under ✓ RGB-D Face (VAP) [69] ✓ FRGCv2 [12] ✓ BU-3DFE [15] ✓ BU-4DFE [17] ✓ 3D-TEC [87] ✓ UMB-DB [88] ✓ SuperFaces [89] ✓ ✓ CurtainFaces [70], [90] ✓ FaceWarehouse [71] ✓ Lock3DFace [43] ✓ IIIT-D RGB-D [72] ✓ UHDB11 [91] ✓ CAS(ME) 3 [92] ✓ 3DMAD [34], [35] ✓ CASIA-SURF [79] ✓ WMCA [74] ✓ HQ-WMCA [93] ✓ CASIA-SURF CeFA [94] ✓ ND-2006 [95] ✓ GavabDB [13] ✓ UoY [96] ✓ BJUT-3D [19] ✓ FRAV3D [14] ✓ Pandora [47] ✓ MPIBC [6] ✓ BIWI [97] ✓ ICT-3DHP [98] ✓ KaspaAROV [82] ✓ ✓ HRRFaceD [83] ✓ IST-EURECOM LFFD [46] ✓ FaceVerse-Detailed [99] ✓ FaceVerse-Coarse [99] ✓ Cui et al. [100] ✓ ESRC3D [49] ✓ SURREY [101] ✓ JNU [101] ✓ 3DWF (2019) [102] ✓ Intellifusion [103], [104] Li et al. [105] ✓ SeetaFace [106] ✓ MotorMark [107] ✓ Sun et al. [108] ✓ IAS-Lab [109] ✓ RGBDFaces [110] ✓ MICC (Florence2D/3D) [111] ✓ Face-Emotion [112] ✓ Florence3D-Re-Id [84] ✓ IKFDB [113] ✓ MMFD [114] ✓ RGB-D-T [115] ✓ FIDENTIS [50] ✓ FaceScape [116] ✓ CASIA HFB [21] ✓ 4DFAB [51] ✓ ✓ ND-Collection-D [10] ✓ 3DFACE-XMU [117] ✓ ZJU-3DFED [118] ✓ FSU [119] ✓ B3D(AC) [25] ✓ ✓ D3DFACS [30] ✓ Hi4D-ADSIP [31] ✓ ADSIP [22] ✓ MAVFER [56] ✓ HeadSpace [57] ✓ MeIn3D [45] ✓ Tuft [58] ✓ UHDB31 [48] ✓ Bechman [120] ✓ Eurocom [53] ✓ ✓ ✓ Msspoof [126] ✓ ✓ SWIR [127] ✓ ✓ ✓ BRSU [128] ✓ ✓ EMSPAD [129] ✓ MLFP [130] ✓ ✓ CASIA-SURF [79] ✓ CIGIT-PPM [131] ✓ ✓ PolyU-HSFD [24] ✓ CMU-HSFD [8] ✓ ND-Collection-C [9] ✓ ✓ ND-NIVL [44] ✓ ✓ CASIA HFB [21] ✓ ✓ CASIA NIR-VIS [32] ✓ ✓ LDHF-DB [36] ✓ ✓ NFRAD [29] ✓ ✓ PolyU-NIRFD [132] ✓ ✓ NVIE [26] ✓ ✓ Liu et al. [42] ✓ ✓ IRIS [133] ✓ ✓ UH [134] ✓ Carl [27] ✓ ✓ ARL-MMFD1 [135] ✓ ✓ ARL-MMFD2 [136] ✓ ✓ UL-FMTV [52] ✓ Eurocom [53] ✓ ✓ Tuft [58] ✓ ✓ ✓ Sejong-A [59] ✓ ✓ ✓ Sejong-B [59] ✓ ✓ ✓ Sober Drunk [38], [39] ✓ PUCV-DTF [54] ✓ TFW [137] ✓ ✓ SpeakingFaces [138] ✓ ✓ KTFE [41] ✓ ✓ NIST/Equinox [139] ✓ ✓ SDFD [55] ✓ CBSR-NIR [140] ✓ ✓ RWTH [141] ✓ UNCC-ThermalFace [60] ✓ IRIS-M3 [16] ✓ UWA-HSFD [142] ✓ an uncontrolled environment with high distortions resulting from changes in illumination. A nighttime situation is an example of a condition where human recognition, based exclusively on visible spectrum pictures, may be impractical. ...
... It was captured employing a Merlin-Uncooled camera in 2002, yielding 2,492 frontal long-wave infrared (LWIR) thermal images sourced from 241 individuals. In 2007, the CSBR-NIR [140] data set was primarily designed to achieve illumination-invariant face verification. This data set encompasses a total of 3,940 near-infrared (NIR) facial images, featuring 197 individuals. ...
Article
Full-text available
Due to their ease-of-use, biometric verification methods to control access to digital devices have become ubiquitous. Many rely on supervised machine learning, a process that is notoriously data-hungry. At the same time, biometric data is sensitive from a privacy perspective, and a comprehensive review from a data set perspective is lacking. In this survey, we present a comprehensive review of multimodal face data sets (e.g., data sets containing RGB color plus other channels such as infrared or depth). This follows a trend in both industry and academia to use such additional modalities to improve the robustness and reliability of the resulting biometric verification systems. Furthermore, such data sets open the path to a plethora of additional applications, such as 3D face reconstruction (e.g., to create avatars for VR and AR environments), face detection, registration, alignment, and recognition systems, emotion detection, anti-spoofing, etc. We also provide information regarding the data acquisition setup and data attributes (ethnicities, poses, facial expressions, age, population size, etc.) as well as a thorough discussion of related applications and state-of-the-art benchmarking. Readers may thus use this survey as a tool to navigate the existing data sets both from the application and data set perspective. To existing surveys we contribute, to the best of our knowledge, the first exhaustive review of multimodalities in these data sets.
... No obstante, la bondad de estos sistemas es dependiente de las condiciones lumínicas exteriores, las cuales ven su efecto paliado con el uso de marcadores infrarrojos (S. Z. Li et al., 2007;Murphy-Chutorian y Trivedi, 2008;Feldstein et al., 2015), siendo esta una solución barata y sencilla. En Murphy-Chutorian y Trivedi (2009) se plantean unos criterios de diseño para sistemas de detección del movimiento de la cabeza en base a una extensa revisión bibliográfica, entre los que se destaca la invariabilidad ante diferentes condiciones lumínicas. ...
Thesis
Full-text available
El factor humano en conducción presenta numerosos desafíos relativos a la seguridad, los cuales podrían ser abordados eficientemente a través de tecnologías relacionadas con la conducción autónoma. En los últimos años ha habido un avance significativo en los sistemas de asistencia al conductor, proporcionando funciones de apoyo y mejorando la seguridad y la comodidad en carretera. En relación con la automatización total, los fabricantes de automóviles han emprendido una carrera tecnológica en busca del vehículo sin conductor invirtiendo recursos significativos en investigación y desarrollo. Sin embargo, existen ciertas barreras que ralentizan la integración de estos vehículos en el parque automovilístico actual. Uno de los factores más determinantes es la aceptación social, condicionada directamente por la confiabilidad de estos sistemas. A pesar de las múltiples pruebas en entornos cerrados y el aumento de sensores, es difícil abarcar el total de la casuística de accidentes de tráfico que se pueden producir en tráfico real. Muchos de los problemas detectados en el ámbito de la conducción autónoma se relacionan con problemas que un conductor humano podría resolver con relativa sencillez, apuntando a una falta de reglas en el sistema de decisión. En este aspecto, los estudios naturalistas desempeñan un papel fundamental en el desarrollo de algoritmos de toma de decisiones basados en el comportamiento humano, ya que los vehículos carecen de cierta información que los conductores adquieren de forma natural. Es por ello que el estudio del comportamiento del conductor es crucial para el desarrollo de sistemas que interactúen con vehículos de conducción manual. Comprender los procesos cognitivos seguidos por un conductor y su estado en diversos entornos perfeccionará el diseño de las reglas de decisión ante diferentes maniobras, optimizando la toma de decisiones en conducción autónoma. El objetivo principal de la tesis es mejorar la caracterización del comportamiento del conductor mediante el análisis de la percepción visual en maniobras complejas realizadas en vías de alta capacidad, como son autovías o autopistas. A lo largo de este estudio, se evalúa la influencia de las variables atencionales del conductor ante diferentes niveles de asistencia a la conducción, observando la repetición de ciertos patrones visuales en función del entorno. La integración de la información visual del conductor en un modelo de toma de decisiones naturalista permitió una validación exitosa del mismo con ensayos experimentales realizados en tráfico real. Previamente, se realizó una fusión sensorial del sistema de percepción del entorno con el sistema de seguimiento visual, permitiendo la proyección automática de la mirada del conductor en el entorno exterior. Los desarrollos realizados generaron adicionalmente conocimiento destacable en relación con la anticipación de la maniobra de cambio de carril y el hueco aceptable para el desarrollo de modelos de conducción. Las conclusiones de la Tesis doctoral contribuyen a una mejora de la modelización del comportamiento del conductor y aportarán un enfoque más naturalista al desarrollo de algoritmos de toma de decisiones, con el objetivo de mejorar la integración de la conducción autónoma en el tráfico mixto.
... The task in HFR is to facilitate cross-domain matching while overcoming the challenges posed by the domain gap. lance cameras, provide superior performance regardless of lighting conditions and are robust against presentation attacks [26,11]. However, training a face recognition system for NIR images will require a large amount of labeled training data which is often not available. ...
Preprint
Full-text available
Heterogeneous Face Recognition (HFR) aims to match face images across different domains, such as thermal and visible spectra, expanding the applicability of Face Recognition (FR) systems to challenging scenarios. However, the domain gap and limited availability of large-scale datasets in the target domain make training robust and invariant HFR models from scratch difficult. In this work, we treat different modalities as distinct styles and propose a framework to adapt feature maps, bridging the domain gap. We introduce a novel Conditional Adaptive Instance Modulation (CAIM) module that can be integrated into pre-trained FR networks, transforming them into HFR networks. The CAIM block modulates intermediate feature maps, to adapt the style of the target modality effectively bridging the domain gap. Our proposed method allows for end-to-end training with a minimal number of paired samples. We extensively evaluate our approach on multiple challenging benchmarks, demonstrating superior performance compared to state-of-the-art methods. The source code and protocols for reproducing the findings will be made publicly available.
... The next unique dataset is CASIA NIR (Li et al, 2007). CASIA NIR (Near-Infrared) is a facial recognition dataset specifically focused on images captured in the near-infrared spectrum. ...
Preprint
Full-text available
This paper presents a comprehensive overview of Convolutional Neural Networks (CNNs) in the context of face recognition. By analyzing 150 research papers, we investigate major publication channels, temporal trends, prevalent CNN models, primary data sets, accuracy levels, primary research focuses, and future prospects for improving CNN-based facial recognition. A major focus is placed on identifying prevalent CNN architectures, techniques used for facial recognition and shedding light on the evolving landscape of CNN designs. Furthermore, we examine the datasets used for training and testing CNNs, and evaluate the accuracy levels achieved by these models. Lastly, we discuss future directions for enhancing CNN-based facial recognition, including addressing bias and fairness, improving robustness to environmental variations, privacy preservation, and exploring transfer learning and multimodal fusion. This paper serves as a valuable resource, summarizing major trends in CNN-based face recognition. It provides insights for researchers and practitioners, guiding future advancements in this rapidly evolving field.
... The IR spectrum, lying just outside the visible light band at 700nm-2000nm, has been of particular interest in the face recognition field. Li et al. 19 showed that effects of lighting such as direction, intensity, and shadows can change the appearance of a face. Unlike traditional visible-light images, IR images exhibit improved illumination invariance, reducing such effects. ...
... In fact, most CCD (charge-coupled device) and CMOS sensors found in digital cameras can detect rays at the NIR (near-IR) spectrum range (700nm-1000nm). 19 Typically, an IR cutoff filter is used to block these components in standard cameras. Researchers have posited that exploiting this extended spectrum could improve face recognition of HPS individuals. ...
... The difference between the effects of shaded and unshaded conditions could not be discerned in our study. However, a previous study, Li et al., 19 showed that the type of illumination used could affect the face recognition system's performance. We found that in outdoor lighting even indirect sunlight is too bright to make a significant difference. ...
Article
Full-text available
Face recognition is widely used for security and access control. Its performance is limited when working with highly pigmented skin tones due to training bias caused by the under-representation of darker-skinned individuals in existing datasets and the fact that darker skin absorbs more light and therefore reflects less discernible detail in the visible spectrum. To improve performance, this work incorporated the infrared (IR) spectrum, which is perceived by electronic sensors. We augmented existing datasets with images of highly pigmented individuals captured using the visible, IR, and full spectra and fine-tuned existing face recognition systems to compare the performance of these three. We found a marked improvement in accuracy and AUC values of the receiver operating characteristic (ROC) curves when including the IR spectrum, increasing performance from 97.5% to 99.0% for highly pigmented faces. Different facial orientations and narrow cropping also improved performance, and the nose region was the most important feature for recognition.
... Planning aesthetic dental treatment has encouraged the creation of a three-dimensional (3D) virtual patient, spurred facial smile analysis and simulation and digital smile design, allowed assessment of clinical outcomes and perhaps most importantly, enhanced coordinated clinician, patient and dental lab communication [1][2][3]. Conventional two-dimensional (2D) photogrammetric facial images can be highly erroneous due to deviations in lighting positioning, parallax, and focal length distances [4], and these are now being superseded by images acquired by 3D imaging technology, such as structured light scanning (SL), laser beam scanning (LB), near infra-red scanning (NIR), and stereophotogrammetry (SP) [5][6][7][8][9][10]. These imaging technologies can easily acquire 3D data and when combined with computer-aided design (CAD) software; create a model corresponding to the facial surface, from which surface measurements can be effectively analyzed. ...
Article
Full-text available
Statement of problem: The development of facial scanners has improved capabilities to create three-dimensional (3D) virtual patients for accurate facial and smile analysis. However, most of these scanners are expensive, stationary and involve a significant clinical footprint. The use of the Apple iPhone and its integrated "TrueDepth" near-infrared (NIR) scanner combined with an image processing application (app) offers the potential to capture and analyze the unique 3D nature of the face; the accuracy and reliability of which are yet to be established for use in clinical dentistry. Purpose: This study was designed to validate both the trueness and precision of the iPhone 11 Pro smartphone TrueDepth NIR scanner in conjunction with the Bellus3D Face app in capturing 3D facial images in a sample of adult participants in comparison to the conventional 3dMDface stereophotogrammetry system. Material and methods: Twenty-nine adult participants were prospectively recruited. Eighteen soft tissue landmarks were marked on each participant's face before imaging. 3D facial images were captured using a 3dMDface system and the Apple iPhone TrueDepth NIR scanner combined with the Bellus3D Face app respectively. The best fit of each experimental model to the 3dMD scan was analyzed using Geomagic Control X software. The root mean square (RMS) was used to measure the "trueness" as the absolute deviation of each TrueDepth scan from the reference 3dMD image. Individual facial landmark deviations were also assessed to evaluate the reliability in different craniofacial regions. The "precision" of the smartphone was tested by taking 10 consecutive scans of the same subject and comparing those to the reference scan. Intra-observer and inter-observer reliabilities were assessed using the intra-class correlation coefficient (ICC). Results: Relative to the 3dMDface system, the mean RMS difference of the iPhone/Bellus3D app was 0.86 ± 0.31 mm. 97% of all the landmarks were within 2 mm of error compared with the reference data. The ICC for intra-observer reproducibility or precision of the iPhone/Bellus3D app was 0.96, which was classified as excellent. The ICC for inter-observer reliability was 0.84, which was classified as good. Conclusions: These results suggest that 3D facial images acquired with this system, the iPhone TrueDepth NIR camera in conjunction with the Bellus3D Face app, are clinically accurate and reliable. Judicious use is advised in clinical situations that require high degrees of detail due to a lack of image resolution and a longer acquisition time. Generally, this system possesses the potential to serve as a practical alternative to conventional stereophotogrammetry systems for use in a clinical setting due to its accessibility and relative ease of use and further research is planned to appraise its updated clinical use.
... The experimental datasets are separated into real face datasets and fake face datasets. The real face datasets are comprised of the CBSR NIR Face Dataset [23], CASIA NIR-VIS 2.0 Face Database [24] and Oulu-CASIA [25], whereas the fake face datasets are constructed in-house.We create Dataset-A, Dataset-B and Dataset-C for training and testing using the aforementioned datasets. The size of each image is 300 × 300. ...
Article
Full-text available
Face recognition is a prevalent identity verification method, but it requires a liveness detection system to guard against face fraud from printed images and mobile phone photographs. It is challenging to guarantee low complexity and high accuracy of the face anti-spoofing model applied on the Android development board at the same time in existing research. On the Android development board, we construct a liveness detection system that is resistant to attacks from printed images and photographs of electronic devices such as tablets. In conjunction with the actual circumstance, we create a lightweight liveness detection algorithm based on near-infrared images. We utilize MTCNN to clip the near-infrared images in order to preserve the facial portions and reduce the calculated cost. After applying Gamma correction to weaken the impact of illumination on the faces, the facial images are subsequently incorporated into the enhanced ShuffleNet V2 model based on the MBConv and squeeze-excitation (SE) modules for secondary classification. We evaluate the performance of the enhanced model on the CBSR NIR Face Dataset, CASIA NIR-VIS 2.0 Face Database and Oulu-CASIA, achieving 98.50%, 99.87% and 100% accuracy, which is superior to the performance of the original ShuffleNet V2 model and state-of-the-art methods. Our model’s FLOPs and Params are 0.28 G and 2.17 M, respectively. Meanwhile, the liveness detection system based on Android has an exceptionally high level of security performance
... One of the advantages is local feature-based approaches are robust to local changes such as occlusion, expression, and pose variations. From the existing literature on local feature-based approaches, one approach is Local Binary Patterns (LBPs) [20]. In LBP, the neighboring pixel changes due to the central pixel. ...
Article
Full-text available
Near sets (also called Descriptively Near Sets) classify nonempty sets of objects based on object feature values. The Near Set Theory provides a framework for measuring the similarity of objects based on features that describe them in much the same way humans perceive the similarity of objects. This paper presents a novel approach for face recognition using Near Set Theory that takes into account variations in facial features due to varying facial expressions, and facial plastic surgery. In the proposed work, we demonstrate two-fold usage of Near set theory; firstly, Near Set Theory as a feature selector to select the plastic surgery facial features with the help of tolerance classes, and secondly, Near Set Theory as a recognizer that uses selected prominent intrinsic facial features which are automatically extracted through the deep learning model. Extensive experimentation was performed on various facial datasets such as YALE, PSD, and ASPS. Experimentation demonstrates 93% of accuracy on the YALE face dataset, 98% of accuracy on the PSD dataset, and 98% of accuracy on the ASPS dataset. A detailed comparative analysis of the proposed work of facial resemblance with other state-of-the-art algorithms is presented in this paper. The experimentation results effectively classify face resemblance using Near Set Theory, which has outperformed several state-of-the-art classification approaches.
... This vector provides an efficient representation of the face which is used to calculate the similarity between images. In the work of Li et al. [28], the authors presented two statistical learning methods for face recognition invariant to indoor lighting using NIR images. With the goal of building face recognition classifiers from a variety of LBP features, they used LDA [29] and AdaBoost [30] to achieve a high accuracy face recognition engine. ...
Article
Full-text available
Besides the many advances made in the facial detection and recognition fields, face recognition applied to visual images (VIS-FR) has received increasing interest in recent years, especially in the field of communication, identity authentication, public safety and to address the risk of terrorism and crime. These systems however encounter important problems in the presence of variations in pose, expression, age, occlusion, disguise, and lighting as these factors significantly reduce the recognition accuracy. To prevent problems in the visible spectrum, several researchers have recommended the use of infrared images. This paper provides an updated overview of deep infrared (IR) approaches in face recognition (FR) and analysis. First, we present the most widely used databases, both public and private, and the various metrics and loss functions that have been proposed and used in deep infrared techniques. We then review deep face analysis and recognition/identification methods proposed in recent years. In this review, we show that infrared techniques have given interesting results for face recognition, solving some of the problems encountered with visible spectrum techniques. We finally identify some weaknesses of current infrared FR approaches as well as many future research directions to address the IR FR limitations.