Fig 2 - uploaded by Jiaying Liu
Content may be subject to copyright.
Facial component decomposition. From left to right: a face image, detected landmarks and facial component regions 

Facial component decomposition. From left to right: a face image, detected landmarks and facial component regions 

Context in source publication

Context 1
... face is annotated by landmark points that locate facial components of interest. As shown in Figure 2, we concentrate on the eyes, nose and mouth regions. The patch features of four component regions are then extracted to form specialized dictionaries Y i , i ∈ {1, 2, 3, 4}. ...

Citations

... Jovanov et al. [73] proposes a time-of-flight depth camera-specific wavelet-based depth video denoising approach based on multi hypothesis motion estimation for facial depth maps. In [74] authors proposed a method and system for super-solving and recovering the facial depth maps. The main idea of this approach is to use a learning-based technique to gather reliable face priors from a high-quality facial depth map to improve the depth images. ...
Article
Full-text available
This article contains all of the information needed to conduct a study on monocular facial depth estimation problems. A brief literature review and applications on facial depth map research were offered first, followed by a comprehensive evaluation of publicly available facial depth datasets and widely used loss functions. The key properties and characteristics of each facial depth map dataset are described and evaluated. Furthermore, facial depth maps loss functions are briefly discussed, which will make it easier to train neural facial depth models on a variety of datasets for both short- and long-range depth maps. The network’s design and components are essential, but its effectiveness is largely determined by how it is trained, which necessitates a large dataset and a suitable loss function. Implementation details of how neural depth networks work and their corresponding evaluation matrices are presented and explained. In addition, an SoA neural model for facial depth estimation is proposed, along with a detailed comparison evaluation and, where feasible, direct comparison of facial depth estimation methods to serve as a foundation for a proposed model that is utilized. The model employed shows better performance compared with current state-of-the-art methods when tested across four datasets. The new loss function used in the proposed method helps the network to learn the facial regions resulting in an accurate depth prediction. The network is trained on synthetic human facial depth datasets whereas for validation purposes real as well as synthetic facial images are used. The results prove that the trained network outperforms current state-of-the-art networks performances, thus setting up a new baseline method for facial depth estimations.