Conference PaperPDF Available

Neural Network-Based Face Detection

Authors:

Abstract

We present a neural network-based upright frontal face detection system. A retinally connected neural network examines small windows of an image and decides whether each window contains a face. The system arbitrates between multiple networks to improve performance over a single network. We present a straightforward procedure for aligning positive face examples for training. To collect negative examples, we use a bootstrap algorithm, which adds false detections into the training set as training progresses. This eliminates the difficult task of manually selecting nonface training examples, which must be chosen to span the entire space of nonface images. Simple heuristics, such as using the fact that faces rarely overlap in images, can further improve the accuracy. Comparisons with several other state-of-the-art face detection systems are presented, showing that our system has comparable performance in terms of detection and false-positive rates.
A preview of the PDF is not available
... Deep learning algorithms were first designed to automatically learn hidden features and relationships in massive datasets, being used for regression, classification, and prediction calculations. As examples, Rowley used a deep learning method for face recognition in 1998 [7], which was introduced into medical diagnosis by Kononenko in 2001 [8]. Amongst the background of the recent developments in big data and artificial intelligence, as well as the significant improvement of computing power supported by GPU hardware, deep learning algorithms have developed rapidly since 2010. ...
Article
Full-text available
Simultaneous acquisition is a construction method that has been proposed in recent years to meet the requirements of ultra-large-scale and high-precision seismic exploration. Such method is highly efficient and can significantly reduce exploration costs by saving manpower and material resources, being extensively used in offshore exploration and several foreign seismic exploration projects. The data deblending step is a significant part of the operation of simultaneous acquisition, which directly affects the acquired data quality, and is a key factor for the success of oil and gas exploration. The simultaneous use of multiple seismic sources can generate blended noise with a random distribution in non-shot-gather datasets. However, the useful signal exhibits strong coherence, making it possible to separate the non-used wavefield from the blended data. Although the blended noise is randomly distributed in non-shot-gather datasets, it also has characteristics that are different from normal ambient noise, and its kinematic and dynamical characteristics are almost similar to the useful signal. As such, traditional filtering methods are not applicable, especially in the case of strong background noise. In the present study, simultaneous acquisition was introduced, the principle of data deblending using CNN was analyzed, and a data deblending method based on an improved version of GoogLeNet was established. The experimental results show that the trained network model could quickly and effectively separate the mixed wavefield from blended data, and achieve the expected useful signal.
... Deep learning is an important branch of machine learning, which is a computational model that imitates the structure of neurons and realizes the learning and understanding of data through neural networks [37]. Deep learning learns from massive amounts of data to determine the parameters in a computational model, resulting in mathematical expressions that can represent the data. ...
Article
Full-text available
With the opening of industrial networks in the information age, the characteristic of Industrial Control Protocols (ICPs) to transmit plaintext without encryption exposes serious security risks, threatening the safe and stable operation of Industrial Control Systems (ICSs). Exploring the work of mining vulnerabilities in ICPs can use fuzzing to mine potential vulnerabilities in protocols to ensure the safe operation of ICS. However, traditional fuzzing methods require the construction of test cases based on expert experience and the format syntax specification of ICPs. This process is time-consuming, labor-intensive, inefficient, and limited when facing unknown ICPs. In response to these issues, this paper proposes an automated fuzzing method for ICPs based on the Denoising Diffusion Probabilistic Model (DDPM). Specifically, DDPM achieves the conversion from noise to data samples, which can easily and quickly generate test cases. On this basis, we designed a universal fuzzing framework, DiffusionFuzz, that can be applied to most ICPs. The experimental results obtained on ICPs such as Modbus/TCP in the Industrial Attack-Defense Range of the Key Laboratory of Information Security for Petrochemical Industry in Liaoning Province demonstrate that the test cases generated by this method are diverse, and outperform other fuzzing methods in terms of acceptance rate and ability to trigger exceptions. Certainly, DiffusionFuzz can enhance the effectiveness of fuzzing, identify vulnerabilities in ICPs, and thereby reduce potential economic risks and impacts.
... Vaillant et al. [4] made groundbreaking strides in face detection by pioneering the use of neural networks, training a convolutional neural network, and employing a sliding window technique to locate faces in images. Likewise, in [5], a connected neural network approach was developed for face detection in images, making a notable contribution to the field. The progress in face detection has been further enhanced by the availability of publicly accessible benchmarks such as the Wilder Face-Face Detection Benchmark [6], PASCAL FACE [7], and Face Detection Database and Benchmark [8]. ...
Article
Full-text available
Computer vision is witnessing a surge of interest in machines accurately recognizing and interpreting human emotions through facial expression analysis. However, variations in image properties such as brightness, contrast, and resolution make it harder for models to predict the underlying emotion accurately. Utilizing a robust architecture of a convolutional neural network (CNN), we designed an efficacious framework for facial emotion recognition that predicts emotions and assigns corresponding probabilities to each fundamental human emotion. Each image is processed with various pre-processing steps before inputting it to the CNN to enhance the visibility and clarity of facial features, enabling the CNN to learn more effectively from the data. As CNNs entail a large amount of data for training, we used a data augmentation technique that helps to enhance the model's generalization capabilities, enabling it to effectively handle previously unseen data. To train the model, we joined the datasets, namely JAFFE and KDEF. We allocated 90% of the data for training, reserving the remaining 10% for testing purposes. The results of the CCN framework demonstrated a peak accuracy of 78.1%, which was achieved with the joint dataset. This accuracy indicated the model's capability to recognize facial emotions with a promising level of performance. Additionally, we developed an application with a graphical user interface for real-time facial emotion classification. This application allows users to classify emotions from still images and live video feeds, making it practical and user-friendly. The real-time application further demonstrates the system's practicality and potential for various real-world applications involving facial emotion analysis.
Article
Full-text available
One of the most difficult tasks in image processing is facial area detection. This study introduces a new face detection method. To improve detection rates, the system incorporates two facial detection algorithms. Gabor wavelets and neural networks are the two algorithms. Convolutional face images undergo initial transformation using Gabor wavelets, with 8 orientations and 5 scales chosen to extract the grey characteristics of the facial region. When added to the original photos, these 40 Gabor wavelets reveal the full extent of the response. We use a second feedforward neural network specifically designed for facial detection. The neural network is trained by backpropagation using the training set of faces and non-faces. Our experiments show that the suggested Gabor wavelet faces, when combined with the neural network feature space classifier, provide very respectable results. Comparing our proposed system to other face detection systems reveals that it performs better in terms of detection and false negative rates.
Article
Full-text available
Cracks are the primary indicator informing the structural health of concrete structures. Frequent inspection is essential for maintenance, and automatic crack inspection offers a significant advantage, given its efficiency and accuracy. Previously, image-based crack detection systems have been utilized for individual images, yet these systems are not effective for large inspection areas. This paper thereby proposes an image-based crack detection system using a Deep Convolution Neural Network (DCNN) to identify cracks in mosaic images composed from UAV photos of concrete footings. UAV images are transformed into 3D footing models, from which the composite images are created. The CNN model is trained on 224 ×\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document} 224 pixel patches, and training samples are augmented by various image transformation techniques. The proposed method is applied to localize cracks on composite images through the sliding window technique. The proposed VGG16 CNN detection system, with 95% detection accuracy, indicates superior performance to feature-based detection systems.
Article
A general purpose computer vision system must be capable of recognizing three-dimensional (3-D) objects. This paper proposes a precise definition of the 3-D object recognition problem, discusses basic concepts associated with this problem, and reviews the relevant literature. Because range images (or depth maps) are often used as sensor input instead of intensity images, techniques for obtaining, processing, and characterizing range data are also surveyed.
Article
Human face detection has always been an important problem for face, expression and gesture recognition. Though numerous attempts have been made to detect and localize faces, these approaches have made assumptions that restrict their extension to more general cases. We identify that the key factor in a generic and robust system is that of using a large amount of image evidence, related and reinforced by model knowledge through a probabilistic framework. In this paper, we propose a feature-based algorithm for detecting faces that is sufficiently generic and is also easily extensible to cope with more demanding variations of the imaging conditions. The algorithm detects feature points from the image using spatial filters and groups them into face candidates using geometric and gray level constraints. A probabilistic framework is then used to reinforce probabilities and to evaluate the likelihood of the candidate as a face. We provide results to support the validity of the approach and demonstrate its capability to detect faces under different scale, orientation and viewpoint.
Article
A novel neural network for distortion-invariant pattern recognition is described. Image regions of interest are determined using a detection stage, each region is then enhanced (the steps used are detailed), features are extracted (new Gabor wavelet features are used), and these features are used to classify the contents of each input region. A new feature space trajectory neural network (FST NN) classifier is used. A new 8 class database is used, a new multilayer NN to calculate the distance measures necessary is detailed, its low storage and on-line computational load requirements are noted. The ability of the adaptive FST algorithm to reduce network complexity while achieving excellent performance is demonstrated. The clutter rejection ability of this neural network to reject false alarm inputs is demonstrated, and time-history processing to further reduce false alarms is discussed. Hardware and commercial realizations are noted.
Article
An abstract is not available.
Article
The human face is a complex pattern. Finding human faces automatically in a scene is a difficult yet significant problem. It is the first important step in a fully automatic human face recognition system. In this paper a new method to locate human faces in a complex background is proposed. This system utilizes a hierarchical knowledge-based method and consists of three levels. The higher two levels are based on mosaic images at different resolutions. In the lower level, an improved edge detection method is proposed. In this research the problem of scale is dealt with, so that the system can locate unknown human faces spanning a wide range of sizes in a complex black-and-white picture. Some experimental results are given.