This paper presents a high-speed and robust dual-band infrared thermal camera based on an ARM CPU.The system consists of a low-resolution long-wavelength infrared detector,a digital temperature and humid⁃ity sensor,an...This paper presents a high-speed and robust dual-band infrared thermal camera based on an ARM CPU.The system consists of a low-resolution long-wavelength infrared detector,a digital temperature and humid⁃ity sensor,and a CMOS sensor.In view of the significant contrast between face and background in thermal infra⁃red images,this paper explores a suitable accuracy-latency tradeoff for thermal face detection and proposes a tiny,lightweight detector named YOLO-Fastest-IR.Four YOLO-Fastest-IR models(IR0 to IR3)with different scales are designed based on YOLO-Fastest.To train and evaluate these lightweight models,a multi-user low-resolution thermal face database(RGBT-MLTF)was collected,and the four networks were trained.Experiments demon⁃strate that the lightweight convolutional neural network performs well in thermal infrared face detection tasks.The proposed algorithm outperforms existing face detection methods in both positioning accuracy and speed,making it more suitable for deployment on mobile platforms or embedded devices.After obtaining the region of interest(ROI)in the infrared(IR)image,the RGB camera is guided by the thermal infrared face detection results to achieve fine positioning of the RGB face.Experimental results show that YOLO-Fastest-IR achieves a frame rate of 92.9 FPS on a Raspberry Pi 4B and successfully detects 97.4%of faces in the RGBT-MLTF test set.Ultimate⁃ly,an infrared temperature measurement system with low cost,strong robustness,and high real-time perfor⁃mance was integrated,achieving a temperature measurement accuracy of 0.3℃.展开更多
Detecting faces under occlusion remains a significant challenge in computer vision due to variations caused by masks,sunglasses,and other obstructions.Addressing this issue is crucial for applications such as surveill...Detecting faces under occlusion remains a significant challenge in computer vision due to variations caused by masks,sunglasses,and other obstructions.Addressing this issue is crucial for applications such as surveillance,biometric authentication,and human-computer interaction.This paper provides a comprehensive review of face detection techniques developed to handle occluded faces.Studies are categorized into four main approaches:feature-based,machine learning-based,deep learning-based,and hybrid methods.We analyzed state-of-the-art studies within each category,examining their methodologies,strengths,and limitations based on widely used benchmark datasets,highlighting their adaptability to partial and severe occlusions.The review also identifies key challenges,including dataset diversity,model generalization,and computational efficiency.Our findings reveal that deep learning methods dominate recent studies,benefiting from their ability to extract hierarchical features and handle complex occlusion patterns.More recently,researchers have increasingly explored Transformer-based architectures,such as Vision Transformer(ViT)and Swin Transformer,to further improve detection robustness under challenging occlusion scenarios.In addition,hybrid approaches,which aim to combine traditional andmodern techniques,are emerging as a promising direction for improving robustness.This review provides valuable insights for researchers aiming to develop more robust face detection systems and for practitioners seeking to deploy reliable solutions in real-world,occlusionprone environments.Further improvements and the proposal of broader datasets are required to developmore scalable,robust,and efficient models that can handle complex occlusions in real-world scenarios.展开更多
Face detection is a critical component inmodern security,surveillance,and human-computer interaction systems,with widespread applications in smartphones,biometric access control,and public monitoring.However,detecting...Face detection is a critical component inmodern security,surveillance,and human-computer interaction systems,with widespread applications in smartphones,biometric access control,and public monitoring.However,detecting faces with high levels of occlusion,such as those covered by masks,veils,or scarves,remains a significant challenge,as traditional models often fail to generalize under such conditions.This paper presents a hybrid approach that combines traditional handcrafted feature extraction technique called Histogram of Oriented Gradients(HOG)and Canny edge detection with modern deep learning models.The goal is to improve face detection accuracy under occlusions.The proposed method leverages the structural strengths of HOG and edge-based object proposals while exploiting the feature extraction capabilities of Convolutional Neural Networks(CNNs).The effectiveness of the proposed model is assessed using a custom dataset containing 10,000 heavily occluded face images and a subset of the Common Objects in Context(COCO)dataset for non-face samples.The COCO dataset was selected for its variety and realism in background contexts.Experimental evaluations demonstrate significant performance improvements compared to baseline CNN models.Results indicate that DenseNet121 combined with HOG outperforms other counterparts in classification metrics with an F1-score of 87.96%and precision of 88.02%.Enhanced performance is achieved through reduced false positives and improved localization accuracy with the integration of object proposals based on Canny and contour detection.While the proposed method increases inference time from 33.52 to 97.80 ms,it achieves a notable improvement in precision from 80.85% to 88.02% when comparing the baseline DenseNet121 model to its hybrid counterpart.Limitations of the method include higher computational cost and the need for careful tuning of parameters across the edge detection,handcrafted features,and CNN components.These findings highlight the potential of combining handcrafted and learned features for occluded face detection tasks.展开更多
Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensi...Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.展开更多
Although important progresses have been already made in face detection,many false faces can be found in detection results and false detection rate is influenced by some factors,such as rotation and tilt of human face,...Although important progresses have been already made in face detection,many false faces can be found in detection results and false detection rate is influenced by some factors,such as rotation and tilt of human face,complicated background,illumination,scale,cloak and hairstyle.This paper proposes a new method called DP-Adaboost algorithm to detect multi-angle human face and improve the correct detection rate.An improved Adaboost algorithm with the fusion of frontal face classifier and a profile face classifier is used to detect the multi-angle face.An improved horizontal differential projection algorithm is put forward to remove those non-face images among the preliminary detection results from the improved Adaboost algorithm.Experiment results show that compared with the classical Adaboost algorithm with a frontal face classifier,the textual DP-Adaboost algorithm can reduce false rate significantly and improve hit rate in multi-angle face detection.展开更多
For face detection under complex background and illumination, a detection method that combines the skin color segmentation and cost-sensitive Adaboost algorithm is proposed in this paper. First, by using the character...For face detection under complex background and illumination, a detection method that combines the skin color segmentation and cost-sensitive Adaboost algorithm is proposed in this paper. First, by using the characteristic of human skin color clustering in the color space, the skin color area in YC b C r color space is extracted and a large number of irrelevant backgrounds are excluded; then for remedying the deficiencies of Adaboost algorithm, the cost-sensitive function is introduced into the Adaboost algorithm; finally the skin color segmentation and cost-sensitive Adaboost algorithm are combined for the face detection. Experimental results show that the proposed detection method has a higher detection rate and detection speed, which can more adapt to the actual field environment.展开更多
In recent years,face detection has attracted much attention and achieved great progress due to its extensively practical applications in the field of face based computer vision.However,the tradeoff between accuracy an...In recent years,face detection has attracted much attention and achieved great progress due to its extensively practical applications in the field of face based computer vision.However,the tradeoff between accuracy and efficiency of the face detectors still needs to be further studied.In this paper,using Darknet-53 as backbone,we propose an improved YOLOv3-attention model by introducing attention mechanism and data augmentation to obtain the robust face detector with high accuracy and efficiency.The attention mechanism is introduced to enhance much higher discrimination of the deep features,and the trick of data augmentation is used in the training procedure to achieve higher detection accuracy without significantly affecting the inference speed.The model has been trained and evaluated on the popular and challenging face detection benchmark,i.e.,the WIDER FACE training and validation subsets,respectively,achieving AP of 0.942,0.919 and 0.821 with the speed of 28FPS.This performance exceeds some existing SOTA algorithms,demonstrating acceptable accuracy and near real time detection for VGA resolution images,even in the complex scenarios.In addition,the proposed model shows good generation ability on another public dataset FDDB.The results indicate the proposed model is a promising face detector with high efficiency and accuracy in the wild.展开更多
Face recognition technology automatically identifies an individual from image or video sources.The detection process can be done by attaining facial characteristics from the image of a subject face.Recent developments...Face recognition technology automatically identifies an individual from image or video sources.The detection process can be done by attaining facial characteristics from the image of a subject face.Recent developments in deep learning(DL)and computer vision(CV)techniques enable the design of automated face recognition and tracking methods.This study presents a novel Harris Hawks Optimization with deep learning-empowered automated face detection and tracking(HHODL-AFDT)method.The proposed HHODL-AFDT model involves a Faster region based convolution neural network(RCNN)-based face detection model and HHO-based hyperparameter opti-mization process.The presented optimal Faster RCNN model precisely rec-ognizes the face and is passed into the face-tracking model using a regression network(REGN).The face tracking using the REGN model uses the fea-tures from neighboring frames and foresees the location of the target face in succeeding frames.The application of the HHO algorithm for optimal hyperparameter selection shows the novelty of the work.The experimental validation of the presented HHODL-AFDT algorithm is conducted using two datasets and the experiment outcomes highlighted the superior performance of the HHODL-AFDT model over current methodologies with maximum accuracy of 90.60%and 88.08%under PICS and VTB datasets,respectively.展开更多
A new kind of region pair grey difference classifier was proposed. The regions in pairs associated to form a feature were not necessarily directly-connected, but were selected dedicatedly to the grey transition betwee...A new kind of region pair grey difference classifier was proposed. The regions in pairs associated to form a feature were not necessarily directly-connected, but were selected dedicatedly to the grey transition between regions coinciding with the face pattern structure. Fifteen brighter and darker region pairs were chosen to form the region pair grey difference features with high discriminant capabilities. Instead of using both false acceptance rate and false rejection rate, the mutual information was used as a unified metric for evaluating the classifying performance. The parameters of specified positions, areas and grey difference bias for each single region pair feature were selected by an optimization processing aiming at maximizing the mutual information between the region pair feature and classifying distribution, respectively. An additional region-based feature depicting the correlation between global region grey intensity patterns was also proposed. Compared with the result of Viola-like approach using over 2 000 features, the proposed approach can achieve similar error rates with only 16 features and 1/6 implementation time on controlled illumination images.展开更多
Biometric applications widely use the face as a component for recognition and automatic detection.Face rotation is a variable component and makes face detection a complex and challenging task with varied angles and ro...Biometric applications widely use the face as a component for recognition and automatic detection.Face rotation is a variable component and makes face detection a complex and challenging task with varied angles and rotation.This problem has been investigated,and a novice algorithm,namely RIFDS(Rotation Invariant Face Detection System),has been devised.The objective of the paper is to implement a robust method for face detection taken at various angle.Further to achieve better results than known algorithms for face detection.In RIFDS Polar Harmonic Transforms(PHT)technique is combined with Multi-Block Local Binary Pattern(MBLBP)in a hybrid manner.The MBLBP is used to extract texture patterns from the digital image,and the PHT is used to manage invariant rotation characteristics.In this manner,RIFDS can detect human faces at different rotations and with different facial expressions.The RIFDS performance is validated on different face databases like LFW,ORL,CMU,MIT-CBCL,JAFFF Face Databases,and Lena images.The results show that the RIFDS algorithm can detect faces at varying angles and at different image resolutions and with an accuracy of 99.9%.The RIFDS algorithm outperforms previous methods like Viola-Jones,Multi-blockLocal Binary Pattern(MBLBP),and Polar HarmonicTransforms(PHTs).The RIFDS approach has a further scope with a genetic algorithm to detect faces(approximation)even from shadows.展开更多
Security access control systems and automatic video surveillance systems are becoming increasingly important recently,and detecting human faces is one of the indispensable processes.In this paper,an approach is presen...Security access control systems and automatic video surveillance systems are becoming increasingly important recently,and detecting human faces is one of the indispensable processes.In this paper,an approach is presented to detect faces in video surveillance.Firstly,both the skin-color and motion components are applied to extract skin-like regions.The skin-color segmentation algorithm is based on the BPNN (back-error-propagation neural network) and the motion component is obtained with frame difference algorithm.Secondly,the image is clustered into separated face candidates by using the region growing technique.Finally,the face candidates are further verified by the rule-based algorithm.Experiment results demonstrate that both the accuracy and processing speed are very promising and the approach can be applied for the practical use.展开更多
The intelligent environment needs Human-Computer Interactive technology (HCI) and a projector projects screen on wall in the intelligent environments. We propose the front-face detection from four captured images re...The intelligent environment needs Human-Computer Interactive technology (HCI) and a projector projects screen on wall in the intelligent environments. We propose the front-face detection from four captured images related to the intelligent room for the deaf. Our proposal purpose is that a deaf user faces wall displaying everywhere. system gets the images from four cameras, and detects the user region from a silhouette image using a different method, detects and cuts a motion body region from a different image, and cuts the vertexchest region from the cut body region image. The system attempts to find front-face using Haar-like feature, and selects a detected front-face image from the vertex-chest region. We estimate the front-face detection of recognition rate, which shows somewhat successfully.展开更多
The spread of social media has increased contacts of members of communities on the lntemet. Members of these communities often use account names instead of real names. When they meet in the real world, they will find ...The spread of social media has increased contacts of members of communities on the lntemet. Members of these communities often use account names instead of real names. When they meet in the real world, they will find it useful to have a tool that enables them to associate the faces in fiont of them with the account names they know. This paper proposes a method that enables a person to identify the account name of the person ("target") in front of him/her using a smartphone. The attendees to a meeting exchange their identifiers (i.e., the account name) and GPS information using smartphones. When the user points his/her smartphone towards a target, the target's identifier is displayed near the target's head on the camera screen using AR (augmented reality). The position where the identifier is displayed is calculated from the differences in longitude and latitude between the user and the target and the azimuth direction of the target from the user. The target is identified based on this information, the face detection coordinates, and the distance between the two. The proposed method has been implemented using Android terminals, and identification accuracy has been examined through experiments.展开更多
Face detection is applied to many tasks such as auto focus control, surveillance, user interface, and face recognition. Processing speed and detection accuracy of the face detection have been improved continuously. Th...Face detection is applied to many tasks such as auto focus control, surveillance, user interface, and face recognition. Processing speed and detection accuracy of the face detection have been improved continuously. This paper describes a novel method of fast face detection with multi-scale window search free from image resizing. We adopt statistics of gradient images (SGI) as image features and append an overlapping cell array to improve detection accuracy. The SGI feature is scale invariant and insensitive to small difference of pixel value. These characteristics enable the multi-scale window search without image resizing. Experimental results show that processing speed of our method is 3.66 times faster than a conventional method, adopting HOG features combined to an SVM classifier, without accuracy degradation.展开更多
One being developed automatic sweep robot, need to estimate if anyone is on a certain range of road ahead then automatically adjust running speed, in order to ensure work efficiency and operation safety. This paper pr...One being developed automatic sweep robot, need to estimate if anyone is on a certain range of road ahead then automatically adjust running speed, in order to ensure work efficiency and operation safety. This paper proposed a method using face detection to predict the data of image sensor. The experimental results show that, the proposed algorithm is practical and reliable, and good outcome have been achieved in the application of instruction robot.展开更多
Background Several face detection and recogni tion methods have been proposed in the past decades that have excellent performance.The conventional face recognition pipeline comprises the following:(1)face detection,(2...Background Several face detection and recogni tion methods have been proposed in the past decades that have excellent performance.The conventional face recognition pipeline comprises the following:(1)face detection,(2)face alignment,(3)feature extraction,and(4)similarity,which are independent of each other.The separate facial analysis stages lead to redundant model calculations,and are difficult for use in end-to-end training.Methods In this paper,we propose a novel end-to-end trainable convolutional network framework for face detection and recognition,in which a geometric transformation matrix is directly learned to align the faces rather than predicting the facial landmarks.In the training stage,our single CNN model is supervised only by face bounding boxes and personal identities,which are publicly available from WIDER FACE and CASIA-WebFace datasets.Our model is tested on Face Detection Dataset and Benchmark(FDDB)and Labeled Face in the Wild(LFW)datasets.Results The results show 89.24%recall for face detection tasks and 98.63%accura cy for face recognition tasks.展开更多
This paper proposes a universal framework,termed as Multi-Task Hybrid Convolutional Neural Network(MHCNN),for joint face detection,facial landmark detection,facial quality,and facial attribute analysis.MHCNN consists ...This paper proposes a universal framework,termed as Multi-Task Hybrid Convolutional Neural Network(MHCNN),for joint face detection,facial landmark detection,facial quality,and facial attribute analysis.MHCNN consists of a high-accuracy single stage detector(SSD)and an efficient tiny convolutional neural network(T-CNN)for joint face detection refinement,alignment and attribute analysis.Though the SSD face detectors achieve promising results,we find that applying a tiny CNN on detections further boosts the detected face scores and bounding boxes.By multi-task training,our T-CNN aims to provide five facial landmarks,facial quality scores,and facial attributes like wearing sunglasses and wearing masks.Since there is no public facial quality data and facial attribute data as we need,we contribute two datasets,namely FaceQ and FaceA,which are collected from the Internet.Experiments show that our MHCNN achieves face detection performance comparable to the state of the art in face detection data set and benchmark(FDDB),and gets reasonable results on AFLW,FaceQ and FaceA.展开更多
This paper presents a method which utilizes color, local symmetry and geometry information of human face based on various models. The algorithm first detects most likely face regions or ROIs (Region-Of-Interest) from ...This paper presents a method which utilizes color, local symmetry and geometry information of human face based on various models. The algorithm first detects most likely face regions or ROIs (Region-Of-Interest) from the image using face color model and face outline model, produces a face color similarity map. Then it performs local symmetry detection within these ROIs to obtain a local symmetry similarity map. The two maps and local similarity map are fused to obtain potential facial feature points. Finally similarity matching is performed to identify faces between the fusion map and face geometry model under affine transformation. The output results are the detected faces with confidence values. The experimental results demonstrate its validity and robustness to identify faces under certain variations.展开更多
This paper provides efficient and robust algorithms for real-time face detection and recognition in complex backgrounds. The algorithms are implemented using a series of signal processing methods including Ada Boost, ...This paper provides efficient and robust algorithms for real-time face detection and recognition in complex backgrounds. The algorithms are implemented using a series of signal processing methods including Ada Boost, cascade classifier, Local Binary Pattern (LBP), Haar-like feature, facial image pre-processing and Principal Component Analysis (PCA). The Ada Boost algorithm is implemented in a cascade classifier to train the face and eye detectors with robust detection accuracy. The LBP descriptor is utilized to extract facial features for fast face detection. The eye detection algorithm reduces the false face detection rate. The detected facial image is then processed to correct the orientation and increase the contrast, therefore, maintains high facial recognition accuracy. Finally, the PCA algorithm is used to recognize faces efficiently. Large databases with faces and non-faces images are used to train and validate face detection and facial recognition algorithms. The algorithms achieve an overall true-positive rate of 98.8% for face detection and 99.2% for correct facial recognition.展开更多
Face detection is considered as a challenging problem in the field of image analysis and computer vision. There are many researches in this area, but because of its importance, it needs to be further developed. Succes...Face detection is considered as a challenging problem in the field of image analysis and computer vision. There are many researches in this area, but because of its importance, it needs to be further developed. Successive Mean Quantization Transform (SMQT) for illumination and sensor insensitive operation and Sparse Network of Winnow (SNoW) to speed up the original classifier based face detection technique presented such a good result. In this paper we use the Mean of Medians of CbCr (MMCbCr) color correction approach to enhance the combined SMQT features and SNoW classifier face detection technique. The proposed technique is applied on color images gathered from various sources such as Internet, and Georgia Database. Experimental results show that the face detection performance of the proposed method is more effective and accurate compared to SFSC method.展开更多
基金Supported by the Fundamental Research Funds for the Central Universities(2024300443)the Natural Science Foundation of Jiangsu Province(BK20241224).
文摘This paper presents a high-speed and robust dual-band infrared thermal camera based on an ARM CPU.The system consists of a low-resolution long-wavelength infrared detector,a digital temperature and humid⁃ity sensor,and a CMOS sensor.In view of the significant contrast between face and background in thermal infra⁃red images,this paper explores a suitable accuracy-latency tradeoff for thermal face detection and proposes a tiny,lightweight detector named YOLO-Fastest-IR.Four YOLO-Fastest-IR models(IR0 to IR3)with different scales are designed based on YOLO-Fastest.To train and evaluate these lightweight models,a multi-user low-resolution thermal face database(RGBT-MLTF)was collected,and the four networks were trained.Experiments demon⁃strate that the lightweight convolutional neural network performs well in thermal infrared face detection tasks.The proposed algorithm outperforms existing face detection methods in both positioning accuracy and speed,making it more suitable for deployment on mobile platforms or embedded devices.After obtaining the region of interest(ROI)in the infrared(IR)image,the RGB camera is guided by the thermal infrared face detection results to achieve fine positioning of the RGB face.Experimental results show that YOLO-Fastest-IR achieves a frame rate of 92.9 FPS on a Raspberry Pi 4B and successfully detects 97.4%of faces in the RGBT-MLTF test set.Ultimate⁃ly,an infrared temperature measurement system with low cost,strong robustness,and high real-time perfor⁃mance was integrated,achieving a temperature measurement accuracy of 0.3℃.
基金funded by A’Sharqiyah University,Sultanate of Oman,under Research Project grant number(BFP/RGP/ICT/22/490).
文摘Detecting faces under occlusion remains a significant challenge in computer vision due to variations caused by masks,sunglasses,and other obstructions.Addressing this issue is crucial for applications such as surveillance,biometric authentication,and human-computer interaction.This paper provides a comprehensive review of face detection techniques developed to handle occluded faces.Studies are categorized into four main approaches:feature-based,machine learning-based,deep learning-based,and hybrid methods.We analyzed state-of-the-art studies within each category,examining their methodologies,strengths,and limitations based on widely used benchmark datasets,highlighting their adaptability to partial and severe occlusions.The review also identifies key challenges,including dataset diversity,model generalization,and computational efficiency.Our findings reveal that deep learning methods dominate recent studies,benefiting from their ability to extract hierarchical features and handle complex occlusion patterns.More recently,researchers have increasingly explored Transformer-based architectures,such as Vision Transformer(ViT)and Swin Transformer,to further improve detection robustness under challenging occlusion scenarios.In addition,hybrid approaches,which aim to combine traditional andmodern techniques,are emerging as a promising direction for improving robustness.This review provides valuable insights for researchers aiming to develop more robust face detection systems and for practitioners seeking to deploy reliable solutions in real-world,occlusionprone environments.Further improvements and the proposal of broader datasets are required to developmore scalable,robust,and efficient models that can handle complex occlusions in real-world scenarios.
基金funded by A’Sharqiyah University,Sultanate of Oman,under Research Project Grant Number(BFP/RGP/ICT/22/490).
文摘Face detection is a critical component inmodern security,surveillance,and human-computer interaction systems,with widespread applications in smartphones,biometric access control,and public monitoring.However,detecting faces with high levels of occlusion,such as those covered by masks,veils,or scarves,remains a significant challenge,as traditional models often fail to generalize under such conditions.This paper presents a hybrid approach that combines traditional handcrafted feature extraction technique called Histogram of Oriented Gradients(HOG)and Canny edge detection with modern deep learning models.The goal is to improve face detection accuracy under occlusions.The proposed method leverages the structural strengths of HOG and edge-based object proposals while exploiting the feature extraction capabilities of Convolutional Neural Networks(CNNs).The effectiveness of the proposed model is assessed using a custom dataset containing 10,000 heavily occluded face images and a subset of the Common Objects in Context(COCO)dataset for non-face samples.The COCO dataset was selected for its variety and realism in background contexts.Experimental evaluations demonstrate significant performance improvements compared to baseline CNN models.Results indicate that DenseNet121 combined with HOG outperforms other counterparts in classification metrics with an F1-score of 87.96%and precision of 88.02%.Enhanced performance is achieved through reduced false positives and improved localization accuracy with the integration of object proposals based on Canny and contour detection.While the proposed method increases inference time from 33.52 to 97.80 ms,it achieves a notable improvement in precision from 80.85% to 88.02% when comparing the baseline DenseNet121 model to its hybrid counterpart.Limitations of the method include higher computational cost and the need for careful tuning of parameters across the edge detection,handcrafted features,and CNN components.These findings highlight the potential of combining handcrafted and learned features for occluded face detection tasks.
文摘Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.
文摘Although important progresses have been already made in face detection,many false faces can be found in detection results and false detection rate is influenced by some factors,such as rotation and tilt of human face,complicated background,illumination,scale,cloak and hairstyle.This paper proposes a new method called DP-Adaboost algorithm to detect multi-angle human face and improve the correct detection rate.An improved Adaboost algorithm with the fusion of frontal face classifier and a profile face classifier is used to detect the multi-angle face.An improved horizontal differential projection algorithm is put forward to remove those non-face images among the preliminary detection results from the improved Adaboost algorithm.Experiment results show that compared with the classical Adaboost algorithm with a frontal face classifier,the textual DP-Adaboost algorithm can reduce false rate significantly and improve hit rate in multi-angle face detection.
基金supported by the National Basic Research Program of China(973 Program)under Grant No.2012CB215202the National Natural Science Foundation of China under Grant No.51205046
文摘For face detection under complex background and illumination, a detection method that combines the skin color segmentation and cost-sensitive Adaboost algorithm is proposed in this paper. First, by using the characteristic of human skin color clustering in the color space, the skin color area in YC b C r color space is extracted and a large number of irrelevant backgrounds are excluded; then for remedying the deficiencies of Adaboost algorithm, the cost-sensitive function is introduced into the Adaboost algorithm; finally the skin color segmentation and cost-sensitive Adaboost algorithm are combined for the face detection. Experimental results show that the proposed detection method has a higher detection rate and detection speed, which can more adapt to the actual field environment.
基金This work is partially supported by National Key Research and Development Project(Grant No.A19808)Fundamental Research Funds for the Central Universities(Grant No.2019JKF225).
文摘In recent years,face detection has attracted much attention and achieved great progress due to its extensively practical applications in the field of face based computer vision.However,the tradeoff between accuracy and efficiency of the face detectors still needs to be further studied.In this paper,using Darknet-53 as backbone,we propose an improved YOLOv3-attention model by introducing attention mechanism and data augmentation to obtain the robust face detector with high accuracy and efficiency.The attention mechanism is introduced to enhance much higher discrimination of the deep features,and the trick of data augmentation is used in the training procedure to achieve higher detection accuracy without significantly affecting the inference speed.The model has been trained and evaluated on the popular and challenging face detection benchmark,i.e.,the WIDER FACE training and validation subsets,respectively,achieving AP of 0.942,0.919 and 0.821 with the speed of 28FPS.This performance exceeds some existing SOTA algorithms,demonstrating acceptable accuracy and near real time detection for VGA resolution images,even in the complex scenarios.In addition,the proposed model shows good generation ability on another public dataset FDDB.The results indicate the proposed model is a promising face detector with high efficiency and accuracy in the wild.
基金Princess Nourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2023R349)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.This study is supported via funding from Prince Sattam bin Abdulaziz University Project Number(PSAU/2023/R/1444).
文摘Face recognition technology automatically identifies an individual from image or video sources.The detection process can be done by attaining facial characteristics from the image of a subject face.Recent developments in deep learning(DL)and computer vision(CV)techniques enable the design of automated face recognition and tracking methods.This study presents a novel Harris Hawks Optimization with deep learning-empowered automated face detection and tracking(HHODL-AFDT)method.The proposed HHODL-AFDT model involves a Faster region based convolution neural network(RCNN)-based face detection model and HHO-based hyperparameter opti-mization process.The presented optimal Faster RCNN model precisely rec-ognizes the face and is passed into the face-tracking model using a regression network(REGN).The face tracking using the REGN model uses the fea-tures from neighboring frames and foresees the location of the target face in succeeding frames.The application of the HHO algorithm for optimal hyperparameter selection shows the novelty of the work.The experimental validation of the presented HHODL-AFDT algorithm is conducted using two datasets and the experiment outcomes highlighted the superior performance of the HHODL-AFDT model over current methodologies with maximum accuracy of 90.60%and 88.08%under PICS and VTB datasets,respectively.
基金Supported by the Joint Research Funds of Dalian University of Technology and Shenyang Automation Institute,Chinese Academy of Sciences
文摘A new kind of region pair grey difference classifier was proposed. The regions in pairs associated to form a feature were not necessarily directly-connected, but were selected dedicatedly to the grey transition between regions coinciding with the face pattern structure. Fifteen brighter and darker region pairs were chosen to form the region pair grey difference features with high discriminant capabilities. Instead of using both false acceptance rate and false rejection rate, the mutual information was used as a unified metric for evaluating the classifying performance. The parameters of specified positions, areas and grey difference bias for each single region pair feature were selected by an optimization processing aiming at maximizing the mutual information between the region pair feature and classifying distribution, respectively. An additional region-based feature depicting the correlation between global region grey intensity patterns was also proposed. Compared with the result of Viola-like approach using over 2 000 features, the proposed approach can achieve similar error rates with only 16 features and 1/6 implementation time on controlled illumination images.
基金The authors would like to thank the Deanship of Scientific Research at Majmaah University for supporting this work under Project Number No-R-2021-154.
文摘Biometric applications widely use the face as a component for recognition and automatic detection.Face rotation is a variable component and makes face detection a complex and challenging task with varied angles and rotation.This problem has been investigated,and a novice algorithm,namely RIFDS(Rotation Invariant Face Detection System),has been devised.The objective of the paper is to implement a robust method for face detection taken at various angle.Further to achieve better results than known algorithms for face detection.In RIFDS Polar Harmonic Transforms(PHT)technique is combined with Multi-Block Local Binary Pattern(MBLBP)in a hybrid manner.The MBLBP is used to extract texture patterns from the digital image,and the PHT is used to manage invariant rotation characteristics.In this manner,RIFDS can detect human faces at different rotations and with different facial expressions.The RIFDS performance is validated on different face databases like LFW,ORL,CMU,MIT-CBCL,JAFFF Face Databases,and Lena images.The results show that the RIFDS algorithm can detect faces at varying angles and at different image resolutions and with an accuracy of 99.9%.The RIFDS algorithm outperforms previous methods like Viola-Jones,Multi-blockLocal Binary Pattern(MBLBP),and Polar HarmonicTransforms(PHTs).The RIFDS approach has a further scope with a genetic algorithm to detect faces(approximation)even from shadows.
基金This work is supported by the National Natural Science
文摘Security access control systems and automatic video surveillance systems are becoming increasingly important recently,and detecting human faces is one of the indispensable processes.In this paper,an approach is presented to detect faces in video surveillance.Firstly,both the skin-color and motion components are applied to extract skin-like regions.The skin-color segmentation algorithm is based on the BPNN (back-error-propagation neural network) and the motion component is obtained with frame difference algorithm.Secondly,the image is clustered into separated face candidates by using the region growing technique.Finally,the face candidates are further verified by the rule-based algorithm.Experiment results demonstrate that both the accuracy and processing speed are very promising and the approach can be applied for the practical use.
基金supported by the Ministry of Knowledge Economy,Korea,the ITRC(Information Technology Research Center)support program(NIA-2009-(C1090-0902-0007))the Contents Technology Research Center support program
文摘The intelligent environment needs Human-Computer Interactive technology (HCI) and a projector projects screen on wall in the intelligent environments. We propose the front-face detection from four captured images related to the intelligent room for the deaf. Our proposal purpose is that a deaf user faces wall displaying everywhere. system gets the images from four cameras, and detects the user region from a silhouette image using a different method, detects and cuts a motion body region from a different image, and cuts the vertexchest region from the cut body region image. The system attempts to find front-face using Haar-like feature, and selects a detected front-face image from the vertex-chest region. We estimate the front-face detection of recognition rate, which shows somewhat successfully.
文摘The spread of social media has increased contacts of members of communities on the lntemet. Members of these communities often use account names instead of real names. When they meet in the real world, they will find it useful to have a tool that enables them to associate the faces in fiont of them with the account names they know. This paper proposes a method that enables a person to identify the account name of the person ("target") in front of him/her using a smartphone. The attendees to a meeting exchange their identifiers (i.e., the account name) and GPS information using smartphones. When the user points his/her smartphone towards a target, the target's identifier is displayed near the target's head on the camera screen using AR (augmented reality). The position where the identifier is displayed is calculated from the differences in longitude and latitude between the user and the target and the azimuth direction of the target from the user. The target is identified based on this information, the face detection coordinates, and the distance between the two. The proposed method has been implemented using Android terminals, and identification accuracy has been examined through experiments.
文摘Face detection is applied to many tasks such as auto focus control, surveillance, user interface, and face recognition. Processing speed and detection accuracy of the face detection have been improved continuously. This paper describes a novel method of fast face detection with multi-scale window search free from image resizing. We adopt statistics of gradient images (SGI) as image features and append an overlapping cell array to improve detection accuracy. The SGI feature is scale invariant and insensitive to small difference of pixel value. These characteristics enable the multi-scale window search without image resizing. Experimental results show that processing speed of our method is 3.66 times faster than a conventional method, adopting HOG features combined to an SVM classifier, without accuracy degradation.
文摘One being developed automatic sweep robot, need to estimate if anyone is on a certain range of road ahead then automatically adjust running speed, in order to ensure work efficiency and operation safety. This paper proposed a method using face detection to predict the data of image sensor. The experimental results show that, the proposed algorithm is practical and reliable, and good outcome have been achieved in the application of instruction robot.
文摘Background Several face detection and recogni tion methods have been proposed in the past decades that have excellent performance.The conventional face recognition pipeline comprises the following:(1)face detection,(2)face alignment,(3)feature extraction,and(4)similarity,which are independent of each other.The separate facial analysis stages lead to redundant model calculations,and are difficult for use in end-to-end training.Methods In this paper,we propose a novel end-to-end trainable convolutional network framework for face detection and recognition,in which a geometric transformation matrix is directly learned to align the faces rather than predicting the facial landmarks.In the training stage,our single CNN model is supervised only by face bounding boxes and personal identities,which are publicly available from WIDER FACE and CASIA-WebFace datasets.Our model is tested on Face Detection Dataset and Benchmark(FDDB)and Labeled Face in the Wild(LFW)datasets.Results The results show 89.24%recall for face detection tasks and 98.63%accura cy for face recognition tasks.
基金supported by ZTE Corporation and State Key Laboratory of Mobile Network and Mobile Multimedia Technology
文摘This paper proposes a universal framework,termed as Multi-Task Hybrid Convolutional Neural Network(MHCNN),for joint face detection,facial landmark detection,facial quality,and facial attribute analysis.MHCNN consists of a high-accuracy single stage detector(SSD)and an efficient tiny convolutional neural network(T-CNN)for joint face detection refinement,alignment and attribute analysis.Though the SSD face detectors achieve promising results,we find that applying a tiny CNN on detections further boosts the detected face scores and bounding boxes.By multi-task training,our T-CNN aims to provide five facial landmarks,facial quality scores,and facial attributes like wearing sunglasses and wearing masks.Since there is no public facial quality data and facial attribute data as we need,we contribute two datasets,namely FaceQ and FaceA,which are collected from the Internet.Experiments show that our MHCNN achieves face detection performance comparable to the state of the art in face detection data set and benchmark(FDDB),and gets reasonable results on AFLW,FaceQ and FaceA.
文摘This paper presents a method which utilizes color, local symmetry and geometry information of human face based on various models. The algorithm first detects most likely face regions or ROIs (Region-Of-Interest) from the image using face color model and face outline model, produces a face color similarity map. Then it performs local symmetry detection within these ROIs to obtain a local symmetry similarity map. The two maps and local similarity map are fused to obtain potential facial feature points. Finally similarity matching is performed to identify faces between the fusion map and face geometry model under affine transformation. The output results are the detected faces with confidence values. The experimental results demonstrate its validity and robustness to identify faces under certain variations.
文摘This paper provides efficient and robust algorithms for real-time face detection and recognition in complex backgrounds. The algorithms are implemented using a series of signal processing methods including Ada Boost, cascade classifier, Local Binary Pattern (LBP), Haar-like feature, facial image pre-processing and Principal Component Analysis (PCA). The Ada Boost algorithm is implemented in a cascade classifier to train the face and eye detectors with robust detection accuracy. The LBP descriptor is utilized to extract facial features for fast face detection. The eye detection algorithm reduces the false face detection rate. The detected facial image is then processed to correct the orientation and increase the contrast, therefore, maintains high facial recognition accuracy. Finally, the PCA algorithm is used to recognize faces efficiently. Large databases with faces and non-faces images are used to train and validate face detection and facial recognition algorithms. The algorithms achieve an overall true-positive rate of 98.8% for face detection and 99.2% for correct facial recognition.
文摘Face detection is considered as a challenging problem in the field of image analysis and computer vision. There are many researches in this area, but because of its importance, it needs to be further developed. Successive Mean Quantization Transform (SMQT) for illumination and sensor insensitive operation and Sparse Network of Winnow (SNoW) to speed up the original classifier based face detection technique presented such a good result. In this paper we use the Mean of Medians of CbCr (MMCbCr) color correction approach to enhance the combined SMQT features and SNoW classifier face detection technique. The proposed technique is applied on color images gathered from various sources such as Internet, and Georgia Database. Experimental results show that the face detection performance of the proposed method is more effective and accurate compared to SFSC method.