KPNet: Towards Minimal Face Detector
Abstract
The small receptive field and capacity of minimal neural networks limit their performance when using them to be the backbone of detectors. In this work, we find that the appearance feature of a generic face is discriminative enough for a tiny and shallow neural network to verify from the background. And the essential barriers behind us are 1) the vague definition of the face bounding box and 2) tricky design of anchor-boxes or receptive field. Unlike most top-down methods for joint face detection and alignment, the proposed KPNet detects small facial keypoints instead of the whole face by in the bottom-up manner. It first predicts the facial landmarks from a low-resolution image via the well-designed fine-grained scale approximation and scale adaptive soft-argmax operator. Finally, the precise face bounding boxes, no matter how we define it, can be inferred from the keypoints. Without any complex head architecture or meticulous network designing, the KPNet achieves state-of-the-art accuracy on generic face detection and alignment benchmarks with only parameters, which runs at 1000fps on GPU and is easy to perform real-time on most modern front-end chips.
Introduction

The performance of face detection has been constantly improved thanks to the anchor-based mechanism (?; ?) with the top-down strategy. By simply assigning dense anchor templates in complex models, we can obtain a face detector with excellent performance. In the current state-of-the-art research, the essence of face detection is how to design the receptive field adaptive to large scale-variance. With the emergence of some seminal works (?; ?; ?; ?; ?) to explore the relationship between the receptive field and the large scale-variance, the performance of face detection is further improved. Inspired by this, the FPN-style framework (?; ?) has become a priority choice for researchers and it can effectively enhance the performance of face detectors to handle faces with different scales. Encouraged by these insights, most of the state-of-the-art algorithms (?; ?; ?) construct adaptive receptive fields to detect faces. They either detect targets with different scales at different levels of the network or detect targets by fusing enhanced features generated by different levels. With the assistance of complex deep backbone (?), improving face detection performance with refined receptive field (?; ?) or embedding new enhanced modules (?; ?; ?) has become the guidance in the field of face detection. However, these top-down approaches with complex backbone networks lead to a heavy computational burden, even though some novel works (?; ?) are proposed to accelerate them. Under the constraints of these current mechanisms, we naturally raise a question: can large scale-variance be solved only through a deeper and more complex backbone with well-designed strategies?
Keep the bottleneck of current research in mind, this paper tries to seek the answer to this question. We re-explore face detection from the two aforementioned essential factors: the receptive field and the large scale-variance. The above question is decomposed into the following more detailed sub-problems and different controlled experiments are performed on FDDB to seek answers.
-
•
It’s a common practice to up-sample the images to or even which facilitates better performance based on a complex backbone network with a large receptive field. Will the lightweight network fail in this configuration?
-
•
How do complex and lightweight networks with different receptive fields perform on low-resolution input images?
We design detection backbones of various depths with different receptive fields by modifying ResNet50 (?). Four detectors CNN-50L, CNN-41L, CNN-26L and CNN-17L with depth 50, 41, 26 and 17 are performed on FDDB. The training set is same with (?) and the recall of top 100 proposals for each image is used for evaluation. Results are shown in Fig. 1 and the anchor setting is that for image resolution [200, 400] to detect face scale [16, 128], for image resolution [400, 600] to detect face scale [32, 256] and for image resolution [800, 1000] to detect face scale [64, 512]. According to the results, the former questions can be explained. When the lightweight network adopts the high-resolution image, even if the appropriate anchor templates are assigned, it still fails in this configuration due to the limitation of the receptive field. It’s worth noting that shrinking the image to low resolution with lightweight backbone can still lead to the comparable performance to the deeper and more complex backbone.
However, potential barriers still exist behind this discovery. The accurate face boxes heavily rely on the tricky design of anchor boxes or receptive field and also, the vague definition of face boxes (e.g. face boxes in FDDB are defined by the ellipse) can easily degrade the performance. Fortunately, bottom-up methods can effectively get rid of the bottleneck in top-down mechanism via converting boxes to keypoints (?; ?). So these ideas naturally lead to a simple, lightweight but accurate framework KPNet where two essential factors are embedded into it, one is to shrink the input image to low resolution with lightweight backbone and the other is to shrink the face concept from box to keypoints to skip the deficiency of general pipeline.
Beyond the general face detection pipeline, the precise facial keypoints can be located first via the carefully designed algorithm and then the accurate face boxes can be inferred by it. So that it can perform as a bottom-up approach to joint face detection and alignment. Different from other top-down algorithms (?; ?) for joint face detection and alignment, KPNet bypasses the vague definition of bounding boxes and takes advantage of the less uncertainty definition of keypoints. Moreover, it’s different from the bottom-up approaches in pose estimation where each landmark is independently predicted and associative embedding (?; ?) is used to group them into an instance. The well-designed fine-grained scale approximation in KPNet can potentially imply the group of landmarks and with the scale adaptive soft-argmax, it can straightforwardly predict the landmarks belonging to the same face.
To summarize, the contributions of this paper are as follows:
1) According to the re-exploration on face detection and the advantages of shrinking target concept from box to keypoints, we propose a novel KPNet with the simple, lightweight but accurate mechanism for the generic face (20 px) detection and alignment.
2) We propose the fine-grained scale approximation and scale-based customization soft-argmax operator to improve the performance by a large margin.
3) Different from all of the joint face detection and alignment methods that adopt the top-down pipeline, KPNet follows the bottom-up mechanism and the more precise definition of landmarks than boxes enables the better performance.
4) Without bells and whistles, KPNet can achieve the SOTA performance on generic face detection benchmarks FDDB, AFW, MALF, and face alignment benchmark AFLW. And also the model inference speed with the offline application can achieve fps on GTX 1080Ti.
Related Works

Face detection. Since the emergence of the powerful CNN (?), the performance of face detection has been improved by a large margin. With the success of anchor-based methods such as Fast RCNN (?) and Faster RCNN (?) on object detection, several different approaches (?; ?) are inspired by them and achieve satisfactory performance on face detection. More recently, the FPN-style framework (?) encourages the researchers to explore the relationship between the receptive field of the face detector and the anchor design skills (?). Benefiting from these explorations, detecting faces with different scales from different layers in a single network (?; ?; ?) has determined its position in the field of face detection. Several works (?; ?) detect a face from the feature fusion of different layers and the enhanced feature make it robust for scale-variance. The deeper and complex backbone with the FPN-style framework achieves the new state-of-the-art and this idea of detector design has dominated face detection for many years. Unfortunately, the accurate face detection heavily relies on the tricky design of anchor boxes or receptive field. Moreover, the vague definition of face boxes makes it hard to generalize to generic face detection.
Face alignment. Face alignment refers to facial landmark detection and it mainly focuses on identifying the geometry structure of the human face. The CNN-based face alignment methods can be divided into two categories, i.e., coordinate regression model and heatmap regression model.
The coordinate regression model directly regresses the facial landmarks from the input image. Many works (?; ?) have the advantage of explicit inference of landmarks without any post-processing. However, these regression-based methods are not performing as well as heatmap regression models. The heatmap regression models generate likelihood heatmaps for each keypoint, respectively. In these heatmap-based methods, hourglass networks (?) becomes the backbone of many works (?; ?) due to its capabilities of obtaining multi-scale information. Some works (?; ?) have adopted facial parts to aid face alignments tasks and LAB (?) uses more precise facial boundary to assist in detecting facial landmarks.
Joint face detection and alignment. Face alignment and face detection are closely related, yet few works (?; ?) jointly perform them. The popular algorithm MTCNN (?) follows the general top-down mechanism which first regresses the face boxes and then generates the corresponding landmarks. However, the accurate face boxes generation relies on the tricky design of anchor boxes and is weak against the vague definition of face boxes. Comparing with it, KPNet first predicts the facial landmarks with unambiguous definition and then inference the face boxes by it. The more precise definition of landmarks than face boxes enables it to achieve better performance than top-down methods.
KPNet
Overview
In KPNet, we adopt the bottom-up mechanism to perform face detection and alignment simultaneously. Fig. 2 provides an overview of KPNet. Firstly, the potential face scale proposals can be predicted by the fine-grained scale approximation. Secondly, the keypoints can be computed by the scale adaptive soft-argmax with the scalemap and the output heatmap of landmarks response generation. Finally, we apply a simple transformation algorithm (?) to infer the final bounding boxes from the landmarks.
Fine-grained scale approximation
To better detect the facial keypoints, we need to locate the focus regions where existing faces are. The anchor-based mechanism is undoubtedly an appropriate solution, but its’ sophisticated design techniques make it a departure from the simplicity and flexibility of our framework. Inspired by (?; ?) where CNN is capable of approximating the scale information in the low-resolution image, we convert the boxes regression to fine-grained face scale classification for each pixel in a feature map. Different from (?) where only the existing scales are predicted to select valid layers from the image pyramid, in KPNet, we add additional spatial information to scale approximation.
The scalemap is generated by fine-grained scale approximation that consists of only one convolutional layer with kernel size 3 is used. It is a probability map with dimension , where is the predefined number of scales. Given an image with face boxes where means the face center and represent the height and width of the face, the is firstly initialized to 0 and then we calculate the active channel index as:
(1) |
where b is the index from 1 to and represents the max edge of . The minimum detected face size for =1280 in 720P image is . In Eq. 1, the face size from to can be mapped into the different channel indexes in . For simplicity, we divide the , into 10 scale bins and thus the total scale number . With the computed , the value at coordinate can be defined as:
(2) |
where means the stride of the network. Encouraged by (?; ?), to alleviate the difficulty of feature learning in the discrete distribution, we introduce the 2D Gaussian function to refine it. Given the radius and the host point , the values of its neighbouring points can be formulated as:
(3) |
where belongs to the neighbour set and is set to 0.1 in our experiments. During training, the input image is resized with the higher dimension equal to 256 and the loss of the scalemap training is a binary multi-class cross entropy loss:
(4) |
where are the ground truth label and prediction of the i-th pixel in . indicates the pixel number in feature map .
For inference, given a threshold, we select all of the valid coordinates from and compute the scale by Eq. 1 based on its channel index . Finally, the scale proposal can be obtained.
Scale adaptive soft-argmax operator
As shown in (?), heatmap regression models can achieve better performance than coordinate regression models do. With this conclusion in mind, instead of regressing the keypoint coordinates, we detect them from the landmark response map. At the end of the backbone, the landmark response generation is used to generate a response map with dimension where K is the number of facial keypoints. It only consists of one convolutional layer with kernel size 3. Instead of the argmax function, which is not differentiable, breaking the learning chain on neural networks (?), we propose the - operator which keeps the properties of specialized part detectors while being fully differentiable. Different from the usage in (?) where it applies to the global response map cooperated with top-down methods, the proposed - is performed on the scale aware locations cooperated with the bottom-up pipeline.
Given a scale proposal where mean the top left and bottom right corner, we define the Softmax operation on it as:
(5) |
where is the value of heat map at location of channel . The coordinates of the landmarks corresponding to the are given by:
(6) |
where indicate the width and height of . returns when and 0 otherwise.
For keypoints regression based on each , we adopt the loss function as follows:
(7) |
where means the ground truth keypoints. After obtaining the facial keypoints, the face boxes can be inferred by it conforming to the definition of us. Define the probability of a scale proposal as , the score of the corresponding face box is formulated as:
(8) |
where means the maximum operation on the spatial resolution and c from 1 to 3 represents the channels corresponding to left eye, right eye, and nose. We empirically utilize the keypoints information to weaken some false positive scale proposals.

Backbone architecture
We use two different CNN architectures as the backbones of KPNet, respectively. One is the robust stacked hourglass network and the other is DRNet with faster inference speed. The hourglass network only consists of one stacked hourglass and the channel number is reduced to 64. The first convolution layer with kernel size , stride 2 is replaced by two small convolution layers without BN and ReLU between them. Kernel size with stride 2 and kernel size with stride 1 are assigned to them, respectively. Furthermore, we upsample the spatial resolution by a factor of 2 (using nearest neighbor upsampling for simplicity) at the end of the hourglass to reduce the bias caused by heavy down-sample operations. Finally, the total stride of the hourglass is 2 and the parameter is . Even though the network has become very lightweight after these specific modifications, abundant operations on large resolution feature maps still limit the inference speed of the network.
Following the principle of reducing the redundant operation on high-resolution feature maps, we design a simple and fast De-redundancy Net (DRNet). It can achieve faster inference speed than the former hourglass with the same number of parameters. The details of DRNet are shown in Fig 3. DRNet is a simple and lightweight network with only 11 layers. Given an input image, we first reduce it via a convolution layer with stride 2 and a max pooling layer with stride 2. Then some residual blocks are followed and two nearest neighbor upsampling layers with factor 2 are used to upsample the spatial resolution. The total stride of DRNet is 2 and the large-span skip layer enables the network to retain as much input information as possible which makes it comfortable for low-resolution input. The lightweight structure without redundant operations on high-resolution feature maps ensures that it can achieve the faster inference speed.
Advantage insights of KPNet
All of the joint face detection and alignment algorithms (?; ?) first generate the face boxes and then predict the corresponding landmarks. Limited by face detection accuracy which heavily relies on the tricky design of anchor boxes and is easily influenced by the ambiguous definition of bounding boxes, it’s hard to achieve better performance with faster speed. Compared with these top-down pipelines, KPNet has the following advantages. First of all, KPNet adopts the bottom-up mechanism that first locates the landmarks with the unambiguous definition, and then the face boxes can be inferred from keypoints. The precise definition of landmarks compared with face boxes allows it to easily achieve higher performance. Secondly, KPNet skips anchor designing and thusly is flexible to deploy this network without complicated design skills. Finally, different from the most face detection methods (?; ?), KPNet does not depend on high-resolution input. Through the combination of low-resolution input and lightweight network, it can achieve SOTA performance on both of the generic face detection (20px) and alignment with fast inference speed (offline application with fps).
Experiments
Implement details
We implement KPNet in PyTorch. Both of the hourglass and DRNet are randomly initialized under the default setting of PyTorch without pretraining on any external dataset. During training, we set the input resolution of the network to , which leads to an output resolution of . For training on generic face detection, we adopt the training set the same as (?) and none of the data augmentations is performed. For joint face detection and alignment, is set to 5 representing the left eye, right eye, nose, left corner of the mouth and right corner of the mouth. We joint optimize the loss function and the with lossweight 1:1 via SGD. Due to the huge pixels in scalemap , the weight of is set to 10000 for faster convergence. Benefiting from the low-resolution input and lightweight backbone, we use a batch size of 128 and train the network on 4 GTX 1080Ti GPUs. For training on AFLW, because of the missing annotation of some face boxes in the training set, we only train the for the annotated facial landmarks. is set to 19 to correspond to the annotated facial landmarks in AFLW. We adopt the data augmentation strategy the same as (?) for preventing over-fitting. We train the network for 150k iterations with a learning rate warmup strategy. The learning rate is linearly increased to 0.01 from 0.00001 in the first 50k iterations and we reduce it to 0.001 for the last 50k iterations.
At the inference stage, we first generate the scale proposals through the predefined threshold from scalemap, and then compute the corresponding keypoints via then scale adaptive soft-argmax according to Eq. 5 and Eq. 6. Finally, NMS with IOU 0.6 is adopted on the face boxes inferred from these keypoints.
Test benchmarks
We evaluate KPNet on the generic face detection benchmarks FDDB (?), AFW, MALF, and face alignment benchmark AFLW (?). We adopt the relabeled version provided by (?) where some missing faces are re-annotated. We follow (?) to adopt the AFLW-Full in our experiments where 20,000 and 4,386 images are used for training and testing, respectively. For face detection, we follow the protocol of (?).
Network | FDDB (%) | AFW (%) | MALF (%) |
---|---|---|---|
RSAbase | 96.0 | 100.0 | 96.49 |
DRNet | 96.6 | 99.6 | 97.1 |
Hourglasslight | 96.7 | 99.8 | 97.59 |
Ablation study
Fine-grained scale approximation. Fine-grained scale approximation is a key component of KPNet. To understand its performance, we directly evaluate its recall on FDDB, AFW, and MALF. Similar to the evaluation metric in former section, we compute the recall of the top 100 scale proposals. Furthermore, we implement the anchor-based detector RSA, which is the SOTA algorithm on generic face detection. According to their claimed configuration, we evaluate the recall for comparison with the same protocol. The result is shown in Tab. 1. Hourglasslight is the modified hourglass network in Sec. Backbone architecture The fine-grained scale approximation can achieve a comparable recall to the SOTA anchor-based algorithm without relying on the experience design.
Advantage of scale adaptive soft-argmax operator. We introduce the scale adaptive soft-argmax (SS) to predict the facial keypoints coordinates from the heatmap. In order to better evaluate the superiority of SS over and , We conduct different experiments with DRNet. For , we replace the SS by a fully connected layer to directly regress the keypoints coordinates. We adopt the global average pooling on the scale proposal indicated in Sec. Backbone architecture to convert it to a feature vector with the fixed size. A fully connected layer with output is applied to regress the keypoint coordinates where means the keypoint number and loss is used for optimization. For , each channel in the specific location of corresponding to a specific keypoint and only the coordinates existing keypoints will be set to 1. The loss function is the same as Eq. 4. In the ablation study, we further conduct FDDB, FDDB and PFDDB as the additional benchmarks. FDDB and FDDB are generated by rotating the FDDB with and , respectively. PFDDB is assigned by all of the images from FDDB which containing profile faces (Roll or Yaw or Pitch ).




Methods | FDDB | FDDB | FDDB | PFDDB |
---|---|---|---|---|
RSAbase | 92.15 | 57.67 | 56.78 | 92.8/59.93 |
90.86 | 68.96 | 67.2 | 90.94/ 47.97 | |
88.98 | 50.65 | 49.9 | 83.51/43.9 | |
SS | 91.6 | 69.32 | 69.97 | 91.29/61.44 |
Backbone | SP | SS | GT box | FDDB |
---|---|---|---|---|
DRNet | ✓ | 90.7/47.3 | ||
DRNet | ✓ | ✓ | 91.6/81.1 | |
DRNet | ✓ | ✓ | 96.6/96.6 | |
Hourglasslight | ✓ | 91.64/12.52 | ||
Hourglasslight | ✓ | ✓ | 92.61/80.9 | |
Hourglasslight | ✓ | ✓ | 96.72/96.72 |
The results are shown in Tab. 2. All of the models are trained on normal faces without rotation augmentation. SS performs better than others, even the anchor-based RSAbase with high-resolution input and image pyramid, KPNet can achieve comparable performance, even more, robust in face-rotation scenarios. In actually, detecting face boxes from key points is important for the lightweight network than directly regressing bounding boxes. Key points have less uncertain information which is easier for lightweight models to fit. Besides, the semantic information of the landmarks is fixed, even if the face angle/pose changes. This ensures its robustness to face angle/ pose variance.
Error analysis. KPNet simultaneously outputs fine-grained scale heatmap and landmarks response map. To understand how each part contributes to the final error, we perform an error analysis by replacing the predicted scale proposal with the ground-truth boxes. Furthermore, to evaluate the SS contribution to KPNet, we detect the faces only by scale proposal.
method | average normalized error |
---|---|
PCD-CNN (?) | 2.40 |
TSR (?) | 2.17 |
LAB (?) | 1.25 |
SAN (?) | 1.91 |
PFLD 1X (?) | 1.88 |
Wing (?) | 1.65 |
CPM+SBR (?) | 2.14 |
KPNet+DRNet | 1.87 |
KPNet+Hourglasslight | 1.45 |
Tab. 3 shows that the proposed SS can effectively improve the quality of scale proposal and provide the more precise facial location, especially the recall at false positive number 1. Replacing the SP by GT box improves the recall by %. This suggests that there is still ample room for improvement in both SP and SS.
Method | Face detection | Face alignment | Online speed | Offline speed | # Param | FDDB | AFLW |
---|---|---|---|---|---|---|---|
RSAbase (?) | ✓ | msTi | W/o | M | 92.15% | W/o | |
(?) | ✓ | msP100 | W/o | M | 93.5% | W/o | |
(?) | ✓ | msXP | - | M | 92.9% | W/o | |
MTCNN (?) | ✓ | ✓ | msTi | - | M | 90.44% | 6.9∗ |
DLIB (?) | ✓ | ✓ | msTi - | - | - | 78.1% | - |
PFLD 1X (?) | ✓ | msTi - | - | M | W/o | 1.88 | |
LAB (?) | ✓ | msX - | - | M | W/o | 1.25 | |
SAN (?) | ✓ | msTi - | - | M | W/o | 1.91 | |
Wing (?) | ✓ | msX - | - | M | W/o | 1.65 | |
KPNet+Hourglasslight | ✓ | ✓ | msTi | msTi | M | 92.61% | 1.45 (4.45∗) |
KPNet+DRNet | ✓ | ✓ | msTi | msTi | M | 91.6% | 1.87 (5.77∗) |
Comparisons with SOTA algorithms
For face detection, we compare our KPNet with state-of-the-art methods (?; ?; ?; ?; ?; ?; ?; ?; ?; ?; ?; ?; ?; ?; ?; ?) and the DLIB c++ library (?), which supports for joint face detection and alignment. Fig. 4 shows the comparison with other approaches on three benchmarks. On AFW, our algorithm KPNet can achieve 99.53% AP and 98.72% AP by Hourglasslight and DRNet, respectively. On FDDB, KPNet+Hourglasslight recalls 92.61% faces with 50 false positives as shown in Fig. 4(a) which outperforms most of the approaches. On MALF, our methods can also achieve a comparable result with the state-of-the-art. It should be noticed that the shape and scale definition of the bounding box on each benchmark varies. KPNet can be easily applied to these benchmarks without complicated design choices.
For face alignment, we compared KPNet with other state-of-the-art methods on AFLW. AFLW is a challenging dataset that has been widely used for evaluating face alignment algorithms. As shown in Tab. 4, our KPNet+Hourglasslight outperforms all of the other approaches and KPNet+DRNet can also achieve comparable performance.
Method | FDDB (%) | AFW (%) | MALF (%) | AFLW |
---|---|---|---|---|
RPN+DRNet | 75.82 | 88.82 | 58.68 | 2.12 |
AE+DRNet | 46.96 | 76.8 | 32.57 | 2.12 |
KPNet+DRNet | 91.6 | 98.72 | 88.92 | 1.87 |
Furthermore, we compare KPNet+DRNet with the bottom-up associative embedding (AE) (?; ?) and top-down RPN+DRNet. As shown in Tab 6, whether adopting anchor-based RPN or AE, the performance is strictly limited by the capacity of lightweight backbone. It’s in stark contrast that KPNet with fine-grained scale approximation and scale adaptive soft-argmax achieve excellent performance. To better understand the performance of KPNet, we visualize some images sampled from FDDB and AFLW in Fig. 5.
Analysis of the inference speed
In this section, we explore the performance and speed in detail compared with other approaches as shown in Tab. 5. We report the recall at false positive number 50 on FDDB and the NME () on AFLW.
In the offline applications, KPNet with DRNet can achieve 1000 fps at GTX 1080Ti, faster than other face detectors with a large margin. Even compared with the state-of-the-art algorithms on face alignment, KPNet still has a faster model inference speed.
Both the MTCNN and DLIB are the popular frameworks for joint face detection and alignment. KPNet outperforms them with a large margin in terms of inference speed and performance. No complex hyperparameter designing is required so that it can be easily applied to different scenarios.
Conclusion
This paper proposes a simple, lightweight but accurate framework KPNet which does away with anchor boxes. It focuses on joint generic face ( 20px) detection and alignment. Unlike most face detection methods and top-down joint face detection and alignment methods, KPNet adopts the bottom-up mechanism. It first predicts the facial landmarks from a low-resolution image via the well-designed fine-grained scale approximation and scale adaptive soft-argmax operator. Finally, the precise face bounding boxes, no matter how we define it, can be inferred from the landmarks. KPNet can effectively alleviate the vague definition of the face bounding box. Without bells and whistles, KPNet achieves state-of-the-art accuracy on generic face detection and alignment benchmarks with only M parameters. The model inference speed can achieve fps on GPU and it’s easily deployed to most modern front-end chips.
References
- [Chen et al. 2016] Chen, D.; Hua, G.; Wen, F.; and Sun, J. 2016. Supervised transformer network for efficient face detection. In ECCV.
- [Chi et al. 2018] Chi, C.; Zhang, S.; Xing, J.; Lei, Z.; Li, S. Z.; and Zou, X. 2018. Selective refinement network for high performance face detection. arXiv preprint arXiv:1809.02693.
- [Deng et al. 2017] Deng, J.; Trigeorgis, G.; Zhou, Y.; and Zafeiriou, S. 2017. Joint multi-view face alignment in the wild. arXiv preprint arXiv:1708.06023.
- [Dong et al. 2018a] Dong, X.; Yan, Y.; Ouyang, W.; and Yang, Y. 2018a. Style aggregated network for facial landmark detection. In CVPR.
- [Dong et al. 2018b] Dong, X.; Yu, S.-I.; Weng, X.; Wei, S.-E.; Yang, Y.; and Sheikh, Y. 2018b. Supervision-by-registration: An unsupervised approach to improve the precision of facial landmark detectors. In CVPR.
- [Farfade, Saberian, and Li 2015] Farfade, S. S.; Saberian, M. J.; and Li, L.-J. 2015. Multi-view face detection using deep convolutional neural networks. In Proceedings of the 5th ACM on International Conference on Multimedia Retrieval.
- [Feng et al. 2018] Feng, Z.-H.; Kittler, J.; Awais, M.; Huber, P.; and Wu, X.-J. 2018. Wing loss for robust facial landmark localisation with convolutional neural networks. In CVPR.
- [Girshick 2015] Girshick, R. 2015. Fast r-cnn. In ICCV.
- [Guo et al. 2019] Guo, X.; Li, S.; Zhang, J.; Ma, J.; Ma, L.; Liu, W.; and Ling, H. 2019. Pfld: A practical facial landmark detector. arXiv preprint arXiv:1902.10859.
- [Hao et al. 2017] Hao, Z.; Liu, Y.; Qin, H.; Yan, J.; Li, X.; and Hu, X. 2017. Scale-aware face detection. In CVPR.
- [He et al. 2016] He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In CVPR.
- [Hu and Ramanan 2017] Hu, P., and Ramanan, D. 2017. Finding tiny faces. In CVPR.
- [Jain and Learned-Miller 2010] Jain, V., and Learned-Miller, E. 2010. Fddb: A benchmark for face detection in unconstrained settings. Technical report, UMass Amherst Technical Report.
- [King 2009] King, D. E. 2009. Dlib-ml: A machine learning toolkit. Journal of Machine Learning Research.
- [Koestinger et al. 2011] Koestinger, M.; Wohlhart, P.; Roth, P. M.; and Bischof, H. 2011. Annotated facial landmarks in the wild: A large-scale, real-world database for facial landmark localization. In ICCV workshops.
- [Kumar and Chellappa 2018] Kumar, A., and Chellappa, R. 2018. Disentangling 3d pose in a dendritic cnn for unconstrained 2d face alignment. In CVPR.
- [Law and Deng 2018] Law, H., and Deng, J. 2018. Cornernet: Detecting objects as paired keypoints. In ECCV.
- [Li et al. 2015] Li, H.; Lin, Z.; Shen, X.; Brandt, J.; and Hua, G. 2015. A convolutional neural network cascade for face detection. In CVPR.
- [Li et al. 2018] Li, J.; Wang, Y.; Wang, C.; Tai, Y.; Qian, J.; Yang, J.; Wang, C.; Li, J.; and Huang, F. 2018. Dsfd: dual shot face detector. arXiv preprint arXiv:1810.10220.
- [Li et al. 2019] Li, H.; Liu, Y.; Ouyang, W.; and Wang, X. 2019. Zoom out-and-in network with map attention decision for region proposal and object detection. International Journal of Computer Vision 127(3):225–238.
- [Li, Liu, and Wang 2019] Li, B.; Liu, Y.; and Wang, X. 2019. Gradient harmonized single-stage detector. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, 8577–8584.
- [Lin et al. 2017] Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; and Belongie, S. 2017. Feature pyramid networks for object detection. In CVPR.
- [Liu et al. 2017] Liu, Y.; Li, H.; Yan, J.; Wei, F.; Wang, X.; and Tang, X. 2017. Recurrent scale approximation for object detection in cnn. In ICCV.
- [Liu et al. 2019] Liu, L.; Li, G.; Xie, Y.; Yu, Y.; Wang, Q.; and Lin, L. 2019. Facial landmark machines: A backbone-branches architecture with progressive representation learning. TMM.
- [Luvizon, Tabia, and Picard 2017] Luvizon, D. C.; Tabia, H.; and Picard, D. 2017. Human pose regression by combining indirect part detection and contextual information. arXiv preprint arXiv:1710.02322.
- [Lv et al. 2017] Lv, J.; Shao, X.; Xing, J.; Cheng, C.; and Zhou, X. 2017. A deep regression architecture with two-stage re-initialization for high performance facial landmark detection. In CVPR.
- [Mathias et al. 2014] Mathias, M.; Benenson, R.; Pedersoli, M.; and Van Gool, L. 2014. Face detection without bells and whistles. In ECCV.
- [Najibi et al. 2017] Najibi, M.; Samangouei, P.; Chellappa, R.; and Davis, L. S. 2017. Ssh: Single stage headless face detector. In ICCV.
- [Newell, Huang, and Deng 2017] Newell, A.; Huang, Z.; and Deng, J. 2017. Associative embedding: End-to-end learning for joint detection and grouping. In NeurIPS.
- [Newell, Yang, and Deng 2016] Newell, A.; Yang, K.; and Deng, J. 2016. Stacked hourglass networks for human pose estimation. In ECCV.
- [Ren et al. 2015] Ren, S.; He, K.; Girshick, R.; and Sun, J. 2015. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in neural information processing systems.
- [Song et al. 2018] Song, G.; Liu, Y.; Jiang, M.; Wang, Y.; Yan, J.; and Leng, B. 2018. Beyond trade-off: Accelerate fcn-based face detector with higher accuracy. In CVPR.
- [Sun, Wang, and Tang 2013] Sun, Y.; Wang, X.; and Tang, X. 2013. Deep convolutional network cascade for facial point detection. In CVPR.
- [Tang et al. 2018] Tang, X.; Du, D. K.; He, Z.; and Liu, J. 2018. Pyramidbox: A context-assisted single shot face detector. In ECCV.
- [Wang et al. 2017a] Wang, H.; Li, Z.; Ji, X.; and Wang, Y. 2017a. Face r-cnn. arXiv preprint arXiv:1706.01061.
- [Wang et al. 2017b] Wang, Y.; Ji, X.; Zhou, Z.; Wang, H.; and Li, Z. 2017b. Detecting faces using region-based fully convolutional networks. arXiv preprint arXiv:1709.05256.
- [Wu et al. 2018] Wu, W.; Qian, C.; Yang, S.; Wang, Q.; Cai, Y.; and Zhou, Q. 2018. Look at boundary: A boundary-aware face alignment algorithm. In CVPR.
- [Yang et al. 2014] Yang, B.; Yan, J.; Lei, Z.; and Li, S. Z. 2014. Aggregate channel features for multi-view face detection. In IEEE international joint conference on biometrics.
- [Yang et al. 2015] Yang, S.; Luo, P.; Loy, C.-C.; and Tang, X. 2015. From facial parts responses to face detection: A deep learning approach. In ICCV.
- [Yang et al. 2017] Yang, S.; Xiong, Y.; Loy, C. C.; and Tang, X. 2017. Face detection through scale-friendly deep convolutional networks. arXiv preprint arXiv:1706.02863.
- [Yu et al. 2016] Yu, J.; Jiang, Y.; Wang, Z.; Cao, Z.; and Huang, T. 2016. Unitbox: An advanced object detection network. In Proceedings of the 24th ACM international conference on Multimedia.
- [Zeng et al. 2017] Zeng, X.; Ouyang, W.; Yan, J.; Li, H.; Xiao, T.; Wang, K.; Liu, Y.; Zhou, Y.; Yang, B.; Wang, Z.; et al. 2017. Crafting gbd-net for object detection. IEEE transactions on pattern analysis and machine intelligence 40(9):2109–2123.
- [Zhang et al. 2016] Zhang, K.; Zhang, Z.; Li, Z.; and Qiao, Y. 2016. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Processing Letters.
- [Zhang et al. 2017a] Zhang, J.; Wu, X.; Zhu, J.; and Hoi, S. C. 2017a. Feature agglomeration networks for single stage face detection. arXiv preprint arXiv:1712.00721.
- [Zhang et al. 2017b] Zhang, K.; Zhang, Z.; Wang, H.; Li, Z.; Qiao, Y.; and Liu, W. 2017b. Detecting faces using inside cascaded contextual cnn. In ICCV, 3171–3179.
- [Zhang et al. 2017c] Zhang, S.; Zhu, X.; Lei, Z.; Shi, H.; Wang, X.; and Li, S. Z. 2017c. Faceboxes: A cpu real-time face detector with high accuracy. In IJCB.
- [Zhang et al. 2017d] Zhang, S.; Zhu, X.; Lei, Z.; Shi, H.; Wang, X.; and Li, S. Z. 2017d. S3fd: Single shot scale-invariant face detector. In ICCV.
- [Zhou, Zhuo, and Krähenbühl 2019] Zhou, X.; Zhuo, J.; and Krähenbühl, P. 2019. Bottom-up object detection by grouping extreme and center points. arXiv preprint arXiv:1901.08043.