Where, What, Whether: Multi-modal Learning Meets Pedestrian Detection
Abstract
Pedestrian detection benefits greatly from deep convolutional neural networks (CNNs). However, it is inherently hard for CNNs to handle situations in the presence of occlusion and scale variation. In this paper, we propose W3Net, which attempts to address above challenges by decomposing the pedestrian detection task into Where, What and Whether problem directing against pedestrian localization, scale prediction and classification correspondingly. Specifically, for a pedestrian instance, we formulate its feature by three steps. i) We generate a bird view map, which is naturally free from occlusion issues, and scan all points on it to look for suitable locations for each pedestrian instance. ii) Instead of utilizing pre-fixed anchors, we model the interdependency between depth and scale aiming at generating depth-guided scales at different locations for better matching instances of different sizes. iii) We learn a latent vector shared by both visual and corpus space, by which false positives with similar vertical structure but lacking human partial features would be filtered out. We achieve state-of-the-art results on widely used datasets (Citypersons and Caltech). In particular. when evaluating on heavy occlusion subset, our results reduce MR-2 from 49.3 to 18.7 on Citypersons, and from 45.18 to 28.33 on Caltech.

1 Introduction
Pedestrian detection is a fundamental topic in computer vision. Generally speaking, the design of pedestrian detector is deeply influenced by the development of object detection, which is used to tell where the object is, and how big it is [23]. Most modern anchor-based detectors [5] [35] [36] get stuck in a paradigm, namely, expertise-based techniques to generate a series of anchors and then identify whether it is a pedestrian or not. It is undeniable that this method has a extensive impact and is widely used as a powerful baseline, but a large number of redundant and low-quality proposals it introduces also limit the accuracy and speed. In contrast, our work falls into the anchor-free fashion, which takes multi-modal data as input to predict where the target is, what the scale is, and whether the target is actually a pedestrian, instead of utilizing pre-fixed anchors.
Where - Object detection actually starts from a basic problem: where is the target? In previous successful practices, such as Faster RCNN [26] or SSD [21], this process is mostly determined by a set of pre-defined anchors, based on a implicit dependency that objects are evenly distributed on the image, which is much similar to exhaustion. The recently popular anchor-free methods shake of the yoke (pre-defined anchors) and detect objects directly from an image, which makes the detection a more natural way. One of the typical and effective anchor-free practice in pedestrian detection is CSP [23], which facilitated the Where problem as a straightforward center prediction task. Our work is conceptually similar to CSP, which is also in the scope of anchor-free, but differs significantly in insights. As the author states in CSP [23] (Section 4.4), the pedestrian center is actually a fuzzy point of semantic information, in which the variation of pedestrian wearing or orientation will have a negative effect on it. In addition, we discover that the pedestrian center is vulnerable to occlusion, that is to say, in some occluded scenes, the pedestrian center is invisible. The above challenges motivate us to find a unified and robust representation for locations of pedestrians. In this paper, we attempt to attribute occlusion issues to the interlocking problem caused by the single image view (front view). If the 2D image is switched to the bird view map, occlusion would be greatly alleviated. As illustrated in Figure 1 (left), even if a pedestrian has occlusion in the front view, the bird view map still performs free from occlusion issues.
What - Another long-standing problem exists in object detection as well as pedestrian detection is what the scale of the target is. For the anchor-based methods, obviously, they rely more on a set of pre-defined scales and aspect ratios. Or, the anchor-free method such as CSP, stacks with convolutions to predict pedestrian scales and set aspect ratio as a uniformed value of 0.41 [23] (Section 3.3). Despite these pipelines have been shown effective in several benchmarks, the exhaustive proposal sizes in anchor-based methods appear redundant, and the uniform scale setting, such as 0.41 in CSP [23], seems to be inflexible. In this paper, the proposed W3Net is motivated by the discovery that the scale distribution of pedestrians in the 2D image is not out of order, on the contrary, is closely related to its geometry. As is shown in Figure 1(middle), intra-image pedestrian scale variations vary along with the estimated depth (the distance from the camera in real world), which means the nearer instances should have lager scales compared to farther one. Following this intuition, we model the interdependency between depth and scale aiming at generating proper scales that flexible and accurate at different locations.
Whether - Neither anchor-based nor anchor-free methods can stay out of the problem: whether the bounding box actually filter the pedestrian. Due to the existence of occlusion issues and the diversity of occlusion patterns, instance features used for the downstream classification show an obvious difference, and consequently make the ”Whether” problem difficult. Some efforts have been made to handle this matter. Part-based methods [35] [36] tends to employ a weakly-supervised manner to perceive the visible body parts, while the dual branch method [39] proposes two sub-networks, one for full body estimation and the other for visible part. These methods, in the final analysis, treat the occluded and non-occluded pedestrians in a split way, which suffers uncertainty of weakly-supervised part labels and imbalance between the number of occluded and non-occluded samples. In contrast, we discover that pedestrians with obvious attributes can be represented by corpus, such as head, arm, body and leg, which offers a possibility to re-encode both occluded and non-occluded instances into a unified corpus space, and thus benefits robust feature generation against ”Whether” problem.
W3Net is evaluated with challenging settings on Citypersons [34] and Caltech [8] pedestrian dataset, and still achieves state-of-the-art performance. In particular, when evaluating heavy occlusion subset on Citypersons, our result reduces MR-2 by a factor of 2 (our model produces 18.7% while prior arts range from 49%-56%).
2 Related Work
With multi-modal learning, the proposed W3Net confronts challenges (occlusion and scale variation), and decouples the task into Where, What and Whether problem directing against pedestrian localization, scale prediction and classification. Therefore, we review recent work on pedestrian detection with or without multi-modal data and compare ours with previous state-of-the-art methods.
Most prevalent pedestrian detectors are on the basis of the framework in general object detection such as Faster RCNN [26] and SSD [21], and leverage pedestrian-oriented traits to tackle challenges such as occlusion and scale variation in the pedestrian detection task. Part-based methods, for example DeepParts [12], FasterRCNN+ATT [35] or OR-CNN [36], fully exploit pedestrian part information especially visible body parts to assist robust feature embeddings. With the careful discovery that pedestrians usually perform up-right posture, TLL [27] escapes from pre-defined anchors and proposes the line localization, which greatly promotes the development of pedestrian detection.

Recently, some researchers have turned attention to multi-modal learning, which offers the possibility of capturing correspondences between multiple modalities and gains an in-depth understanding of natural phenomena. SDS-RCNN [3] proposes a segmentation infusion network to enable joint supervision on semantic segmentation and pedestrian detection. F-DNN+SS [9] uses a derivation of the Faster RCNN framework, and further incorporates pixel-wise semantic segmentation in a post-processing manner to suppress background proposals. F-DNN2+SS [10] employs an ensemble learning approach, and semantic segmentation network to adjust the confidence in the detector proposals.
Motivated by above, the proposed W3Net takes a further step in three-folds:
1) We first attempt to explore a new possibility that pedestrians can be effectively described in the bird view map, which bypasses the limitation of 2D (front) images and benefits the generation of robust features against various occlusion patterns.
2) We first attempt to model interdependency between depth and scale for pedestrian detection, which predicts flexible and proper proposals other than predefined ones.
3) We first attempt to propose a method to embed pedestrian features in corpus space, which aligns both occluded and non-occluded instances.
3 Methodology
W3Net falls into the anchor-free fashion, which decouples the pedestrian detection task into three sub-problems. With the input of multi-modal data, including bird view map, depth and corpus information, the three problems ”Where”, ”What” and ”Whether” target at pedestrian localization, scale prediction and classification to benefit robust feature generation against occlusion and scale variation.
The pipeline is shown in Figure 2. Specially, taking the front view image (Domain_A) as input, the network first generates the corresponding bird view map (Domain_C) through two Cycle GAN [40]. The is then processed by the forward feature extractor to predict probability of pedestrians at each location and the corresponding width, denoted as ”Where” branch. The connecting ”What” branch, supported by depth relation, aligns results from ”Where” branch and generating reasonable proposals for the following ”Whether” branch without any conv or fc effort. Proposal features extracted by feature extractor are subsequently fed into the encoder-decoder, in which the latent space is occupied by corpus and promotes both occluded or non-occluded instances to be re-encoded by common attributes of pedestrians. Overall, the three ”W” branches link together, complement each other, and constitute the proposed W3Net.
3.1 Bird view: Where the target is
Compared with front view, bird view owns the unique advantages: 1) The bird view map is naturally free from occlusion issues. We discover that occlusion, whether it is intra-class or inter-class [30], could be mostly attributed to the single detection view (front view). 2) Pedestrians are more likely to be condensed into a unified feature representation through bird view, namely ’head’. Previous anchor-free method, such as CSP which suffers difficulty to decide an ’exact’ center point during training, the proposed ’head’ prediction on bird view map appends more specific and accurate classification targets for the ”Where” branch.
However, how to obtain the bird view map still remains to be solved, which we treat as an image-to-image generation process in this paper. Due to the lack of real bird view map in most of existing pedestrian datasets, such as Citypersons [34] and Caltech [8], we attempt to introduce synthetic data, which directly captured from 3D games to train a bird view generation model. As the previous practice in [1], synthetic data such as [25] [18] faces a huge challenge: the performance would be limited if the model trained on synthetic data is directly used on real-world data. Inspired by above, we design our front-to-bird view generation network by two steps. The first includes a CycleGAN [40] [17], denoted as in Figure 2, to transfer real-world data (Domain_A) to the synthetic images (Domain_B). The following step introduces the other CycleGAN, denoted as , to train a bird view generation model over synthetic data captured from the game of GTA5, which consists of a total 50,000 pairs of images including front view and bird view from virtual car. Based on the well-trained Generative Adversarial Network (GAN) [15] and , the input front view map could be transformed to the bird view map .
Upon , a detection head, which is composed of one conv and two conv layers, is attached behind the feature extractor to predict probability of pedestrians at each location and the corresponding width. With its inherent superiority on occlusion, the bird view map can bypass the limitation of front view images, and thus benefit the generation of robust features. However, it is not well performed yet. On the one hand, bird view map pays for the loss of pedestrian height information, which brings about unsatisfied height prediction results on the single bird view map, and thus derives the following ”What” branch to estimate flexible and accurate scales for different instances. On the other hand, bird view map suffers from false positives caused by suspected objects, such as the top of street lamps vs. pedestrian ’head’, thus the introduced ”Whether” branch is directed against this issue.
3.2 Depth: What the scale is
As is known to us, the bounding box has four-degree-of-freedom, that is . During the process of proposal generating, are already known, which are evenly distributed on the feature map, and the corresponding has been predicted on the bird view. Therefore, the ”What” branch only includes two tasks: 1) What the width is on the front view map. 2) What the corresponding height is.
Subsequently, to address above tasks, we start from a more general situation with the following definition.
Definition: We define as the front view when the image is captured at the front horizontally, and as the view when the image is captured with the depression angle at the height of . Specially, if is , the view looks towards perpendicular to horizontal.
On the above basis, we could build up a set of geometric relations. We suppose the real world 3D coordinates of the pedestrian P under the view of is , in which could be also considered as the distance from the camera, that is depth,while the corresponding 2D coordinates on the image is , in which given the camera intrinsic matrix , the projection from can be formulated as the following:
(1) |
Following the same steps, 3D coordinates under the view of can also be transformed to 2D coordinates correspondingly, denoted as . All of this, the ultimate goal is to build the relation between and . After the view transformation, there is a relationship that has already been known, which can be written as:
(2) |
Besides, same as Equation 1, can be achieved by:
(3) |
From Equation 1 to Equation 3, the j-th location with the coordinates on the front view map could be related to the corresponding location on the bird view map , that is . Besides, as discussed in Section 3.1, each location has been assigned with the probability score of pedestrian location and the corresponding width, described as , which could be radiated to .
Above-mentioned formulations are designed to tackle the first task in ”What” branch: what the width of each location is on the front view map. There remains the other issue: what about the height is. Informally speaking, for the specific task of pedestrian detection, the difference in pedestrian height is, to a great extent, depth difference, in which the scale of an instance in the image is inversely proportional to the distance from the camera [37]. In another word, it means the farther away from the camera, the smaller the scale of the target will be. This fact allows us to model interdependency between depth and scale, and treat the process of proposal generation as an uneven distribution problem of scale.
First, still starting from Equation 1, the pedestrian height in one image could be modeled as the distance from head to foot , formulated as , while the real height in the real world could be mapped to , in which is a constant composed of camera intrinsic parameters and is the depth we mentioned. As analyzed above, (depth) and (height) are completely inverse when is the fixed value. However, due to the influence of gender and age in pedestrian height, it is slightly unreasonable to set as a fixed value directly. In our height estimates, we assume that the height of all pedestrians obeys the uniform distribution and we analyze the error of the assumption. Previous studies of 63,000 European adults have shown that the average height of males and females is 178cm and 165 cm respectively, with a standard deviation of 7cm in both cases [28]. Besides, the distribution of human stature follows a Gaussian distribution for male and female populations [13], denoted as: and , respectively. Specifically, for each point in the image , its pedestrian height in real world is a sample from the above distribution, and we utilize the normal distribution function to evaluate the uncertainty of the sampled height . Take the male distribution as an example. if the sampling height is , the corresponding uncertainty can be formulated as:
(4) |
in which is function of standard normal distribution. The above models the estimation uncertainty of due to variation of pedestrian height in real world, in which goes smaller when gets closer to , and instead goes larger. This estimation uncertainty also servers as the confidence value to re-weight loss function, described in Section 3.4. With the above formulated and , the corresponding proposal in the location of could be generated, and subsequently leads to the following ”Whether” branch.
3.3 Corpus: Whether the target is a pedestrian
After the ”Where” and ”What” branch, a large number of proposals have been generated, and the corresponding possibility score , which indicates whether there is an pedestrian instance or not in the bird view map, could also be projected to the front view. However, the single bird view prediction has the problem: the increase of false positives. Therefore, we introduce visual and attribute embedding, namely ”Whether” branch, to alleviate the above problem. This branch is based on two observations: one is that many false positives are mainly with human-like vertical structures, such as railings or tree trunks [33]; another is the fact that pedestrians, which perform walking or standing with up-straight pose, have one specific body structure from top to bottom: Head-Arm-Body-Leg. With this pedestrian-specific structure, we attempt to re-encode both occluded and non-occluded ones into the unified representation to distinguish those false positives with global vertical structure but lacking human-body based partial features.
The component of the proposed ”Whether” is shown in Figure 2. It consists of a variational autoencoder (VAE) [7], which is optimized by to learn a lantent space shared by both visual features and attribute embeddings. More specifically, the embedding network takes the attributes embedding vector as input, and after passing through two fully connected (fc) layers and the Rectified Linear Unit (ReLU), outputs a visual embedding vector, which has the same dimensions as the visual feature vector of each proposal extracted from the encoder, optimizing:
(5) |
where is the total number of all proposals, is the visual feature of the i-th proposal, represents the word vector of pedestrian attributes including Head, Arm, Body and Leg, represents the visual feature encoder and is the embedding network. The output feature via the decoder network used for the following classification and regression predicts the possibility score of whether there is a pedestrian or not, denoted as , which is combined with and thus the overall possibility score of the j-th proposal could be formulated as:
(6) |
3.4 Training Details
Towards this task, we resort to depth maps derived from a CNN-based monocular depth prediction model [1]. The feature generator consists of FPN [20] with a powerful backbone ResNet-50 [16], and the overall framework is optimized by the following loss function.
In the ”Where” branch, according to Equation 1 to 3, the horizontal and vertical central axes of each ground truth are mapped to the bird view, and thus constitute the bird view ground truth. Same as the loss function in CSP [23], we also formulate possibility and width prediction as a classification and regression task via the cross-entropy and smooth L1 loss, respectively, denoted as and .
In the ”What” branch, it is actually a geometric transformation without extra conv layers introduced. It is worth noting that in order to guarantee the robustness of the detector, pedestrian height in the real world is sampled from a distribution rather than using a fixed height value. Therefore, in order to evaluate the uncertainty of different samples, we introduce in Equation 4 and consequently re-weight it on the following classification and regression loss.
In the ”Whether” branch, features extracted from each proposal are first encoded to a latent space, and subsequently decoded as the final classification and regression features, optimized by the loss and . The overall loss can be formulated as:
(7) | ||||
where represents the i-th proposal and is the total number of all proposals. , , and are the weights for each loss, which are experimentally set as 0.01, 0.1, 0.1 and 1, respectively.
5 | 21.3 | 63.4 | |
---|---|---|---|
20 | 20.8 | 60.8 | |
5 | 18.2 | 50.6 | |
20 | 14.3 | 33.6 | |
5 | 15.5 | 34.9 | |
20 | 9.3 | 18.7 | |
5 | 16.7 | 40.9 | |
20 | 13.5 | 26.9 |
✓ | 9.3 | 18.7 | |||
✓ | 10.5 | 29.0 | |||
✓ | ✓ | 10.3 | 25.8 | ||
✓ | ✓ | 9.8 | 20.1 | ||
✓ | ✓ | 12.3 | 43.4 | ||
✓ | ✓ | 11.0 | 38.6 |
4 Experiments
We assess the effectiveness of our proposed method for pedestrian detection on widely used datasets Cityperson [34] and Caltech [8]. Results are the MR-2 evaluation metric, in which lower is better.
4.1 Experiment Setup
Datasets Citypersons [34] is a diverse dataset built upon the Cityscapes data, which includes 5000 images (2975 for training, 500 for validation, and 1525 for testing). In a total of 5 000 images, it has 35k person and 13k ignore region annotations. And it notices the density of persons are consistent across train/validation/test subsets. The Caltech Dataset [8] consists of approximately 10 hours of 640x480 30Hz video taken from a vehicle driving through regular traffic in an urban environment. About 250,000 frames with a total of 350,000 bounding boxes and 2300 unique pedestrians were annotated. All of the datasets contain challenging settings, denoted as Heavy occlusion, in which the propotion of visible parts of pedestrians is less than .
Implementation We implemented the proposed method in Pytorch with the backbone ResNet50 [16] and the Nvidia GTX1080Ti. We optimize the network using the Stochastic Gradient Descent (SGD) algorithm with 0.9 momentum and 0.0005 weight decay, respectively. For Citypersons, the mini-batch contains 2 images and we train the network for iterations with the initial learning rate of and decay it to for another iterations. For Caltech, the mini-batch contains 8 images and we train the network for iterations with the initial learning rate of and decay it to for another iterations.
4.2 Ablation
In this section, we evaluate how each significant component of our network contributes to performance using the Citypersons dataset under the Reasonable and Heavy settings, which are specification for the non-occluded and occluded pedestrian. It is worth noting that to evaluate the performance of each branch, such as results in Table 3, we take each ”W” step as one enhanced component, and replace the corresponding module in baseline (Faster RCNN, Table 3 Line 1). For example, to get the result of ”Where”, the process of proposal generation in RPN is replaced by ’head’ and width prediction in the bird view map together with fixed aspect ratio. Or to evaluate the ”What” step, we replace the fixed aspect ratio with the depth based scale estimation. The same is true for experiments in Table 4.
14.6 | 60.6 | |||
✓ | 11.0 | 23.3 | ||
✓ | 12.4 | 52.6 | ||
✓ | 10.8 | 30.2 | ||
✓ | ✓ | 9.9 | 21.7 | |
✓ | ✓ | 10.3 | 20.5 | |
✓ | ✓ | ✓ | 9.3 | 18.7 |
✓ | 10.9 | 24.6 | ||
✓ | 10.3 | 45.2 | ||
✓ | 11.0 | 31.0 |
W3Net (with ) | 13.4 | 32.1 |
---|---|---|
W3Net (with and ) | 9.3 | 18.7 |
Why is the bird view? Occlusion is one of the great challenges entrenched in pedestrian detection, which is ultimately caused by isolated view point, especially the front view. However, we could actually discover that in the real 3D world, even if a pedestrian is occluded from the initial view, bird view still performs free from occlusion issues. Besides, pedestrian instances are condensed to a point on the bird view, namely ”head”, which can be seamlessly connected with the anchor-free detectors. Inspired by above, experiments are conducted and comparisons are reported in Table 3. The method combined with ”Where” branch achieves 23.3 on Heavy occlusion subset, which is an absolute 37.3-point improvement over our baseline 60.6%. When the ”Where” branch is added to CSP, as is shown in Table 4, the performance of CSP is boosted by a large margin from 49.3 to 24.6. It is in accordance with our intuition that ”Where” branch is specially designed for the long-standing occlusion problem. In addition, ablations are carried out in Table 1 and 5 to investigate whether the difference in generation of bird view map will influence the results. Comparisons reported in Table 1 show that the view under and meters achieves the best performance, and we believe that more detailed exploration of the parameters and can further improve the performance, but it is not in the scope of this work. In Table 5, the method without domain transfer directly generate bird view images from the real-world data suffering a great decline on both Reasonable and Heavy subsets, which demonstrates the effectiveness of the proposed .

TLL [27] | 15.5 | 53.6 | - |
FRCNN [35] | 15.4 | - | - |
TLL+MRF [27] | 14.4 | 52.0 | - |
RepLoss [30] | 13.2 | 56.9 | - |
OR-CNN [36] | 12.8 | 55.7 | - |
ALFNet [22] | 12.0 | 51.9 | 0.27s/img |
CSP [23] | 11.0 | 49.3 | 0.33s/img |
W3Net(ours) | 9.3 | 18.7 | 0.31s/img |
Why is the depth? Depth is a useful auxiliary information, however, it has not been fully identified and utilized in pedestrian detection. One is that depth implies scale cues, especially for pedestrians, which share strong intra-class similarity. The other is that depth plays a connecting role in our framework. Depth, that is ”What” branch, projects detection results from ”Where” branch to the front view, and on the basis of interdependency between depth and scale, generating reasonable proposals for the following ”Whether” branch. As shown in Table 3 and 4, compared with the baseline or the bare CSP, the proposed ”What” branch goes a step further and achieves superior performance on both subsets. It is still worth noting that ”What” branch is just a kind of matrix transformation without any conv or fc layers introduced, which preserves the effectiveness of the detector while ensuring better performance. To further investigate the effect, we also conduct experiments including a prediction network to estimate pedestrian height, as reported in Table 2, in which the single means the detector combines width prediction on bird view with height estimation by depth-scale relations, while combines both width and height prediction on bird view and front view, respectively. It can be observed that other methods can achieve comparable but suboptimal results to prediction. This result may be attributed to the accurate and consistent height estimation with less noise during training, which has also been proved in Figure 3. Proposals generated by the proposed method shares better iou performance, in which the proportion of is improved greatly.

Why is the corpus? A natural question is, for occlusion, is the bird view enough? The bird view map can effectively provide the locations of all pedestrians. However, due to ambiguous pedestrian features, suspected objects, such as railings, which are similar to pedestrian ’head’ in the bird view, could also be falsely classified as the pedestrian. In order to tackle this problem, we introduce the ”Whether” branch, which falls into the proposal detection fashion on the basis of the front view map. The difficulty lies in proposal classification is that features of occluded pedestrians usually perform incomplete compared with those of non-occluded instances, that is to say, these two types have different feature distribution. Therefore, we re-encode proposals into a unified corpus space, which with fixed dimension and clear semantic information generates robust relations between visual and semantic cues and thus benefits robust feature representation. Results are reported in Table 3 and 4. It can be seen that although ”What” and ”Whether” branch both focus on occluded pedestrian detection, their detection uses different modal input, which maximises the complementary information of different views and thus achieve the best performance finally.
ACF++ [24] | 17.7 | 79.51 | 100 | 14.68 |
DeepParts [12] | 11.89 | 60.42 | 100 | 12.9 |
FasterRCNN+ATT [35] | 10.33 | 45.18 | 90.94 | 8.11 |
MS-CNN [4] | 9.95 | 59.94 | 97.23 | 8.08 |
RPN+BF [32] | 9.58 | 74.36 | 100 | 7.28 |
TLL [27] | 8.45 | - | 68.03 | - |
SDS-RCNN [3] | 7.36 | 58.55 | 100 | 6.44 |
RepLoss [30] | - | - | - | 4.0 |
OR-CNN [36] | - | - | - | 4.1 |
W3Net(Ours) | 6.37 | 28.33 | 51.05 | 3.82 |
4.3 Comparisons with the State-of-the-arts
Performance compared with state-of-the-art methods on Citypersons Reasonable, Partial and Heavy dataset is shown in Table 6, while the result on Caltech is reported in Table 7. It can be observed that 1) W3Net leads a new state-of-the-art result of 9.3% (Citypersons) and 3.82% (Caltech) on both Reasonable subsets.It is note worthy that at the same time, the proposed method also achieves promising performance on occluded instances. In particular,when evaluating heavy occlusion subset on Citypersons, our model produces18.7% while prior arts range from 49%-56%. 2) Result on Far subset of Caltech outperforms previous advanced detectors, such as TLL [27] by 16.98 which is specially designed for small-scale targets, demonstrates the superiority on scale variation. 3) Despite introducing multi-modal data, the careful discovery of depth relation actually reduces computational cost without extra conv or fc layers, and thus the inference time on one GTX 1080Ti with the 1x. image scale is still in line with the state-of-the-arts.
5 Conclusion
In this paper, we propose a novel network confronts challenges (occlusion and scale variation) preoccupied in pedestrian detection, which decouples the task into Where, What and Whether problem directing against pedestrian localization, scale estimation and classification correspondingly. As a result, the detector achieves the new state-of-the-art performance under various challenging settings. The framework is a customized design for pedestrian detection, but it is easy to extend into other tasks like face or vehicle detection, which need re-model interdependency between depth and scale, and is already in our future plans.
6 Acknowledgments
This work was partially funded by the National Science Fund of China under Grant No.61971281, the National Key Research and Development Program No.2017YFB1002401, STCSM No.18DZ1112300 and No.18DZ2270700.
References
- [1] A. Atapour-Abarghouei and T.P. Breckon. Real-time monocular depth estimation using synthetic data with domain adaptation. In Proc. Computer Vision and Pattern Recognition, pages 1–8. IEEE, June 2018.
- [2] Tadas Baltrušaitis, Chaitanya Ahuja, and Louis Philippe Morency. Multimodal machine learning: A survey and taxonomy. IEEE Transactions on Pattern Analysis & Machine Intelligence, PP(99):1–1, 2017.
- [3] G. Brazil, X. Yin, and X. Liu. Illuminating pedestrians via simultaneous detection segmentation. In ICCV, 2017.
- [4] Zhaowei Cai, Quanfu Fan, Rogerio S. Feris, and Nuno Vasconcelos. A unified multi-scale deep convolutional neural network for fast object detection. In ECCV, 2016.
- [5] Hao Cheng, Chong Yang Zhang, Wenjuan Song, Yan Li, and You Ping Zhong. Pedestrian detection with multi-scale context-embedded feature learning. In ACPR, 2017.
- [6] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, and Fei-Fei Li. Imagenet: a large-scale hierarchical image database. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, 2009.
- [7] Max Welling Diederik P Kingma. Auto-encoding variational bayes. In arXiv preprint arXiv:1312.6114., 2013.
- [8] P. Dollar, C. Wojek, B. Schiele, and P. Perona. Pedestrian detection: A benchmark. In CVPR, pages 304–311, 2009.
- [9] Xianzhi Du, Mostafa El-Khamy, Jungwon Lee, and Larry Davis. Fused dnn: A deep neural network fusion approach to fast and robust pedestrian detection. In Applications of Computer Vision, 2017.
- [10] Xianzhi Du, Mostafa El-Khamy, Vlad I. Morariu, Jungwon Lee, and Larry Davis. Fused deep neural networks for efficient pedestrian detection. 2018.
- [11] M Enzweiler and D. M. Gavrila. A multilevel mixture-of-experts framework for pedestrian classification. IEEE Transactions on Image Processing, 20(10):2967–2979, 2011.
- [12] P. Felzenszwalb, R. Girshick, D. McAllester, and D. Ramanan. Object detection with discriminatively trained part based models. In Pattern Analysis and Machine Intelligence, volume 32, 2010.
- [13] J V Freeman, T J Cole, S Chinn, P R Jones, E M White, and M A Preece. Cross sectional stature and weight reference curves for the uk, 1990. Archives of Disease in Childhood, 73(1):17–24.
- [14] A Gonzalez, D Vazquez, A. M. Lopez, and J Amores. On-board object detection: Multicue, multimodal, and multiview random forest of local experts. IEEE Transactions on Cybernetics, PP(99):1–11, 2016.
- [15] Ian J Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Xu Bing, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In International Conference on Neural Information Processing Systems, 2014.
- [16] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016.
- [17] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. Image-to-image translation with conditional adversarial networks. In Computer Vision and Pattern Recognition (CVPR), 2017 IEEE Conference on, 2017.
- [18] Tuan Anh Le, Atilim Gunes Baydin, Robert Zinkov, and Frank Wood. Using synthetic data to train neural networks is model-based reasoning. In arXiv preprint arXiv:1703.00868., 2017.
- [19] Jianan Li, Xiaodan Liang, Sheng Mei Shen, Tingfa Xu, and Shuicheng. Yan. Scale-aware fast r-cnn for pedestrian detection. In IEEE Transactions on Multimedia, volume PP, pages 1–1, 2015.
- [20] Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In CVPR, 2017.
- [21] Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng Yang Fu, and Alexander C. Berg. Ssd: Single shot multibox detector. In ECCV, 2016.
- [22] Wei Liu, Shengcai Liao, Weidong Hu, Xuezhi Liang, and Xiao Chen. Learning efficient single-stage pedestrian detectors by asymptotic localization fitting. In The European Conference on Computer Vision (ECCV), September 2018.
- [23] Wei Liu, Shengcai Liao, Weiqiang Ren, Weidong Hu, and Yinan Yu. High-level semantic feature detection: A new perspective for pedestrian detection. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [24] Eshed OhnBar and Mohan M. Trivedi. To boost or not to boost? on the limits of boosted trees for object detection. In ICPR, 2016.
- [25] Ravi S. Hegde Param S. Rajpura, Hristo Bojinov. Object detection using deep cnns trained on synthetic images. In arXiv preprint arXiv:1706.06782., 2017.
- [26] S. Ren, K. He, R Girshick, and J. Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In IEEE Transactions on Pattern Analysis and Machine Intelligence, volume 39, pages 1137–1149, 2017.
- [27] T. Song, L. Sun, D. Xie, H. Sun, and S. Pu. Small-scale pedestrian detection based on somatic topology localization and temporal feature aggregation. In ECCV, 2018.
- [28] Visscher and Peter M. Sizing up human height variation. 40(5):489–490.
- [29] Jiaqi Wang, Kai Chen, Shuo Yang, Chen Change Loy, and Dahua Lin. Region proposal by guided anchoring. In IEEE Conference on Computer Vision and Pattern Recognition, 2019.
- [30] Xinlong Wang, Tete Xiao, Yuning Jiang, Shao Shuai, Sun Jian, and Chunhua Shen. Repulsion loss: Detecting pedestrians in a crowd. In CVPR, 2017.
- [31] Wei Yin, Yifan Liu, Chunhua Shen, and Youliang Yan. Enforcing geometric constraints of virtual normal for depth prediction. In The IEEE International Conference on Computer Vision (ICCV), 2019.
- [32] Liliang Zhang, Lin Liang, Xiaodan Liang, and Kaiming He. Is faster r-cnn doing well for pedestrian detection? In ECCV, 2016.
- [33] Shanshan Zhang, Rodrigo Benenson, Mohamed Omran, Jan Hosang, and Bernt. Schiele. How far are we from solving pedestrian detection? In CVPR, 2016.
- [34] Shanshan Zhang, Rodrigo Benenson, and Bernt. Schiele. Citypersons: A diverse dataset for pedestrian detection. In CVPR, 2017.
- [35] Shanshan Zhang, Jian Yang, and Bernt Schiele. Occluded pedestrian detection through guided attention in cnns. In CVPR, 2018.
- [36] Shifeng Zhang, Longyin Wen, Xiao Bian, Zhen Lei, and Stan Z. Li. Occlusion-aware r-cnn: Detecting pedestrians in a crowd. In ECCV, 2018.
- [37] Muming Zhao, Jian Zhang, Chongyang Zhang, and Wenjun Zhang. Leveraging heterogeneous auxiliary tasks to assist crowd counting. In CVPR, 2019.
- [38] Chengju Zhou, Meiqing Wu, and Siew Kei Lam. Ssa-cnn: Semantic self-attention cnn for pedestrian detection. 2019.
- [39] Chunluan Zhou and Junsong Yuan. Bi-box regression for pedestrian detection and occlusion estimation. In ECCV, 2018.
- [40] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networkss. In Computer Vision (ICCV), 2017 IEEE International Conference on, 2017.