This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Estimating Distances Between People using a Single Overhead Fisheye Camera with Application to Social-Distancing Oversight

Zhangchi Lu, Mertcan Cokbas, Prakash Ishwar, Janusz Konrad
Department of Electrical and Computer Engineering, Boston University
8 Saint Mary’s Street, Boston, MA 02215
[zhchlu, mcokbas, pi, jkonrad]@bu.edu
This work was supported by ARPA-E (agreement DE-AR0000944) and by Boston University Undergraduate Research Opportunities Program.
Abstract

Unobtrusive monitoring of distances between people indoors is a useful tool in the fight against pandemics. A natural resource to accomplish this are surveillance cameras. Unlike previous distance estimation methods, we use a single, overhead, fisheye camera with wide area coverage and propose two approaches. One method leverages a geometric model of the fisheye lens, whereas the other method uses a neural network to predict the 3D-world distance from people-locations in a fisheye image. To evaluate our algorithms, we collected a first-of-its-kind dataset using single fisheye camera, that comprises a wide range of distances between people (1–58 ft) and will be made publicly available. The algorithms achieve 1–2 ft distance error and over 95% accuracy in detecting social-distance violations.

1 Introduction

The general problem of depth/distance estimation in 3D world has been studied in computer vision from its beginnings. However, the narrower problem of estimating the distance between people has gained attention only recently. In particular, the COVID pandemic has sparked interest in inconspicuous monitoring of social-distance violations (e.g., less than 6ft) [8, 10, 18, 21, 11, 1, 16]. A natural, cost-effective resource that can be leveraged to accomplish this goal are the surveillance cameras widely deployed in commercial, office and academic buildings.

Recent methods developed for the estimation of 3D distance have typically used 2 cameras (stereo) equipped with either rectilinear [5, 12] or fisheye [14, 20] lenses. Stereo-based methods, however, require careful camera calibration (both intrinsic and extrinsic parameters) and are very sensitive to misalignments between cameras (translation and rotation) after calibration. Although methods have been proposed using single rectilinear-lens camera [10, 18, 11, 16, 1], that do not suffer from stereo calibration and misalignment shortcomings, usually one such camera can cover only a fragment of a large space. While multiple cameras can be deployed, this increases the cost and complexity of the system.

In this paper, we focus on estimating the distance between people indoors using a single overhead fisheye camera with 360×{}^{\circ}\times180 field of view. Such a camera can effectively cover a room up to 2,000ft2 greatly reducing deployment costs compared to multiple rectilinear-lens cameras. However, fisheye cameras introduce geometric distortions so methods developed for rectilinear-lens cameras are not directly applicable; the geometric distortions must be accounted for when estimating distances in 3D space.

We propose two methods to estimate the distance between people using a single fisheye camera. The first method leverages a fisheye-camera model and its calibration methodology developed by Bone et al.  [3] to inverse-project location of a person from fisheye image to 3D world. This inverse projection suffers from scale (depth) ambiguity that we address by using a human-height constraint. Knowing the 3D-world coordinates of two people we can easily compute the distance between them. Unlike the first method based on camera geometry, the second method uses the Multi-Layer Perceptron (MLP) and is data-driven. In order to train the MLP, we collected training data using a large chess mat. For testing both methods, we collected another dataset with people placed in various locations of a 72×\times28-foot room. The dataset includes over 300 pairs of people with over 70 different distances between them. Unlike other inter-people distance-estimation datasets, our dataset, due to the large field of view, includes a wide range of distances between people (from 1 ft to 58 ft). We call this dataset Distance Estimation between People from Overhead Fisheye cameras (DEPOF).

The main contributions of this work are:

  1. 1.

    We propose two approaches for distance estimation between people using a single overhead fisheye camera. To the best of our knowledge no such approach has been developed to date.

  2. 2.

    We created a fisheye-camera dataset for the evaluation of inter-people distance-estimation methods. This is the first dataset of its kind that is publicly available at vip.bu.edu/depof

2 Related Work

In the last two years, spurred by the COVID pandemic, a number of methods have been developed to estimate distances between people indoors. Most of these methods comprise two key steps: the detection of people in an image, and estimation of the 3D-world distance between people.

In order to detect people/objects, some methods [21, 15] rely on YOLO, other methods [18, 10] use Faster R-CNN and still other methods [8] use GMM-based foreground detection. However, this is not the focus of this paper; we assume that bounding boxes around people are available.

To estimate the distance between detected people, a number of approaches have emerged that use a single camera with rectilinear lens. Some approaches rely on typical dimensions of various body parts (e.g., shoulder width) [1, 16], while others perform a careful camera calibration [10, 11, 18] to infer inter-person distances. Also, stereo-based methods (two cameras) have been recently proposed to estimate the distance to a person/object [5, 12], but they require very precise camera calibration and are sensitive to post-calibration misalignments.

Very recently, a single overhead fisheye camera was proposed to detect social distance violations in buses (which is coarser goal than distance estimation), but no quantitative results were published [19]. Fisheye-stereo is often used in front-facing configuration for distance estimation in autonomous navigation [14, 20], but recently it was proposed in overhead configuration for person re-identification indoors based on location rather than appearance [3]. To accomplish this, the authors developed a novel calibration method to determine both intrinsic and extrinsic fisheye-camera parameters. We leverage this study to calibrate our single fisheye camera and we use a geometric model developed therein.

In terms of benchmark datasets for estimating distances between people, Epfl-Mpv-VSD, Epfl-Wildtrack-VSD, OxTown-VSD [1] and KORTE [16] are prime examples. Out of them only Epfl-Mpv-VSD and KORTE include some indoor scenes. More importantly, however, all of them have been collected with rectilinear-lens cameras, and are not useful for our study. In fact, the first three datasets have been derived by Aghaei et al. [1] from Epfl-Mpv[7], Epfl-Wildtrack [4] and OxTown [2] datasets that are designed for people detection and tracking. In contrast, our datasets have been specifically designed for the estimation of distances between people in an indoor, large-space setting under a variety of occlusion scenarios.

3 Methodology

We focus on large indoor spaces monitored by a single, overhead, fisheye camera. An example of an image captured in this scenario is shown in Fig. 1. We propose two methods to measure the distance between two people visible in such an image. One method uses a geometric model of a previously calibrated camera while the other makes no assumptions about the camera and is data-driven. Although these methods are well-known, we apply them in a unique way to address the distance estimation problem using a single fisheye camera.

In this work, we are not concerned with the detection of people; this can be accomplished by any recent method developed for overhead fisheye cameras [17, 13, 6]. Therefore, we assume that tight bounding boxes around people are given. Furthermore, we assume that the center of a bounding box defines the location of the detected person.

Let 𝒙A,𝒙B2\bm{x}_{A},\bm{x}_{B}\in\mathbb{Z}^{2} be the pixel coordinates of bounding-box centers for person AA and person BB, respectively. Given a pair (𝒙A,𝒙B)(\bm{x}_{A},\bm{x}_{B}), the task is to estimate the 3D-world distance between people captured by the respective bounding boxes. Below, we describe two methods to accomplish this.

Refer to caption
Figure 1: Field of view from an Axis M3057-PLVE camera mounted on the ceiling of a 72×\times28 ft2 classroom and illustration of height adjustment (see Section 3.3 for details).

3.1 Geometry-based method

In this approach, to estimate the 3D-world distance between two people we adopt the unified spherical model (USM) proposed by Gayer and Danilidis [9] for fisheye cameras and a calibration methodology to find this model’s parameters developed by Bone et al. [3]. This model, derived in [3], enables computation of an inverse mapping from image coordinates to 3D space as described next.

Consider the scenario in Fig. 2 where the center of the 3D-world coordinate system is at the optical center of a fisheye camera mounted overhead at height BB above the floor and a person of height HH stands on the floor. Let a 3D-world point 𝑷=[Px,Py,Pz]T3\bm{P}=[P_{x},P_{y},P_{z}]^{T}\in\mathbb{R}^{3} be located on this person’s body at half-height and let 𝑷\bm{P} appear at 2D coordinates 𝒙\bm{x} in the fisheye image.

Bone et al. [3] showed that the 3D-world coordinates 𝑷\bm{P} can be recovered from 𝒙\bm{x} with knowledge of PzP_{z} and a 5-vector of USM parameters 𝝎{\bm{\omega}} via a non-linear function GG:

𝑷=G(𝒙,Pz;𝝎),\bm{P}=G(\bm{x},P_{z};{\bm{\omega}}), (1)

In order to estimate 𝝎{\bm{\omega}}, an automatic calibration method using a moving LED light was developed in [3]. In addition to 𝝎\bm{\omega}, the value of PzP_{z} is needed since this is a 2D-to-3D mapping. However, based on Fig. 2 we see that Pz=BH/2P_{z}=B-H/2.

Refer to caption
Figure 2: Illustration of the relationship between PzP_{z} and person’s height HH.

In practice, we can only get a pixel-quantized estimate 𝒙^\widehat{\bm{x}} of 𝒙\bm{x} from which we can compute an estimate 𝑷^\widehat{\bm{P}} of 𝑷\bm{P} using (1). Let 𝑷^A\widehat{\bm{P}}_{A} and 𝑷^B\widehat{\bm{P}}_{B} denote the estimated 3D-world coordinates of person AA and person BB, respectively, based on the centers of their bounding boxes 𝒙^A\widehat{\bm{x}}_{A} and 𝒙^B\widehat{\bm{x}}_{B}. Then, we can estimate the 3D-world Euclidean distance d^AB\widehat{d}_{AB} between them via:

d^AB=𝑷^A𝑷^B2.\widehat{d}_{AB}=||\widehat{\bm{P}}_{A}-\widehat{\bm{P}}_{B}||_{2}. (2)

3.2 Neural-network approach

In this approach, we train a neural network to estimate the distance between person AA and person BB. Since the distance between two points in a fisheye image is invariant to rotation, we pre-process locations 𝒙A\bm{x}_{A} and 𝒙B\bm{x}_{B} before feeding them into the network. First, we convert 𝒙A\bm{x}_{A} and 𝒙B\bm{x}_{B} to polar coordinates: 𝒙A(rA,θA)\bm{x}_{A}\rightarrow(r_{A},\theta_{A}) and 𝒙B(rB,θB)\bm{x}_{B}\rightarrow(r_{B},\theta_{B}), where rr_{\cdot} denotes radius and θ\theta_{\cdot} denotes angle. Then, we compute the angle between normalized locations as follows:

θ:=(θAθB) mod π.\theta:=(\theta_{A}-\theta_{B})\textrm{ mod }\pi. (3)

Note that by its definition, 0θπ0\leq\theta\leq\pi. Finally, we form a feature vector associated with locations 𝒙A\bm{x}_{A} and 𝒙B\bm{x}_{B} as follows: 𝑽=[rA,rB,θ]T\bm{V}=[r_{A},r_{B},\theta]^{T}. We chose a regression Multi-Layer Perceptron (MLP) to estimate the 3D-world distance between people (in lieu of a CNN) since the input vector is a 3-vector with no required ordering of coordinates for which convolution would be beneficial. We collected a training set of images, where for each vector 𝑽\bm{V} we know the ground-truth distance dABd_{AB}, and trained the MLP, F:3F:\mathbb{R}^{3}\mapsto\mathbb{R}, as a regression model that performs the following mapping:

d^AB=F(𝑽).\widehat{d}_{AB}=F(\bm{V}). (4)

In training, we used the mean squared-error (MSE) loss:

=1Mi=1Md^ABidABi2,\mathcal{L}=\frac{1}{M}\sum_{i=1}^{M}||\widehat{d}_{AB_{i}}-d_{AB_{i}}||^{2}, (5)

where MM is the batch size.

3.3 Person’s height adjustment

While the geometry-based approach can be tuned for specific height of a person through PzP_{z} (1), the neural-network approach would require a training dataset having a large number of annotated examples at multiple heights. Since this is extremely labor intensive, we train the MLP at a single height of 32.5 inches (details in Section 4.1) which corresponds to one-half of H=H= 65 in, an average person’s height. Recall that we assume a person’s location in the image is the center-point of the person’s bounding box. Therefore, for a standing, fully-visible 65-inch person this center-point matches the 32.5-inch training height well. However, there would be a mismatch for people of other heights or when a person is partially occluded, for example by a table. In the latter case, the detected bounding box would be above the table and so would be the bounding-box center. To compensate for this height mismatch between the training and testing data, we propose a test-time adjustment in the MLP approach.

This height adjustment can be thought of as lowering the center-point of a person in terms of pixel coordinates. An illustration of this idea is shown in Fig. 1 where the red point represents the center of the red bounding box and hh its height. In the process of height adjustment during test time, we move the actual center (red point) of the bounding box along the box’s axis pointing to the center of the image (white-dashed line) to produce an adjusted center (green point). This displacement is defined as α×h2\alpha\times\frac{h}{2} and we consider a range of values for α\alpha (see Fig. 5). A value of α>0\alpha>0 corresponds to moving the bounding-box center towards the image center, i.e., we reduce the height of a detected person.

Refer to caption

        (a) Layout of chessboard mats          (b) Chessboard mat at position #4             (c) Chessboard mat at position #13

Figure 3: Illustration of chessboard-mat layout used for training the MLP model.

4 Datasets

We introduce a unique dataset, Distance Estimation between People from Overhead Fisheye cameras (DEPOF) 111vip.bu.edu/depof which was collected with Axis M3057-PLVE cameras at 2,048×\times2,048-pixel resolution.

4.1 Training dataset

In order to train the MLP, we need ground-truth distance data. We placed a 9 ft ×\times 9 ft chessboard mat on classroom tables of equal height (32.5 inches) in 19 different locations as shown in Fig. 3. The mat at location #2 was carefully placed with sides parallel to those of the mat at location #1 and forming a contiguous extension of mat #1 (as if two mats were placed abutting each other). Similarly, the mat at location #3 was placed contiguously with respect to the mat at location #2 and so on until location #8. The same procedure was repeated for locations #9-#16 with the mat at location #9 carefully aligned with the mat at location #1 and the distance between them carefully measured (121.5 inches). In order to provide ground-truth data in the center of camera’s field of view, the mat was also placed directly under each of 3 cameras (locations #17, #18, #19 shown in orange) with no alignment to mats at other locations.

Refer to caption
Figure 4: Spatial layout of locations in testing datasets.

All the black/white corners of chessboard images were annotated, resulting in numerous (𝒙A,𝒙B)(\bm{x}_{A},\bm{x}_{B}) pairs. Since we know that each square is of size 12.5 inches and the neighboring chessboards are abutting and aligned, we could accurately compute the 3D distances between physical-mat points corresponding to 𝒙A\bm{x}_{A} and 𝒙B\bm{x}_{B} The overall process can be thought of as creating a virtual grid with 12.5-inch spacing placed 32.5 inches above the floor of the classroom.

4.2 Testing datasets

In order to test both of the proposed approaches, we collected a dataset with people in a 72 ft ×\times 28 ft classroom. First, we marked locations on the floor where individuals would stand (Fig. 4). We measured distances between all locations marked by a letter (green disk) which gives us (102)=45{10\choose 2}=45 distances which turned out to be distinct. For locations marked by a number (yellow squares), we measured the distances along the dashed lines (20 distinct distances). Using this spatial layout, we collected and annotated two sets of data.

  • Fixed-height dataset: One person of height H=H= 70.08 inches moves from one marked location to another and an image is captured at each location. This allows us to evaluate our algorithms on people of the same known height.

  • Varying-height dataset: Several people of different heights stand at different locations in various permutations to capture multiple heights at each location. We use this dataset to evaluate sensitivity of our algorithms to a person’s height variations.

In addition to the 65 distances (45+2045+20), we performed 88 additional measurements for the fixed-height dataset and 22 additional measurements for the varying-height dataset.

Depending on their location with respect to the camera, a person may be fully visible or partially occluded (e.g., by a table or chair). In order to understand the impact of occlusions on distance estimation, we grouped all the pairs in both testing datasets into 4 categories as follows:

  • Visible-Visible (V-V): Both people are fully visible.

  • Visible-Occluded (V-O): One person is visible while the other one is partially occluded.

  • Occluded-Occluded (O-O): Both people are partially occluded.

  • All: All pairs, regardless of the occlusion status.

Table 1 shows various statistics for both datasets: the number of pairs in each category, the number of distances measured and their range as well as the number of pairs with distance in three ranges: 0ft–6ft, 6ft–12ft and >>12ft.

Table 1: Statistics of the testing datasets.
Fixed- Varying-
height height
dataset dataset
Number of V-V pairs 35 100
Number of V-O pairs 32 126
Number of O-O pairs 6 30
Number of All pairs 73 256
Number of distances 73 67
Smallest distance (G to 11) 11.63 inches
Largest distance (A to J) 701.96 inches
Number of pairs: 0 ft to 6 ft 25 45
Number of pairs: 6 ft to 12 ft 15 73
Number of pairs: above 12 ft 33 138

In both datasets, we found bounding boxes of people using a state-of-the-art people-detection algorithm for overhead fisheye images [6] from which we computed their locations (bounding-box centers). To measure the real-world distances between people, we used a laser tape measure.

5 Experimental Results

In this section, we evaluate both of the proposed methods. First, we describe the experimental setup. Then, we compare the methods in terms of distance estimation accuracy and evaluate the impact of person’s height adjustment on performance. Finally, we assess the ability of both methods to detect social-distancing violations.

5.1 Experimental setup

In the geometry-based approach, to learn parameters 𝝎{\bm{\omega}} of the inverse mapping GG (1) we used the method described by Bone et al. [3]. This method requires the use of 2 fisheye cameras, but is largely automatic and has to be applied only once for a given camera type (model and manufacturer). In the experiments, we used one camera at a time (3 cameras are installed in the test classroom – locations #17-#19 in Fig. 3) and report the results only for the center camera due to space constraints. Results for other cameras are similar.

In the neural-network approach, we used an MLP with 4 hidden layers and 100 nodes per layer. In training, we used MSE loss (5) and Adam optimizer with 0.001 learning rate.

5.2 Distance estimation evaluation

In Tables 2 and 3, we compare the performance of both methods when estimating the distance between people on the fixed-height and varying-height datasets, respectively. We report the mean absolute error (MAE) between the estimated and ground-truth distances:

MAE=1Ni=1N|d^ABidABi|\textrm{MAE}=\frac{1}{N}\sum_{i=1}^{N}|\widehat{d}_{AB_{i}}-d_{AB_{i}}| (6)

where NN is the number of pairs in the dataset while d^ABi\widehat{d}_{AB_{i}} and dABid_{AB_{i}} are the estimated and ground-truth distances for the ii-th pair ABAB, respectively.

It is clear from Table 2 that the geometry-based approach using H/2=H/2= 35.04 inches (to compute PzP_{z}) consistently outperforms the same approach using H/2=H/2= 32.5 inches, which, in turn, significantly outperforms the neural-network approach trained on chess mats placed at the height of 32.5 inches. While it is not surprising that knowing a test-person’s height of H=H= 70.08 inches improves geometry-based method’s accuracy, it is interesting that even assuming H/2=H/2= 32.5 inches the geometry-based approach significantly outperforms the MLP optimized for a fixed height of 32.5 inches during training.

Table 2: Mean-absolute distance error between two people for the fixed-height dataset.
MAE [in]
V-V V-O O-O All
Geometry-based 9.85 31.69 32.30 21.27
(H/2=H/2= 35.04 in)
Geometry-based 12.20 39.90 42.64 26.84
(H/2=H/2= 32.5 in)
Neural network 17.72 48.84 56.58 34.56
(trained on 32.5 in)

Similar performance trends can be observed in Table 3 for the varying-height dataset but with larger distance-error values than in Table 2. This is due to the fact that in the varying-height dataset people have different heights, so a selected parameter HH in the geometry-based algorithm or a training height in the neural-network algorithm cannot match all people’s heights at the same time.

Table 3: Mean-absolute distance error between two people for the varying-height dataset.
MAE [in]
V-V V-O O-O All
Geometry-based 14.70 41.37 55.55 32.62
(H/2=H/2= 35.04 in)
Geometry-based 20.18 51.14 67.61 40.98
(H/2=H/2= 32.5 in)
Neural network 24.64 55.88 70.87 45.43
(trained on 32.5 in)

Note that for two fully-visible people of the same and known height (Table 2), the geometry-based algorithm has an average distance error of less than 10 inches. This error grows to about 21 inches for all pairs (visible and occluded). For people of different and unknown heights (Table 3), the average error for pairs of fully-visible individuals (for H/2=H/2= 35.04 inches) is slightly above 1 ft and for all pairs it is less than 3 ft. While these might seem to be fairly large distance errors, one has to note that the distances between people are as large as 58.5 ft (702 inches).

In terms of the computational complexity, on an Intel(R) Xeon(R) CPU E5-2680 [email protected] both algorithms can easily support real time operation although the geometry-based algorithm is significantly faster. For example, suppose 3D-world distances are to be computed between all pairs of 100 image locations. The geometry-based algorithm can first map all pixel coordinates to 3D world coordinates (1) and then compute the Euclidean distance for all (1002)={100\choose 2}= 4,950 pairs. This, on average, takes 4 μs\mu s. The neural-network algorithm has to apply the MLP to all 4,950 pairs separately taking on average 949 μs\mu s.

5.3 Impact of a person’s height adjustment

As we discussed in Section 3.3, the centers of the detected bounding boxes may not reflect the true height of a person due to occlusions. In this context, we proposed a method to adjust a bounding-box center location during testing to compensate for the occlusion effect.

Here, we evaluate the impact of this height adjustment on each method’s performance. For a fair comparison, we use H/2=H/2= 32.5 inches (table height) in the geometry-based method. Recall that the neural-network approach was trained on chess mats placed at this height.

The value of MAE as a function of height-adjustment parameter α\alpha is shown in Figs. 5(a) and 5(b) for the fixed-height dataset. When the true bounding-box centers are used (α=0\alpha=0), the MAE for the neural-network approach and V-V pairs (blue line) is close to 20 inches. However, when the centers are lowered by about 10% (α\alpha\approx 0.10), the MAE for V-V pairs drops to about 9 in. Similar trends can be observed in Figs. 5(c) and 5(d) for the varying-height dataset. If the true bounding-box centers are used, the MAE for V-V pairs is above 20 inches for the neural-network approach. However, when the centers are lowered by about 15% (α\alpha\approx 0.15), the MAE for V-V pairs drops to around 12 inches. Analogous trends can be seen for other types of pairs and for all pairs, as well as for the geometry-based approach.

Refer to caption
(a) Neural-network algorithm, fixed-height dataset
Refer to caption
(b) Geometry-based algorithm, fixed-height dataset
Refer to caption
(c) Neural-network algorithm, varying-height dataset
Refer to caption
(d) Geometry-based algorithm, varying-height dataset
Figure 5: MAE for height adjustments: 0.1α<1.0-0.1\leq\alpha<1.0.

In Tables 4 and 5, we show the lowest MAE values for each pair type along with the corresponding value of α\alpha. The two methods perform quite similarly (except for O-O pairs in the fixed-height dataset on which the geometry-based method performs better). For example, for the fixed-height dataset in Table 4 MAE for the best α\alpha for V-V pairs drops to about 9 inches for both algorithms compared to 12-18 inches seen in Table 2. Overall, in both datasets, with the right choice of α\alpha, the MAE is well below 2 ft, which can be argued to be a reasonable result considering that the inter-people distances in our dataset are up to 58.5 ft.

Looking at Fig. 5 and Tables 4-5, one notes that MAE is minimized for much smaller values of α\alpha for V-V pairs (α=\alpha= 0.08-0.17) than for O-O pairs (α=\alpha= 0.32-0.51). This is because the majority of occlusions happen in the lower half of people’s bodies in the testing dataset. When a person is blocked in the bottom half, the bounding-box center radially shifts away from the image center. An example of this can be seen in Fig. 1, where the person delineated by the yellow bounding box would have been delineated by the blue bounding box had there been no occlusion. Due to the occlusion, however, the bounding-box center shifts from the blue point to the yellow point. Therefore, the O-O pairs need to be compensated more than the V-V pairs, i.e., a higher value of α\alpha is needed.

Table 4: Best mean-absolute distance error between two people for both methods using height adaptation for the fixed-height dataset. Under each MAE value, the corresponding optimum “α\alpha” value used is written in parentheses.
MAE [in]
(α\alpha)
V-V V-O O-O All
Geometry-based 9.36 21.07 8.31 18.10
(H/2=H/2= 32.5in) (0.08) (0.28) (0.32) (0.18)
Neural network 8.79 22.20 11.24 19.27
(Trained on 32.5 in) (0.12) (0.33) (0.42) (0.26)
Table 5: Best mean-absolute distance error between two people for both methods using height adaptation for the varying-height dataset. Under each MAE value, the corresponding optimum “α\alpha” value used is written in parentheses.
MAE [in]
(α\alpha)
V-V V-O O-O All
Geometry-based 12.76 20.49 18.66 21.48
(H/2=H/2= 32.5in) (0.12) (0.41) (0.48) (0.38)
Neural network 11.62 21.30 18.60 21.47
(Trained on 32.5 in) (0.17) (0.45) (0.51) (0.41)

In results reported thus far, the same value of α\alpha was used for both people in every pair. In the V-O and ‘All’ categories, however, it could be advantageous to use different values of α\alpha for the visible and occluded person. To verify this hypothesis, we applied α=0.1\alpha=0.1 to all visible bounding boxes and α=0.5\alpha=0.5 to all occluded bounding boxes in the fixed-height dataset. This α\alpha adjustment per person gave an MAE of 12.80 inches (down from 18.10 inches) for the geometry-based algorithm and 11.06 inches (down from 19.27 inches) for the neural-network approach. The corresponding MAE values for the varying-height dataset were: 13.97 inches (down from 21.48 inches) and 13.14 inches (down from 21.47 inches). Clearly, an automatic detection of body occlusions and a suitable adjustment of parameter α\alpha can further improve the distance estimation accuracy. This could be a fruitful direction for future work.

5.4 Application in the context of social distancing

One very practical application of the proposed methods is to detect situations when social-distancing recommendations (typically 6 ft) are being violated. This problem can be cast as binary classification: two people closer to each other than 6 ft are considered to be a “positive” case (violation takes place) whereas two people more than 6 ft apart are considered to be a “negative” case (no violation). To measure performance, we use Correct Classification Rate (CCR) and F1-score. Table 6 shows their values for both algorithms applied in this context to “All” pairs. We report results for α=\alpha= 0.5 which gives the lowest MAE for pairs with occlusions on the varying-height dataset (Table 5).

On the fixed-height dataset, the neural-network approach slightly outperforms the geometry-based algorithm: by 1.37% points in terms of CCR and by 0.63% points in terms of F1-score. The methods perform identically on the varying-height dataset, achieving CCR value close to 95% and F1-score close to 80%. These results suggest that despite the presence of people of different heights both approaches achieve high enough CCR and F1-score values to be potentially useful in practice for the detection of social-distance violations in the wild.

Table 6: Social-distance violation detection results (α\alpha=0.5).
Fixed-height Varying-height
dataset dataset
CCR F1 CCR F1
[%] [%] [%] [%]
Geometry-based 94.52 91.14 94.53 79.69
(H/2H/2=32.5in)
Neural network 95.89 91.77 94.53 79.69
(Trained on 32.5in)

6 Concluding Remarks

We developed two methods (the first of their kind) for estimating the distance between people in indoor scenarios based on a single image from a single overhead fisheye camera. Demonstrating the ability to accurately measure the distance between people from a single overhead fisheye camera (with its wide FOV) has practical utility since it can decrease the number of cameras (and cost) needed to monitor a given area. A novel methodological contribution of our work is the use of a height-adjustment test-time pre-processing operation which makes the distance estimates resilient to height variation of individuals as well as body occlusions. We demonstrated that both methods can achieve errors on the order of 10-20in for suitable choices of height-adjustment tuning parameters. We also showed that both of our methods can predict social distance violation with a high F1-score and accuracy.

References

  • [1] M. Aghaei, M. Bustreo, Y. Wang, G. Bailo, P. Morerio, and A. D. Bue. Single image human proxemics estimation for visual social distancing. 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), pages 2784–2794, 2021.
  • [2] B. Benfold and I. Reid. Stable multi-target tracking in real-time surveillance video. In CVPR 2011, pages 3457–3464, 2011.
  • [3] J. Bone, M. Cokbas, O. Tezcan, J. Konrad, and P. Ishwar. Geometry-based person re-identification in fisheye stereo. In 2021 17th IEEE Int. Conf. on Advanced Video and Signal Based Surveillance (AVSS), pages 1–10, 2021.
  • [4] T. Chavdarova, P. Baqué, S. Bouquet, A. Maksai, C. Jose, T. Bagautdinov, L. Lettry, P. Fua, L. Van Gool, and F. Fleuret. Wildtrack: A multi-camera hd dataset for dense unscripted pedestrian detection. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5030–5039, 2018.
  • [5] E. Dandil and K. K. Cevik. Computer vision based distance measurement system using stereo camera view. In 2019 3rd International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT), pages 1–4, 2019.
  • [6] Z. Duan, M. Ozan T., H. Nakamura, P. Ishwar, and J. Konrad. Rapid: Rotation-aware people detection in overhead fisheye images. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 2700–2709, 2020.
  • [7] F. Fleuret, J. Berclaz, R. Lengagne, and P. Fua. Multicamera people tracking with a probabilistic occupancy map. IEEE Transactions on Pattern Analysis and Machine Intelligence, 30(2):267–282, 2008.
  • [8] A. Gad, G. ElBary, M. Alkhedher, and M. Ghazal. Vision-based approach for automated social distance violators detection. In 2020 Int. Conf. on Innovation and Intelligence for Informatics, Computing and Technologies (3ICT), pages 1–5, 2020.
  • [9] C. Geyer and K. Danilidis. Catadioptric projective geometry. Int. J. Computer Vision, 45:223–243, 2001.
  • [10] S. Gupta, R. Kapil, G. Kanahasabai, S. Joshi, and A. Joshi. Sd-measure: A social distancing detector. In 2020 Int. Conf. on Computational Intelligence and Communication Networks (CICN), pages 306–311, 2020.
  • [11] Y. C. Hou, M. Z. Baharuddin, S. Yussof, and S. Dzulkifly. Social distancing detection with deep learning model. In 2020 8th Int. Conf. on Information Technology and Multimedia (ICIMU), pages 334–338, 2020.
  • [12] P. N. Huu, T. Tran Van, and N. G. Thi. Proposing distortion compensation algorithm for determining distance using two cameras. In 2019 NAFOSTED Conf. on Information and Computer Science (NICS), pages 172–177, 2019.
  • [13] S. Li, M. O. Tezcan, P. Ishwar, and J. Konrad. Supervised people counting using an overhead fisheye camera. In 2019 16th IEEE Int. Conf. on Advanced Video and Signal Based Surveillance (AVSS), pages 1–8, 2019.
  • [14] A. Ohashi, Y. Tanaka, G. Masuyama, K. Umeda, D. Fukuda, T. Ogata, T. Narita, S. Kaneko, Y. Uchida, and K. Irie. Fisheye stereo camera using equirectangular images. In 2016 11th France-Japan and 9th Europe-Asia Congress on Mechatronics / 17th Int. Conf. on Research and Education in Mechatronics (REM), pages 284–289, 2016.
  • [15] X. Pan, Z. Yi, and J. Tao. The research on social distance detection on the complex environment of multi-pedestrians. In 2021 33rd Chinese Control and Decision Conference (CCDC), pages 763–768, 2021.
  • [16] M. Seker, A. Männistö, A. Iosifidis, and J. Raitoharju. Automatic social distance estimation from images: Performance evaluation, test benchmark, and algorithm. CoRR, abs/2103.06759, 2021.
  • [17] M. Tamura, S. Horiguchi, and T. Murakami. Omnidirectional pedestrian detection by rotation invariant training. In 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), pages 1989–1998, 2019.
  • [18] J. M. Tellis, S. Jaiswal, R. Kabra, and P. Mehta. Monitor social distance using convolutional neural network and image transformation. In 2021 IEEE Bombay Section Signature Conference (IBSSC), pages 1–4, 2021.
  • [19] D. Tsiktsiris, A. Lalas, M. Dasygenis, K. Votis, and D. Tzovaras. An efficient method for addressing covid-19 proximity related issues in autonomous shuttles public transportation. In I. Maglogiannis, L. Iliadis, J. Macintyre, and P. Cortez, editors, Artificial Intelligence Applications and Innovations, pages 170–179. Springer International Publishing, 2022.
  • [20] F. Yamano, H. Iida, K. Umeda, A. Ohashi, D. Fukuda, S. Kaneko, J. Murayama, and Y. Uchida. Improving the accuracy of a fisheye stereo camera with a disparity offset map. In 2018 12th France-Japan and 10th Europe-Asia Congress on Mechatronics, pages 94–97, 2018.
  • [21] M. Yeshasvi, V. Bind, and S. T. Social distance capturing and alerting tool. In 2021 3rd Int. Conf. on Signal Processing and Communication (ICPSC), pages 568–572, 2021.