This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Unveiling personnel movement in a larger indoor area with a non-overlapping multi-camera system

Ping Zhang Zhenxiang Tao Wenjie Yang Minze Chen Shan Ding Xiaodong Liu Rui Yang [email protected] Hui Zhang Department of Engineering Physics, Tsinghua University, Beijing 100084, China Institute of Public Safety Research, Tsinghua University, Beijing 100084, China Public Order School, People’s Public Security University of China, 100038, Beijing, China
Abstract

Surveillance cameras are widely applied for indoor occupancy measurement and human movement perception, which benefit for building energy management and social security. To address the challenges of limited view angle of single camera as well as lacking of inter-camera collaboration, this study presents a non-overlapping multi-camera system to enlarge the surveillance area and devotes to retrieve the same person appeared from different camera views. The system is deployed in an office building and four-day videos are collected. By training a deep convolutional neural network, the proposed system first extracts the appearance feature embeddings of each personal image, which detected from different cameras, for similarity comparison. Then, a stochastic inter-camera transition matrix is associated with appearance feature for further improving the person re-identification ranking results. Finally, a noise-suppression explanation is given for analyzing the matching improvements. This paper expands the scope of indoor movement perception based on non-overlapping multiple cameras and improves the accuracy of pedestrian re-identification without introducing additional types of sensors.

keywords:
surveillance videos , multi-camera system , computer vision , person re-identificaiton , indoor person movement
journal: Journal of  Templates

1 Introduction

For intelligent buildings, accurate indoor occupancy measurement and individual moving information acquisition are indispensable prerequisites to improve energy efficiency and security management. Related applications, like customer counting in shopping malls, space utilization analysis in multi-story buildings, and passenger behavior analytic in transportation hubs, all require indoor personal location and movement information. Considering the easy installation and decreased costs of surveillance cameras, an increasing number of research have been devoted to treating the video processing as an active perception method for indoor personal sensing [1, 2, 3]. Meanwhile, computer vision techniques, including person detection, recognition, and re-identification are applied to indoor surveillance video analysis for occupant detection, zonal person counting, and movement pattern analyzing.

Considering the dynamic characteristics of indoor personal movements, when people moving inside buildings, they will be captured by multiple cameras which always have no field-of-view overlap. To figure out the individuals’ fully moving patterns, it is necessary to associate the incomplete information captured by each single camera with cross-view person re-identification techniques. Person re-identification can be treated as retrieving the same people who appeared in other cameras by comparing appearance feature similarity. However, the main encountered challenges include: (1) people dressing looks alike which bring difficulties to distinguish different individuals; (2) the camera view angle and personal pose changed when person moving among different indoor surveillance spaces; (3) the complete track acquisition are impeded when people moving across the blind area between two neighboring cameras with no filed-of-view overlap.

To conquer the difficulties mentioned above, this study first deploys multiple non-overlapping cameras in different locations of an office building. Correspondingly, a graph-based camera linkage model is established to capture indoor personal movement over a specific floor area. Based on the detected personal images from different camera videos, distinctive appearance feature embeddings are extracted by a deep convolutional neural network for similarity comparison. Furthermore, statistical indoor transition patterns, including stochastic transition choice matrix, are coupled with appearance features for improving the cross-camera person matching accuracy. The main contributions to the community include:

  • 1.

    A multi-camera video-based person re-identification framework is proposed for enlarging surveillance areas and unveiling indoor person moving patterns in larger scale.

  • 2.

    Stochastic indoor personal transition matrix is integrated with appearance features for improving cross-camera person re-identification accuracy.

  • 3.

    A noise suppression-based theoretical analysis is given for explaining the accuracy improvements.

2 Related Works

Surveillance cameras can be treated as visual sensors for indoor personal monitoring and sensing [4, 5, 6]. The uses of video surveillance to analyze the distribution of indoor occupants are passive detection scheme, which do not requires the cooperation of pedestrians, and has advantage in reflecting their daily movement characteristics. By deploying ceiling-mounted Kinect camera at the room entrance areas, Petersen et al. [7] presented image processing techniques to detect, track and count indoor occupants. To avoid the body parts occlusion problem happens in camera views, Zou et al. [8] and Chen et al. [9] proposed a head detection algorithm for measuring indoor occupancy. Gao et al. [10] further applied the head detection-based people counting method to crowded classroom surveillance environment. Combing with computer vision-based tracking method, Kuipers et al. [11] deployed four cameras in an atrium and carried out several tests to study the dynamics of mobility.

However, the above-mentioned research mainly focuses on single distinct camera cases inside partial indoor areas without taking the information fusion between neighboring cameras into account. When confronted with larger indoor area which cannot fully covered by a single camera, these methods will fail to restore the complete personal moving activities. To enlarge the coverage area, several researches deploy PTZ camera for indoor occupant counting and monitoring [12, 13]. However, the physical limitation of the complicated indoor structure impedes the widely usage of PTZ cameras. Instead, researchers are devoted to deploying camera networks, which consists of multiple non-overlapping cameras to maximize the coverage of personal movements. To measure the indoor occupancy, Liu et al. [14] proposed a dynamic Bayesian network-based method to combine the detection results from multiple vision sensors, which deployed at the room entrance and interior. Camps et al. [15] deployed a human re-identification system inside an airport, which composed of three cameras. In addition, several public indoor video data sets are released for multi-camera person re-identification and tracking. Figueira et al. [16] deployed a heterogeneous camera network in an indoor office, which consists of 13 cameras distributed over three floors. Bialkowski et al. [17] proposed a multi-camera database, which consists of 8 cameras inside building environment, for the task of person re-identification in surveillance networks. Marroquin et al. [18] created WiseNET dataset for people detection and tracking, which composed of 6 indoor cameras in a building floor with contextual information and annotations. By deploying 15 synchronized cameras mounted in corridor and junction, Styles et al. [19] devote to predicting the personal location within camera networks.

With the ever increasing of big visual data and the rising of computing power, much research efforts are now dedicated to performing deep learning based methods for extracting more distinctive appearance feature embeddings, which applied for learning feature distance between image pairs [20, 21, 22, 23, 24]. To fill the gap of the discontinuity of pedestrian data across non-overlapping cameras, indoor spatial-temporal information, including camera topology, cross-camera transition choice, and travel time distributions are exploited to improve the accuracy of person re-identification results [25, 26, 27, 28]. In this work, we propose an indoor camera linkage model, and formulate the movement of pedestrian in buildings as the information flow among camera nodes. By training a deep convolutional neural network-based feature extractor, and leveraging building information as well as the transition matrix of pedestrian movement, the cross-camera pedestrian matching is realized.

3 Experiment and Methodology

3.1 Framework and modules

An indoor multiple camera system, which composed of four modules, is deployed for capturing personal movement in larger indoor area, as shown in Figure 1. First, integrating with the prior knowledge about building structure and surveillance camera placement, topological camera linkage is established, and raw videos are collected from these non-overlapping cameras. Secondly, by using an off-the-shelf person detection approach, all of the person who appeared in these cameras are detected. The detected images are then cropped from each video and annotated with global unique labels. By using the supervised learning method, a deep convolutional neural network (ConvNets) is trained to extract appearance features for similarity comparison. Furthermore, by ranking the similarity distance, same individual who appeared at different cameras are matched. Finally, we devoted to recognizing indoor personal moving patterns from the person re-identification results combined with spatial-temporal constrains.

Refer to caption
Figure 1: Framework and pipeline of this study

3.2 Data preparation

Refer to caption
Figure 2: Schematic diagram of five cameras deployed inside the office building

Five synchronized 4K UHD ceiling-mounting cameras are installed on the 10th10^{\text{th}} floor of Liuqing Building, Tsinghua University. As shown in Figure 2, the floor area is 1,200 m2m^{2} and the coverage area of each camera do not overlap with each other. For each video, its frame rate is 25 frames per second, and with High Efficiency Video Coding (HEVC). All the five cameras are linked to a Digital Video Recorder (DVR) with cable connections. In this paper, four business day video data were recorded, and the total length is 480 hours. Undirected camera link graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) is established, where 𝒱\mathcal{V} denotes each camera node, and \mathcal{E} is the connectivity between different cameras. Indoor personal movements are treated as transition process among different camera nodes.

This work uses an off-the-shelf person detector [29] to get all the bounding boxes of each person who appeared in different videos. To reduce the labor-consumption of image labeling, the detected personal images are cropped from videos in every 25 frames. We devoted to match the same person appeared in different field-of-views, therefore, person who appeared in only one camera is excluded. Video data in the four different business-day are separated, and all the detected images in each day are manually annotated with a globally-unique identifier, i.e. 𝒟={(𝐱𝟏,y1),(𝐱𝟐,y2),,(𝐱𝐍,yN)}\mathcal{D}=\{(\mathbf{x_{1}},y_{1}),(\mathbf{x_{2}},y_{2}),...,(\mathbf{x_{N}},y_{N})\} and N{128,142,159,138}N\in\{128,142,159,138\}, where 𝐱𝐢\mathbf{x_{i}} denotes the detected images of one person from different camera views and yiy_{i} is the corresponding label. For each day, people was chosen with a random number 80%, and the corresponding detected images are chosen as training set. The left people and images are as testing set, i.e. 𝒟=𝒟training𝒟testing\mathcal{D}=\mathcal{D}_{training}\bigcup\mathcal{D}_{testing} where |𝒟training|=M|\mathcal{D}_{training}|=M, |𝒟testing|=NM,M{104,112,131,111}|\mathcal{D}_{testing}|=N-M,M\in\{104,112,131,111\}. Descriptions for the prepared data set are presented in Table 1. To keep consistency with training data, we random leave one picture out from each training image library as the validation set. In data preparations, each pedestrian appears in at least two cameras, and at most five cameras. The statistics of the number of each single pedestrian appearing in different cameras are shown in Figure 3. For the given training set 𝒟training\mathcal{D}_{training}, a supervised deep learning technique, named deep convolutional neural networks (deep ConvNets), is applied for training an appearance feature extractor, which aims for clustering the same person’s images into one group and dividing different person’s images into distinct partitions. Data preparation and training process are depicted in Figure 4.

Table 1: Training and testing data set preparation in four days
Day #Training IDs #Training Images #Testing IDs #Testing Images
Day 1 104 19,313 24 4,247
Day 2 112 34,403 30 6,052
Day 3 131 39,150 28 4,592
Day 4 111 31,072 27 4,083
Refer to caption
Figure 3: Statistics of the number of single pedestrians appearing in different cameras

3.3 Appearance based person re-identification

3.3.1 Training processes

Refer to caption
Figure 4: Data preparation and training process

Given training set 𝒟training\mathcal{D}_{training}, the objective of deep ConvNets is to learn high-level feature embedding function y^i=ϕ(xi,θ)\hat{y}_{i}=\phi(x_{i},\theta) to minimize the distance between the estimated value y^i\hat{y}_{i} and the true label yiy_{i}. In this study, we represent the categorical data y^i\hat{y}_{i} and yiy_{i} as one-hot encoding and the cross entropy between each pair of estimated value and true label is defined as loss function. Therefore, the objective function in the training process is minimize the summation of cross-entropy loss.

(θ)=minθi=1Ml(ϕ(xi;θ),yi)\mathcal{L}(\theta)=\min_{\theta}{\sum_{i=1}^{M}l(\phi(x_{i};\theta),y_{i})} (1)

In the structure of the deep ConvNets, PCB [30] and ResNet-50 [31] are used as pretrained model. In order to perform transfer learning, all the layers are frozen except the output layer, where the total neuron number is modified to equal with the person count MM in the training set of different days, i.e. 104, 112, 131, and 111 respectfully. For the training strategies, an stochastic gradient descent approach is applied with back-propagation to update the parameters in each layer of the deep ConvNets. To satisfy the input requirements, all the images in 𝒟training\mathcal{D}_{training} are resized to 384×\times192×\times3. The mini-batch size in each iteration is 128 and the total epoch is 50 with learning rate decay in each 10 epochs. Algorithm 1 gives the the training algorithm for indoor person appearance feature extraction. The loss and rank-1 accuracy in training and validation processes are shown in Figure 5 and Figure 6.

Input: person count M, person detection results with annotations 𝒟training\mathcal{D}_{training}, max number of epochs (T=50T=50), number of the deep ConvNets updates per step (K=|𝒟training|/batch_sizeK=|\mathcal{D}_{training}|/batch\_size)
Output: feature embedding function ϕ\phi and parameters θ\theta
1 for epoch=1,2,…,T do
2       for step=1,2,…,K do
3             1. Sample a mini-batch size of person images in 𝒟training\mathcal{D}_{training};
4            2. Update ϕ\phi and θ\theta by taking SGD step on mini-batch loss (θ)\mathcal{L}(\theta) in (1);
5            if step % 10 == 0 then
6                   3. decay the learning rate by a factor of 0.1 in every 10 epochs;
7            
8      
9
Algorithm 1 Training algorithm for indoor person appearance feature extraction
Refer to caption
Figure 5: Loss in training and validation processes
Refer to caption
Figure 6: Rank-1 accuracy in training and validation processes

3.3.2 Testing processes

In the testing stage, we apply the trained feature embedding function ϕ\phi to 𝒟testing\mathcal{D}_{testing} and take the output from the fully connected layer as the finally feature embedding for each testing image. Considering PCB and ResNet-50 are used, each feature of testing image is a column vector with size 12,288 (6 parts multiple by 2048, with each part is normalized to 1). The cosine distance between each image pair {xi,xj}\{x_{i},x_{j}\} is calculated as appearance similarity between two feature embeddings:

simapp(xi,xj)=ϕ(xi;θ)ϕ(xj;θ)ϕ(xi;θ)22ϕ(xi;θ)22sim_{app}(x_{i},x_{j})=\frac{\phi(x_{i};\theta)\cdot\phi(x_{j};\theta)}{\left\|\phi(x_{i};\theta)\right\|_{2}^{2}\cdot\left\|\phi(x_{i};\theta)\right\|_{2}^{2}} (2)

where xi,xj𝒟testingx_{i},x_{j}\in\mathcal{D}_{testing} and ϕ(xi;θ),ϕ(xj;θ)𝐑12288\phi(x_{i};\theta),\phi(x_{j};\theta)\in\mathbf{R}^{12288}, ϕ(xi;θ)22=ϕ(xj;θ)22=6\left\|\phi(x_{i};\theta)\right\|_{2}^{2}=\left\|\phi(x_{j};\theta)\right\|_{2}^{2}=6.

For each image (query image) in the testing set, we calculate cosine distance to depict the similarity between all the other images (gallery images) and feature similarity matrix are therefore obtained. Meanwhile, to match the indoor person who appeared in different cameras, image pairs of the same individual appeared in the same camera are excluded. Furthermore, by sorting the similarity score and comparing with the image annotations, the Cumulative Matching Characteristic (CMC) curves are obtained, which will be shown by the red curves in Figure 9.

Refer to caption
Figure 7: Daily arrive and departure in four different weekdays

3.4 Transition Pattern Modeling

Considering the dynamic change of personal pose and moving orientations, the appearance feature only is not sufficient to discriminate different occupants. Spatial-temporal constraints are further required to refine the appearance-based person re-identification results. Statistical transition patterns are the repetitive patterns that person takes, which relates to the properties of spatial layout, camera topology, and individual movements. The daily entry and exit count and probability density estimation are shown in Figure 7. In this study, personal transition between different cameras can be formulated as stochastic matrix. By using the training set, the number and proportion of pedestrians passing between different cameras are calculated. The walk-through selection here is a right stochastic matrix, which means the total of transition probability from camera ii to other cameras equals to 1. The first-order state-transition matrix is calculated as follows:

pj|i=Nj|ij=1𝒩(i)Nj|ip_{j|i}=\frac{N_{j|i}}{\sum_{j=1}^{\mathcal{N}(i)}{N_{j|i}}} (3)

s.t. j=1𝒩(i)pj|i=1\sum_{j=1}^{\mathcal{N}(i)}{p_{j|i}}=1, where Nj|iN_{j|i}, pj|ip_{j|i} denotes the number and proportion of person transition from camera ii to camera jj respectfully. 𝒩(i)\mathcal{N}(i) means the neighbor cameras of camera ii, which can be obtain from undirected camera link graph 𝒢\mathcal{G}. The transition count and state-transition matrix between among different cameras are therefore calculated and shown in Figure 8.

Refer to caption
Figure 8: The transition count and state-transition matrix

3.5 Results

Coupling appearance feature with the state-transition matrix, similarity between the query image xix_{i} and gallery image xjx_{j} is further formulated as:

sim(xi,xj)=simapp(xi,xj)pj|isim(x_{i},x_{j})=sim_{app}(x_{i},x_{j})\cdot p_{j|i} (4)
sim(xi,xj)=ϕ(xi;θ)ϕ(xj;θ)ϕ(xi;θ)22ϕ(xi;θ)22Nj|ij=1𝒩(i)Nj|isim(x_{i},x_{j})=\frac{\phi(x_{i};\theta)\cdot\phi(x_{j};\theta)}{\left\|\phi(x_{i};\theta)\right\|_{2}^{2}\cdot\left\|\phi(x_{i};\theta)\right\|_{2}^{2}}\cdot\frac{N_{j|i}}{\sum_{j=1}^{\mathcal{N}(i)}{N_{j|i}}} (5)

Table 2,3,4,5 demonstrates several representative values in the Cumulative Matching Characteristic curve and the improvements when adding state-transition matrix to sole appearance feature. Experimental results show that the Rank-1 accuracy has been improved by 7.72%, 8.21%, 6.62% and 0.22% respectfully. The average improvement is 5.69%. The full ranking results in the Cumulative Matching Characteristic curve is shown in Figure 9, where the red curves shows the sole-appearance feature used, and the green curves represent the improvements after adding spatial constrains. The purple arrows demonstrate the corresponding rank-1 improvements.

Table 2: The value of CMC changes before and after adding state-transition matrix (Day 1)
Day 1 Rank-1 Rank-5 Rank-10 Rank-15
Appearance 0.7907 0.8959 0.9242 0.9388
w/ transition choice 0.8679 (↑) 0.9211 (↑) 0.9352 (↑) 0.9421 (↑)
Table 3: The value of CMC changes before and after adding state-transition matrix (Day 2)
Day 2 Rank-1 Rank-5 Rank-10 Rank-15
Appearance 0.7024 0.8574 0.8972 0.9147
w/ transition choice 0.7845 (↑) 0.8947 (↑) 0.9245 (↑) 0.9354 (↑)
Table 4: The value of CMC changes before and after adding state-transition matrix (Day 3)
Day 3 Rank-1 Rank-5 Rank-10 Rank-15
Appearance 0.7803 0.9438 0.9625 0.9686
w/ transition choice 0.8465 (↑) 0.9323 0.9595 0.9702
Table 5: The value of CMC changes before and after adding state-transition matrix (Day 4)
Day 4 Rank-1 Rank-5 Rank-10 Rank-15
Appearance 0.8567 0.928 0.9481 0.9603
w/ transition choice 0.8589 (↑) 0.9226 0.9402 0.9495
Refer to caption
Figure 9: Rank-K accuracy improvements in person re-identification results

We normalized the final pedestrian similarity matrix, and visualized the changes in the appearance similarity heat map after transition choice is added. As shown Figure 10, the similairty matrix is symmetric, where the image count equals with the total testing images number as shown in Table 1, i.e., 4247, 6052, 4592, and 4083 respectfully. The value of each pixel in the similarity heat map indicates the level of final similarity score between query image xix_{i} and gallery image xjx_{j}, where 1 indicate the same person, and 0 is the different. Each person has multiple images under different cameras, and the small rectangles in the main diagonal area express the similarity of the same pedestrians appeared in other different cameras. As can be seen from the figure in the right-hand side, by adding state transition constraints, the ranking algorithm actually do a step further for filtering and suppressing the matching score when people walking among different cameras with low probabilities. Relatively, the similarity of the same pedestrian under different cameras are improved.

Refer to caption
Figure 10: Comparison between similarities when adding transition choice

A specific person re-identification example is demonstrated. Given a probe image, the ranking algorithm retrieves the best 10 matches from the gallery, i.e. the personnel images captured by different cameras. The digital number in green color representdongs the correct matching, and the red is the wrong matching. From the figure we can see that more correct match is retrieved when transition choice is combined with appearance feature embeddings.

4 Discussion

In this paper, a multi-camera system, which consists of five non-overlapping surveillance cameras, is deployed inside an office building. A graph-based camera linkage is further established to extend the range of the pedestrian movement monitoring. For matching the person images captured by different cameras, a deep convolutional neural network is trained to extract robust appearance feature embeddings and similarity comparison is therefore applied. The ratio of the correct matching in rank-1 accuracy over Cumulative Matching Characteristics curve reaches to 78.25% on average. By leveraging statistical transition patterns, the false match with low similarity score in the ranking list is further filtered out. Coupling the state-transition matrix constraints with appearance feature similarities, the rank-1 accuracy increased by 5.69% on average, which shows potential for further improvements when more video data and transition choice are accumulated.

5 Conclusions

To compensate for the insufficient coverage of single camera, this study demonstrates a multi-camera system to achieve a wider range of scene capturing, which can be applied to depict a larger-scale indoor personal movement monitoring. More than 120 pedestrians moving inside an building in each 4 different weekdays are monitored by the proposed system.

Meanwhile, there is no additional type of sensors required, and does not need active cooperation from indoor occupants. The appearance feature embeddings are associated with the state-transition matrix among multiple non-overlapping cameras to improve the accuracy of pedestrian re-identification. By adding the selection of transition choice to the appearance feature embeddings, the rank-1 accuracy of cross-camera pedestrian matching is improved to 83.94%, which can be treated as a baseline for further improvements. At the same time, these cameras are installed in public indoor area and do not rely on high-definition human faces, which maximize privacy protection.

Our future research is devoted to unveiling the transition time distribution when pedestrians passing among different cameras. By sorting the similarity score by time sequence, we aim to further restoring the indoor personal trajectory in multi-camera circumstance.

Acknowledgments

This work is supported by National Key Research and Development Project of China (No.2017YFC0704100, and No.2017YFC0803300) and National Natural Science Foundation of China (Grant U2033206).

References

  • [1] K. Sun, Q. Zhao, J. Zou, A review of building occupancy measurement systems, Energy and Buildings 216 (2020) 109965. doi:10.1016/j.enbuild.2020.109965.
  • [2] H. Saha, A. R. Florita, G. P. Henze, S. Sarkar, Occupancy sensing in buildings: A review of data analytics approaches, Energy and Buildings 188-189 (2019) 278–285. doi:10.1016/j.enbuild.2019.02.030.
  • [3] L. Rueda, K. Agbossou, A. Cardenas, N. Henao, S. Kelouwani, A comprehensive review of approaches to building occupancy detection, Building and Environment 180 (2020) 106966. doi:10.1016/j.buildenv.2020.106966.
  • [4] Y. Benezeth, H. Laurent, B. Emile, C. Rosenberger, Towards a sensor for detecting human presence and characterizing activity, Energy and Buildings 43 (2-3) (2011) 305–314. doi:10.1016/j.enbuild.2010.09.014.
  • [5] J. Yang, M. Santamouris, S. E. Lee, Review of occupancy sensing systems and occupancy modeling methodologies for the application in institutional buildings, Energy and Buildings 121 (2016) 344–349. doi:10.1016/j.enbuild.2015.12.019.
  • [6] X. Wang, Intelligent multi-camera video surveillance: A review, Pattern Recognition Letters 34 (1) (2013) 3–19. doi:10.1016/j.patrec.2012.07.005.
  • [7] S. Petersen, T. H. Pedersen, K. U. Nielsen, M. D. Knudsen, Establishing an image-based ground truth for validation of sensor data-based room occupancy detection, Energy and Buildings 130 (2016) 787–793. doi:10.1016/j.enbuild.2016.09.009.
  • [8] J. Zou, Q. Zhao, W. Yang, F. Wang, Occupancy detection in the office by analyzing surveillance videos and its application to building energy conservation, Energy and Buildings 152 (2017) 385–398. doi:10.1016/j.enbuild.2017.07.064.
  • [9] L. Chen, F. Chen, X. Guan, A video-based indoor occupant detection and localization algorithm for smart buildings, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 5754 LNCS (2009) 565–573. doi:10.1007/978-3-642-04070-2_62.
  • [10] C. Gao, P. Li, Y. Zhang, J. Liu, L. Wang, People counting based on head detection combining Adaboost and CNN in crowded surveillance environment, Neurocomputing 208 (2016) 108–116. doi:10.1016/j.neucom.2016.01.097.
  • [11] M. Kuipers, A. Tomé, T. Pinheiro, M. Nunes, T. Heitor, Building space-use analysis system - A multi location/multi sensor platform, Automation in Construction 47 (2014) 10–23. doi:10.1016/j.autcon.2014.07.001.
  • [12] H. C. Shih, A robust occupancy detection and tracking algorithm for the automatic monitoring and commissioning of a building, Energy and Buildings 77 (May 2013) (2014) 270–280. doi:10.1016/j.enbuild.2014.03.069.
  • [13] C. Ding, B. Song, A. Morye, J. A. Farrell, A. K. Roy-Chowdhury, Collaborative sensing in a distributed ptz camera network, IEEE Transactions on Image Processing 21 (7) (2012) 3282–3295. doi:10.1109/TIP.2012.2188806.
  • [14] D. Liu, X. Guan, Y. Du, Q. Zhao, Measuring indoor occupancy in intelligent buildings using the fusion of vision sensors, Measurement Science and Technology 24 (7). doi:10.1088/0957-0233/24/7/074023.
  • [15] O. Camps, M. Gou, T. Hebble, S. Karanam, O. Lehmann, Y. Li, R. J. Radke, Z. Wu, F. Xiong, From the Lab to the Real World: Re-identification in an Airport Camera Network, IEEE Transactions on Circuits and Systems for Video Technology 27 (3) (2017) 540–553. doi:10.1109/TCSVT.2016.2556538.
  • [16] D. Figueira, M. Taiana, A. Nambiar, J. Nascimento, A. Bernardino, The HDA+ data set for research on fully automated re-identification systems, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 8927 (2015) 241–255. doi:10.1007/978-3-319-16199-0_17.
  • [17] A. Bialkowski, S. Denman, S. Sridharan, C. Fookes, P. Lucey, A database for person re-identification in multi-camera surveillance networks, 2012 International Conference on Digital Image Computing Techniques and Applications, DICTA 2012doi:10.1109/DICTA.2012.6411689.
  • [18] R. Marroquin, J. Dubois, C. Nicolle, WiseNET: An indoor multi-camera multi-space dataset with contextual information and annotations for people detection and tracking, Data in Brief 27 (2019) 104654. doi:10.1016/j.dib.2019.104654.
  • [19] O. Styles, T. Guha, V. Sanchez, A. Kot, Multi-camera trajectory forecasting: Pedestrian trajectory prediction in a network of cameras, IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops 2020-June (2020) 4379–4382. arXiv:2005.00282, doi:10.1109/CVPRW50498.2020.00516.
  • [20] A. Bedagkar-Gala, S. K. Shah, A survey of approaches and trends in person re-identification, Image and Vision Computing 32 (4) (2014) 270–286. doi:10.1016/j.imavis.2014.02.001.
  • [21] L. Zheng, Y. Yang, A. G. Hauptmann, Person Re-identification: Past, Present and Future 14 (8) (2016) 1–20. arXiv:1610.02984.
    URL http://arxiv.org/abs/1610.02984
  • [22] S. Karanam, M. Gou, Z. Wu, A. Rates-Borras, O. Camps, R. J. Radke, A Systematic Evaluation and Benchmark for Person Re-Identification: Features, Metrics, and Datasets, IEEE Transactions on Pattern Analysis and Machine Intelligence 41 (3) (2019) 523–536. arXiv:1605.09653, doi:10.1109/TPAMI.2018.2807450.
  • [23] Q. Leng, M. Ye, Q. Tian, A Survey of Open-World Person Re-Identification, IEEE Transactions on Circuits and Systems for Video Technology 30 (4) (2020) 1092–1108. doi:10.1109/TCSVT.2019.2898940.
  • [24] M. Ye, J. Shen, G. Lin, T. Xiang, L. Shao, S. C. H. Hoi, Deep Learning for Person Re-identification: A Survey and Outlook (2020) 1–20arXiv:2001.04193.
    URL http://arxiv.org/abs/2001.04193
  • [25] K. W. Chen, C. C. Lai, P. J. Lee, C. S. Chen, Y. P. Hung, Adaptive learning for target tracking and true linking discovering across multiple non-overlapping cameras, IEEE Transactions on Multimedia 13 (4) (2011) 625–638. doi:10.1109/TMM.2011.2131639.
  • [26] X. Chen, K. Huang, T. Tan, Object tracking across non-overlapping views by learning inter-camera transfer models, Pattern Recognition 47 (3) (2014) 1126–1137. doi:10.1016/j.patcog.2013.06.011.
  • [27] O. Javed, K. Shafique, Z. Rasheed, M. Shah, Modeling inter-camera space-time and appearance relationships for tracking across non-overlapping views, Computer Vision and Image Understanding 109 (2) (2008) 146–162. doi:10.1016/j.cviu.2007.01.003.
  • [28] G. Wang, J. Lai, P. Huang, X. Xie, Spatial-temporal person re-identification, 33rd AAAI Conference on Artificial Intelligence, AAAI 2019, 31st Innovative Applications of Artificial Intelligence Conference, IAAI 2019 and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019 2 (2019) 8933–8940. arXiv:1812.03282, doi:10.1609/aaai.v33i01.33018933.
  • [29] J. Redmon, A. Farhadi, YOLOv3: An incremental improvement, arXivarXiv:1804.02767.
  • [30] Y. Sun, L. Zheng, Y. Li, Y. Yang, Q. Tian, S. Wang, Learning Part-based Convolutional Features for Person Re-identification, IEEE Transactions on Pattern Analysis and Machine Intelligence PP (c) (2019) 1–1. doi:10.1109/tpami.2019.2938523.
  • [31] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition 2016-December (2016) 770–778. arXiv:1512.03385, doi:10.1109/CVPR.2016.90.