Adapting the re-ID challenge for static sensors
Abstract
The Grévy’s zebra, an endangered species native to Kenya and southern Ethiopia, has been the target of sustained conservation efforts in recent years. Accurately monitoring Grévy’s zebra populations is essential for ecologists to evaluate ongoing conservation initiatives. Recently, in both 2016 and 2018, a full census of the Grévy’s zebra population has been enabled by the Great Grévy’s Rally (GGR), a citizen science event that combines teams of volunteers to capture data with computer vision algorithms that help experts estimate the number of individuals in the population. A complementary, scalable, cost-effective, and long-term Grévy’s population monitoring approach involves deploying a network of camera traps, which we have done at the Mpala Research Centre in Laikipia County, Kenya. In both scenarios, a substantial majority of the images of zebras are not usable for individual identification, due to "in-the-wild" imaging conditions — occlusions from vegetation or other animals, oblique views, low image quality, and animals that appear in the far background and are thus too small to identify. Camera trap images, without an intelligent human photographer to select the framing and focus on the animals of interest, are of even poorer quality, with high rates of occlusion and high spatio-temporal similarity within image bursts. We employ an image filtering pipeline incorporating animal detection, species identification, viewpoint estimation, quality evaluation, and temporal subsampling to compensate for these factors and obtain individual crops from camera trap and GGR images of suitable quality for re-ID. We then employ the Local Clusterings and their Alternatives (LCA) algorithm, a hybrid computer vision & graph clustering method for animal re-ID, on the resulting high-quality crops. Our method processed images taken during GGR-16 and GGR-18 in Meru County, Kenya, into 4,142 highly-comparable annotations, requiring only 120 contrastive same-vs-different-individual decisions from a human reviewer to produce a population estimate of 349 individuals (within 4.6 of the ground-truth count in Meru County). Our method also efficiently processed 8.9M unlabeled camera trap images from 70 camera traps at Mpala over two years into 685 encounters of 173 unique individuals, requiring only 331 contrastive decisions from a human reviewer.
1 Introduction
The population of Grévy’s zebras experienced a dramatic decline beginning in the 1970s largely due to hunting and competition for food and water resources with local pastoral communities. Estimates have placed the number of Grévy’s zebras remaining in the wild at under 2,000, with the vast majority in the Samburu region of central Kenya. Due to extensive conservation efforts by the Kenyan and Ethiopian governments as well as environmental NGOs, the population of Grévy’s zebras have stabilized in recent years Rubenstein et al. [2016].
Accurately censusing Grévy’s zebra populations is critical for ecologists to evaluate these existing conservation efforts. Developing and maintaining a census of known individuals requires effective animal re-identification methods to ensure that only unique individuals are included in population counts Tuia et al. [2022]. A popular method for population size estimation is “capture-mark-recapture" Pradel [1996]. By this method, a set of animals in the target population is first captured and marked, then a second set of animals is independently recaptured, and finally a population estimate formed from the number of animals captured twice. This method, however, proves difficult to scale to large populations and territories, and may lead to inaccurate population estimates when animals are not confined to the study area and are capable of evading tagging Lee et al. [2014]; Tilling and Sterne [1999]. Further, the tagging methods

used in population estimation studies (such as ear tags and radio collars Alexander and Appel [1994]; Thouless [1995]; Collins et al. [2014]) can be excessively expensive and time-consuming to implement in the field, and overly harmful to the animal Witmer [2005].
A modern alternative to manual mark-recapture studies that does not have these limitations is exemplified in the Great Grévy’s Rally (GGR) events of 2016 and 2018 where volunteers spread over the range of Grévy’s zebra to photograph as many animals as possible over two consecutive days. Relying on the distinctive appearance of an animal itself as a means for identification, a combination of algorithmic and human curation efforts produced population estimates that have been accepted as definitive by the Kenyan Wildlife Service and the IUCN Rubenstein et al. [2018a]. Despite the success of the GGR, the curation effort of ~50,000 images per event is formidable, since the dataset must be narrowed to the recognizable subset of images of Grévy’s and a streamlined curation procedure must be developed, all-the-while preserving the accuracy of the resulting census. Parham [2021] demonstrates how to ensure accuracy while reducing human effort.


Complementary to the focused burst of human effort for the GGR photographic events is the possibility of using a fleet of camera traps to collect the data needed for population monitoring. For animals that allow for sight-based ID, images taken by camera traps are a potentially cost-effective and non-invasive method to re-identify individuals and obtain robust population estimates. Unlike alternative methods such as tagging or even the field photography of the GGR events, camera traps do not require the physical presence of field researchers for animal re-identification; this ensures that the natural behaviors of animals are not disrupted and saves considerable time and resources Delisle et al. [2021]; Schneider et al. [2019].
However, analysis of camera trap images by human observation alone is not practical. In recent years, computer vision algorithms have been shown to present a highly accurate and standardized method for camera trap image analysis. Computer vision techniques have seen significant success in automated species detection & identification from camera trap images, especially with advances in deep learning Norouzzadeh et al. [2019]; Beery et al. [2019a, b, c]; Tabak et al. [2020]. Accurate species ID from camera trap images has laid the foundation for the next step in an end-to-end photographic censusing pipeline from camera trap images: automated individual re-ID (for reviews, see: Schneider et al. [2019]; Ravoor and T. S. B. [2020]; Vidal et al. [2021]). However, the challenging nature of the automatically-captured data (motion blur, occlusion, poor lighting, far-away animals) often leads to flawed animal re-identification even by human experts Meek et al. [2013]. Additionally, images captured in a burst from the same motion trigger are often highly repetitive, leading to potential bias in automatic re-identification if not handled carefully.
In this work, we seek to adapt existing techniques to the combined challenges of animal re-identification in GGR-style rally events and in the especially challenging static camera trap paradigm.
Unlike many other classification tasks with numerous and highly similar classes, animal re-identification is an open-set classification problem Geng et al. [2020], with the need to assign every unique individual to its own class and to recognize novel individuals unseen during training. There are two key categories of animal re-ID algorithms. Ranking algorithms for re-ID query an image of the target individual against an existing database to obtain a ranking of the most confident matches. Hotspotter Crall et al. [2013] is one such texture-based ranking algorithm that uses the SIFT Lowe [2004] algorithm to extract salient features from the query image and subsequently a nearest neighbor search to match the query image against the database. The algorithm is specialized for striped and spotted animals, and it has been used for re-ID of Grévy’s zebra Crall et al. [2013]; Rubenstein et al. [2018b] and several other animal species from camera trap data Park et al. [2019]; Nipko et al. [2020]; Blount et al. [2022]. Verification algorithms for re-ID, on the other hand, do not require querying an existing database; instead, a verification algorithm simply decides whether two images contain the same individual. An example of a verification algorithm is the Verification Algorithm for Match Probabilities (VAMP) Crall [2017], a random forest classifier Breiman [2001]; Pal [2005] that receives two images and decides if they contain the same animal, different animals, or are incomparable. Lastly, contrastive deep learning algorithms, such as the Pose Invariant Embedding (PIE) network Zheng et al. [2017]; Moskvyak et al. [2021], can learn a global feature embedding (instead of handcrafted features with Hotspotter & VAMP) for a particular image, allowing for distance-based comparisons with the feature embeddings of other images in a database; this allows PIE to serve as both a ranking and verification algorithm simultaneously. However, unlike Hotspotter and VAMP, PIE (and deep learning algorithms in general) requires a significant amount of training data. Instead, when considering a population without individual-level ground truth labels readily available, it is preferable to use classical computer vision for ranking and verification that can be easily bootstrapped with minimal human supervision.
Equally important to the choice of re-ID algorithms is determining which annotations — bounding boxes within images drawn around Grévy’s zebras by a detection algorithm — should even be considered for re-ID. Annotations that are of poor quality, are partially occluded, are from an indistinguishable viewpoint, or show an uncommon viewpoint lead to significantly increased work by humans and can cause over-counting. To avoid this, we introduce the notion of a "census annotation" to restrict the attention of re-ID algorithms to annotations that should be universally recognizable.
In this paper, we use census annotations, Hotspotter and VAMP in conjunction with the Local Clusterings and their Alternatives (LCA) decision management algorithm Stewart et al. [TBD]; Parham [2021] for Grévy’s zebra re-ID. Like other animal and human re-ID systems Kulits et al. [2021]; Bodesheim et al. [2022]; Wang et al. [2016], the LCA algorithm offers a human-in-the-loop approach to dynamically cluster annotations by individual and request human reviews for verification hard cases. This paper is an extended version of work presented at the 3rd International Workshop on Camera Traps, AI, and Ecology Sundaresan et al. [2023]. We extend our previous work by broadening the scope to include human-captured images from the Great Grévy’s Rally alongside static camera trap data and elaborate on our methodology for ensuring annotation comparability across encounters. To the best of our knowledge, this marks the first use of interactive & error-driven clustering algorithms like LCA for animal re-ID for either events like the GGR or camera trap data for any species.
2 Methods
2.1 GZCD Dataset
The GZCD dataset is comprised of 5,464 images sourced from Meru County, Kenya, taken over four days of GGR-16 and GGR-18 by 13 photographers (see Fig. 3). The photographers were trained to capture a consistent right-side viewpoint for Grévy’s zebra, and only images taken of the intended side were kept in the dataset by human reviewers. The dataset is highly curated: bounding boxes and labels (species, viewpoint, and quality) were set by human reviewers for all animals. The 7,372 right-view Grévy’s annotations were further filtered based on these quality labels, resulting in a set of 4,269 “quality baseline" right-view Grévy’s annotations. Note that the spatial subset for Meru County, Kenya is geographically isolated by mountains from neighboring conservation areas, giving the expectation that the population is largely self-contained. Thus, the GZCD dataset is an excellent testbed for our censusing pipeline.

2.2 Camera Trap Dataset
We use images collected from a network of 70 camera traps distributed around the Mpala Research Centre in Kenya’s Laikipia Plateau. The network has collected 8.9 million images over the past two years of deployment. There are four types of camera trap placement schemes in the network: systematically in a grid, at “magnet" sites (e.g. salt licks), as well as expert-targeted & random placement along roads (see Fig. 2).

2.3 Automated Species Identification
Prior to re-ID, the raw camera trap images were first passed through a YOLO v2 species detection model Parham et al. [2018a] to localize all zebra with a bounding box (both Grévy’s and plains zebra species), and crop the region of the image within each bounding box for downstream use. Localized bounding boxes crop out irrelevant and potentially distracting background information and yield distinct annotations of independent individuals from images that feature several animals. YOLO v2 has shown to be more accurate for animal detection than alternative object detection models, such as Faster R-CNN Parham and Stewart [2016]. Next, the cropped regions are classified to zebra species — Grévy’s vs. plains zebra — and viewpoint — left vs. right — by a DenseNet model. Only right, front-right, and back-right viewpoints are considered for identification; differing viewpoints cannot be matched with one another, as the right and left sides of Grevy’s zebra are distinct. Both the annotation localization and classification networks were trained on the WILD dataset Parham et al. [2018a].
2.4 Census Annotations
Beyond assigning viewpoint and species to each annotation, we wish to ensure that these annotations — from both human-captured and camera trap images — would be universally comparable. In principle, when such an annotation fails to match the other annotations we should be able to safely conclude that it is the only annotation from that animal. Furthermore, focusing on these annotations (a) should make verification decisions easier both for an algorithm such as VAMP and for a human reviewer ("verifier"), (b) should allow recovery from mistakes such as incidental matching 111As examples, a matching algorithm may recognize animals seen in the background (i.e., a “photobomb” Crall [2017]); camera traps are acutely affected by a similar scenario by accidentally matching background textures, and (c) should avoid increased human effort to address matching ambiguities.222See time and effort studies of animal ID curation in Parham [2021]. For Grévy’s zebras, this entails annotations that have both the distinctive hip and shoulder chevron (see Fig. 4, which we will call “census annotations" (CA) Parham [2021].
In order to train a census annotation network, we started by curating 10,229 Grévy’s zebra annotations from GGR-18. These were arranged in a series of grids and presented via a web interface to two reviewers, who selected the annotations that were Grévy’s CAs. Then, a DenseNet with a linear classification layer was trained on this CA dataset to decide whether an annotation was indeed a CA, producing an associated CA confidence score (see Fig. 5). Finally, a regression network was trained to narrow the annotation region to only surround the hip and shoulder chevron and minimize any distracting background information. Each of these new annotations was saved as a “Census Annotation Region”, or CA-R Parham [2021]. The CA and CA-R networks were then used to obtain the CA-Rs and corresponding CA scores for both GZCD and the camera trap dataset.
2.5 Filtering Pipeline
Mark-recapture statistical models are highly sensitive to inaccurate identifications, but are built with the assumption that not all individuals are seen. Thus, it is important to prioritize the precision of our re-identification pipeline over assigning an ID to each annotation. One significant factor in inaccurate identification, by both humans and algorithms, is image quality — blurry, or poorly lit annotations are difficult or impossible to reliably identify. Filtering annotations by quality is therefore a critical component of the re-ID pipeline, and automated quality filtering reduces the amount of human time and work required to rectify matching errors caused by undesirable annotations. Here, we devise an annotation filtering scheme suitable for re-ID from camera traps (Fig. 6).
As already described, we only consider census annotation regions (CA-Rs), those the show right-side viewpoints of Grévy’s zebra, including both the shoulder and hip chevron. Beyond this, we further filter by time of day: only annotations from images taken between 6:30 AM and 7:00 PM (sunrise and sunset at Mpala) were kept for re-ID. Due to the camera trap settings, images taken during the daytime are optical (RGB) and are of higher resolution (13 MP). Nighttime images are taken with an infrared flash at lower resolution (9 MP). Qualitatively, nighttime images are more difficult for human reviewers than daytime images; without modification, we believed the VAMP verification algorithm was likely to perform better on higher contrast and quality images taken during the day.
Next, the annotations were filtered by encounter. To define the encounters, we used an agglomerative clustering approach: for each camera, annotations from images taken within the same minute and in consecutive minutes were grouped together in the same encounter. Next, the annotation with the highest Census Annotation confidence score was selected from each encounter, and the rest were discarded. As the camera traps take images in bursts, the images in a encounter are nearly identical, and hence the corresponding annotations very likely must feature the same individuals. This step encourages matching across the best representatives from every encounter, reducing the amount of data required but preserving the relevant information. Experiments in Parham [2021] show that filtering in this manner, in particular filtering for CA-Rs, does not have a significant impact on the individual count, but dramatically reduces the human effort required to produce it (Fig. 7). While the GZCD used mobile photographers, the encounters it used were based on location and time, and often included many Census Annotations; the camera trap dataset used a shorter time-only decision to reduce trivial duplicates.
Lastly, the annotations were filtered for CA confidence score. Annotations above a 0.31 confidence score threshold were kept for re-ID. In previous work Parham [2021], a lower threshold was shown to add incomparable annotations to the re-ID database and encourage spurious matches, while a higher threshold was shown to eliminate relevant sightings without a considerable increase in matching speed and accuracy. The filtering pipeline yielded 685 right-view Grévy’s zebra annotations across the reserve, with each representing a single encounter (see Fig. 2).



2.6 Zebra ID Curation
The goal of the LCA decision management algorithm is to build an identification graph with the filtered annotations, using automated ranking & verification and delaying human intervention as much as possible. In the identification graph, annotations are represented as vertices, and relationships between annotations (whether positive or negative matches) are represented as edges. The LCA algorithm seeks to group annotations into clusters corresponding to individuals by maximizing positive edge weights within clusters and negative weights between clusters. This simplifies the identity labeling process significantly: instead of being asked to identify an annotation as one of up to 150 known individuals, reviewers are asked to compare pairs of annotations and determine whether they belong to the same individual or not. This contrastive task can be accomplished even by non-experts due to innate human capacity for pattern matching — it is similar to a game of spot the difference.
The ID graph is initialized with no edges, with every annotation constituting its own cluster. For every annotation, the Hotspotter ranking algorithm returns a list of its most confident matched annotations, and LCA forms edges between the annotation and its potential matches. Next, the VAMP verification algorithm evaluates every pair of vertices connected by an edge and assigns an edge weight (positive or negative) based on its confidence in the match. The LCA algorithm has two main phases. In the scoring phase, it keeps the initial edges and weights intact. LCA iterates through every local clustering (a single cluster or a pair of clusters) and checks to see if there exists an alternative clustering that increases a score measure defined as the sum of intra-cluster edge weights minus the sum of inter-cluster edge weights. If such an alternative clustering exists, the new arrangement is adopted. Once there are no longer any better local alternative clusterings, LCA proceeds to the stability phase. In this phase, LCA considers local clusterings for which the difference between it and its next best alternative are sufficiently small that changes in edge weights can potentially introduce a superior alternative clustering. For such local clusterings, LCA requests additional reviews to VAMP and the human (if VAMP cannot litigate the review on its own) to determine if certain edge weights must change. Once all local clusterings are significantly better than their next best alternative, the algorithm is considered to have converged.
3 Results
3.1 Results on GZCD
As discussed above, there were 4,269 “Quality Baseline” right-view Grévy’s annotations selected for identification from GZCD. Additionally, from the initial set of 7,372 right-view Grévy’s annotations in the GZCD, 4,142 were classified as CAs by our CA model using a score quality threshold of 0.31. Furthermore, each of these CAs were manually annotated to yield 4,142 corresponding CA-Rs. It should be noted that the CAs are not a strict subset with the Quality Baseline set; the two sets overlap significantly for trivially good sightings but each have annotations that are not present in the other. These three annotations sets — quality baseline, CAs, and CA-Rs — were each passed into the LCA algorithm for identification.
As the GZCD is composed of images taken during GGR-16 and GGR-18 in Meru County, we are able to compare the ID curation results with ground-truth estimates obtained from the two rallies. The Lincoln-Petersen index Petersen [1896] for the “Quality Baseline” is in 2016 and for 2018. The population estimate based on only CA was for GGR 2016 and for GGR 2018, whereas the CA-R estimate was zebra in 2016 and animals in 2018 Parham et al. [2018b].
With the Quality Baseline set, the LCA algorithm requested 22,972 decisions in total, with 420 from the human reviewer (yielding an automation rate of 98.2%). However, LCA converged faster with the CA set, with 18,607 total number of reviews, 352 of which were performed by the human reviewer (yielding an automation rate of 98.1%). We observe that LCA converged with 4.4 decisions per annotation with the CA set compared to 5.3 with the quality baseline set, indicating that the census annotations were more discriminative and required fewer alternative clusterings for LCA to consider. Finally, LCA converged the fastest with the CA-R set, with 13,427 reviews requested in total and only 120 performed by the human reviewer (yielding an automation rate of 99.1%). The results with the CA-R set were accurate within 4.6% on GGR-16 data (predicting 34926 animals against a ground-truth value of 36627) and within 0.5% (37130 animals predicted against 37329 ground-truth). In comparison, the quality baseline LCA results underestimated the ground-truth counts by only 12 individuals for GGR-16 and 1 individual for GGR-18, but required nearly four times the number of human reviews – therefore, considering CA-Rs relative to the quality baseline results in a 71.4% reduction in human work, while the resulting estimates still remain within the confidence interval of the established baseline count. These results are summarized in Table 4. While the CA-R result was obtained with a major reduction in human effort, Figure 7 also demonstrates that the amount of time spent on each CA-R review is substantially decreased (and more accurate).
3.2 Results on Camera Trap Dataset
8.9 million camera trap images from the initial dataset were passed through the species ID pipeline, yielding 84,383 zebra images (including both plains and Grévy’s zebra). Following species and viewpoint filtering, 23,512 right-view Grévy’s zebra annotations remained across 3,338 distinct encounters. After excluding nighttime encounters, 1,138 daytime encounters remained. We then sampled the single highest-quality annotation from each of the 1,138 encounters, as determined by our CA model, to avoid matching on near-identical images from the same burst. We further used a score quality threshold of 0.31 on census annotations that reduced the number of annotations to be identified to 734. As a last step, we used a blurriness filter to remove any additional challenging annotations, resulting in a final set of 685 CA-Rs.
In order to associate these 685 annotations to clusters of individuals, the LCA algorithm requested 5,403 automated reviews by VAMP and 331 additional human reviews; pairwise comparisons of annotations were performed largely automatically using VAMP, with a automation rate of 93.9%. The converged ID graph had 173 clusters, each corresponding to an individual zebra ID within the 685 annotations. On average, each individual was sighted across 1.9 static camera traps and 3.96 encounters, indicating significant matching across space and time. See Supplementary Fig. 8 for the distribution of encounters for each individual, and Supplementary Fig. 9 for the distribution of static camera traps that sighted each individual. The number of individuals sighted by each camera also differed based on its placement strategy (random grid, known Grévy’s territories along roads, randomly along roads, and both timelapse & motion-trigger traps at magnet sites). On average, camera traps at a magnet site (e.g. salt lick, dam, etc.) sighted the most individuals on average (8.5 per trap) and also produced the most first individual sightings on average (5.85 per trap). See Table 4 and Supplementary Fig. 10 for the distribution of total individuals sighted by cameras of each placement strategy. Particularly, the encounters for large individual clusters were spread across many months, with some spanning more than a year (see Supplementary Fig. 11). Based on the distinct encounters and cameras associated with each individual, we can generate spatial maps of encounters for any individual across the reserve: see Fig. 1 for a map of encounters for a particular zebra (Individual 32).
4 Discussion
Census Annotations and Census Annotation Regions are critical to automated photographic census because they 1) speed up human verification of match pairs and reduce the number of manual decision errors, 2) better separate the positive and negative scores predicted by algorithmic verifiers like VAMP, 3) reduce the amount of incidental matching Parham [2021], and 4) drastically reduce the amount of human interaction needed during manual review. Census Annotation Regions are powerful because they force a photographic census to consider only the most critical information for matching. Furthermore, our results indicate that using Census Annotations and Census Annotation Regions results in consistent population estimates on both the known GZCD baseline and camera trap dataset.
For our static camera dataset, the proposed combined method of subsampling and curation has enabled us to efficiently combine the 685 high-quality encounters to form 173 clusters, each representing an individual Grévy’s zebra. The use of LCA for semi-automatic, interactive decision-making resulted in fewer than 0.5 human reviews per annotation. By contrast, when working only with ranking algorithms, typically several — five to ten — potential matches must be examined, and LCA includes consistency checking implicitly.
LCA review requests for GZCD Set Annotations VAMP reviews Human reviews Total reviews Automation rate Quality Baseline 4,269 22,552 420 22,972 98.2% CA 4,142 18,255 352 18,607 98.1% CA-R 4,142 13,307 120 13,427 99.1% \botrule
Number of individuals sighted by camera placement strategy Total Avg. New Random grid 16 2.29 1.43 Roadside, known territories 105 2.76 2.26 Roadside, random 9 3 0.33 Magnet 107 8.5 5.85 \botrule
Each annotation in the cluster is representative of an encounter with that individual at a known position and time, and enables us to convert a set of 8.9 million unlabeled camera trap images into identified encounters that can be used as the input into a spatially-explicit mark recapture model Green et al. [2020] to estimate the total Grévy’s population. Because camera traps are monitoring long-term, in this case over two years unlike the contained 2-day period captured by each Great Grevy’s Rally, assumptions about no births and no deaths and no transients in the population are invalid. It is important to note that the ecologists at the Laikipia Zebra Project have confirmed that the Grévy’s population at Mpala has been an open set over the period in question: of the estimated 150 Grévy’s individuals currently seen over repeat seasons at Mpala, only 4 are “stable" and reside on the reserve long-term, while the vast majority (130 expert re-identified individuals) have been seen to migrate and return periodically, driven by water scarcity and threat of predation Service [2017]. It is thus likely that the camera network captured some number of transient individuals over the two year sampling span.
Limitations of the method may include imperfect CA labeling, where annotations with seemingly high CA scores are in actuality difficult to match, leading to two (or more) clusters in place of a single cluster. Additionally, we observe that errors could also occur via potential failures in any one of the algorithmic components whether at the species ID or at the individual re-ID level. Human decision failures in particular may inflate the number of individual clusters, with ambiguous matches mislabeled by the human reviewer resulting in more clusters than necessary. A significant focus of our ongoing work is to further analyze these results in order to tease these factors apart.
Lastly, while the method is currently limited to images taken during daylight hours, an approximately equal-sized set of images and annotations is collected at night. In recent preliminary tests, we selected the twenty-five night time annotations having the best combination of census annotation score and image contrast, and matched them against the clustered daytime annotations produced as a result of the work described here. For twenty of them (80%) Hotspotter produced correct matches in the top few. This shows potential for extending the overall method to the night-time, but many details must be revisited due to the overall lower quality of nocturnal images before they can be reliably included in generating enhanced census results.
5 Conclusion
In this paper, we perform efficient semi-automated Grévy’s zebra re-ID from camera trap data, with an algorithmic tool chain using Hotspotter for ranking, VAMP for verification, and LCA for decision management. Our method can be used even by non-zebra experts Parham [2021], as it only requires contrastive comparisons between pairs of individuals instead of matching each individual into the previously identified population. Ultimately, this curational re-ID process found 173 distinct individuals in camera trap data collected across a two year period at the Mpala Research Centre.
Going forward, we aim to further refine the filtering pipeline to improve matching with fewer human reviews and reduce attrition of data through the pipeline. We also are excited to further explore the results of the method, for example by adjusting the CA score threshold and observing how this affects LCA convergence behavior & results. We also hope to further adapt LCA to enable temporal sequences and spatial relationships to be taken into account, to enable us to better make use of the additional images of the individuals from each camera trap image burst and the spatial structure of the habits of individuals over time. Finally, and along similar lines, we hope to introduce “short-circuiting" to reduce human reviews when spatio-temporal constraints imply that there is no possibility of a match.
References
- Rubenstein et al. [2016] D. Rubenstein, B Low Mackey, ZD. Davidson, F. Kebebe, and S.R.B. King. Equus grevyi. The IUCN Red List of Threatened Species 2016, e.T7950A89624491, 2016. http://dx.doi.org/10.2305/IUCN.UK.2016-3.RLTS.T7950A89624491.en.
- Tuia et al. [2022] Devis Tuia, Benjamin Kellenberger, Sara Beery, Blair R Costelloe, Silvia Zuffi, Benjamin Risse, Alexander Mathis, Mackenzie W Mathis, Frank van Langevelde, Tilo Burghardt, Roland Kays, Holger Klinck, Martin Wikelski, Iain D Couzin, Grant van Horn, Margaret C Crofoot, Charles V Stewart, and Tanya Berger-Wolf. Perspectives in machine learning for wildlife conservation. Nat. Commun., 13(1):792, February 2022.
- Pradel [1996] R Pradel. Utilization of Capture-Mark-Recapture for the study of recruitment and population growth rate. Biometrics, 52(2):703–709, 1996.
- Lee et al. [2014] K A Lee, C Huveneers, O Gimenez, V Peddemors, and R G Harcourt. To catch or to sight? a comparison of demographic parameter estimates obtained from mark-recapture and mark-resight models. Biodivers. Conserv., 23(11):2781–2800, October 2014.
- Tilling and Sterne [1999] K Tilling and J A Sterne. Capture-recapture models including covariate effects. Am. J. Epidemiol., 149(4):392–400, February 1999.
- Alexander and Appel [1994] K A Alexander and M J Appel. African wild dogs (lycaon pictus) endangered by a canine distemper epizootic among domestic dogs near the masai mara national reserve, kenya. J. Wildl. Dis., 30(4):481–485, October 1994.
- Thouless [1995] C R Thouless. Long distance movements of elephants in northern kenya. Afr. J. Ecol., 33(4):321–334, December 1995.
- Collins et al. [2014] Gail H Collins, Steven L Petersen, Craig A Carr, and Leon Pielstick. Testing VHF/GPS collar design and safety in the study of free-roaming horses. PLoS One, 9(9):e103189, September 2014.
- Witmer [2005] Gary W Witmer. Wildlife population monitoring: some practical considerations. Wildl. Res., 32(3):259–263, June 2005.
- Rubenstein et al. [2018a] Daniel Rubenstein, Jason Parham, Charles Stewart, Tanya Berger-Wolf, Jason Holmberg, Jon Crall, Belinda Mackey, Sheila Funnel, Kasmira Cockerill, Zeke Davidson, Lizbeth Mate, Cosmas Nzomo, Rosemary Warungu, Dino Martins, Vincent Ontita, Joy Omulupi, Jennifer Weston, George Anyona, Geoffrey Chege, David Kimiti, Kaia Tombak, Andrew Gersick, and Nancy Rubenstein. The state of kenya’s grevy’s zebras and reticulated giraffes: Results of the great grevy’s rally 2018. Technical report, Grevy’s Zebra Trust, 2018a.
- Parham [2021] J. R. Parham. Animal detection for photographic censusing. PhD thesis, Dept. Comput. Sci., Rensselaer Polytechnic Inst., Troy, NY, USA, 2021.
- Delisle et al. [2021] Zackary J Delisle, Elizabeth A Flaherty, Mackenzie R Nobbe, Cole M Wzientek, and Robert K Swihart. Next-generation camera trapping: Systematic review of historic trends suggests keys to expanded research applications in ecology and conservation. Front. Ecol. Evol., 9, February 2021.
- Schneider et al. [2019] Stefan Schneider, Graham W Taylor, Stefan Linquist, and Stefan C Kremer. Past, present and future approaches using computer vision for animal re-identification from camera trap data. Methods Ecol. Evol., 10(4):461–470, April 2019.
- Norouzzadeh et al. [2019] Mohammad Sadegh Norouzzadeh, Dan Morris, Sara Beery, Neel Joshi, Nebojsa Jojic, and Jeff Clune. A deep active learning system for species identification and counting in camera trap images. arXiv, October 2019.
- Beery et al. [2019a] Sara Beery, Yang Liu, Dan Morris, Jim Piavis, Ashish Kapoor, Markus Meister, Neel Joshi, and Pietro Perona. Synthetic examples improve generalization for rare classes. arXiv, April 2019a.
- Beery et al. [2019b] Sara Beery, Dan Morris, and Siyu Yang. Efficient pipeline for camera trap image review. arXiv, July 2019b.
- Beery et al. [2019c] Sara Beery, Guanhang Wu, Vivek Rathod, Ronny Votel, and Jonathan Huang. Context R-CNN: Long term temporal context for Per-Camera object detection. arXiv, December 2019c.
- Tabak et al. [2020] Michael A Tabak, Mohammad S Norouzzadeh, David W Wolfson, Erica J Newton, Raoul K Boughton, Jacob S Ivan, Eric A Odell, Eric S Newkirk, Reesa Y Conrey, Jennifer Stenglein, Fabiola Iannarilli, John Erb, Ryan K Brook, Amy J Davis, Jesse Lewis, Daniel P Walsh, James C Beasley, Kurt C VerCauteren, Jeff Clune, and Ryan S Miller. Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2. Ecol. Evol., 10(19):10374–10383, October 2020.
- Ravoor and T. S. B. [2020] Prashanth C Ravoor and Sudarshan T. S. B. Deep learning methods for Multi-Species animal re-identification and tracking – a survey. Computer Science Review, 38:100289, November 2020.
- Vidal et al. [2021] Maxime Vidal, Nathan Wolf, Beth Rosenberg, Bradley P Harris, and Alexander Mathis. Perspectives on individual animal identification from biology and computer vision. Integr. Comp. Biol., 61(3):900–916, October 2021.
- Meek et al. [2013] Paul Douglas Meek, Karl Vernes, and Greg Falzon. On the reliability of expert identification of small-medium sized mammals from camera trap photos. Wildlife Biology in Practice, 9(2):1–19, 2013.
- Geng et al. [2020] Chuanxing Geng, Sheng-jun Huang, and Songcan Chen. Recent advances in open set recognition: A survey. IEEE transactions on pattern analysis and machine intelligence, 43(10):3614–3631, 2020.
- Crall et al. [2013] Jonathan P. Crall, Charles V. Stewart, Tanya Y. Berger-Wolf, Daniel I. Rubenstein, and Siva R. Sundaresan. Hotspotter — patterned species instance recognition. In 2013 IEEE Workshop on Applications of Computer Vision (WACV), pages 230–237, 2013. 10.1109/WACV.2013.6475023.
- Lowe [2004] David G Lowe. Distinctive image features from Scale-Invariant keypoints. Int. J. Comput. Vis., 60(2):91–110, November 2004.
- Rubenstein et al. [2018b] D. Rubenstein, C. Stewart J. Parham, T. Y. Berger-Wolf, J. Holmberg, J. Crall, B. L. Mackey, S. Funnel, K. Cockerill, Z. Davidson, L. Mate, C. Nzomo, R. Warungu, D. Martins, V. Ontita, J. Omulupi, J. Weston, G. Anyona, G. Chege, D. Kimiti, K. Tombak, A. Gersick, and N. Rubenstein. The state of kenya’s grevy’s zebras and reticulated giraffes: Results of the great grevy’s rally 2018. Wildlife Direct, 2018b.
- Park et al. [2019] Heebok Park, Anya Lim, Tae-Young Choi, Seung-Yoon Baek, Eui-Geun Song, and Yung Chul Park. Where to spot: individual identification of leopard cats (prionailurus bengalensis euptilurus) in south korea. Hangug hwangyeong saengtae haghoeji, 43(1):1–5, December 2019.
- Nipko et al. [2020] Robert B Nipko, Brogan E Holcombe, and Marcella J Kelly. Identifying individual jaguars and ocelots via pattern-recognition software: Comparing HotSpotter and wild-ID. Wildl. Soc. Bull., 44(2):424–433, June 2020.
- Blount et al. [2022] Drew Blount, Eve Bohnett, Jason Holmberg, Jason Parham, Sorosh Poya Faryabi, Örjan Johansson, Li An, Bilal Ahmad, Wajid Khan, and Stephane Ostrowski. Comparison of two individual identification algorithms for snow leopards after automated detection. bioRxiv, pages 2022–01, 2022. Publisher: Cold Spring Harbor Laboratory.
- Crall [2017] J. P. Crall. Identifying individual animals using ranking, verification, and connectivity. PhD thesis, Dept. Comput. Sci., Rensselaer Polytechnic Inst., Troy, NY, USA, 2017.
- Breiman [2001] Leo Breiman. Random forests. Mach. Learn., 45(1):5–32, October 2001.
- Pal [2005] M Pal. Random forest classifier for remote sensing classification. Int. J. Remote Sens., 26(1):217–222, January 2005.
- Zheng et al. [2017] Liang Zheng, Yujia Huang, Huchuan Lu, and Yi Yang. Pose invariant embedding for deep person re-identification. arXiv, January 2017.
- Moskvyak et al. [2021] Olga Moskvyak, Frederic Maire, Feras Dayoub, Asia O Armstrong, and Mahsa Baktashmotlagh. Robust re-identification of manta rays from natural markings by learning pose invariant embeddings. In 2021 Digital Image Computing: Techniques and Applications (DICTA), pages 1–8, November 2021.
- Stewart et al. [TBD] C. Stewart, T. Y. Berger-Wolf, J. Parham, A. Sundaresan, S. Beery, and J. Crall. Local clusterings and their alternatives. In preparation, TBD.
- Kulits et al. [2021] Peter Kulits, Jake Wall, Anka Bedetti, Michelle Henley, and Sara Beery. ElephantBook: A Semi-Automated Human-in-the-Loop system for elephant Re-Identification. arXiv, June 2021.
- Bodesheim et al. [2022] Paul Bodesheim, Jan Blunk, Matthias Körschens, Clemens-Alexander Brust, Christoph Käding, and Joachim Denzler. Pre-trained models are not enough: active and lifelong learning is important for long-term visual monitoring of mammals in biodiversity research—individual identification and attribute prediction with image features from deep neural networks and decoupled decision models applied to elephants and great apes. Mamm. Biol., 102(3):853–875, June 2022.
- Wang et al. [2016] Hanxiao Wang, Shaogang Gong, Xiatian Zhu, and Tao Xiang. Human-In-The-Loop person Re-Identification. arXiv, December 2016.
- Sundaresan et al. [2023] Avirath Sundaresan, Jason Parham, Jonathan Crall, Rosemary Warungu, Timothy Muthami, Jackson Miliko, Margaret Mwangi, Jason Holmberg, Tanya Berger-Wolf, Daniel Rubenstein, Charles Stewart, and Sara Beery. Adapting the re-id challenge for static sensors. 3rd International Workshop for Camera Traps, AI, and Ecology, 2023.
- Parham et al. [2018a] Jason Parham, Charles Stewart, Jonathan Crall, Daniel Rubenstein, Jason Holmberg, and Tanya Berger-Wolf. An animal detection pipeline for identification. In 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), pages 1075–1083, March 2018a.
- Parham and Stewart [2016] Jason Parham and Charles Stewart. Detecting plains and grevy’s zebras in the realworld. In 2016 IEEE Winter Applications of Computer Vision Workshops (WACVW), pages 1–9, March 2016.
- Petersen [1896] Carl Georg Johannes Petersen. The yearly immigration of young plaice into the Limfjord from the German Sea. Rep. Danish Biol. Station, 6(1):1–48, Jan. 1896.
- Parham et al. [2018b] Jason Parham, Charles Stewart, Tanya Berger-Wolf, Daniel Rubenstein, and Jason Holmberg. The great grevy’s rally: A review on procedure. In AI for Wildlife Conservation Workshop, 2018b.
- Green et al. [2020] Austin M Green, Mark W Chynoweth, and Çağan Hakkı Şekercioğlu. Spatially explicit capture-recapture through camera trapping: a review of benchmark analyses for wildlife density estimation. Frontiers in Ecology and Evolution, 8:563477, 2020.
- Service [2017] Kenya Wildlife Service. Recovery and action plan for grevy’s zebra (equus grevyi) in kenya (2017-2026). Technical report, Kenya Wildlife Service, Nairobi, Kenya, 2017.
Supplementary Figures
Additional data visualization and analysis figures have been included in this section to give additional context to the reader.



