CAMERA TRAP STUDIES AS GENERATORS OF BIG DATA: THE NEED FOR EFFICIENT AND STANDARDIZED DATA MANAGEMENT
Camera traps—remotely triggered cameras to photograph wildlife—were not widely used to monitor or detect wildlife until toward the end of the 20th century when the technology was adopted by hunters to assist in tracking game. This new interest created a growing market for camera trap technologies, bringing costs down and increasing the variety of available equipment (Sanderson & Trolle, ). In addition to this, the move from film to digital cameras and advances in infrared sensors meant that ecological monitoring using camera traps became more feasible, accessible, and affordable (Burton et al., ; Kucera & Barrett, ). Now camera traps are used for a wide range of activities with different purposes such as biodiversity inventories, biodiversity and population monitoring, ecological and behavioral research, monitoring of human impact on ecosystems, and effectiveness of conservation interventions (Burton et al., ; O'Connell, ; Rovero, Tobler, & Sanderson, ; Rovero, Zimmermann, Berzi, & Meek, ; Rowcliffe & Carbone, ; Steenweg et al., ; Swann & Perkins, ). The large amount of information on camera trap technology, survey design, and statistical analysis shared in publications and conferences mirrors this interest (O'Connell, ). Numbers of published studies using camera traps have increased rapidly (Burton et al., ; Rowcliffe & Carbone, ), doubling every 2.9 years (Steenweg et al., ). Steenweg et al. () estimated that camera traps have been employed at a minimum of 160,000 individual sites globally, with an average of 78 cameras used per study. Camera traps nowadays generate a vast amount of data (Rovero et al., ; Sanderson & Harris, ), with some studies recording up to 2.6 million images (McShea, Forrester, Costello, He, & Kays, ).
Once camera trap images have accumulated, a process starts that ultimately leads to data ready for analysis (e.g., see Harris, Thompson, Childs, & Sanderson, ; Krishnappa & Turner, ; Niedballa, Sollmann, Courtiol, & Wilting, ). First, images have to be retrieved from camera traps and stored securely. Secondly, files may need to be organized and labeled. Third—and often most time‐consuming—is image content identification and information management, which is also called image interpretation or annotation. Image annotation can be performed completely manually, but is becoming increasingly facilitated by technology, such as extraction of metadata from images for certain data (e.g., date and time). This whole process will be called “data management” hereafter. The final step of data analysis is covered elsewhere (e.g., O'Connell, Nichols, & Karanth, ; Rovero et al., ).
Several problems arise during camera trap data management. First, processing the massive amounts of data accumulated in camera trap studies requires a high amount of resources in terms of time and man power. This leads to a situation where data management rather than data collection is the limiting factor in the completion of studies (Barrueto, Clevenger, Dorsey, & Ford, ; Bubnicki, Churski, & Kuijper, ). Cataloging and classification of data often lag behind data acquisition, and sometimes a large amount of data remains unused and ultimately lost for science and conservation management (Harris et al., ). Further, such conservation and research projects are often tax‐payer funded through sources such as governmental grants, and uncataloged and analyzed data lead to a loss of public funds. In addition to this, such funds are often tied to a specific topic or research question, and so the cataloging and analyzing of images focus solely on the target species, for example. If all data were cataloged, however, then more conservation‐relevant outcomes in relation to funds could be generated, representing better “value for money”.
Secondly, as retrieval, storage and extraction of data from images is still mostly performed manually (except for standardized image metadata tags such as date, time, and label), human errors are introduced into data management and can lead to unintended data loss and incorrect data extraction from images (Krishnappa & Turner, ; Maydanchik, ; Sanderson & Harris, ).
Secondly, as retrieval, storage and extraction of data from images is still mostly performed manually (except for standardized image metadata tags such as date, time, and label), human errors are introduced into data management and can lead to unintended data loss and incorrect data extraction from images (Krishnappa & Turner, ; Maydanchik, ; Sanderson & Harris, ).
Finally, different choices in the process of data management, such as data coding, labeling of images, or data storage, may lead to a lack of accessibility, transparency, and inconsistency between projects. This prevents between‐project cooperation or effective data sharing (Harris et al., ; Meek et al., ; Rowcliffe & Carbone, ). Indeed, Chaudhary, Walters, Bever, Hoeksema, and Wilson () found that across‐site comparisons and meta‐analyses are almost absent from the literature. Although two‐thirds of camera trap studies assessed by Burton et al. () focus on multiple species as opposed to single species, often only a small proportion of these images contain the information required for a project, and a project may only annotate in respect to a limited amount of certain data variables, but not in others (Bubnicki et al., ; Wong & Kachel, ). However, in order to easily reuse the same dataset for other research areas (e.g., examining interspecific interactions and human‐wildlife conflict) or conservation management, all original images should be cataloged immediately, and in a transparent and consistent way (Burton et al., ; Chaudhary et al., ; Forrester et al., ). Having data accessible and in a standardized format makes it easier to share data and contribute to other projects which may have different focal species, as well as being able to compare sites (Wong & Kachel, ). A lot of research is conducted as short‐term projects or by individual researcher, so after their completion potentially important data may be lost for further future analysis (Hampton et al., ). The importance of sharing data in a consistent way does not only apply to the effective cooperation between research projects, but also within projects that involve large teams, often in different locations of the world, or projects relying on citizen science. In summary, standardized and transparent data management is essential to drive science and conservation management forward. This will enable researchers and practitioners to be successful in answering pressing ecological questions and guide conservation policy on a meaningful, possibly global, level (Meek et al., ; Wildlife Insights, ).
With this background situation, it is widely acknowledged that there is a requirement for a universal, user‐friendly, and standardized way to manage, store, classify, and share camera trap data (Sanderson & Harris, ; Barrueto et al., ; Forrester et al., ). Data management and processing facilitated by technology would decrease required resources, significantly decrease the risk of human error, and make sharing of data between and within projects easy.
With this background situation, it is widely acknowledged that there is a requirement for a universal, user‐friendly, and standardized way to manage, store, classify, and share camera trap data (Sanderson & Harris, ; Barrueto et al., ; Forrester et al., ). Data management and processing facilitated by technology would decrease required resources, significantly decrease the risk of human error, and make sharing of data between and within projects easy.
The aim of this paper was to give an overview of available programs for camera trap studies that facilitate efficient and standardized data management. We assessed whether the available software is suitable for the various needs of different camera trap projects and identified gaps that should be addressed in further developments of comprehensive software that meet the most common and important needs.
The data collection for this review was conducted from January to April 2017. To identify relevant software, we initially used the search term “camera trap data management” on Google and Google Scholar. This approach identified most existent software, while Agouti was identified by utilizing our professional contacts.
We conducted informal surveys (Barrueto et al., ) among colleagues and other researchers, enabling us to determine features and specifications that are perceived to be important to camera trap data management, as well as identify important data variables recorded during camera trap studies. Colleagues and researchers were identified using a snowball method, for example, each contact person was asked to suggest other experts in the field. In addition to experts’ advice, we included software characteristics identified by Ivan and Newkirk (), Niedballa et al. () and Scotson et al. (). Using an existing set of camera trap images, we tested the available software, recorded each software's features and characteristics (Ivan & Newkirk, ; Niedballa et al., ; Scotson et al., ), and assessed its ability to record the data variables we had identified. If more features were found during this testing, we added them to the list and tested all software against them. Next, we compared all features between programs and to the potential needs of projects. Finally, we explored potentially useful features that have not yet been incorporated into available software, such as automatic subject recognition (He et al., ; McShea et al., ; Wang, ; Yu et al., ).
To simplify, we did not include applications that are used to primarily allow researchers to crowdsource image classification and annotation. These web applications allow citizen scientists to perform basic classification and annotation of images via a website (O'Connell, ), such as with Snapshot Serengeti (Hines, Swanson, Kosmala, & Lintott, ; Swanson et al., ) which uses the platform Zooniverse (Zooniverse ). Such applications are designed to be easy‐to‐use for nonexperts and thus do not usually include many of the features that may be necessary for a comprehensive standardized software.
CURRENTLY AVAILABLE CAMERA TRAP DATA MANAGEMENT SOFTWARE AND THEIR FEATURES
The vast amount of data being generated by camera trap studies around the world has led to an increasing number of programs to manage and process the generated image data. The number of available software increased from four programs identified in 2013 (Barrueto et al., ), five and seven, respectively, in 2016 (Ivan & Newkirk, ; Niedballa et al., ) and eight in 2017 (Scotson et al., ) to twelve published or otherwise available programs identified by us (Table ). As long as no standardized and widely available software is produced and offered, the number of different, specifically tailored pieces of software may increase with the number of projects generating data.
Software characteristics and features of tested camera trap data management programs, as well as image variables that tested programs are able to record and process. —represents “no”, or the absence of a featureRenamer & CamTrapa | ViXeNb | Aardwolfc | Camelotd | Snoopye | Wild.IDf | Camera Baseg | CPW Photo Warehouseh | eMammal | camtrapRj | TRAPPERk | Agoutil | |
General features | ||||||||||||
Operating system | Windows | Windows, MacOS | Windows, MacOS, Linux | Windows, Linux | Windows, MacOS | Windows | Windows | Windows | Windows, MacOS | Windows, MacOS, Linux | Windows, MacOS, Linux | Windows |
Installation requirements | .exe | .exe | mySQL | mySQL | mySQL | Java | MS Access | MS Access | Internet access | R | Internet access, see website | Internet access |
Requires coding skills | Yes | – | – | – | – | – | – | – | – | Yes (R) | –m | – |
Open source | Yes | Yes | Yes | – | – | – | – | – | – | Yes | Yes | – (but open to partners) |
Web‐based | – | – | – | – | – | – | – | – | Yes | – | Yes | Yes |
Data storage | Local | Local | Local | Local | Local | Local | Local | Local | Cloud | Local | Server–based | Cloud |
Image storage capacity | Unlimited | c. 1,000,000 | Unlimited | Unlimited | Unlimited | Unlimited | c. 2,000,000 | c. 2,000,000 | Unlimited | Unlimited | Unlimited | Unlimited |
Functionality | ||||||||||||
Automatic metadata import | – | – | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Still/moving images | Still | Still | Still | Both | Both | Still | Both | Still | Still | Still | Both | Still |
In‐built media viewer | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | – | Yes | Yes |
Batch ID | – | – | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Capture intervals | – | – | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Filter/query data | – | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Record active days | Yes | – | – | Yes | – | Yes | Yes | Yes | – | Yes | Yes | Yes |
Automatic subject detection | – | – | – | – | Planned | – | – | – | Yes | – | – | – |
Automatic species recognition | – | – | – | – | Planned | – | – | – | Yes | – | – | – |
In‐built mapping | – | – | – | – | – | – | Yes | – | – | Yes | Yes | Yes |
In app analysis | – | – | – | – | – | – | – | – | – | Yes | –m | – |
Generation of standard reports | Yes | – | Yes | Yes | – | – | Yes | Yes | Yes | – | –m | – |
Generate input files | .csvn; PRESENCE | – | – | .csvn; PRESENCE; R | Yes | – | .csvn; Excel; MARK; CAPTURE; DENSITY; PRESENCE; EstimateS | .csvn; Excel; MARK; PRESENCE; DENSITY; R | .csvn; PRESENCE, R | .csvn | .csvn | .csvn |
Recordable data | ||||||||||||
Camera make/model | Yesp | Yesp | Yesp | – | Yes | Yes | Yes | – | – | Yes | Yeso | Yes |
Drop down species list | – | – | Yes | Yes | Yes | Yes | Yes | Yes | Yes | – | Yeso | Yes |
Multiple species | – | Yes | Yes | Yes | Yes | Yes | – | Yes | Yes | Yes | Yeso | Yes |
ID individuals | – | – | Yes | – | Yes | Yes | Yes | Yes | Yes | Yes | Yeso | Yes |
Group size | – | – | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yeso | Yes |
Age/sex classes per ID | – | – | Yes | Yes | Yes | – | Sex only | Notes | Yes | Yes | Yeso | Yes |
Behavior per ID | – | – | Yes | – | – | – | – | – | – | Yes | Yeso | Yes |
Weather variables | – | – | Yes | – | Yes | – | – | – | – | Yes | Yeso | – |
Moon phase | – | – | Yes | – | Yes | Yes | – | – | – | Yes | Yeso | – |
Sunrise/sunset | – | – | Yes | – | Yes | – | Yes | – | – | Yes | Yeso | – |
Location variables | – | – | Yes | Yes | Yes | – | – | – | – | Yes | Yeso | – |
Latitude/longitude | – | – | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yeso | Yes |
Altitude | – | – | Yes | – | Yes | – | – | – | – | Yes | Yeso | Yes |
Spatial (habitat) characteristics | – | – | Yes | Yes | – | – | – | – | – | Yes | Yesp | – |
Support | ||||||||||||
Multiple/shareable users | – | – | – | Yes | – | – | – | Yes | Yes | Yes | Yes | Yes |
Crowd source IDs | – | – | – | – | – | – | – | Yes | – | Yes | Yes | Yes |
Help/support | User manual | User manual | – | User manual, forum | User manual | User manual | User manual | User manual | Forum | Forum | Forum | User manual |
Subscription fee? | – | – | – | – | – | – | – | – | Yes | – | – | Yes (to be discussed) |
Reference | Harris et al. (); Sanderson and Harris () | Ramachandran and Devarajan () | Krishnappa and Turner () | Hendry and Mann () | Smedley and Terdal () | Fegraus et al. () | Tobler () | Ivan and Newkirk () | Forrester et al. () | Niedballa et al. () | Bubnicki et al. () |
aG. Harris & J. Sanderson; smallwildcats.com/camera‐trap‐instructions bK. Devarajan & P. Ramachandran; github.com/vixen‐project/vixen cUniversity of Oslo; github.com/yathin/aardwolf2 dFlora & Fauna International – Vietnam; gitlab.com/camelot‐project eR. Smedley; tulsasoftdb.com/snoopy fTEAM Network; wildid.teamnetwork.org gSan Diego Zoo Global Institute for Conservation Research;
During the testing phase, we checked all available programs against software characteristics they had to offer and the image‐related data variables that programs were able to record (Table ). These features are discussed below, structured according to steps being involved in data management (Ivan & Newkirk, ).
Metadata import
The very first step of downloading and storing data from camera traps still needs to be performed manually. This preprocess can include the organization of directories, for instance. Some programs require a specific folder path and folder names to be able to process the data further (e.g., CamTrap), while others only require a basic folder structure with locations and camera trap ID. Another step may be the conversion of files, for example videos into web‐friendly formats if web applications are used. Extraction of metadata, such as date and time of image, is a crucial step of data management as it significantly decreases data entry errors. Some software are restricted to facilitating image metadata extraction (e.g., for date, time), but have no other application (Table ). Other available metadata depends on the camera trap model; some models are able to record temperature, location, moon phase etc. It is important to note that while image standards such as Exif standardize certain metadata tags (e.g., time, date, and camera settings), other camera trap specific tags are not standardized (e.g., temperature, location). Therefore, a standardization of metadata by camera trap manufacturers would be of advantage (Forrester et al., ).
Facilitation of classification and annotation of images
Most available programs were developed for a specific project (but see e.g., Agouti, eMammal, TRAPPER), and accordingly, image classification is tailored to the respective project focus. Although some software is openly available and has been used for other projects, the tailored range of features and variables compromises the use by a wider audience. This is mainly because of a fixed, predetermined range of survey, camera trap location, and image variables that can be entered (Table ). A range of variables are listed in Table . For instance, Wild.ID is highly suited for use on general biodiversity surveys, as it allows tagging of species in images; however, it does not offer the flexibility to gather more in‐depth data for each image (e.g., behavior). Camera Base was originally developed for tiger surveys and facilitates the (manual) identification of individuals in images in order to perform capture–mark–recapture analysis for population estimates. However, if the same image dataset were to be used for another research focus, for example, a project focussed on behavior instead of population size surveys, annotation of images would be difficult. One exception is for instance TRAPPER, an open‐source program designed to address a larger range of species and topics, and with an associated forum where users can discuss the use and further development of the program. Open‐source programs (Aardwolf, camtrapR, TRAPPER, ViXeN) are relatively rare, but may offer the opportunity to tailor image classification, survey characteristics, and other functions to specific projects. A disadvantage may be that this individualization of program features can lead to new problems in standardizing data management. Some projects may need an even more flexible design; ex situ research projects involving camera trapping may seek to study enclosure use (e.g., frequency of using certain enclosure parts in relation to visitor numbers) or very specific behaviors (e.g., stereotyping or aggression), while projects focussing on the monitoring of illegal human activities, for instance, may need to record those activities in relation to wildlife populations (e.g., presence and activity of people in images). In conclusion, the facilitation of image classification and annotation is extremely important and widely recognized in most programs, but can still be restricted to the specific variables targeted in a given project. A start for standardizing data management could be made by agreeing on certain standards within distinct camera trap communities (e.g., taxonomic, geographical, or thematic focus), as an immediate standardization of all processes for all current camera trap programs may be impractical and impossible.
Almost all programs have an in‐built media viewer that makes it easy to go through images in order to classify them. Usually, the images that can be uploaded into these viewers are “still images”, but a few programs allow “moving images” (videos) as well. One problem with videos is that metadata are usually not stored on the file; for example, original date and time may get overwritten during image storing processes. Videos however may yield more information, for example in behavioral research (Kuijper, Bubnicki, Churski, Mols, & van Hooft, ; Swinnen, Reijniers, Breno, & Leirs, ) and particularly in marine studies (Bond et al., ; Ebner et al., ). As a result, some programs (e.g., eMammal) allow loading a series of photos taken from the same event and classify it as one data point, effectively creating a low frame rate video (McShea et al., ). Other programs (Camelot, CameraBase, Snoopy and TRAPPER) are able to deal with video file types (e.g., .avi) directly.
Camera Base, camtrapR, and TRAPPER feature various mapping capabilities. Camera Base and camtrapR can generate simple maps of camera locations and species records within the programs, while camtrapR can generate shapefiles to use with GIS software (Niedballa et al., ), and CameraBase can generate custom lists for use with GIS. TRAPPER features an interface for mapping within the program, while the provision of an Application Programming Interface (API) means that there is functionality with other open‐source software such as QGIS (Bubnicki et al., ).
Most of the tested programs offer the ability to sort and query data according to assigned tags. This function can be used to explore images for specific data occurrences (behaviors, species etc.) or to filter images for use in communications or outreach. Further, this can allow researchers to classify certain photos that may need to be reassessed—for example Wild.ID offers a “certainty” field, allowing users to tag and filter images according to the confidence of their assessment.
Generating export files and automatic analyses
All software tested offer the generation of a range of export files that can be used by a variety of analysis software, such as MARK (White & Burnham, ), PRESENCE (Hines, ), DENSITY (Efford, Dawson, & Robbins, ), and R packages (e.g., unmarked [Fiske & Chandler, ], secr [Efford, ] or RMark [Laake, ]). Usually, these are comma separated values (.csv) or text (.txt) files, but some software also allow more specialized input files—for example, the generation of shapefiles for GIS programs by camtrapR. In‐built analysis of data is offered by camtrapR; however, knowledge of the coding language R is necessary for analysis.
Other characteristics
Ease of use
Some programs tested are based on MS Access (Camera Base and CPW Photo Warehouse), and their manipulation requires respective skills. While MS Access skills are still relatively common, more complicated IT or programming skills are less widespread. The software TRAPPER is comprehensive, but it has one major limitation, which is acknowledged by its developers; the installation and maintenance require good IT knowledge for initial server configuration and updating of the source code (Bubnicki et al., ). camtrapR is a further open‐source program; however, its use requires knowledge and understanding of the R programming language. For a program to be used by a wide range of projects, especially projects led by local practitioners with limited IT and programming skills, the installation and tailoring of certain features to specific needs (e.g., variables needed for data collection) should be as easy as possible for the project administrator. In this respect, the inclusion of project configuration wizards into the software would be of great advantage.
Web‐based programs and multi‐user options
In large multi‐site projects, various people may work from different locations on the same data set. In order to make work‐flow more effective, web‐ or cloud‐based software as well as the login of multiple users should be possible. Agouti, eMammal, and TRAPPER are web‐based, while several other programs allow multiple users. However, few programs so far allow the definition of the role of users; Agouti has two user levels, Project Investigator and other workers, that allow different activities. Assigning user roles can be very useful, for example, the set‐up of the study and data variables to be collected can only be set‐up by an administrator, while short‐term staff may just be allowed to have access to a certain part of the data processing. Where programs allow multiple users, a useful feature would be the ability to track changes in data annotation and image processing.
Costs and memberships
Of the programs tested, eMammal has subscription and ongoing costs, leveled according to the number of images that need to be processed and stored, while for Agouti costs can be discussed. Both are among the most advanced and comprehensive of the programs assessed here. Assuming that maintenance and service costs increase with wider use, more comprehensive and widely available programs may be costlier for users (unless alternative means of funding, maintenance, and/or user service are used). All other programs tested are freely available, either by publicly available download or request.
FUTURE FEATURES
While technology and informatics in general is an extremely fast developing area, technology used in camera trap data management is moving relatively slowly. Often relevant technology exists, but a lack of interdisciplinary approaches as well as openness of technology means its application in camera trap data management is limited.
A less common, but useful, feature is automatic subject or movement detection, which is used by eMammal (He et al., ) and MotionMerkat (Weinstein, ). Generally, pixel values of a frame are compared to the distribution of pixels in the previous frame, adjusted for a variance, and either classified as background if no change occurred and foreground if a change occurred. In eMammal, this foreground is then extracted and displayed, which helps to identify small or well camouflaged animals that triggered the camera trap but may be easily overlooked when classifying through images (He et al., ; McShea et al., ). Following the automatic detection of the subject, the species can be manually identified.
It is recognized that species identification requires comprehensive taxonomic knowledge and is one of the most time‐consuming areas of work for nonexperts such as citizen scientists (He et al., ). Thus, a potential future feature that could further enhance data analysis and management is automatic species recognition. Various approaches for similar classification programs exist for other purposes, such as facial recognition software, and it is surprising that image classification in camera trapping is still performed manually (Yu et al., ). A few authors have attempted to develop species recognition processes, with varying success (He et al., ; McShea et al., ; Norouzzadeh et al., ; Villa, Salazar, & Vargas, ; Wang, ; Yu et al., ). Yu et al. (), for instance, developed a mechanism that extracts the foreground (the animal) from the background, analyses the features of the object, and finally classifies the images by a linear support vector machine algorithm. Using 7,000 camera trap images of 18 species from two different field sites, the authors report a classification accuracy of 82%. Wang () reaches similar accuracies ranging between 77% and 87%, depending on the exact classification method and dataset, while He et al. () reports lower levels of accuracy (34% and 38%; depending on method used). Most recently, Villa et al. () and Norouzzadeh et al. () have reported up to 98.1% and 96.6% accuracy, respectively, however, this drops significantly when incorporating an unbalanced dataset including uncommon species. Indeed, a disadvantage is that for each species, a high number of “practice” images must be available for the algorithm to learn, and this may not be possible for rare species (Norouzzadeh et al., ). Thus, a further development of automatic species recognition software would be extremely useful, for instance combined with the option to allow human assessments for low accuracy annotations or rare species.
A similar direction, but more specific to a single method, is computer‐assisted data extraction for population size estimations. Some software are able to match the same individuals based on natural individual markers, such as fur patterns, and use capture–mark–recapture methods to estimate population sizes (reviewed in Bolger, Morrison, Vance, Lee, & Farid, ). This approach is being developed for use in Snoopy (Smedley & Terdal, ). Further, McShea et al. () aim to develop algorithms for eMammal that will extract information for use in Random Encounter Modeling, such as rate of movement (Rowcliffe, Carbone, Jansen, Kays, & Kranstauber, ), as well as biometric and behavioral data such as body size or group size. The integration of such algorithms into comprehensive software offering a range of different methods used in ecology seems sensible.
The burgeoning field of citizen science is being slowly integrated into camera trap data management software. While some specific websites and underlying processes exist that allow for basic image classification (e.g., SnapshotSerengeti, using Zooniverse), CPW Photo Warehouse and camtrapR are the only comprehensive programs that have specific features that facilitate the setting up of user IDs for crowd sourcing image classification.
In the future, high quality sound recording should be integrated into a camera trapping approach and respective software options developed, such as automatic recognition of sound patterns (Zaragozí, Belda, Giménez, Navarro, & Bonet, ). Soundscape ecology is defined as the collection of biological, geophysical, and anthropogenic sounds that emanate from a landscape and which vary over space and time reflecting important ecosystem processes and human activities and can be used to answer a variety of research questions (Pijanowski et al., ). Sound collection is more dynamic and may cover a wider range around the recording device in comparison with more spatially restricted image collection. By adding high quality audio sensors, a wider range of species may be detected (e.g., arboreal species), and the value of effort spent on data collection is maximized.
As a result of a more widely used comprehensive data management software, standardized data can be collected and used to answer spatially and temporally broader questions (Steenweg et al., ). To be able to make these data accessible, the development of software must go hand‐in‐hand with a data archive. Such databanks already exist for ecological research. Examples include EURODEER, a spatial database that stores shared roe deer (Capreolus capreolus) movement data and enables researchers to collaborate and produce better science (EURODEER, ). MOVEBANK is not restricted to a taxonomic group and archives animal movement data (Wikelski & Kays, ). The federated Wildlife Insights project has the potential for a similar platform for camera trap data (Wildlife Insights, ); for instance, the TEAM network that developed Wild.ID is working with this platform (Steenweg et al., ). Further, the Data Observation Network for Earth (DataONE) federation (Allard ) and its associated “nodes”, for example the Dryad Digital Repository (White, Carrier, Thompson, Greenberg, & Scherle, ), offer long‐term storage for scientific data. Using these repositories, standardized ecological data and metadata can be readily stored and accessed (Reichman, Jones, & Schildhauer, ), making them a valuable resource in standardizing camera trap data management.
CONCLUSIONS
Considering the increase of camera trap studies in the last few decades and the amount of data generated, it is surprising that so few comprehensive camera trap data management programs have been developed. Instead project‐specific software solutions are still dominant in the camera trapping community (Bubnicki et al., ). As these programs are not widely used across the camera trap community yet, they do not require high resources necessary for maintenance and service.
If databases are to be widely used and have impact on regional and global decision‐making, data management software should be easy‐to‐use, accessible, and in an open‐source format (Steenweg et al., ). So far, we have not been able to identify a single piece of software that would cover all possible needs that a variety of projects may have. Although we recognize that it may be challenging to develop a “one size fits all” solution (Barrueto et al., ), a comprehensive, yet flexible, program would be very beneficial in order to decrease the amount of resources needed to manage and classify camera trap data, decrease the amount of human error, and increase transparency and repeatability of projects. In order to serve this purpose, the respective program would include many of the features identified in this review (Table ). Being open source may allow tailoring of the available features to the projects’ needs easily, but should do so in a transparent way, so data can still be shared. An interdisciplinary approach would be needed to facilitate the development of such a program, as the input of ecologists and computer programmers is needed. Such a program would complement the recommendation made by Scotson et al. () that researchers should “adopt a standardized, nonproprietary, and transferable data storage format to store all camera trap data”.
Extending from recommendations by other authors to standardize camera trap methods and study designs (e.g., Ahumada et al., ; Scotson et al., ), we encourage a higher transparency in camera trap data management, processing, and storage in order to make datasets easily available for other purposes. This can be performed by developing more comprehensive and user‐friendly software. In agreement with Nichols, Karanth, and O'Connell (), we stress that the generation of big data is not the end purpose, but the understanding of ecological systems (science) and the effort to conserve and improve these ecosystems by informing decision‐making (conservation, management) should be the end purpose of these studies (Rowcliffe & Carbone, ; Steenweg et al., ).
ACKNOWLEDGMENTS
This work was supported by the Biotechnology and Biological Sciences Research Council [grant number BB/J014508/1], a Doctoral Training Programme studentship awarded to SY. The authors would like to thank the editor and two anonymous reviewers for their valuable comments and feedback.
CONFLICT OF INTEREST
The authors have no conflict of interests.
AUTHORS’ CONTRIBUTIONS
JRM and SY developed the concept of the review and collected information; all authors analyzed the findings. JRM and SY led the writing of the manuscript. All authors contributed critically to the drafts and gave final approval for publication.
DATA ACCESSIBILITY
No data were collected.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2018. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Improving technology and increasing affordability mean that camera trapping—the use of remotely triggered cameras to photograph wildlife—is becoming an increasingly common tool in the monitoring and conservation of wild populations. Each camera trap study generates a vast amount of data, which need to be processed and labeled before analysis. Traditionally, processing camera trap data has been performed manually by entering data into a spreadsheet. This is time‐consuming, prone to human error, and data management may be inconsistent between projects, hindering collaboration. Recently, several programs have become available to facilitate and quicken data processing. Here, we review available software and assess their ability to better standardize camera trap data management and facilitate data sharing and collaboration. To identify available software for camera trap data management, we used internet searches and contacted researchers and practitioners working on large camera trap projects, as well as software developers. We tested all available programs against a range of software characteristics in addition to their ability to record a suite of important data variables extracted from images. We identified and reviewed 12 available programs for the management of camera trap data. These ranged from simple software assisting with the extraction of metadata from an image, through to comprehensive programs that facilitate data entry and analysis. Many of the programs tested were developed for use on specific studies and so do not cover all possible software or data collection requirements that different projects may have. We highlight the importance of a standardized software solution for camera trap data management. This approach would allow all possible data to be collected, enabling researchers to share data and contribute to other studies, as well as facilitating multi‐project comparisons. By standardizing camera trap data collection and management in this way, future studies would be better placed to guide conservation policy on a global level.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details

1 School of Life Sciences, University of Nottingham, Nottingham, UK; IUCN/SSC Asian Wild Cattle Specialist Group, c/o Chester Zoo, Chester, UK
2 The North of England Zoological Society/Chester Zoo, Chester, UK
3 Zoological Society of London, London, UK