Processing citizen science- and machine-annotated time-lapse imagery for biologically meaningful metrics

12Citations
Citations of this article
71Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Time-lapse cameras facilitate remote and high-resolution monitoring of wild animal and plant communities, but the image data produced require further processing to be useful. Here we publish pipelines to process raw time-lapse imagery, resulting in count data (number of penguins per image) and ‘nearest neighbour distance’ measurements. The latter provide useful summaries of colony spatial structure (which can indicate phenological stage) and can be used to detect movement – metrics which could be valuable for a number of different monitoring scenarios, including image capture during aerial surveys. We present two alternative pathways for producing counts: (1) via the Zooniverse citizen science project Penguin Watch and (2) via a computer vision algorithm (Pengbot), and share a comparison of citizen science-, machine learning-, and expert- derived counts. We provide example files for 14 Penguin Watch cameras, generated from 63,070 raw images annotated by 50,445 volunteers. We encourage the use of this large open-source dataset, and the associated processing methodologies, for both ecological studies and continued machine learning and computer vision development.

Cite

CITATION STYLE

APA

Jones, F. M., Arteta, C., Zisserman, A., Lempitsky, V., Lintott, C. J., & Hart, T. (2020). Processing citizen science- and machine-annotated time-lapse imagery for biologically meaningful metrics. Scientific Data, 7(1). https://doi.org/10.1038/s41597-020-0442-6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free