Edge Computing for Vision-Based, Urban-Insects Traps in the Context of Smart Cities

17Citations
Citations of this article
48Readers
Mendeley users who have this article in their library.

Abstract

Our aim is to promote the widespread use of electronic insect traps that report captured pests to a human-controlled agency. This work reports on edge-computing as applied to camerabased insect traps. We present a low-cost device with high power autonomy and an adequate picture quality that reports an internal image of the trap to a server and counts the insects it contains based on quantized and embedded deep-learning models. The paper compares different aspects of performance of three different edge devices, namely ESP32, Raspberry Pi Model 4 (RPi), and Google Coral, running a deep learning framework (TensorFlow Lite). All edge devices were able to process images and report accuracy in counting exceeding 95%, but at different rates and power consumption. Our findings suggest that ESP32 appears to be the best choice in the context of this application according to our policy for low-cost devices.

Cite

CITATION STYLE

APA

Saradopoulos, I., Potamitis, I., Ntalampiras, S., Konstantaras, A. I., & Antonidakis, E. N. (2022). Edge Computing for Vision-Based, Urban-Insects Traps in the Context of Smart Cities. Sensors, 22(5). https://doi.org/10.3390/s22052006

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free