Semantic Scene Understanding with Large Language Models on Unmanned Aerial Vehicles

15Citations
Citations of this article
28Readers
Mendeley users who have this article in their library.

Abstract

Unmanned Aerial Vehicles (UAVs) are able to provide instantaneous visual cues and a high-level data throughput that could be further leveraged to address complex tasks, such as semantically rich scene understanding. In this work, we built on the use of Large Language Models (LLMs) and Visual Language Models (VLMs), together with a state-of-the-art detection pipeline, to provide thorough zero-shot UAV scene literary text descriptions. The generated texts achieve a GUNNING Fog median grade level in the range of 7–12. Applications of this framework could be found in the filming industry and could enhance user experience in theme parks or in the advertisement sector. We demonstrate a low-cost highly efficient state-of-the-art practical implementation of microdrones in a well-controlled and challenging setting, in addition to proposing the use of standardized readability metrics to assess LLM-enhanced descriptions.

Cite

CITATION STYLE

APA

de Curtò, J., de Zarzà, I., & Calafate, C. T. (2023). Semantic Scene Understanding with Large Language Models on Unmanned Aerial Vehicles. Drones, 7(2). https://doi.org/10.3390/drones7020114

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free