AI sensing for robotics using deep learning based visual and language modeling

1Citations
Citations of this article
58Readers
Mendeley users who have this article in their library.

Abstract

An artificial intelligence(AI) system should be capable of processing the sensory inputs to extract both task-specific and general information about its environment. However, most of the existing algorithms extract only task specific information. In this work, an innovative approach to address the problem of processing visual sensory data is presented by utilizing convolutional neural network (CNN). It recognizes and represents the physical and semantic nature of the surrounding in both human readable and machine processable format. This work utilizes the image captioning model to capture the semantics of the input image and a modular design to generate a probability distribution for semantic topics. It gives any autonomous system the ability to process visual information in a human-like way and generates more insights which are hardly possible with a conventional algorithm. Here a model and data collection method are proposed.

Cite

CITATION STYLE

APA

Singh, Y., & Kameshwar Rao, J. V. (2020). AI sensing for robotics using deep learning based visual and language modeling. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 60–63). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.challengehml-1.8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free