An artificial intelligence(AI) system should be capable of processing the sensory inputs to extract both task-specific and general information about its environment. However, most of the existing algorithms extract only task specific information. In this work, an innovative approach to address the problem of processing visual sensory data is presented by utilizing convolutional neural network (CNN). It recognizes and represents the physical and semantic nature of the surrounding in both human readable and machine processable format. This work utilizes the image captioning model to capture the semantics of the input image and a modular design to generate a probability distribution for semantic topics. It gives any autonomous system the ability to process visual information in a human-like way and generates more insights which are hardly possible with a conventional algorithm. Here a model and data collection method are proposed.
CITATION STYLE
Singh, Y., & Kameshwar Rao, J. V. (2020). AI sensing for robotics using deep learning based visual and language modeling. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 60–63). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.challengehml-1.8
Mendeley helps you to discover research relevant for your work.