Abstract
Semantic models derived from visual information have helped to overcome some of the limitations of solely text-based distributional semantic models. Researchers have demonstrated that text and image-based representations encode complementary semantic information, which when combined provide a more complete representation of word meaning, in particular when compared with data on human conceptual knowledge. In this work, we reveal that these vision-based representations, whilst quite effective, do not make use of all the semantic information available in the neural network that could be used to inform vector-based models of semantic representation. Instead, we build image-based meta-embeddings from computer vision models, which can incorporate information from all layers of the network, and show that they encode a richer set of semantic attributes and yield a more complete representation of human conceptual knowledge.
Cite
CITATION STYLE
Derby, S., Miller, P., & Devereux, B. (2020). Encoding Lexico-Semantic Knowledge using Ensembles of Feature Maps from Deep Convolutional Neural Networks. In COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference (pp. 1906–1921). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.173
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.