Gaussian visual-linguistic embedding for zero-shot recognition

29Citations
Citations of this article
98Readers
Mendeley users who have this article in their library.

Abstract

An exciting outcome of research at the intersection of language and vision is that of zero-shot learning (ZSL). ZSL promises to scale visual recognition by borrowing distributed semantic models learned from linguistic corpora and turning them into visual recognition models. However the popular word-vector DSM embeddings are relatively impoverished in their expressivity as they model each word as a single vector point. In this paper we explore word- distribution embeddings for ZSL. We present a visual-linguistic mapping for ZSL in the case where words and visual categories are both represented by distributions. Experiments show improved results on ZSL benchmarks due to this better exploiting of intra-concept variability in each modality.

References Powered by Scopus

Learning to detect unseen object classes by between-class attribute transfer

2047Citations
N/AReaders
Get full text

Learning deep representations of fine-grained visual descriptions

720Citations
N/AReaders
Get full text

Multimodal distributional semantics

685Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Zero-shot learning-a comprehensive evaluation of the good, the bad and the ugly

1117Citations
N/AReaders
Get full text

A survey of zero-shot learning: Settings, methods, and applications

544Citations
N/AReaders
Get full text

Learning a deep embedding model for zero-shot learning

541Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Mukherjee, T., & Hospedales, T. (2016). Gaussian visual-linguistic embedding for zero-shot recognition. In EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 912–918). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d16-1089

Readers over time

‘16‘17‘18‘19‘20‘21‘22‘23‘2406121824

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 34

74%

Researcher 8

17%

Professor / Associate Prof. 2

4%

Lecturer / Post doc 2

4%

Readers' Discipline

Tooltip

Computer Science 43

80%

Linguistics 5

9%

Engineering 4

7%

Social Sciences 2

4%

Save time finding and organizing research with Mendeley

Sign up for free
0