High-risk learning: Acquiring new word vectors from tiny data

55Citations
Citations of this article
201Readers
Mendeley users who have this article in their library.

Abstract

Distributional semantics models are known to struggle with small data. It is generally accepted that in order to learn ‘a good vector’ for a word, a model must have sufficient examples of its usage. This contradicts the fact that humans can guess the meaning of a word from a few occurrences only. In this paper, we show that a neural language model such as Word2Vec only necessitates minor modifications to its standard architecture to learn new terms from tiny data, using background knowledge from a previously learnt semantic space. We test our model on word definitions and on a nonce task involving 2-6 sentences’ worth of context, showing a large increase in performance over state-of-the-art models on the definitional task.

References Powered by Scopus

A Neural Probabilistic Language Model

5160Citations
N/AReaders
Get full text

From frequency to meaning: Vector space models of semantics

2004Citations
N/AReaders
Get full text

The waCky wide web: A collection of very large linguistically processed web-crawled corpora

854Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Tail-GNN: Tail-Node Graph Neural Networks

88Citations
N/AReaders
Get full text

A la carte embedding: Cheap but effective induction of semantic feature vectors

79Citations
N/AReaders
Get full text

An Empirical Survey of Data Augmentation for Limited Data Learning in NLP

77Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Herbelot, A., & Baroni, M. (2017). High-risk learning: Acquiring new word vectors from tiny data. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 304–309). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1030

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 80

69%

Researcher 23

20%

Professor / Associate Prof. 9

8%

Lecturer / Post doc 4

3%

Readers' Discipline

Tooltip

Computer Science 100

78%

Linguistics 15

12%

Engineering 10

8%

Neuroscience 4

3%

Save time finding and organizing research with Mendeley

Sign up for free