Learning to Respond to Mixed-code Queries using BilingualWord Embeddings

1Citations
Citations of this article
73Readers
Mendeley users who have this article in their library.

Abstract

We present a method for learning bilingual word embeddings in order to support second language (L2) learners in finding recurring phrases and example sentences that match mixed-code queries (e.g., "sentence") composed of words in both target language and native language (L1). In our approach, mixed-code queries are transformed into target language queries aimed at maximizing the probability of retrieving relevant target language phrases and sentences. The method involves converting a given parallel corpus into mixed-code data, generating word embeddings from mixed-code data, and expanding queries in target languages based on bilingual word embeddings. We present a prototype search engine, x.Linggle, that applies the method to a linguistic search engine for a parallel corpus. Preliminary evaluation on a list of common word-translation shows that the method performs reasonably well.

Cite

CITATION STYLE

APA

Ho, C. F., Chen, J. J., Yang, C. Y., & Chang, J. S. (2019). Learning to Respond to Mixed-code Queries using BilingualWord Embeddings. In NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Demonstrations Session (pp. 24–28). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/n19-4005

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free