Most sentence embedding models typically represent each sentence only using word surface, which makes these models indiscriminative for ubiquitous homonymy and polysemy. In order to enhance discriminativeness, we employ concept conceptualization model to assign associated concepts for each sentence in the text corpus, and learn conceptual sentence embedding (CSE). Hence, the sentence representations are more expressive than some widely-used document representation models such as latent topic models, especially for short text. In the experiments, we evaluate the CSE models on two tasks, text classification and information retrieval. The experimental results show that the proposed models outperform typical sentence embedding models.
CITATION STYLE
Wang, Y., Huang, H., Feng, C., Zhou, Q., & Gu, J. (2016). Conceptual sentence embeddings. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9658, pp. 390–401). Springer Verlag. https://doi.org/10.1007/978-3-319-39937-9_30
Mendeley helps you to discover research relevant for your work.