Relation classification via recurrent neural network with attention and tensor layers

52Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

Relation classification is a crucial component in many Natural Language Processing (NLP) systems. In this paper, we propose a novel bidirectional recurrent neural network architecture (using Long Short-Term Memory, LSTM, cells) for relation classification, with an attention layer for organizing the context information on the word level and a tensor layer for detecting complex connections between two entities. The above two feature extraction operations are based on the LSTM networks and use their outputs. Our model allows end-to-end learning from the raw sentences in the dataset, without trimming or reconstructing them. Experiments on the SemEval-2010 Task 8 dataset show that our model outperforms most state-of-the-art methods.

Cite

CITATION STYLE

APA

Zhang, R., Meng, F., Zhou, Y., & Liu, B. (2018). Relation classification via recurrent neural network with attention and tensor layers. Big Data Mining and Analytics, 1(3), 234–244. https://doi.org/10.26599/BDMA.2018.9020022

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free