Relation extraction with deep reinforcement learning

4Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

In recent years, deep learning has been widely applied in relation extraction task. The method uses only word embeddings as network input, and can model relations between target named entity pairs. It equally deals with each relation mention, so it cannot effectively extract relations from the corpus with an enormous number of non-relations, which is the main reason why the performance of relation extraction is significantly lower than that of relation classification. This paper designs a deep reinforcement learning framework for relation extraction, which considers relation extraction task as a two-step decision-making game. The method models relation mentions with CNN and Tree-LSTM, which can calculate initial state and transition state for the game respectively. In addition, we can tackle the problem of unbalanced corpus by designing penalty function which can increase the penalties for first-step decision-making errors. Finally, we use Q-Learning algorithm with value function approximation to learn control policy π for the game. This paper sets up a series of experiments in ACE2005 corpus, which show that the deep reinforcement learning framework can achieve state-of-the-art performance in relation extraction task.

Cite

CITATION STYLE

APA

Zhang, H., Feng, Y., Hao, W., Chen, G., & Jin, D. (2017). Relation extraction with deep reinforcement learning. In IEICE Transactions on Information and Systems (Vol. E100D, pp. 1893–1902). Maruzen Co., Ltd. https://doi.org/10.1587/transinf.2016EDP7450

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free