Sentence semantic matching is one of the fundamental tasks in natural language processing, which requires an agent to determine the semantic relation among input sentences. Recently, deep neural networks have achieved impressive performance in this area, especially BERT. Despite their effectiveness, most of these models treat output labels as meaningless one-hot vectors, underestimating the semantic information and guidance of relations that these labels reveal, especially for tasks with a small number of labels. To address this problem, we propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching. Specifically, we first employ BERT to encode the input sentences from a global perspective. Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective. To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task for guiding R2-Net to consider more about relations. Meanwhile, a triplet loss is employed to distinguish the intra-class and inter-class relations in a finer granularity. Empirical experiments on two sentence semantic matching tasks demonstrate the superiority of our proposed model.
CITATION STYLE
Zhang, K., Wu, L., Lv, G., Wang, M., Chen, E., & Ruan, S. (2021). Making the Relation Matters: Relation of Relation Learning Network for Sentence Semantic Matching. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 16, pp. 14411–14419). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i16.17694
Mendeley helps you to discover research relevant for your work.