Attention-over-attention field-aware factorization machine

10Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

Factorization Machine (FM) has been a popular approach in supervised predictive tasks, such as click-through rate prediction and recommender systems, due to its great performance and efficiency. Recently, several variants of FM have been proposed to improve its performance. However, most of the state-of-the-art prediction algorithms neglected the field information of features, and they also failed to discriminate the importance of feature interactions due to the problem of redundant features. In this paper, we present a novel algorithm called Attention-over-Attention Field-aware Factorization Machine (AoAFFM) for better capturing the characteristics of feature interactions. Specifically, we propose the field-aware embedding layer to exploit the field information of features, and combine it with the attention-over-attention mechanism to learn both feature-level and interaction-level attention to estimate the weight of feature interactions. Experimental results show that the proposed AoAFFM improves FM and FFM with large margin, and outperforms state-of-the-art algorithms on three public benchmark datasets.

Cite

CITATION STYLE

APA

Wang, Z., Ma, J., Zhang, Y., Wang, Q., Ren, J., & Sun, P. (2020). Attention-over-attention field-aware factorization machine. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 6323–6330). AAAI press. https://doi.org/10.1609/aaai.v34i04.6101

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free