We present a general framework for association learning, where entities are embedded in a common latent space to express relatedness via geometry-an approach that underlies the state of the art for link prediction, relation learning, multi-label tagging, relevance retrieval and ranking. Although current approaches rely on local training methods applied to non-convex formulations, we demonstrate how general convex formulations can be achieved for entity embedding, both for standard multi-linear and prototype-distance models. We investigate an efficient optimization strategy that allows scaling. An experimental evaluation reveals the advantages of global training in different case studies.
CITATION STYLE
Mirzazadeh, F., Guo, Y., & Schuurmans, D. (2014). Convex co-embedding. In Proceedings of the National Conference on Artificial Intelligence (Vol. 3, pp. 1989–1996). AI Access Foundation. https://doi.org/10.1609/aaai.v28i1.8976
Mendeley helps you to discover research relevant for your work.