Multimodal sequential fashion attribute prediction

14Citations
Citations of this article
25Readers
Mendeley users who have this article in their library.

Abstract

We address multimodal product attribute prediction of fashion items based on product images and titles. The product attributes, such as type, sub-type, cut or fit, are in a chain format, with previous attribute values constraining the values of the next attributes. We propose to address this task with a sequential prediction model that can learn to capture the dependencies between the different attribute values in the chain. Our experiments on three product datasets show that the sequential model outperforms two non-sequential baselines on all experimental datasets. Compared to other models, the sequential model is also better able to generate sequences of attribute chains not seen during training. We also measure the contributions of both image and textual input and show that while text-only models always outperform image-only models, only the multimodal sequential model combining both image and text improves over the text-only model on all experimental datasets.

Cite

CITATION STYLE

APA

Arslan, H. S., Sirts, K., Fishel, M., & Anbarjafari, G. (2019). Multimodal sequential fashion attribute prediction. Information (Switzerland), 10(10). https://doi.org/10.3390/info10100308

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free