Unified Image Aesthetic Prediction via Scanpath-Guided Feature Aggregation Network

0Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The performance of automatic aesthetic prediction has achieved significant improvement by utilizing deep convolutional neural networks (CNNs). However, existing CNN methods can only achieve limited success because (1) most of the methods take one fixed-size patch as the training example, which loses the fine-grained details and the holistic layout information, and (2) most of the methods ignore the biologically cues such as the gaze shifting sequence in image aesthetic assessment. To address these challenges, we propose a scanpath-guided feature aggregation model for aesthetic prediction. In our model, human fixation map and the view scanpath are predicted by a multi-scale network. Then a sequence of regions are adaptively selected according to the scanpath. These attended regions are then progressively fed into the CNN and LSTM network to accumulate the information, yielding a compact image level representation. Extensive experiments on the large scale aesthetics assessment benchmark AVA and Photo.net data set thoroughly demonstrate the efficacy of our approach for unified aesthetic prediction tasks: (i) aesthetic quality classification; (ii) aesthetic score regression; and (iii) aesthetic score distribution prediction.

Cite

CITATION STYLE

APA

Zhang, X., Gao, X., Lu, W., Yu, Y., & He, L. (2019). Unified Image Aesthetic Prediction via Scanpath-Guided Feature Aggregation Network. In Communications in Computer and Information Science (Vol. 1072, pp. 259–271). Springer. https://doi.org/10.1007/978-981-15-1398-5_19

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free