Attention-based two-layer long short-term memory model for automatic essay scoring

6Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

Abstract

We propose a neural network architecture-based automatic essay scoring model which contains a two-layer long short-term memory (LSTM) and an attention mechanism layer. The Google word vector dataset, which includes the richer word information and contextual information than the local-trained word vector dataset, is used to generate the embedding word vector of the input layer of the model by pre-training. The lower layer of the two-layer LSTM network captures the context semantic information and hidden context dependency, and the upper layer extracts the deeper context dependency. The attention mechanism layer focuses on the information extracted from the upper hidden layer of two-layer LSTM and calculates the attention probability to highlight the importance of key information in the text. The dataset used for automatic essay scoring task is provided by the Hewlett Foundation, and the quadratic weighted kappa coefficient is used as the evaluation index of the model. The experimental results show that the proposed method outperforms other automatic essay scoring baseline models such as bidirectional LSTM, SKIPFLOW-LSTM, and so on, in terms of the value of quadratic weighted kappa coefficient.

Cite

CITATION STYLE

APA

Xia, L., Luo, D., Liu, J., Guan, M., Zhang, Z., & Gong, A. (2020). Attention-based two-layer long short-term memory model for automatic essay scoring. Shenzhen Daxue Xuebao (Ligong Ban)/Journal of Shenzhen University Science and Engineering, 37(6), 559–566. https://doi.org/10.3724/SP.J.1249.2020.06559

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free