Abstract
We present a novel fine-tuning algorithm in a deep hybrid architecture for semisupervised text classification. During each increment of the online learning process, the fine-tuning algorithm serves as a top-down mechanism for pseudo-jointly modifying model parameters following a bottom-up generative learning pass. The resulting model, trained under what we call the Bottom-Up-Top-Down learning algorithm, is shown to outperform a variety of competitive models and baselines trained across a wide range of splits between supervised and unsupervised training data.
Cite
CITATION STYLE
Ororbia, A. G., Giles, C. L., & Reitter, D. (2015). Learning a deep hybrid model for semi-supervised text classification. In Conference Proceedings - EMNLP 2015: Conference on Empirical Methods in Natural Language Processing (pp. 471–481). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d15-1053
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.