Cuckoo feature hashing: Dynamic weight sharing for sparse analytics

3Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Feature hashing is widely used to process large scale sparse features for learning of predictive models. Collisions inherently happen in the hashing process and hurt the model performance. In this paper, we develop a new feature hashing scheme called Cuckoo Feature Hashing (CCFH), which treats feature hashing as a problem of dynamic weight sharing during model training. By leveraging a set of indicators to dynamically decide the weight of each feature based on alternative hash locations, CCFH effectively prevents the collisions between important features to the model, i.e. predictive features, and thus avoid model performance degradation. Experimental results on prediction tasks with hundred-millions of features demonstrate that CCFH can achieve the same level of performance by using only 15%-25% parameters compared with conventional feature hashing.

Cite

CITATION STYLE

APA

Gao, J., Ooi, B. C., Shen, Y., & Lee, W. C. (2018). Cuckoo feature hashing: Dynamic weight sharing for sparse analytics. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2018-July, pp. 2135–2141). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2018/295

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free