Visible to the public KATE: K-Competitive Autoencoder for Text

TitleKATE: K-Competitive Autoencoder for Text
Publication TypeConference Paper
Year of Publication2017
AuthorsChen, Yu, Zaki, Mohammed J.
Conference NameProceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
PublisherACM
Conference LocationNew York, NY, USA
ISBN Number978-1-4503-4887-4
Keywordsautoencoders, competitive learning, composability, Human Behavior, human factors, Metrics, pubcrawl, representation learning, Scalability, text analytics
Abstract

Autoencoders have been successful in learning meaningful representations from image datasets. However, their performance on text datasets has not been widely studied. Traditional autoencoders tend to learn possibly trivial representations of text documents due to their confoundin properties such as high-dimensionality, sparsity and power-law word distributions. In this paper, we propose a novel k-competitive autoencoder, called KATE, for text documents. Due to the competition between the neurons in the hidden layer, each neuron becomes specialized in recognizing specific data patterns, and overall the model can learn meaningful representations of textual data. A comprehensive set of experiments show that KATE can learn better representations than traditional autoencoders including denoising, contractive, variational, and k-sparse autoencoders. Our model also outperforms deep generative models, probabilistic topic models, and even word representation models (e.g., Word2Vec) in terms of several downstream tasks such as document classification, regression, and retrieval.

URLhttps://dl.acm.org/citation.cfm?doid=3097983.3098017
DOI10.1145/3097983.3098017
Citation Keychen_kate:_2017