Long Tail Constraint On Non-Negative Matrix Factorization

2018 2ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE APPLICATIONS AND TECHNOLOGIES (AIAAT 2018)(2018)

Cited 2|Views4
No score
Abstract
The topic distribution in text generally has long tail effect, but few people do research on how to dig out long tail topics from matrix factorization. So we propose a method, this is Non-negative Matrix Factorization with Long-tail Constraint (LTNMF). LTNMF adds the soft orthogonal constraints to the feature matrix to ensure the independence of the topics on the basis of the non-negative matrix factorization. The sparse constraints and long tail constraints are added to the topic document matrix to enhance the robustness of the model and the characterization of the long tail features of the topic distribution. The combination of soft orthogonal constraints, sparse constraints and long tail constraints enables the model to extract the long tail topic information in the data and ensure the quality of the topic. We use Sougou and 20newsgroup datasets to experiment, and the results show that LTMNF can dig more topic words and improve the accuracy and the standard mutual information of clustering in text classification.
More
Translated text
Key words
non-negative
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined