Deep Embeddings For Brand Detection In Product Titles

ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS, AIST 2019(2019)

引用 0|浏览0
暂无评分
摘要
In this paper, we compare various techniques to learn expressive product title embeddings starting from TF-IDF and ending with deep neural architectures. The problem is to recognize brands from noisy retail product names coming from different sources such as receipts and supply documents. In this work we consider product titles written in English and Russian. To determine the state-of-the-art on openly accessed "Universe-HTT barcode reference" dataset, traditional machine learning models, such as SVMs, were compared to Neural Networks with classical softmax activation and cross entropy loss. Furthermore, the scalable variant of the problem was studied, where new brands are recognized without retraining the model. The approach is based on k-Nearest Neighbors, where the search space could be represented by either TFIDF vectors or deep embeddings. For the latter we have considered two solutions: (1) pretrained FastText embeddings followed by LSTM with Attention and (2) character-level Convolutional Neural Network. Our research shows that deep embeddings significantly outperform TF-IDF vectors. Classification error was reduced from 13.2% for TF-IDF approach to 8.9% and to 7.3% for LSTM embeddings and character-level CNN embeddings correspondingly.
更多
查看译文
关键词
Text classification, FastText, LSTM with Attention, Embeddings, Triplet loss, Similarity learning, Brand classification
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要