Learning Word Representations with Deep Neural Networks for Turkish

Signal Processing and Communications Applications Conference(2019)

引用 1|浏览5
暂无评分
摘要
We test different word embedding methods in Turkish. The goal is to represent related words in a high dimensional space such that their positions reflect this relationship. We compare word2vec, fastText, and ELMo on three Turkish corpora of different sizes. Word2vec works at the word level, fastText works at the character level; ELMo, unlike the other two, is context dependent. Our experiments show that fastText is better on name and verb inflection, and word2vec is better on semantic/syntactic analogy tasks. Bag-of-words model is better than most trained word embedding models on classification.
更多
查看译文
关键词
word representation,natural language processing
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要