How to Generate a Good Word Embedding
IEEE Intelligent Systems2016Vol. 31(6), pp. 5–14
Citations Over TimeTop 1% of 2016 papers
Abstract
The authors analyze three critical components in training word embeddings: model, corpus, and training parameters. They systematize existing neural-network-based word embedding methods and experimentally compare them using the same corpus. They then evaluate each word embedding in three ways: analyzing its semantic properties, using it as a feature for supervised tasks, and using it to initialize neural networks. They also provide several simple guidelines for training good word embeddings.
Related Papers
- → Towards the Importance of the Type of Deep Neural Network and Employment of Pre-trained Word Vectors for Toxicity Detection: An Experimental Study(2021)6 cited
- → Unsupervised Learning: Word Vector(2021)1 cited
- → Deconstructing Word Embeddings(2019)1 cited
- → NMT Multi-Sense Embeddings per Word(2019)
- → Dual embedding with input embedding and output embedding for better word representation(2022)