This tool provides an efficient implementation of the continuous bag-of-words and skip-gram architectures for computing vector representations of words. These representations can be subsequently used in many natural language processing applications and for further research. The word2vec tool takes a text corpus as input and produces the word vectors as output. It first constructs a vocabulary from the training text data and then learns vector representation of words. The resulting word vector file can be used as features in many natural language processing and machine learning applications. ..
Keywords for this software
References in zbMATH (referenced in 8 articles )
Showing results 1 to 8 of 8.
- Chen, Wenliang; Zhang, Min; Zhang, Yue; Duan, Xiangyu: Exploiting meta features for dependency parsing and part-of-speech tagging (2016)
- Gallay, Ladislav; Šimko, Marián: Utilizing vector models for automatic text lemmatization (2016)
- Joshi, Shalmali; Ghosh, Joydeep; Reid, Mark; Koyejo, Oluwasanmi: Rényi divergence minimization based co-regularized multiview clustering (2016)
- McQueen, James; Meilä, Marina; VanderPlas, Jacob; Zhang, Zhongyue: Megaman: scalable manifold learning in python (2016)
- Bengio, Yoshua (ed.): Editorial introduction to the neural networks special issue on deep learning of representations (2015)
- Derrac, Joaquín; Schockaert, Steven: Inducing semantic relations from conceptual spaces: a data-driven approach to plausible reasoning (2015)
- Dhillon, Paramveer S.; Foster, Dean P.; Ungar, Lyle H.: Eigenwords: spectral word embeddings (2015)
- Pilehvar, Mohammad Taher; Navigli, Roberto: From senses to texts: an all-in-one graph-based approach for measuring semantic similarity (2015)