Word2vec Concept from scratch – part 1

In this post I will be implementing the Word2vec algorithm with negative sampling from scratch using python. Implementing this from scratch allow to have a better grasp the inner operations of word2vec’s skip-gram and negative sampling approach. For our analysis, we will use the Open American National Corpus (http://www.anc.org/), which consists of roughly 15 million…

Text generation using Markov chain models

Markov chains is one of the oldest ways to generate quite believable text. Rather than generating text by randomly selecting characters (possible but completely impractical), we use the Markov property that follows the chain of linked events, where what get generated next depends only on what is generated currently. Typically generating text trough, a Markov process…