Neural Language Models

From feedforward through stacked recurrent

Arun Jagota
Towards Data Science
36 min readMay 7, 2021

--

Photo by Markus Spiske on Unsplash

In NLP, a language model is a probability distribution over sequences on an alphabet of tokens. A central problem in language modeling is to learn a language model from examples, such as a model of English sentences from a training set of sentences.

Language models have many uses. Such as…

--

--

PhD, Computer Science, neural nets. 14+ years in industry: data science algos developer. 24+ patents issued. 50 academic pubs. Blogs on ML/data science topics.