cdw / ngram

n-gram language models

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

N-gram language models

Language modeling — that is, predicting the probability of a word in a sentence — is a fundamental task in natural language processing. It is used in many NLP applications such as autocomplete, spelling correction, or text generation.

Currently, language models based on neural networks, and especially transformers, are the state of the art: they predict very accurately the next word based on previous words. However, in this project, I will revisit the most classic of language models: the n-gram models.

Project structure

About

n-gram language models

License:MIT License


Languages

Language:Jupyter Notebook 96.2%Language:Python 3.8%