Next Word Prediction using n-gram Probabilistic Model with various Smoothing Techniques
-
Updated
Jul 23, 2018 - Jupyter Notebook
Next Word Prediction using n-gram Probabilistic Model with various Smoothing Techniques
NLP course - language models - word tokenization - Leventsheim distance - Naive Bayes example
Sample project for next word predictions using n-grams
The autocomplete system model for Indonesian was built using the perplexity score approach and n-grams count probability in determining the next word.
a probabilistic language identification system that identifies the language of a sentence
Add a description, image, and links to the ngram-probabilistic-model topic page so that developers can more easily learn about it.
To associate your repository with the ngram-probabilistic-model topic, visit your repo's landing page and select "manage topics."