Skip to main content

NGram library

Project description

N-Gram ============

    An N-gram is a sequence of N words: a 2-gram (or bigram) is a two-word sequence of words like “lütfen ödevinizi”, “ödevinizi çabuk”, or ”çabuk veriniz”, and a 3-gram (or trigram) is a three-word sequence of words like “lütfen ödevinizi çabuk”, or “ödevinizi çabuk veriniz”.
    
    ## Smoothing
    
    To keep a language model from assigning zero probability to unseen events, we’ll have to shave off a bit of probability mass from some more frequent events and give it to the events we’ve never seen. This modification is called smoothing or discounting.
    
    ### Laplace Smoothing
    
    The simplest way to do smoothing is to add one to all the bigram counts, before we normalize them into probabilities. All the counts that used to be zero will now have a count of 1, the counts of 1 will be 2, and so on. This algorithm is called Laplace smoothing.
    
    ### Add-k Smoothing
    
    One alternative to add-one smoothing is to move a bit less of the probability mass from the seen to the unseen events. Instead of adding 1 to each count, we add a fractional count k. This algorithm is therefore called add-k smoothing.
    
    Video Lectures
    ============
    
    [<img src="https://github.com/StarlangSoftware/NGram/blob/master/video1.jpg" width="50%">](https://youtu.be/oNWKVUdPUJY)[<img src="https://github.com/StarlangSoftware/NGram/blob/master/video2.jpg" width="50%">](https://youtu.be/ZG5m6OFdudI)
    
    For Developers
    ============
    
    You can also see [Python](https://github.com/starlangsoftware/NGram-Py), [Java](https://github.com/starlangsoftware/NGram), [C](https://github.com/starlangsoftware/NGram-C), [C++](https://github.com/starlangsoftware/NGram-CPP), [Swift](https://github.com/starlangsoftware/NGram-Swift), [Js](https://github.com/starlangsoftware/NGram-Js), or [C#](https://github.com/starlangsoftware/NGram-CS) repository.
    
    ## Requirements
    
    * [Python 3.7 or higher](#python)
    * [Git](#git)
    
    ### Python 
    
    To check if you have a compatible version of Python installed, use the following command:
    
        python -V
        
    You can find the latest version of Python [here](https://www.python.org/downloads/).
    
    ### Git
    
    Install the [latest version of Git](https://git-scm.com/book/en/v2/Getting-Started-Installing-Git).
    
    ## Pip Install
    
    	pip3 install NlpToolkit-NGram-Cy
    
    ## Download Code
    
    In order to work on code, create a fork from GitHub page. 
    Use Git for cloning the code to your local or below line for Ubuntu:
    
    	git clone <your-fork-git-link>
    
    A directory called NGram will be created. Or you can use below link for exploring the code:
    
    	git clone https://github.com/starlangsoftware/NGram-Cy.git
    
    ## Open project with Pycharm IDE
    
    Steps for opening the cloned project:
    
    * Start IDE
    * Select **File | Open** from main menu
    * Choose `NGram-CY` file
    * Select open as project option
    * Couple of seconds, dependencies will be downloaded. 
    
    Detailed Description
    ============
    
    + [Training NGram](#training-ngram)
    + [Using NGram](#using-ngram)
    + [Saving NGram](#saving-ngram)
    + [Loading NGram](#loading-ngram)
    
    ## Training NGram
         
    To create an empty NGram model:
    
    	NGram(N: int)
    
    For example,
    
    	a = NGram(2)
    
    this creates an empty NGram model.
    
    To add an sentence to NGram
    
    	addNGramSentence(self, symbols: list)
    
    For example,
    
    	nGram = NGram(2)
    	nGram.addNGramSentence(["jack", "read", "books", "john", "mary", "went"])
    	nGram.addNGramSentence(["jack", "read", "books", "mary", "went"])
    
    
    with the lines above, an empty NGram model is created and two sentences are
    added to the bigram model.
    
    NoSmoothing class is the simplest technique for smoothing. It doesn't require training.
    Only probabilities are calculated using counters. For example, to calculate the probabilities
    of a given NGram model using NoSmoothing:
    
    	a.calculateNGramProbabilities(NoSmoothing())
    
    LaplaceSmoothing class is a simple smoothing technique for smoothing. It doesn't require
    training. Probabilities are calculated adding 1 to each counter. For example, to calculate
    the probabilities of a given NGram model using LaplaceSmoothing:
    
    	a.calculateNGramProbabilities(LaplaceSmoothing())
    
    GoodTuringSmoothing class is a complex smoothing technique that doesn't require training.
    To calculate the probabilities of a given NGram model using GoodTuringSmoothing:
    
    	a.calculateNGramProbabilities(GoodTuringSmoothing())
    
    AdditiveSmoothing class is a smoothing technique that requires training.
    
    	a.calculateNGramProbabilities(AdditiveSmoothing())
    
    ## Using NGram
    
    To find the probability of an NGram:
    
    	getProbability(self, *args) -> float
    
    For example, to find the bigram probability:
    
    	a.getProbability("jack", "reads")
    
    To find the trigram probability:
    
    	a.getProbability("jack", "reads", "books")
    
    ## Saving NGram
        
    To save the NGram model:
    
    	saveAsText(self, fileName: str)
    
    For example, to save model "a" to the file "model.txt":
    
    	a.saveAsText("model.txt");              
    
    ## Loading NGram            
    
    To load an existing NGram model:
    
    	NGram(fileName: str)
    
    For example,
    
    	a = NGram("model.txt")
    
    this loads an NGram model in the file "model.txt".

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

nlptoolkit_ngram_cy-1.0.13.tar.gz (891.6 kB view details)

Uploaded Source

File details

Details for the file nlptoolkit_ngram_cy-1.0.13.tar.gz.

File metadata

  • Download URL: nlptoolkit_ngram_cy-1.0.13.tar.gz
  • Upload date:
  • Size: 891.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.0

File hashes

Hashes for nlptoolkit_ngram_cy-1.0.13.tar.gz
Algorithm Hash digest
SHA256 8c6b4008b08b087e100d9a12911e99dc16ef5e91c084fadc1de95ddd6f1b83b4
MD5 42cefc87e2ed09ae1250890a38969394
BLAKE2b-256 97a96047ae3dacb13319d29408815e155f9e084a05c976d75938e1e7c81cdc62

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page