Skip to main content

Sanitise text while keeping your sanity

Project description

# Saniti

Sanitise lists of text documents quickly, easily and whilst maintaining your sanity

The aim was to streamline processing lists of documents into the same outputs into simply specifying the list of texts and defining the sanitization pipeline.

### Usage:

As a function-ish

>original_text = [“I like to moves it, move its”, “I likeing to move it!”, “the of”]

>text = saniti(original_text, [“token”, “destop”, “depunct”, “unempty”, “stem”, “out_corp_dict”]) #sanitise the text while initalising the class

>print(text.text)

As a class

>sani1 = saniti() # initialise the santising class

>text = sani1.process(original_text, [“token”, “destop”, “depunct”, “unempty”, “lemma”, “out_tag_doc”]) # sanitise the text

>print(text)

## Pipeline Components

  • “token” - tokenise texts

  • “depunct” - remove punctuation

  • “unempty” - remove empty words within documents

  • “lemma” - lemmatize text

  • “destop” - remove stopwords

  • “stem” - stem texts

  • “out_tag_doc” - turns the texts into gensim tagged documents for Doc2Vec

  • “out_corp_dict” - turns the texts into gensim corpus and dictionary

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

saniti-0.1.22.tar.gz (2.7 kB view hashes)

Uploaded Source

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page