Semantic library of the Data Services agency
Project description
WORDS N FUN : Semantic analysis module
The purpose of this project is two folds:
- To normalize tools and how tos of semantic analysis projects
- To offer end to end pipelines to speed up the time to market of NLP services
Licence
This project is distributed under the GNU AFFERO GENERAL PUBLIC LICENSE V3.0. Please check the LICENSE file.
Philosophy of this package
This package used to contain two main parts:
- Preprocessing : to speed up the application of numerous preprocessing routines to text data
- Modelling : to speed up the developpement of NLP machine learning models The later has been refactored into what is now called the NLP framework.
Preprocessing
One of the main goals of the project was to build a pipeline that could accomodate several data types:
- str
- list
- np.array
- pd.Series
- pd.DataFrame
- csv file (where the supplied data is the path to the actual file)
Whatever the supplied data type, it will kept along the pipeline, meaning that if a csv file is given, a new csv file will be created as the output. If the input is a pd.Series, another pd.Series of the same shape will be returned. This feature is controled by the utils.data_agnostic function.
Main scripts
utils.py
Contains utils functions and tools, mainly the data agnostic feature.
basic.py
Contains the main data transformation functions.
api.py
Contains the main entry point of the package controlling how one could create an end to end transformation pipeline.
Getting started
Requirements & setup
Coming soon: CI pipeline to push this package to pypi
- Python 3.X (tests are ran on python 3.8)
- Set up module dependencies :
pip install -r requirements.txt
- Set up the actual module :
pip install -e words_n_fun
orpython setup.py develop
- Download nltk stopwords data:
import nltk
nltk.download('stopwords')
How to use this module
A notebook tutorial is provided in the tutorial directory. Roughly speaking, her is what a basic use case should look like:
# Module import
from words_n_fun.preprocessing import api
# Definition of the desired transformation pipeline
pipeline = ['remove_non_string', 'get_true_spaces', 'to_lower', 'remove_punct',
'remove_numeric', 'remove_stopwords', 'lemmatize', 'remove_accents',
'trim_string', 'remove_leading_and_ending_spaces']
#### example 1 : csv file ####
# Input file
input_file = "path/to/my/file.csv"
# Column containing the text to preprocess
col = "myText"
# Separator character
sep = ';'
# Instanciation of a preprocessor object
preprocessor = api.get_preprocessor(pipeline=pipeline, prefered_column=col, sep=sep)
# Process data
output_file = preprocessor.transform(input_file)
#### example 2 : list ####
# Input data
input_list = ["First text to transform", "Second text to pocess"]
# Instanciation of a preprocessor object
preprocessor = api.get_preprocessor(pipeline=pipeline)
# Process data
output_list = preprocessor.transform(input_list)
Best Practices & guidelines
Contributors must try their best to follow these mainstream guidelines :
- Module management https://docs.python.org/3.9/tutorial/modules.html
- Python development https://github.com/google/styleguide/blob/gh-pages/pyguide.md, https://www.python.org/dev/peps/pep-0008/
- All the preprocessing functions must take pd.Series data as input. The built-in utils.data_agnostic decorator allows the user to send different types of data but the actual function carrying out these transformation should be built around pd.Series.
Numerous unit tests have been included in tests
:
- Please make sure to run them before trying to submit a merge request
- New features must be shipped with their corresponding tests
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file words_n_fun-1.6.1.tar.gz
.
File metadata
- Download URL: words_n_fun-1.6.1.tar.gz
- Upload date:
- Size: 42.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.8.19
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 2782e9ba9b5d49851dcc552a81ea85e6bc3aa5726437c7567c321f22cfb2c7fc |
|
MD5 | 093c7bf72d2c4460d7d3fd8a71b4659c |
|
BLAKE2b-256 | ec77e7f67caa074410f0ad0dea47fdc5fbcf4e0ec17e9d61eadc2491b7a76a3d |
File details
Details for the file words_n_fun-1.6.1-py3-none-any.whl
.
File metadata
- Download URL: words_n_fun-1.6.1-py3-none-any.whl
- Upload date:
- Size: 47.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.8.19
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | a6d661fed8b441d002c7ff9d4ec92ba98ddf8f6798aa3d8ebe8605d4985fe4c1 |
|
MD5 | 8a0b7bb6043169f4223597e88deeeea1 |
|
BLAKE2b-256 | 1904bbf472ac89588ddef121bc9e63d9684f8f758b617026a8417905db3ca701 |