Skip to main content

Preprocessing and Extraction of Linguistic Information for Computational Analysis

Project description

pelican_nlp stands for “Preprocessing and Extraction of Linguistic Information for Computational Analysis - Natural Language Processing”. This package enables the creation of standardized and reproducible language processing pipelines, extracting linguistic features from various tasks like discourse, fluency, and image descriptions.

PyPI version License Supported Python Versions

Installation

Install the package using pip:

pip install pelican_nlp

For the latest development version:

pip install https://github.com/ypauli/pelican_nlp/releases/tag/v0.1.2-alpha

Usage

To use the pelican_nlp package:

Adapt your configuration file to your needs. ALWAYS change the specified project folder location.

from pelican_nlp.main import Pelican

configuration_file = "/path/to/your/config/file"
pelican = Pelican(configuration_file)
pelican.run()

For reliable operation, data must be stored in the Language Processing Data Structure (LPDS) format, inspired by brain imaging data structure conventions.

Text and audio files should follow this naming convention:

[subjectID]_[sessionID]_[task]_[task-supplement]_[corpus].[extension]

  • subjectID: ID of subject (e.g., sub-01), mandatory

  • sessionID: ID of session (e.g., ses-01), if available

  • task: task used for file creation, mandatory

  • task-supplement: additional information regarding the task, if available

  • corpus: (e.g., healthy-control / patient) specify files belonging to the same group, mandatory

  • extension: file extension (e.g., txt / pdf / docx / rtf), mandatory

Example filenames:

  • sub-01_interview_schizophrenia.rtf

  • sub-03_ses-02_fluency_semantic_animals.docx

To optimize performance, close other programs and limit GPU usage during language processing.

Features

  • Feature 1: Cleaning text files
    • Handles whitespaces, timestamps, punctuation, special characters, and case-sensitivity.

  • Feature 2: Linguistic Feature Extraction
    • Extracts semantic embeddings, logits, distance from optimality, and semantic similarity.

Examples

You can find example setups in the [examples/](https://github.com/ypauli/pelican_nlp/examples) folder. ALWAYS change the path to the project folder specified in the configuration file to your specific project location.

Contributing

Contributions are welcome! Please check out the contributing guide.

License

This project is licensed under Attribution-NonCommercial 4.0 International. See the LICENSE file for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pelican_nlp-0.1.3.tar.gz (318.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

pelican_nlp-0.1.3-py3-none-any.whl (307.1 kB view details)

Uploaded Python 3

File details

Details for the file pelican_nlp-0.1.3.tar.gz.

File metadata

  • Download URL: pelican_nlp-0.1.3.tar.gz
  • Upload date:
  • Size: 318.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.10.16

File hashes

Hashes for pelican_nlp-0.1.3.tar.gz
Algorithm Hash digest
SHA256 ed3e200f49670b34e71cfb8086a17f8a62fd1cec23565778f8b17cc2be321bc5
MD5 f60b90dda6eb7eef79968ca79ad45d2f
BLAKE2b-256 fd872a65e6a4958e5fd42fba7f46fe4d049c4abbc02e96ba8c4f2f0c203324b0

See more details on using hashes here.

File details

Details for the file pelican_nlp-0.1.3-py3-none-any.whl.

File metadata

  • Download URL: pelican_nlp-0.1.3-py3-none-any.whl
  • Upload date:
  • Size: 307.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.10.16

File hashes

Hashes for pelican_nlp-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 11f4932e8dca68780d6862d8d30bc85c11550e1a8fbfd75d686006ba4b5cc24d
MD5 2efc8c6c9ac0b3a253d475f4c0c5f598
BLAKE2b-256 ce2d4f1b9261d01ef71b3db0fe15c0365eaf1b4f2a19b8c1af54501a2f616add

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page