Skip to main content

COMBO-NLP - A library for Morphosyntactic Tagging and Dependency Parsing.

Project description

COMBO-NLP

A library for Morphosyntactic Tagging and Dependency Parsing.

Installation

pip install combo-nlp

LAMBO segmenter (optional)

A segmenter is only needed when passing raw text strings to COMBO. If you provide pre-tokenized input (list[str] or list[list[str]]), no segmenter is required.

When you initialize COMBO with a language name (e.g. COMBO("Polish")), it automatically loads a LAMBO segmenter. If LAMBO is not installed, an ImportError is raised. LAMBO is hosted on a custom PyPI index and must be installed separately:

pip install --index-url https://pypi.clarin-pl.eu/ lambo

Usage

Full text input

from combo import COMBO

# Load by HuggingFace model ID:
nlp = COMBO.from_pretrained("clarin-pl/combo-nlp-xlm-roberta-base-polish-pbd-ud2.17")
result = nlp("Ala ma kota.")

# Or load by language name (with Lambo segmenter):
nlp = COMBO("Polish")
result = nlp("Ala ma kota.")

# Or use the Language enum:
from combo import Language
nlp = COMBO(Language.POLISH)
result = nlp("Ala ma kota.")

# Multiple sentences (joined with newlines):
sentences = ["Ala ma kota.", "Pies je."]
result = nlp("\n".join(sentences))

# Access results:
for sentence in result:
    for token in sentence:
        print(token.form, token.upos, token.head, token.deprel, token.lemma)

Pre-tokenized input

from combo import COMBO

nlp = COMBO.from_pretrained("clarin-pl/combo-nlp-xlm-roberta-base-polish-pbd-ud2.17")

# Single sentence:
result = nlp(["Ala", "ma", "kota", "."])

# Multiple sentences:
result = nlp([["Ala", "ma", "kota", "."], ["Pies", "je", "."]])

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

combo_nlp-4.0.0.post1.tar.gz (80.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

combo_nlp-4.0.0.post1-py3-none-any.whl (88.9 kB view details)

Uploaded Python 3

File details

Details for the file combo_nlp-4.0.0.post1.tar.gz.

File metadata

  • Download URL: combo_nlp-4.0.0.post1.tar.gz
  • Upload date:
  • Size: 80.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.10

File hashes

Hashes for combo_nlp-4.0.0.post1.tar.gz
Algorithm Hash digest
SHA256 7152c8e992069da50c6ec1cf236e991300732f9cc1d062500a2a468cc2d66d30
MD5 dae0f56ebca930ac0e068f8fdd1055df
BLAKE2b-256 1366de531af1d4d44ecc0a81a45056048fe2d6d9a99919b1e3c9342389255c27

See more details on using hashes here.

File details

Details for the file combo_nlp-4.0.0.post1-py3-none-any.whl.

File metadata

File hashes

Hashes for combo_nlp-4.0.0.post1-py3-none-any.whl
Algorithm Hash digest
SHA256 f58fbfb41e293321c09db8796241e2d612f502c5f3dbf6b8db0c62092964a72c
MD5 f0f8994ecadd165962285b7c51c21f36
BLAKE2b-256 0ee33c7703cc1dea81210502752c99ff1d9557fd854ffa8519a8487a14e55403

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page