Skip to main content

Collection of high-level tools to simplify everyday development tasks, with a focus on AI/ML

Project description

fmtr.tools

A collection of high-level tools to simplify everyday development tasks, with a slight focus on full-stack AI/ML.

This repository is an attempt to provide a one-stop source for a wide range of utilities and tools designed to streamline a typical, modern development workflow. There is an emphasis on a lean and nimble approach to dependencies, which tries to strike a balance between powerful functionality while avoiding unnecessary bloat.

Why?

Personally, I'm grossly impatient, and simply resent writing the same code, however simple, in multiple projects.

This could be trivial stuff like reading an integer from an environment variable (while handling errors gracefully) - or more complex ones (like just wanting a simple parallel-processing function without writing Queues, or remembering which libraries you need to do it for you).

At the same time, I find that traditional tools collections inevitably become bloated and unwieldy over time, so wanted something with a somewhat sophisticated approach to dependencies.

Key Features

  • Wide-Ranging Utilities: The collection includes tools for configuration, data types, environment management, functions, hashing, importing, iterating, JSON handling, path manipulation, platform-specific operations, randomness, and string operations.
  • Lean Dependencies: Dependencies are managed via extras, allowing you to install only what you need. Missing dependencies are handled in a clear way, telling you what's missing and how to install it.

Installing

The base library can be installed like this:

pip install fmtr.tools

Usage

Some simple import and usage examples

Read an integer from an environment variable and write it to a (human-readable) JSON file

from fmtr import tools
from fmtr.tools import Path

value=tools.env.get_int('MY_VALUE',default=None)
data=dict(value=value)
Path('data.json').write_json(data)

Zero-faff parallel multi-processing

Install the extra:

pip install fmtr.tools[parallel] --upgrade
from fmtr.tools import parallel

def expensive_computation(n):
    import math
    result = 0
    for i in range(1, n + 1):
        result += math.sqrt(i) * math.sin(i) * math.log(i)
    return result

if __name__ == '__main__':
    results=parallel.apply(expensive_computation, [10_000] * 1_000)

Extras

Most tools require no additional dependencies, but for any that do, you can add them like this:

pip install fmtr.tools[<extra>] --upgrade

If you try to use a module without the required extras, you'll get a message telling you which one is needed:

MissingExtraError: The current module is missing dependencies. To install them, run: `pip install fmtr.tools[logging] --upgrade`

Modules

The included modules, plus any extra requirements, are as follows:

  • tools.ai: Manages bulk inference for LLMs using dynamic batching. Includes classes for managing prompt encoding, generating outputs, and handling tool calls, with support for both local and remote models. Uses Pytorch and Transformers for model operations, and provides functionality for encoding prompts, generating responses, and applying tool functions.
    • Extras: ai
  • tools.config: Base config class with overridable field processors.
    • Extras: None
  • tools.dataclass: Utilities for extracting and filtering fields and metadata from dataclasses, with support for applying filters and retrieving enabled fields based on metadata attributes.
    • Extras: None
  • tools.datatype
    • Extras: None
  • tools.dm: Defines custom data modelling base classes for creating Pydantic models with error-tolerant deserialization from JSON (e.g. when output from an LLM).
    • Extras: dm
  • tools.environment: Tools for managing environment variables, including functions to retrieve variables with type conversions and default values. Features include environment variable fetching, handling missing variables, and creating type-specific getters for integers, floats, booleans, dates, and paths.
    • Extras: None
  • tools.env: Alias of tools.environment.
    • Extras: None
  • tools.function: Utilities for combining and splitting arguments and keyword arguments.
    • Extras: None
  • tools.hash: String hashing
    • Extras: None
  • tools.hfh: Utilities for caching and managing Hugging Face model repositories: setting tokens, downloading snapshots, tagging repositories, and retrieving local cache paths.
    • Extras: hfh
  • tools.html: Utilities for converting HTML documents to plain text.
    • Extras: html
  • tools.interface: Provides a base class for building Streamlit interfaces with a class-based structure.
    • Extras: interface
  • tools.iterator: Pivoting/unpivoting data structures
    • Extras: None
  • tools.json: Serialisation/deserialisation to human-readable, unicode JSON.
    • Extras: None
  • tools.merge: Utility for recursively merging multiple dictionaries or objects using the DeepMerge library.
    • Extras: merge
  • tools.name: Generates random memorable names (similar to Docker Container names) by combining an adjective with a surname.
    • Extras: None
  • tools.openai: Utilities for interacting with the OpenAI API, simple text-to-text output, etc.
    • Extras: openai.api
  • tools.Path: Enhanced pathlib.Path object with additional functionality for Windows-to-Unix path conversion, reading/writing JSON and YAML files, and convenient directory creation with parent directories. Includes methods for obtaining paths to modules and temporary directories.
    • Extras: None
  • tools.platform: Detecting if host is WSL, Docker etc.
    • Extras: None
  • tools.ContextProcess: Manages a function running in a separate process using a context manager. Provides methods to start, stop, and restart the process, with configurable restart delays. Useful for ensuring clean process management and automatic stopping when the context manager exits.
    • Extras: None
  • tools.random: Provides additional functions for random number generation and selection, useful for data augmentation.
    • Extras: None
  • tools.semantic: Manages semantic similarity operations using Sentence Transformers: loading a pre-trained model, vectorizing a text corpus, and retrieving the top matches based on similarity scores for a given query string.
    • Extras: semantic
  • tools.string: Provides utilities for handling string formatting.
    • Extras: None
  • tools.logging: Configures and initializes a logger using the Loguru library. Provides customizable logging formats for time, level, file, function, and message components.
    • Extras: logging
  • tools.logger: Prefabricated logger object, suitable for most projects, timestamped, color-coded, etc.
    • Extras: logging
  • tools.augmentation: Data augmentation stub.
    • Extras: augmentation
  • tools.Container: Runs a Docker container within a context manager, ensuring the container is stopped and removed when the context is exited.
    • Extras: docker.api
  • tools.parallel: Provides utilities for parallel computation using Dask. Supports executing functions across multiple workers or processes, handles different data formats, and options for progress display and parallelism configuration.
    • Extras: parallel
  • tools.profiling: Context-based code timing.
    • Extras: profiling
  • tools.tokenization: Provides utilities for creating and configuring tokenizers using the Tokenizers library. Iincludes functions for training both word-level and byte-pair encoding (BPE) tokenizers, applying special formatting and templates, and managing tokenizer configurations such as padding, truncation, and special tokens.
    • Extras: tokenization
  • tools.unicode: Simple unicode decoding (via Unidecode).
    • Extras: unicode

Contribution

Any contributions would be most welcome! If you have a utility that fits well within this collection, or improvements to existing tools, feel free to open a pull request.

License

This project is licensed under the Apache License Version 2.0. See the LICENSE file for more details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fmtr.tools-1.0.2.tar.gz (36.2 kB view details)

Uploaded Source

Built Distribution

fmtr.tools-1.0.2-py3-none-any.whl (44.3 kB view details)

Uploaded Python 3

File details

Details for the file fmtr.tools-1.0.2.tar.gz.

File metadata

  • Download URL: fmtr.tools-1.0.2.tar.gz
  • Upload date:
  • Size: 36.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.10.12

File hashes

Hashes for fmtr.tools-1.0.2.tar.gz
Algorithm Hash digest
SHA256 bd6ea2b33f3d112d28397b3c232c98c08faa6796ddb790ae048d85187bc84ab1
MD5 005542c2ab95c62c01fe63d666d8382b
BLAKE2b-256 1c636b988ae7404be26efe6ec5e9d6a764ed0172ebec9f1237642e63ee0160a0

See more details on using hashes here.

File details

Details for the file fmtr.tools-1.0.2-py3-none-any.whl.

File metadata

  • Download URL: fmtr.tools-1.0.2-py3-none-any.whl
  • Upload date:
  • Size: 44.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.10.12

File hashes

Hashes for fmtr.tools-1.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 b71ce6c1d3be80fc3e7014296a378b8b32d5b704bc347dbd05cbd0751cb49198
MD5 f17d2b600b2f6b92421e016a09f38cf1
BLAKE2b-256 6276bcba38be9fe0652f75fcc04fe8f5785a0a92f9363ba7339d6185a6e9103a

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page