Skip to main content

A collection of tools, and helpers that I usually want for a handful of projects, so to avoid rewriting them every time, I decided to create this package.

Project description

Raccoon Tools

This is a collection of tools that I regularly use on several projects. To stop duplicating and to (hopefully) help someone, I decided to make them into a public package.

Functionalities

Decorators

retry

A decorator that retries a function call a specified number of times before giving up. It logs each attempt and the final failure if all retries are exhausted.

Parameters:

  • retries: Maximum number of retries before giving up (default: 3).
  • delay: Delay in seconds between each retry (default: 1).
  • delay_is_exponential: If True, the delay between retries will increase exponentially (default: False).
  • only_exceptions_of_type: A list of exception types to catch and retry on. If None, all exceptions are caught.
  • log_level: The log level used by the decorator (default: logging.ERROR).

Example:

from raccoontools.decorators.retry import retry

attempts = {"count": 0}

@retry(retries=4, delay=0.25)
def flaky_operation():
    attempts["count"] += 1
    if attempts["count"] < 3:
        raise RuntimeError("Still failing...")
    return "Success!"

print(flaky_operation())

retry_request

Like the previous decorator, but for HTTP requests. It logs each attempt and the final failure if all retries are exhausted. It also provides options to handle specific HTTP status codes.

Parameters:

  • retries: Maximum number of retries before giving up (default: 3).
  • delay: Delay in seconds between each retry (default: 1).
  • delay_is_exponential: If True, the delay between retries will increase exponentially (default: False).
  • skip_retry_on_404: If True, the decorator will not retry on 404 responses (default: False).
  • retry_only_on_status_codes: A list of HTTP status codes to retry on. If None, no retries will be made.
  • get_new_token_on_401: An optional callable to execute and get a new token when a 401 response is received.
  • get_new_token_on_403: An optional callable to execute and get a new token when a 403 response is received.
  • log_level: The log level used by the decorator (default: logging.ERROR).

Example:

import requests
from raccoontools.decorators.retry import retry_request

@retry_request(retries=5, delay=1, skip_retry_on_404=True)
def call_service():
    return requests.get("https://httpbin.org/status/500")

response = call_service()
print(response.status_code)

benchmark

A decorator that benchmarks the execution time of a function. The results are logged using the logging module at the INFO level. The decorated function can also provide benchmark information via the get_benchmark_info method.

Example:

import time
from raccoontools.decorators.benchmark import benchmark

@benchmark
def slow_function():
    time.sleep(0.5)
    return "done"

slow_function()
print(slow_function.get_benchmark_info())

Generators

infinite_iterator

Generates an infinite iterator from a list.

Parameters:

  • list_to_iterate_over: The list to iterate over.

Example:

from raccoontools.generators.misc_generators import infinite_iterator

rotating_status = infinite_iterator(["⏳", "⌛", "✅"])
for _ in range(6):
    print(next(rotating_status), end=" ")

read_line

Reads a file line by line.

Parameters:

  • file: Path to the file.
  • strip_line: Strip whitespace from the beginning and end of each line (default: True).
  • encoding: File encoding (default: 'utf-8').
  • buffer_size: Size of the read buffer in bytes. If None, the default system buffer is used.

Example:

from pathlib import Path
from raccoontools.generators.file_ops_generators import read_line

for line in read_line(Path("logs/app.log"), strip_line=True):
    print(line)

read_csv

Reads a CSV file line by line, yielding each line as a dictionary + metadata.

Parameters:

  • file: Path to the CSV file.
  • encoding: File encoding (default: 'utf-8').
  • has_headers: If True, the first line of the CSV file is treated as a header.
  • buffer_size: Size of the read buffer in bytes. If None, the default system buffer is used.

Example:

from raccoontools.generators.file_ops_generators import read_csv

for row, metadata in read_csv("data/report.csv", has_headers=True):
    print(metadata.index, row)

id_guid_generator

Generates unique GUID (Globally Unique Identifier) strings.

Parameters:

  • ids_to_generate: The number of GUIDs to generate. If None, generates indefinitely.

Example:

from raccoontools.generators.sequence_generators import id_guid_generator

guid_gen = id_guid_generator(ids_to_generate=2)
print(list(guid_gen))

id_int_generator

Generates integer IDs with optional validation.

Parameters:

  • ids_to_generate: The number of IDs to generate. If None, generates indefinitely.
  • start_at: The starting value for the ID sequence (default: 0).
  • validate_id: A function to validate each ID. If None, all IDs are considered valid.

Example:

from raccoontools.generators.sequence_generators import id_int_generator

only_even_ids = id_int_generator(ids_to_generate=3, start_at=10, validate_id=lambda x: x % 2 == 0)
print(list(only_even_ids))  # [10, 12, 14]

timestamp_generator

Generates Unix timestamps.

Parameters:

  • timestamps_to_generate: The number of timestamps to generate. If None, generates indefinitely.

Example:

from raccoontools.generators.sequence_generators import timestamp_generator

timestamps = []
for ts in timestamp_generator(3):
    timestamps.append(ts)
print(timestamps)

sentence_generator

Generates Lorem Ipsum sentences with lengths ranging from min_length to max_length.

Parameters:

  • sentences_to_generate: The number of sentences to generate. If None, generates indefinitely.
  • min_length: The minimum length of each sentence (default: 1).
  • max_length: The maximum length of each sentence. If None, a random value between 10 and 512 is used for each sentence.

Example:

from raccoontools.generators.sequence_generators import sentence_generator

for sentence in sentence_generator(2, min_length=40, max_length=80):
    print(sentence)

Shared Utilities

file_ops

Provides functions to load and save JSON data to and from files.

  • load_json_from_file(file: Path, encoding: str = "utf-8", object_hook: Optional[Callable] = obj_dump_deserializer) -> Union[dict, List[dict]]: Loads a JSON file and returns the data as a dictionary or list of dictionaries. By default, uses obj_dump_deserializer to reconstruct types (datetime, int, float, Path). Pass object_hook=None for raw JSON parsing with no type coercion.
  • save_json_to_file(data: Union[dict, List[dict]], target_file_or_folder: Path, dump_kwargs: dict = None, encoding: str = "utf-8") -> Path: Saves a dictionary or list of dictionaries to a JSON file.

Example:

from pathlib import Path
from raccoontools.shared.file_ops import load_json_from_file, save_json_to_file

payload = {"status": "ok"}
target_dir = Path("artifacts")
target_dir.mkdir(exist_ok=True)
saved_file = save_json_to_file(payload, target_dir)
loaded_payload = load_json_from_file(saved_file)
print(saved_file, loaded_payload)

file_utils

Provides utility functions for file operations.

  • get_filename_for_new_file(file_extension: str, prefix: str = None, add_current_datetime_as_format: str = "%Y%m%d%H%M%S%f", use_utc: bool = True, unique_identifier: Union[str, bool] = True, part_separator: str = "-", suffix: str = None) -> str: Generates a unique filename for a new file.

Example:

from raccoontools.shared.file_utils import get_filename_for_new_file

filename = get_filename_for_new_file("json", prefix="snapshot", suffix="v1")
print(filename)  # snapshot-20240201130501999999-...-v1.json

http

Provides utility functions for HTTP headers.

  • get_headers(token: str, content_type: str = "application/json", user_agent: str = None, fake_browser_user_agent: bool = False, extra_args: Dict[str, str] = None) -> Dict[str, str]: Generates headers for an HTTP request.

Example:

from raccoontools.shared.http import get_headers

headers = get_headers(
    token="abc123",
    fake_browser_user_agent=True,
    extra_args={"X-Trace-Id": "req-1"}
)
print(headers)

requests_with_retry

A wrapper around requests using the retry_request decorator. Can be used as a drop-in replacement for requests.

  • get(url, params=None, **kwargs) -> requests.Response: Sends a GET request with retry functionality.
  • options(url, **kwargs) -> requests.Response: Sends an OPTIONS request with retry functionality.
  • head(url, **kwargs) -> requests.Response: Sends a HEAD request with retry functionality.
  • post(url, data=None, json=None, **kwargs) -> requests.Response: Sends a POST request with retry functionality.
  • put(url, data=None, **kwargs) -> requests.Response: Sends a PUT request with retry functionality.
  • patch(url, data=None, **kwargs) -> requests.Response: Sends a PATCH request with retry functionality.
  • delete(url, **kwargs) -> requests.Response: Sends a DELETE request with retry functionality.

Example:

import raccoontools.shared.requests_with_retry as requests

response = requests.get("https://httpbin.org/status/500")
print(response.status_code)

serializer

Provides functions to serialize and deserialize objects.

  • serialize_to_dict(obj) -> Union[dict, List[dict], None]: Serializes an object to a dictionary or list of dictionaries.
  • parse_csv(csv_data: str) -> List[dict]: Parses a CSV string and returns a list of dictionaries.
  • csv_string_to_dict_list(data: Union[str, List[str], dict, List[dict]], no_data_return: str = "No data available") -> Union[List[dict], str]: Converts a CSV string to a list of dictionaries.
  • dataset_to_prompt_text(dataset: List[dict]) -> str: Converts a dataset to a prompt text.
  • obj_dump_serializer(obj): Serializes objects for saving data to a file.
  • obj_dump_deserializer(obj): Deserializes objects when loading data from a file.

Example:

from pydantic import BaseModel
from raccoontools.shared.serializer import serialize_to_dict, csv_string_to_dict_list

class User(BaseModel):
    name: str
    active: bool

payload = serialize_to_dict(User(name="Ada", active=True))
print(payload)  # {'name': 'Ada', 'active': True}

rows = csv_string_to_dict_list("name,score\nAda,10\nBob,7\n")
print(rows)

Changelog

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

raccoontools-2.0.0.tar.gz (29.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

raccoontools-2.0.0-py3-none-any.whl (20.5 kB view details)

Uploaded Python 3

File details

Details for the file raccoontools-2.0.0.tar.gz.

File metadata

  • Download URL: raccoontools-2.0.0.tar.gz
  • Upload date:
  • Size: 29.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.3

File hashes

Hashes for raccoontools-2.0.0.tar.gz
Algorithm Hash digest
SHA256 40e5973ec8b88ace382f5b1641d1868f7146bda4d05e96d563415802450ff01a
MD5 ed1c4ce5ee50d3d9b3a8e7bdd0653df4
BLAKE2b-256 456b6d4aea79a45c50166f9fb513706a67ed2bdef816a4d04dd177d5beb6079a

See more details on using hashes here.

File details

Details for the file raccoontools-2.0.0-py3-none-any.whl.

File metadata

  • Download URL: raccoontools-2.0.0-py3-none-any.whl
  • Upload date:
  • Size: 20.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.3

File hashes

Hashes for raccoontools-2.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 b8559399fcf70ec400a70a042e70e674c6e3113a83bc782ebb2e141fbeb1d1ad
MD5 d20bc9c71b33c281e15f855d754b7a18
BLAKE2b-256 c0da4acb1ef2a4b65aeac3d21c564b7b57144c8e4394e49c5f02b87b271e4b21

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page