Skip to main content

Collection of common utilities.

Project description

WVUtils

WVUtils is a collection of utilities that are shared across multiple Phosmic projects.


Requirements:

WVUtils requires Python 3.10 or higher and is platform independent.

Issue reporting

If you discover an issue with WVUtils, please report it at https://github.com/Phosmic/wvutils/issues.

License

This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version.

This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.

You should have received a copy of the GNU General Public License along with this program. If not, see https://www.gnu.org/licenses/.


Installing

Most stable version from PyPi:

PyPI PyPI - Python Version PyPI - License

python3 -m pip install wvutils

Development version from GitHub:

GitHub Workflow Status Codecov GitHub

git clone git+https://github.com/Phosmic/wvutils.git
cd wvutils
python3 -m pip install -e .

Library

wvutils.aws

Utilities for interacting with AWS services.

This module provides utilities for interacting with AWS services.

get_boto3_session

def get_boto3_session(region_name: AWSRegion) -> Session

Get the globally shared Boto3 session for a region (thread-safe).

Todo:

  • Add support for other session parameters.

Arguments:

  • region_name AWSRegion - Region name for the session.

Returns:

  • Session - Boto3 session.

clear_boto3_sessions

def clear_boto3_sessions() -> int

Clear all globally shared Boto3 sessions (thread-safe).

Returns:

  • int - Number of sessions cleared.

boto3_client_ctx

@contextmanager
def boto3_client_ctx(service_name: str, region_name: AWSRegion)

Context manager for a Boto3 client (thread-safe).

Todo:

  • Add support for other session parameters.

Arguments:

  • service_name str - Name of the service.
  • region_name AWSRegion - Region name for the service.

Yields:

  • Any - Boto3 client.

Raises:

  • ClientError - If an error occurs.

parse_s3_uri

def parse_s3_uri(s3_uri: str) -> tuple[str, str]

Parse the bucket name and path from a S3 URI.

Arguments:

  • s3_uri str - S3 URI to parse.

Returns:

  • tuple[str, str] - Bucket name and path.

download_from_s3

def download_from_s3(file_path: FilePath,
                     bucket_name: str,
                     bucket_path: str,
                     region_name: AWSRegion,
                     overwrite: bool = False) -> None

Download a file from S3.

Arguments:

  • file_path FilePath - Output path to use while downloading the file.
  • bucket_name str - Name of the S3 bucket containing the file.
  • bucket_path str - Path of the S3 bucket containing the file.
  • region_name AWSRegion - Region name for S3.
  • overwrite bool - Overwrite file on disk if already exists. Defaults to False.

Raises:

  • FileExistsError - If the file already exists and overwrite is False.

upload_file_to_s3

def upload_file_to_s3(file_path: FilePath, bucket_name: str, bucket_path: str,
                      region_name: AWSRegion) -> None

Upload a file to S3.

Arguments:

  • file_path FilePath - Path of the file to upload.
  • bucket_name str - Name of the S3 bucket to upload the file to.
  • bucket_path str - Path in the S3 bucket to upload the file to.
  • region_name AWSRegion - Region name for S3.

Raises:

  • FileNotFoundError - If the file does not exist.

upload_bytes_to_s3

def upload_bytes_to_s3(raw_b: bytes, bucket_name: str, bucket_path: str,
                       region_name: AWSRegion) -> None

Write bytes to a file in S3.

Arguments:

  • raw_b bytes - Bytes of the file to be written.
  • bucket_name str - Name of the S3 bucket to upload the file to.
  • bucket_path str - Path in the S3 bucket to upload the file to.
  • region_name AWSRegion - Region name for S3.

secrets_fetch

def secrets_fetch(
        secret_name: str,
        region_name: AWSRegion) -> str | int | float | list | dict | None

Request and decode a secret from Secrets.

Arguments:

  • secret_name str - Secret name to use.
  • region_name AWSRegion - Region name for Secrets.

Returns:

  • str | int | float | list | dict | None - Secret string.

Raises:

  • ClientError - If an error occurs while fetching the secret.
  • ValueError - If the secret is not valid JSON.

athena_execute_query

def athena_execute_query(query: str, database_name: str,
                         region_name: AWSRegion) -> str | None

Execute a query in Athena.

Arguments:

  • query str - Query to execute.
  • database_name str - Name of database to execute the query against.
  • region_name AWSRegion - Region name for Athena.

Returns:

  • str | None - Query execution ID of the query.

athena_retrieve_query

def athena_retrieve_query(qeid: str, database_name: str,
                          region_name: AWSRegion) -> str | None

Retrieve the S3 URI for results of a query in Athena.

Arguments:

  • qeid str - Query execution ID of the query to fetch.
  • database_name str - Name of the database the query is running against (for debugging).
  • region_name AWSRegion - Region name for Athena.

Returns:

  • str | None - Current status of the query, or S3 URI where results are stored.

Raises:

  • ValueError - If the query execution ID is unknown or missing.

athena_stop_query

def athena_stop_query(qeid: str, region_name: AWSRegion) -> None

Stop the execution of a query in Athena.

Arguments:

  • qeid str - Query execution ID of the query to stop.
  • region_name AWSRegion - Region name for Athena.

wvutils.errors

Custom errors.

This module contains custom exceptions that are used throughout the package.

WVUtilsError Objects

class WVUtilsError(Exception)

Base class for all wvutils exceptions.

JSONError Objects

class JSONError(Exception)

Base class for all JSON exceptions.

JSONEncodeError Objects

class JSONEncodeError(JSONError, TypeError)

Raised when JSON serializing fails.

JSONDecodeError Objects

class JSONDecodeError(JSONError, ValueError)

Raised when JSON deserializing fails.

PickleError Objects

class PickleError(Exception)

Base class for all pickle exceptions.

PickleEncodeError Objects

class PickleEncodeError(PickleError, TypeError)

Raised when pickle serializing fails.

PickleDecodeError Objects

class PickleDecodeError(PickleError, ValueError)

Raised when unpickling fails.

HashError Objects

class HashError(Exception)

Base class for all hashing exceptions.

HashEncodeError Objects

class HashEncodeError(HashError, TypeError)

Raised when hashing fails.

wvutils.path

Utilities for working with paths.

This module provides utilities for working with paths.

is_pathlike

def is_pathlike(potential_path: Any) -> bool

Check if an object is path-like.

An object is path-like if it is a string or has a __fspath__ method.

Arguments:

  • potential_path Any - Object to check.

Returns:

  • bool - True if the object is path-like, otherwise False.

stringify_path

def stringify_path(file_path: FilePath) -> str

Stringify a path-like object.

The path-like object is first converted to a string, then the user directory is expanded.

An object is path-like if it is a string or has a __fspath__ method.

Arguments:

  • file_path FilePath - Path-like object to stringify.

Returns:

  • str - Path-like object as a string.

Raises:

  • TypeError - If the object is not path-like.

ensure_abspath

def ensure_abspath(file_path: str) -> str

Make a path absolute if it is not already.

Arguments:

  • file_path str - Path to ensure is absolute.

Returns:

  • str - Absolute path.

resolve_path

def resolve_path(file_path: FilePath) -> str

Stringify and resolve a path-like object.

The path-like object is first converted to a string, then the user directory is expanded, and finally the path is resolved to an absolute path.

An object is path-like if it is a string or has a __fspath__ method.

Arguments:

  • file_path FilePath - Path-like object to resolve.

Returns:

  • str - Absolute path of the path-like object as a string.

Raises:

  • TypeError - If the object is not path-like.

xdg_cache_path

def xdg_cache_path() -> str

Base directory to store user-specific non-essential data files.

This should be '${HOME}/.cache', but the 'HOME' environment variable may not exist on non-POSIX-compliant systems. On POSIX-compliant systems, the XDG base directory specification is followed exactly since '~' expands to '$HOME' if it is present.

Returns:

  • str - Path for XDG cache.

wvutils.proxies

Utilities for working with proxies.

This module provides utilities for working with proxies.

ProxyManager Objects

class ProxyManager()

Manages a list of proxies.

This class manages a list of proxies, allowing for randomization, re-use, etc.

ProxyManager.add_proxies

def add_proxies(proxies: list[str], include_duplicates: bool = False) -> None

Add additional proxy addresses.

All proxy addresses added will be added to the end of the list.

Arguments:

  • proxies list[str] - List of proxy addresses.
  • include_duplicates bool, optional - Whether to include duplicates. Defaults to False.

ProxyManager.set_proxies

def set_proxies(proxies: list[str]) -> None

Set the proxy addresses.

Note: This will clear all existing proxies.

Arguments:

  • proxies list[str] - List of proxy addresses.

ProxyManager.can_cycle

@property
def can_cycle() -> bool

Check if can cycle to the next proxy address.

Returns:

  • bool - True if can cycle, False otherwise.

ProxyManager.cycle

def cycle() -> None

Attempt to cycle to the next proxy address.

ProxyManager.proxy

@property
def proxy() -> str | None

Current proxy address.

Returns:

  • str | None - Current proxy, or None if no proxies.

https_to_http

def https_to_http(address: str) -> str

Convert a HTTPS proxy address to HTTP.

Arguments:

  • address str - HTTPS proxy address.

Returns:

  • str - HTTP proxy address.

Raises:

  • ValueError - If the address does not start with 'http://' or 'https://'.

prepare_http_proxy_for_requests

def prepare_http_proxy_for_requests(address: str) -> dict[str, str]

Prepare a HTTP(S) proxy address for use with the 'requests' library.

Arguments:

  • address str - HTTP(S) proxy address.

Returns:

  • dict[str, str] - Dictionary of HTTP and HTTPS proxy addresses.

Raises:

  • ValueError - If the address does not start with 'http://' or 'https://'.

wvutils.args

Utilities for parsing arguments from the command line.

This module provides utilities for parsing arguments from the command line.

nonempty_string

def nonempty_string(name: str) -> Callable[[str], str]

Ensure a string is non-empty.

Example:

subparser.add_argument(
    "hashtag",
    type=nonempty_string("hashtag"),
    help="A hashtag (without #)",
)

Arguments:

  • name str - Name of the function, used for debugging.

Returns:

  • Callable[[str], str] - The decorated function.

safechars_string

def safechars_string(
        name: str,
        allowed_chars: Collection[str] | None = None) -> Callable[[str], str]

Ensure a string contains only safe characters.

Example:

subparser.add_argument(
    "--session-key",
    type=safechars_string,
    help="Key to share a single token across processes",
)

Arguments:

  • name str - Name of the function, used for debugging.
  • allowed_chars Collection[str] | None, optional - Custom characters used to validate the function name. Defaults to None.

Returns:

  • Callable[[str], str] - The decorated function.

Raises:

  • ValueError - If empty collection of allowed characters is provided.

wvutils.general

General utilities for working with Python.

This module provides general utilities for working with Python.

is_readable_iolike

def is_readable_iolike(potential_io: Any) -> bool

Check if an object is a readable IO-like.

An object is readable IO-like if it has one of the following:

  • callable readable method that returns True

Or if it has all of the following:

  • callable read method.
  • string attribute mode that contains "r".
  • callable seek method.
  • callable close method.
  • callable __enter__ method.
  • callable __exit__ method.

Arguments:

  • potential_io Any - Object to check.

Returns:

  • bool - True if the object is a readable IO-like, otherwise False.

is_writable_iolike

def is_writable_iolike(potential_io: Any) -> bool

Check if an object is a writable IO-like.

An object is writable IO-like if it has one of the following:

  • callable writable method that returns True

Or if it has all of the following:

  • callable write method.
  • string attribute mode that contains "w".
  • callable seek method.
  • callable close method.
  • callable __enter__ method.
  • callable __exit__ method.

Arguments:

  • potential_io Any - Object to check.

Returns:

  • bool - True if the object is a writable IO-like, otherwise False.

is_iolike

def is_iolike(potential_io: Any) -> bool

Check if an object is IO-like.

An object is IO-like if it has one of the following:

Arguments:

  • potential_io Any - Object to check.

Returns:

  • bool - True if the object is IO-like, otherwise False.

count_lines_in_file

def count_lines_in_file(file_path: FilePath) -> int

Count the number of lines in a file.

Notes:

All files have at least 1 line (# of lines = # of newlines + 1).

Arguments:

  • file_path FilePath - Path of the file to count lines in.

Returns:

  • int - Total number of lines in the file.

sys_set_recursion_limit

def sys_set_recursion_limit() -> None

Raise recursion limit to allow for more recurse.

gc_set_threshold

def gc_set_threshold() -> None

Reduce Number of GC Runs to Improve Performance

Notes:

Only applies to CPython.

chunker

def chunker(seq: Sequence[Any],
            n: int) -> Generator[Sequence[Any], None, None]

Iterate a sequence in size n chunks.

Arguments:

  • seq Sequence[Any] - Sequence of values.
  • n int - Number of values per chunk.

Yields:

  • Sequence[Any] - Chunk of values with length <= n.

Raises:

  • ValueError - If n is 0 or negative.

is_iterable

def is_iterable(obj: Any) -> bool

Check if an object is iterable.

Arguments:

  • obj Any - Object to check.

Returns:

  • bool - Whether the object is iterable.

rename_key

def rename_key(obj: dict,
               src_key: str,
               dest_key: str,
               in_place: bool = False) -> dict | None

Rename a dictionary key.

Todo:

  • Add support for nested keys.
  • Add support for renaming multiple keys at once.
  • Add support for non-string (built-in) key types.

All of the following are True:

isinstance(True, bool)
isinstance(True, int)
1 == True
1 in {1: "a"}
True in {1: "a"}
1 in {True: "a"}
True in {True: "a"}
1 in {1: "a", True: "b"}
True in {1: "a", True: "b"}

Arguments:

  • obj dict - Reference to the dictionary to modify.
  • src str - Name of the key to rename.
  • dest str - Name of the key to change to.
  • in_place bool, optional - Perform in-place using the provided reference. Defaults to False.

Returns:

  • dict | None - Copy of the dictionary if in_place is False, otherwise None.

unnest_key

def unnest_key(obj: dict, *keys: str) -> Any | None

Fetch a value from a deeply nested dictionary.

Arguments:

  • obj dict - Dictionary to recursively iterate.
  • *keys str - Ordered keys to fetch.

Returns:

  • Any | None - The result of the provided keys, or None if any key is not found.

sort_dict_by_key

def sort_dict_by_key(obj: dict,
                     reverse: bool = False,
                     deep_copy: bool = False) -> dict | None

Sort a dictionary by key.

Arguments:

  • obj dict - Dictionary to sort.
  • reverse bool, optional - Sort in reverse order. Defaults to False.
  • deep_copy bool, optional - Return a deep copy of the dictionary. Defaults to False.

Returns:

  • dict | None - Dictionary sorted by key. If in_place is True, None is returned.

Raises:

  • ValueError - If the dictionary keys are not of the same type.

dedupe_list

def dedupe_list(values: list[Any], raise_on_dupe: bool = False) -> list[Any]

Remove duplicate values from a list.

Example:

dedupe_list([1, 2, 3, 1, 2, 3])
# [1, 2, 3]

Arguments:

  • values list[Any] - List of values to dedupe.
  • raise_on_dupe bool, optional - Raise an error if a duplicate is found. Defaults to False.

Returns:

  • list[Any] - List of unique values.

Raises:

  • ValueError - If a duplicate is found and raise_on_dupe is True.

dupe_in_list

def dupe_in_list(values: list[Any]) -> bool

Check if a list has duplicate values.

Arguments:

  • values list[Any] - List of values to check.

Returns:

  • bool - Whether the list has duplicate values.

invert_dict_of_str

def invert_dict_of_str(obj: dict[Any, str],
                       deep_copy: bool = False,
                       raise_on_dupe: bool = False) -> dict

Invert a dictionary of strings.

Notes:

The value of the last key with a given value will be used.

Example:

invert_dict_of_str({"a": "b", "c": "d"})
# {"b": "a", "d": "c"}

Arguments:

  • obj dict[Any, str] - Dictionary to invert.
  • deep_copy bool, optional - Return a deep copy of the dictionary. Defaults to False.
  • raise_on_dupe bool, optional - Raise an error if a duplicate is found. Defaults to False.

Returns:

  • dict - Inverted dictionary.

Raises:

  • ValueError - If a duplicate is found and raise_on_dupe is True.

get_all_subclasses

def get_all_subclasses(cls: type) -> list[type]

Get all subclasses of a class.

Arguments:

  • cls type - Class to get subclasses of.

Returns:

  • list[type] - List of subclasses.

wvutils.dt

Datetime utilities.

This module contains functions and classes that are used to work with datetimes.

dtformats Objects

class dtformats()

Datetime formats.

Attributes:

  • twitter str - Twitter datetime format.
  • reddit str - Reddit datetime format.
  • general str - General datetime format.
  • db str - Database datetime format.
  • date str - Date format.
  • time_12h str - 12-hour time format with timezone.
  • time_24h str - 24-hour time format with timezone.

num_days_in_month

def num_days_in_month(year: int, month: int) -> int

Determine the number of days in a month.

Arguments:

  • year int - Year to check.
  • month int - Month to check.

Returns:

  • int - Number of days in the month.

wvutils.parquet

Parquet utilities.

This module provides utilities for working with Parquet files.

get_parquet_session

def get_parquet_session(use_s3: bool,
                        region_name: AWSRegion | None = None) -> fs.FileSystem

Get the globally shared Parquet filesystem session.

Todo:

  • Add support for other session parameters.

Arguments:

  • use_s3 bool - Use S3 if True, otherwise uses local filesystem.
  • region_name AWSRegion | None, optional - AWS region name. Defaults to None.

Returns:

  • pyarrow.fs.FileSystem - The filesystem session.

clear_parquet_sessions

def clear_parquet_sessions() -> int

Clear all globally shared Parquet filesystem sessions.

Returns:

  • int - Number of sessions cleared.

create_pa_schema

def create_pa_schema(schema_template: dict[str, str]) -> pa.schema

Create a parquet schema from a template.

Example:

{
    "key_a": "string",
    "key_b": "integer",
    "key_c": "float",
    "key_d": "bool",
    "key_e": "timestamp[s]",
    "key_f": "timestamp[ms]",
    "key_g": "timestamp[ns]",
}

becomes

pa.schema([
    ("key_a", pa.string()),
    ("key_b", pa.int64()),
    ("key_c", pa.float64()),
    ("key_d", pa.bool_()),
    ("key_e", pa.timestamp("s",  tz=utc)),
    ("key_f", pa.timestamp("ms", tz=utc)),
    ("key_g", pa.timestamp("ns", tz=utc)),
])

Arguments:

  • schema_template Sequence[Sequence[str]] - Data names and parquet types for creating the schema.

Returns:

  • pa.schema - Final parquet schema.

Raises:

  • ValueError - If an unknown type name is encountered.

force_dataframe_dtypes

def force_dataframe_dtypes(dataframe: pd.DataFrame,
                           template: dict[str, str]) -> pd.DataFrame

Force the data types of a dataframe using a template.

Arguments:

  • dataframe pd.DataFrame - Dataframe to force types.
  • template dict[str, str] - Template to use for forcing types.

Returns:

  • pd.DataFrame - Dataframe with forced types.

Raises:

  • ValueError - If an unknown type name is encountered.

export_dataset

def export_dataset(data: list[dict] | deque[dict] | pd.DataFrame,
                   output_location: str | FilePath,
                   primary_template: dict[str, str],
                   partitions_template: dict[str, str],
                   *,
                   basename_template: str | None = None,
                   use_s3: bool = False,
                   region_name: AWSRegion | None = None,
                   use_threads: bool = False,
                   overwrite: bool = False) -> None

Write to dataset to local filesystem or AWS S3.

Arguments:

  • data list[dict] | deque[dict] | pd.DataFrame] - List, deque, or dataframe of objects to be written.
  • output_location str | FilePath - Location to write the dataset to.
  • primary_template dict[str, str] - Parquet schema template to use for the table (excluding partitions). partitions_template(dict[str, str]): Parquet schema template to use for the partitions.
  • basename_template str, optional - Filename template to use when writing to S3 or locally. Defaults to None.
  • use_s3 bool, optional - Use S3 as the destination when exporting parquet files. Defaults to False.
  • region_name AWSRegion, optional - AWS region to use when exporting to S3. Defaults to None.
  • use_threads bool, optional - Use multiple threads when exporting. Defaults to False.
  • overwrite bool, optional - Overwrite existing files. Defaults to False.

wvutils.restruct

Utilities for restructuring data.

This module provides utilities for restructuring data, including serialization and hashing.

JSON

Python JSON
dict object
list, tuple array
str string
int, float, int- & float-derived enums number
True true
False false
None null

Hash

No content.

Pickle

An important difference between cloudpickle and pickle is that cloudpickle can serialize a function or class by value, whereas pickle can only serialize it by reference. Serialization by reference treats functions and classes as attributes of modules, and pickles them through instructions that trigger the import of their module at load time. Serialization by reference is thus limited in that it assumes that the module containing the function or class is available/importable in the unpickling environment. This assumption breaks when pickling constructs defined in an interactive session, a case that is automatically detected by cloudpickle, that pickles such constructs by value.

Read more: https://github.com/cloudpipe/cloudpickle/blob/master/README.md#overriding-pickles-serialization-mechanism-for-importable-constructs

json_dumps

def json_dumps(obj: JSONSerializable) -> str

Encode an object as JSON.

Arguments:

  • obj JSONSerializable - Object to encode.

Returns:

  • str - Object encoded as JSON.

Raises:

  • JSONEncodeError - If the object could not be encoded.

jsonl_dumps

def jsonl_dumps(objs: Iterable[JSONSerializable]) -> str

Encode objects as JSONL.

Arguments:

  • objs Iterable[JSONSerializable] - Objects to encode.

Returns:

  • str - Objects encoded as JSONL.

Raises:

  • JSONEncodeError - If the objects could not be encoded.

json_dump

def json_dump(obj: JSONSerializable, file_path: str) -> None

Encode an object as JSON and write it to a file.

Arguments:

  • file_path str - Path of the file to open.

Raises:

  • JSONEncodeError - If the object could not be encoded.

jsonl_dump

def jsonl_dump(objs: Iterable[JSONSerializable], file_path: str) -> None

Encode objects as JSONL and write them to a file.

Arguments:

  • objs Iterable[JSONSerializable] - Objects to encode.
  • file_path str - Path of the file to open.

Raises:

  • JSONEncodeError - If the objects could not be encoded.

json_loads

def json_loads(encoded_obj: str) -> JSONSerializable

Decode a JSON-encoded object.

Arguments:

  • encoded_obj str - Object to decode.

Returns:

  • JSONSerializable - Decoded object.

Raises:

  • JSONDecodeError - If the object could not be decoded.

json_load

def json_load(file_path: FilePath) -> JSONSerializable

Decode a file containing a JSON-encoded object.

Arguments:

  • file_path FilePath - Path of the file to open.

Returns:

  • JSONSerializable - Decoded object.

Raises:

  • JSONDecodeError - If the file could not be decoded.

jsonl_loader

def jsonl_loader(
    file_path: FilePath,
    *,
    allow_empty_lines: bool = False
) -> Generator[JSONSerializable, None, None]

Decode a file containing JSON-encoded objects, one per line.

Arguments:

  • file_path FilePath - Path of the file to open.
  • allow_empty_lines bool, optional - Whether to allow (skip) empty lines. Defaults to False.

Yields:

  • JSONSerializable - Decoded object.

Raises:

  • JSONDecodeError - If the line could not be decoded, or if an empty line was found and allow_empty_lines is False.

squeegee_loader

def squeegee_loader(
        file_path: FilePath) -> Generator[JSONSerializable, None, None]

Automatically decode a file containing JSON-encoded objects.

Supports multiple formats (JSON, JSONL, JSONL of JSONL, etc).

Todo:

  • Add support for pretty-printed JSON that has been appended to a file.

Arguments:

  • file_path FilePath - Path of the file to open.

Yields:

  • JSONSerializable - Decoded object.

Raises:

  • JSONDecodeError - If the line could not be decoded.

gen_hash

def gen_hash(obj: MD5Hashable) -> str

Create an MD5 hash from a hashable object.

Note: Tuples and deques are not hashable, so they are converted to lists.

Arguments:

  • obj MD5Hashable - Object to hash.

Returns:

  • str - MD5 hash of the object.

Raises:

  • HashEncodeError - If the object could not be encoded.

pickle_dump

def pickle_dump(obj: PickleSerializable, file_path: FilePath) -> None

Serialize an object as a pickle and write it to a file.

Arguments:

  • obj JSONSerializable - Object to serialize.
  • file_path FilePath - Path of the file to write.

Raises:

  • PickleEncodeError - If the object could not be encoded.

pickle_dumps

def pickle_dumps(obj: PickleSerializable) -> bytes

Serialize an object as a pickle.

Arguments:

  • obj PickleSerializable - Object to serialize.

Returns:

  • bytes - Serialized object.

Raises:

  • PickleEncodeError - If the object could not be encoded.

pickle_load

def pickle_load(file_path: FilePath) -> PickleSerializable

Deserialize a pickle-serialized object from a file.

Note: Not safe for large files.

Arguments:

  • file_path FilePath - Path of the file to open.

Returns:

  • PickleSerializable - Deserialized object.

Raises:

  • PickleDecodeError - If the object could not be decoded.

pickle_loads

def pickle_loads(serialized_obj: bytes) -> PickleSerializable

Deserialize a pickle-serialized object.

Arguments:

  • serialized_obj bytes - Object to deserialize.

Returns:

  • PickleSerializable - Deserialized object.

Raises:

  • PickleDecodeError - If the object could not be decoded.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

wvutils-1.0.4.tar.gz (90.9 kB view details)

Uploaded Source

Built Distribution

wvutils-1.0.4-py3-none-any.whl (52.7 kB view details)

Uploaded Python 3

File details

Details for the file wvutils-1.0.4.tar.gz.

File metadata

  • Download URL: wvutils-1.0.4.tar.gz
  • Upload date:
  • Size: 90.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.12

File hashes

Hashes for wvutils-1.0.4.tar.gz
Algorithm Hash digest
SHA256 eef282c577d094c7c4150175318eab5f8b454794edf6dbbc87143cacaedb8c6f
MD5 20364eb856e976016698cff875d1f525
BLAKE2b-256 4e42d39fe933895a68879c6398ee4170884b8cbc5660f2d5b6a9e280c1ff3817

See more details on using hashes here.

Provenance

File details

Details for the file wvutils-1.0.4-py3-none-any.whl.

File metadata

  • Download URL: wvutils-1.0.4-py3-none-any.whl
  • Upload date:
  • Size: 52.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.12

File hashes

Hashes for wvutils-1.0.4-py3-none-any.whl
Algorithm Hash digest
SHA256 e45011cd97a71ffc1c85b080e7e2cdffae19f4013eabfd6717f5e05461f6df83
MD5 bdc37b5054a4b06e97df32f8a1922d53
BLAKE2b-256 978b9ef899e3b9cc8573526f6c03e196e00587cc56466a2de0babe0f4371f6a2

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page