Common webapp scaffolding.
Project description
lassen
40.4881° N, 121.5049° W
Core utilities for MonkeySee web applications.
Not guaranteed to be backwards compatible, use at your own risk.
Structure
Stores: Each datamodel is expected to have its own store. Base classes that provide standard logic are provided by lassen.store
- StoreBase: Base class for all stores
- StoreFilterMixin: Mixin for filtering stores that specify an additional schema to use to filter
Schemas: Each datamodel should define a Model class (SQLAlchemy base object) and a series of Schema objects (Pydantic) that allow the Store to serialize the models. These schemas are also often used for direct CRUD referencing in the API layer.
We use a base Stub
file to generate these schemas from a centralized definition. When defining generators you should use a path that can be fully managed by lassen, since we will remove and regenerate these files on each run.
STORE_GENERATOR = StoreGenerator("models/auto")
SCHEMA_GENERATOR = SchemaGenerator("schemas/auto")
poetry run generate-lassen
Datasets: Optional huggingface datasets
processing utilities. Only installed under the lassen[datasets]
extra. These provide support for:
- batch_to_examples: Iterate and manipulate each example separately, versus over nested key-based lists.
- examples_to_batch: Takes the output of a typehinted element-wise batch and converts into the format needed for dataset insertion. If datasets can't automatically interpret the type of the fields, also provide automatic casting based on the typehinted dataclass.
from lassen.datasets import batch_to_examples, examples_to_batch
import pandas as pd
@dataclass
class BatchInsertion:
texts: list[str]
def batch_process(examples):
new_examples : list[BatchInsertion] = []
for example in batch_to_examples(examples):
new_examples.append(
BatchInsertion(
example["raw_text"].split()
)
)
# datasets won't be able to typehint a dataset that starts with an empty example, so we use our explicit schema to cast the data
return examples_to_batch(new_examples, BatchInsertion, explicit_schema=True)
df = pd.DataFrame(
[
{"raw_text": ""},
{"raw_text": "This is a test"},
{"raw_text": "This is another test"},
]
)
dataset = Dataset.from_pandas(df)
dataset = dataset.map(
batch_process,
batched=True,
batch_size=1,
num_proc=1,
remove_columns=dataset.column_names,
)
Migrations: Lassen includes a templated alembic.init and env.py file. Client applications just need to have a migrations
folder within their project root. After this you can swap poetry run alembic
with poetry run migrate
.
poetry run migrate upgrade head
Settings: Application settings should subclass our core settings. This provides a standard way to load settings from environment variables and includes common database keys.
from lassen.core.config import CoreSettings, register_settings
@register_settings
class ClientSettings(CoreSettings):
pass
Schemas: For helper schemas when returning results via API, see lassen.schema.
Development
poetry install --extras "datasets"
createuser lassen
createdb -O lassen lassen_db
createdb -O lassen lassen_test_db
Unit Tests:
poetry run pytest
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file lassen-0.3.0.tar.gz
.
File metadata
- Download URL: lassen-0.3.0.tar.gz
- Upload date:
- Size: 49.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.2.1 CPython/3.10.4 Darwin/22.5.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | cf2b5c28176260004541f520537c3a71c393f8f3d3d86fb521eae6ae7d7c1003 |
|
MD5 | 58bd6a36ccbbbe95069dfa757b1a6eb8 |
|
BLAKE2b-256 | eb1dd8ec8d2063328322c0282377f49de76c91b89010852111fe65088217e346 |
File details
Details for the file lassen-0.3.0-py3-none-any.whl
.
File metadata
- Download URL: lassen-0.3.0-py3-none-any.whl
- Upload date:
- Size: 54.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.2.1 CPython/3.10.4 Darwin/22.5.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4682c75af67ff01fc648653e70958fed44fd3365ffeb30b96861a09ae0982e5b |
|
MD5 | 2150c51bac8faafd780fe22344fa810b |
|
BLAKE2b-256 | 2c83fdcfc64e718458e0b5e1ef956086fc99c6150b6b51c1ffcea9ce9226cbbb |