A cowsay clone for python in one file.
Project description
What is Bombilla?
Bombilla is a configuration format for describing python objects and executions in plain json. Bombilla is compatible with any python framework, e.g pytorch lightning, keras and jax. You can use bombilla to define an experiment with a json file, and then execute it.
Installation
bombilla can be installed with pip:
pip install bomba
API
Example of using bombilla API:
from bombilla import Bombilla
bombilla = Bombilla(bombilla_object_descriptor_dict)
# parses dict and loads modules, does not executes anything yet
bombilla.load()
# executes everything in the dictionary, creates objects
bombilla.execute()
# you can pass argument if you want to execute a function on a specific object (e.g. train a model)
bombilla.execute_method("trainer", "fit", *args, **kwargs)
# you can get the object by key
bombilla.find("resnet")
Object descriptor format
An object descriptor is a dictionary that describes python objects and executions. The dictionary can contain the following keys:
module
: the python module where the object is definedclass_name
: the name of the classobject_key
: the key of the object for dynamic referencingparams
: the parameters for creating the objectfunction
: the function to be executedmethod_args
: arguments for calling a specific method on an object
Note that all the arguments are directly passed to the object constructor, so you can use any argument that is accepted by the function's singnature.
For example, here an object is described from torchvision.models
python module, and used to load a pre-trained resnet.
"resnet":{
"module":"torchvision.models",
"class_name":"resnet18",
"object_key":"resnet",
"params":{
"pretrained":true
}
}
Or a simple jax execution configuration file that executes a function:
{
"data": {
"module": "data_loader.datasets",
"function": "mnist",
"object_key": "data",
"params": {
"permute_train": true
}
},
"train_function": {
"module": "trainers.mnist_classifier",
"function": "train",
"params": {
"datasets": "{data}",
"step_size": 0.001,
"num_epochs": 10,
"batch_size": 128,
"momentum_mass": 0.9
}
}
}
Or you can call methods on objects:
{
"data": {
"module": "data_loaders.cifar10.data_loader",
"function": "get_train_data_loader",
"object_key": "data",
"params": {
"batch_size": 128
}
},
"trainer": {
"module": "modules.resnet.resnet",
"function": "get_model",
"params": {
"input_shape": [
180,
180,
3
],
"num_classes": 2
},
"object_key": "model",
"method_args": [
{
"function": "compile",
"params": {
"optimizer": {
"module": "tensorflow.keras.optimizers",
"class_name": "Adam",
"params": {
"learning_rate": 0.001
}
},
"loss": "binary_crossentropy",
"metrics": [
"accuracy"
]
}
},
{
"function": "fit",
"params": {
"": "{data}",
"epochs": 10
}
}
]
}
}
And describe a nested dictionary of any depth:
{
"data": {
"module": "data_loaders.cifar10.data_loader",
"class_name": "CifarLightningDataModule",
"object_key": "data",
"params": {
"location": "./data/cifar10",
"batch_size": 128,
"image_size": [
256,
256
],
"crop_size": 4
}
},
"pytorch_lightning_module": {
"module": "base_classification",
"class_name": "LightningClassificationModule",
"object_key": "pl_model",
"params": {
"classifier": {
"module": "modules.resnet.resnet",
"object_key": "classifier",
"class_name": "ResNet",
"params": {
"block": "BasicBlock",
"layers": [
3,
4,
6,
3
],
"num_classes": 10,
"in_channels": 3,
"zero_init_residual": false,
"groups": 1,
"width_per_group": 64,
"replace_stride_with_dilation": [
false,
false,
false
],
"norm_layer": "BatchNorm2d"
}
},
"optimizers": {
"optimizer": {
"module": "torch.optim",
"class_name": "Adam",
"object_key": "optimizer",
"params": {
"lr": 0.0004,
"betas": [
0.5,
0.999
],
"params": {
"function_call": "parameters",
"reference_key": "classifier",
"params": {}
}
}
},
"lr_scheduler": {
"monitor": "val_loss",
"scheduler": {
"module": "torch.optim.lr_scheduler",
"class_name": "ReduceLROnPlateau",
"params": {
"optimizer": "{optimizer}",
"mode": "min",
"factor": 0.5,
"threshold": 1e-08,
"threshold_mode": "rel",
"patience": 0,
"verbose": true
}
}
}
}
}
},
"trainer": {
"module": "pytorch_lightning",
"class_name": "Trainer",
"params": {
"gpus": 1,
"max_epochs": 100,
"precision": 16,
"gradient_clip_val": 0.5,
"enable_checkpointing": true,
"callbacks": [
{
"module": "pytorch_lightning.callbacks",
"class_name": "EarlyStopping",
"params": {
"monitor": "val_loss",
"patience": 10,
"mode": "min"
}
},
{
"module": "pytorch_lightning.callbacks",
"class_name": "ModelCheckpoint",
"params": {
"dirpath": "{save_dir}/checkpoints",
"monitor": "val_loss",
"save_top_k": 1,
"verbose": true,
"save_last": true,
"mode": "min"
}
}
],
"logger": {
"module": "pytorch_lightning.loggers",
"class_name": "CSVLogger",
"params": {
"save_dir": "./logs"
}
}
},
"method_args": [
{
"function": "fit",
"params": {
"model": "{pl_model}",
"datamodule": "{data}"
}
},
{
"function": "test",
"params": {
"model": "{pl_model}",
"datamodule": "{data}"
}
}
]
}
}
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.