Stores mosaik simulation data in an HDF5 database.
Project description
mosaik-hdf5
Store mosaik simulation data in an HDF5 database.
Mosaik-hdf5 stores the relations-graph of your simulation, timeseries for entities connected to it and optionally static entity and simulation meta data. The database structure usually looks like this:
/ [meta data] | +- Relations | | | +- Sim-0.Entity-1 | | | +- PyPower-0.1-Node-2 | +- Series | +- Sim-0.Entiy-1 [static data] | | | +- val_out | +- PyPower-0.1-Node-2 [static data] | +- P | +- Q
The Relations group contains one dataset for every entity. For each of the entity’s relations, the dataset has one tuple (path_to_relation, path_to_relatio_series).
The Series group contains (by default) one group for every entity. Each of these group as one dataset for every attribute.
Static entity data is stored as attributes in the entity groups. Simulation meta data is stored as attributes of the root group.
You can, optionally, create a more nested structure for the series, for example, if you want to group entities by simulator and/or simulator instance. This is done via regular expression replacements based on the entity ID.
Installation
mosaik-hdf5 uses the h5py module. If you get an error during installation that hdf5.h is missing, install the HDF5 headers (e.g., sudo apt-get install libhdf5-dev or brew install hdf5) or use a binary package (e.g., for Windows)
$ pip install mosaik-hdf5
Usage
You can run mosaik-hdf5 as a sub-process or your simulation or in-process with it. Here are example configurations for both variants:
sim_config = {
'HDF5-inproc': {
'python': 'mosaik_hdf5:MosaikHdf5',
},
'HDF5-subproc': {
'cmd': 'mosaik-hdf5 %(addr)s',
},
}
Initialization
When you start mosaik-hdf5, you have to provide a step_size and a duration argument. The step_size defines how often data will be collected. The duration is the simulation end time in seconds. It is used to calculate the dataset size for every time series. For example, if duration is half-an-hour (1800s) and step_size is 60, each dataset will have a length of 30.
You can optionally pass a series_path tuple which contains a regular expression pattern and replacement string (see the Python docs for details).
For example, by default the entity IDs Sim-0.Entity-1 and PyPower-0.1-Node-2 would map to the series paths /Series/Sim-0.Entity-1 and /Series/PyPower-0.1-Node-2. But you want to group the entities by simulator type and simulator instance. Also, since one mosaik-pypower instance can contain multiple grids, you also want to take care of that. So what you want is something like that: /Series/Sim/Sim-0/Sim-0.Entity-1 and /Series/PyPower/PyPower-0.1/PyPower-0.1-Node-2. In this (rather complex) case, series_path can be (r'(((\w+)-(\d+.\d+|\d+))[.-](.*))', r'\3/\2/\1'). Easy, isn’t it?
Here are two examples for this:
a = world.start('HDF5', step_size=60, duration=1800)
pattern = r'(((\w+)-(\d+.\d+|\d+))[.-](.*))'
repl = r'\3/\2/\1'
b = world.start('HDF5', step_size=1, duration=10,
series_path=(pattern, repl))
Model instantiation
Every instance of mosaik-hdf5 allows you to create exactly one instance of its Database model (which is also the only model provided). The Database has the following parameters:
gridfile is the filename of the HDF5 database that will be created.
buf_size (default: 1000) is the size of the internal data buffer for each series dataset. Mosaik-hdf5 buffers the data for every dataset and only writes larger chunks of data to the disk in order to improve the writing performance. If you have a lot of entities (> 100k) and only little memory, you may reduce this number. If you have lots of RAM, you can play with larger buffer sizes and see if it improves the performance for you.
dataset_opts (default: None) is a dictionary of arguments that get passed to h5py’s create_dataset() method.
This can, for example, be used to enable compression (note, that the lzf compression is not supported by all HDF5 viewers).
Examples:
# Basic usage
hdf5 = world.start('HDF5', step_size=1, duration=1)
db = hdf5.Database('data.hdf5')
# Use gzip compression
hdf5 = world.start('HDF5', step_size=1, duration=1)
db = hdf5.Database('data.hdf5', dataset_opts={
'compression': 'gzip',
'compression_opts': 9,
})
# Use lzf compression and a larger buffer
hdf5 = world.start('HDF5', step_size=1, duration=1)
db = hdf5.Database('data.hdf5', buf_size=1336,
dataset_opts={'compression': 'lzf'})
Storing data
The Database model has no attributes, but it accepts any inputs. This means that you can just connect anything to it. For each entity and attribute that is connected to the database, a corresponding dataset will be created in the database.
Mosaik-hdf5 also provides to extra methods that allow you to store some simulation meta data and static entity data. You can only use these methods once you created an instance of the Database model. The method set_meta_data() takes a single dict with an arbitrary amount of key-values pairs. The method set_static_data() takes a dict of entities and data dicts.
In the following example, we’ll create some (fake) PV entities and a power grid (with nodes and lines). We want to store the PV’s active and reactive power (P, Q), the node voltage and angle (Vm, Va) for all nodes and the complex current (I_real, I_imag) of all branches:
pv_pmax = 10
pvs = make_pvs(pv_pmax, ...) # A list of PV entities
nodes, lines = make_grid(...) # Lists of nodes/lines of a power grid
hdf5 = world.start('HDF5', step_size=1, duration=10)
db = hdf5.Database('data.hdf5')
# Store meta and static data
hdf5.set_meta_data({'duration': 10, 'description': 'hdf5 demo'}
hdf5.set_static_data({pv.full_id: {'p_max': pv_pmax} for pv in pvs})
# Connect inputs to database
mosaik.util.connect_many_to_one(world, pvs, db, 'P', 'Q')
mosaik.util.connect_many_to_one(world, nodes, db, 'Vm', 'Va')
mosaik.util.connect_many_to_one(world, lines, db, 'I_real', 'I_imag')
For a real example, you can take a look at the mosaik-demo.
Getting help
If you need, please visit the mosaik-users mailing list .
Changelog
0.4 - 2021-06-01
[CHANGE] Updated to mosaik-api 3.0.
0.3 - 2016-02-15
[NEW] Implemented the new “setup_done()” method.
0.2 – 2014-10-29
0.1.2 – 2014-09-22
[CHANGE] Updated to mosaik-api 2.0.
0.1.1 – 2014-07-31
[FIX] Fixed a regression in 0.1.
0.1 – 2014-07-31
Initial release
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file mosaik-hdf5-0.4.tar.gz
.
File metadata
- Download URL: mosaik-hdf5-0.4.tar.gz
- Upload date:
- Size: 16.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.1 importlib_metadata/4.5.0 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.61.0 CPython/3.8.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 050ba63b62958f41f8eb936c7247264958da8155bba436e848d65c3f7f946744 |
|
MD5 | 728125f36679b71fa2b2f373cdc5f64d |
|
BLAKE2b-256 | c8434079967ed1a6d495e434240b0a6d9c88f99a30046292ab015310ac2fd9fe |
File details
Details for the file mosaik_hdf5-0.4-py2.py3-none-any.whl
.
File metadata
- Download URL: mosaik_hdf5-0.4-py2.py3-none-any.whl
- Upload date:
- Size: 15.9 kB
- Tags: Python 2, Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.1 importlib_metadata/4.5.0 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.61.0 CPython/3.8.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6797d9d6204ce2029c70ee575e99620de626efd5562c5a3a187993c76e4a56ff |
|
MD5 | 3f05e2dfbcae664c8271f2f89e6fb5af |
|
BLAKE2b-256 | 7343f644a75eb96c89223ebf4671df8a630a82c9760f2bfafc5d33574cb76387 |