Python client library for the HYDWS web service
Project description
HYDWS Client
Hydraulic Web Service Client This client can be used to access the data from a Hydraulic Webservice more comfortably.
Installation
Currently the package needs to be installed via the github repository:
pip install hydws-client
Usage
Imports
from datetime import datetime
from hydws.client import HYDWSDataSource
from hydws.parser import BoreholeHydraulics, SectionHydraulics
TLDR; How to get the data?
hydws_url = 'http://example.api.com/hydws/v1'
hydws = HYDWSDataSource(hydws_url)
# Or with a custom timeout (in seconds)
hydws = HYDWSDataSource(hydws_url, timeout=30)
Assuming you know exactly what you want, you can use the following methods to get the data you need.
Throughout, you can usually use the name and the id interchangeably.
Let's assume we have the borehole id, and the section name.
borehole_id = 'caf65646-8093-4aaf-989c-1c837f497667'
section_name = '16A-32/section_03'
hydraulics_start = datetime(2024, 4, 6, 1, 0, 0)
hydraulics_end = datetime(2024, 4, 6, 1, 1, 0)
The fastest way to get to the data, without caring about any metadata:
hydraulics = hydws.get_section_hydraulics(borehole_id, section_name, hydraulics_start, hydraulics_end, format='pandas')
hydraulics
| topflow | toppressure | |
|---|---|---|
| datetime | ||
| 2024-04-06 01:00:00 | 0.212990 | 4.710498e+07 |
| 2024-04-06 01:00:01 | 0.212990 | 4.719461e+07 |
| 2024-04-06 01:00:02 | 0.212990 | 4.721530e+07 |
| 2024-04-06 01:00:03 | 0.212831 | 4.716703e+07 |
| 2024-04-06 01:00:04 | 0.212646 | 4.714635e+07 |
| ... | ... | ... |
| 2024-04-06 01:00:56 | 0.213653 | 4.716703e+07 |
| 2024-04-06 01:00:57 | 0.213308 | 4.717393e+07 |
| 2024-04-06 01:00:58 | 0.213308 | 4.710498e+07 |
| 2024-04-06 01:00:59 | 0.213308 | 4.714635e+07 |
| 2024-04-06 01:01:00 | 0.213494 | 4.716014e+07 |
61 rows × 2 columns
Boreholes and Sections
Otherwise, if you want the metadata, or need to parse a file containing hydws, or prefer an object with the borehole and/or section structure, you can use the hydws client together with BoreholeHydraulics and SectionHydraulics classes:
If you just want to consider one Section, it's easiest to directly use the SectionHydraulics class:
section_json = hydws.get_section(borehole_id, section_name, hydraulics_start, hydraulics_end)
section = SectionHydraulics(section_json)
section.metadata # to access the metadata of the section
section.hydraulics # to access the hydraulic data as a dataframe
To filter the data to a narrower time range without another API call:
filtered = section.query_datetime(
starttime=datetime(2024, 4, 6, 1, 0, 30),
endtime=datetime(2024, 4, 6, 1, 0, 45)
)
filtered.hydraulics # returns a new SectionHydraulics with filtered data
| topflow | toppressure | |
|---|---|---|
| datetime | ||
| 2024-04-06 01:00:00 | 0.212990 | 4.710498e+07 |
| 2024-04-06 01:00:01 | 0.212990 | 4.719461e+07 |
| 2024-04-06 01:00:02 | 0.212990 | 4.721530e+07 |
| 2024-04-06 01:00:03 | 0.212831 | 4.716703e+07 |
| 2024-04-06 01:00:04 | 0.212646 | 4.714635e+07 |
| ... | ... | ... |
| 2024-04-06 01:00:56 | 0.213653 | 4.716703e+07 |
| 2024-04-06 01:00:57 | 0.213308 | 4.717393e+07 |
| 2024-04-06 01:00:58 | 0.213308 | 4.710498e+07 |
| 2024-04-06 01:00:59 | 0.213308 | 4.714635e+07 |
| 2024-04-06 01:01:00 | 0.213494 | 4.716014e+07 |
61 rows × 2 columns
If you prefer to directly get all the metadata and hydraulic data of a Borehole for a given time, use the following methods:
# Get Borehole, containing all of its sections and their hydraulic data for the given time range
borehole_json = hydws.get_borehole(borehole_id, hydraulics_start, hydraulics_end)
borehole = BoreholeHydraulics(borehole_json)
borehole.metadata # to access the metadata of the borehole
{'publicid': 'caf65646-8093-4aaf-989c-1c837f497667',
'description': 'Well 16A-32',
'name': '16A-32',
'location': 'FORGE',
'institution': 'FORGE Utah',
'measureddepth': {'value': 3339.1},
'bedrockaltitude': {'value': 0.0},
'altitude': {'value': 1650.02},
'latitude': {'value': -112.906857},
'longitude': {'value': 38.506874},
'creationinfo': {'creationtime': '2024-04-01T22:40:47.911589'}}
Accessing the sections inside a borehole, can be done using the publicid as a dict key, or by using the section name and the nloc attribute:
section_id = '37801a57-90b9-4fb5-83d7-506ee9166acf'
section = borehole[section_id] # use the section id as a key to access the section
section = borehole.nloc[section_name] # use the section name as a key to access the section
section.hydraulics
You can also filter all sections in a borehole at once:
filtered_borehole = borehole.query_datetime(starttime=start, endtime=end)
.dataframe tbody tr th {
vertical-align: top;
}
.dataframe thead th {
text-align: right;
}
</style>
| topflow | toppressure | |
|---|---|---|
| datetime | ||
| 2024-04-06 01:00:00 | 0.212990 | 4.710498e+07 |
| 2024-04-06 01:00:01 | 0.212990 | 4.719461e+07 |
| 2024-04-06 01:00:02 | 0.212990 | 4.721530e+07 |
| 2024-04-06 01:00:03 | 0.212831 | 4.716703e+07 |
| 2024-04-06 01:00:04 | 0.212646 | 4.714635e+07 |
| ... | ... | ... |
| 2024-04-06 01:00:56 | 0.213653 | 4.716703e+07 |
| 2024-04-06 01:00:57 | 0.213308 | 4.717393e+07 |
| 2024-04-06 01:00:58 | 0.213308 | 4.710498e+07 |
| 2024-04-06 01:00:59 | 0.213308 | 4.714635e+07 |
| 2024-04-06 01:01:00 | 0.213494 | 4.716014e+07 |
61 rows × 2 columns
Exporting Data
To convert your data back to HYDWS JSON format (e.g., for saving or uploading):
# Export section with all hydraulic samples
section_json = section.to_json()
# Export with resampled data (every 60 seconds)
section_json_resampled = section.to_json(resample=60)
# Export entire borehole with all sections
borehole_json = borehole.to_json()
Write Operations (POST/DELETE)
The client supports write operations when an API key is provided:
# Initialize with API key for write operations
hydws = HYDWSDataSource(hydws_url, api_key='your-api-key')
Posting Borehole Data
You can post borehole data using either a BoreholeHydraulics object or a raw dict:
# Create and populate a borehole
borehole = BoreholeHydraulics()
borehole.metadata['name'] = 'New Borehole'
section_id = borehole.section_from_dataframe(df)
# Post to the API
result = hydws.post_borehole(borehole)
# Or with merge options (to merge with existing data)
result = hydws.post_borehole(borehole, merge=True, merge_limit=60)
Deleting Data
# Delete a borehole (by name or UUID)
hydws.delete_borehole('Borehole Name')
# Delete hydraulic samples for a section (optionally within a time range)
hydws.delete_section_hydraulics(
'Borehole Name',
'Section Name',
starttime=datetime(2024, 1, 1),
endtime=datetime(2024, 1, 31)
)
Dry-Run Mode (test=True)
All write operations support a test parameter for dry-run simulation:
# Preview what would be posted (validates and prints stats)
hydws.post_borehole(borehole, test=True)
# Output:
# Borehole: New Borehole
# Sections: 2
# Total hydraulics: 1000
# - Section 1: 500 hydraulic samples
# - Section 2: 500 hydraulic samples
# Preview what would be deleted
hydws.delete_borehole('Borehole Name', test=True)
# Output:
# Borehole: Borehole Name
# Sections: 3
# Would be deleted.
# Check how many hydraulic samples would be deleted
hydws.delete_section_hydraulics('Borehole', 'Section', test=True)
# Output:
# Hydraulic samples to delete: 150
Loading Data from Files
You can load borehole data directly from a JSON file:
from hydws.parser import BoreholeHydraulics
# Load borehole from a HYDWS JSON file
borehole = BoreholeHydraulics.from_file('path/to/borehole.json')
Creating Data Programmatically
You can create hydraulic data from your own DataFrames or JSON.
Valid column names for hydraulic DataFrames:
- Flow/Pressure/Temperature:
topflow,toppressure,toptemperature,bottomflow,bottompressure,bottomtemperature - Fluid properties:
fluiddensity,fluidviscosity,fluidph,fluidcomposition
import pandas as pd
from hydws.parser import BoreholeHydraulics, SectionHydraulics
# Create an empty borehole
borehole = BoreholeHydraulics()
borehole.metadata['name'] = 'My Borehole'
# Option 1: Add a section from a DataFrame
df = pd.DataFrame({
'topflow': [1.2e-6, 1.3e-6],
'toppressure': [4.1e6, 4.2e6]
}, index=pd.to_datetime(['2024-01-01 00:00', '2024-01-01 00:01']))
df.index.name = 'datetime'
section_id = borehole.section_from_dataframe(df)
# Option 2: Add a section from JSON (e.g., from a file or API response)
section_json = {'publicid': '...', 'name': 'Section 1', 'hydraulics': [...], ...}
section_id = borehole.section_from_json(section_json)
# Option 3: Add an empty section and populate it later
section_id = borehole.add_empty_section()
borehole[section_id].metadata['name'] = 'New Section'
borehole[section_id].hydraulics = df
You can also work directly with SectionHydraulics:
section = SectionHydraulics()
# Load hydraulic data from JSON (list of samples)
hydraulic_samples = [
{'datetime': {'value': '2024-01-01T00:00:00'}, 'topflow': {'value': 1.2e-6}},
{'datetime': {'value': '2024-01-01T00:01:00'}, 'topflow': {'value': 1.3e-6}}
]
section.load_hydraulic_json(hydraulic_samples)
# Export to JSON
section_json = section.to_json()
Navigating Boreholes, Sections and Metadata
If you don't exactly know which names or publicids there are, you can use the following methods to display the available boreholes and sections.
borehole_metadata_all = hydws.list_boreholes() # returns all metadata
borehole_names = hydws.list_borehole_names() # only returns name, publicid or both (default is name)
borehole_names
['16A-32', '16B-32']
The first column is the name of the borehole, the second one its public ID. It's generally possible to use either of the two for all of the following functions. The same applies to the section names and their IDs.
section_metadata_all = hydws.list_sections(borehole_names[0]) # returns all metadata
section_names = hydws.list_section_names(borehole_names[0]) # only returns name, publicid or both (default is name)
section_names
['16A-32/section_01', '16A-32/section_02', '16A-32/section_03']
It is also possible to use theget_borehole_metadata or the get_section_metadata functions to get the metadata of a specific borehole or section.
borehole_metadata = hydws.get_borehole_metadata(borehole_names[0])
section_metadata = hydws.get_section_metadata(borehole_names[0], section_names[0])
section_metadata
{'publicid': '8eb6ce9a-247d-4675-8647-880841bb9531',
'starttime': '2022-04-17T02:35:57',
'endtime': '2022-04-17T05:50:13',
'topclosed': True,
'bottomclosed': True,
'description': '200 ft long open hole section at the toe of the well',
'name': '16A-32/section_01',
'hydraulics': [],
'casingdiameter': {},
'holediameter': {},
'bottommeasureddepth': {'value': 3339.09},
'topmeasureddepth': {'value': 3278.13},
'bottomaltitude': {'value': -958.71},
'bottomlatitude': {'value': 38.504462},
'bottomlongitude': {'value': -112.893086},
'topaltitude': {'value': -936.3276802186539},
'toplatitude': {'value': 38.50454148787154},
'toplongitude': {'value': -112.89372764195421}}
Coordinate Transformations
When working with local coordinate systems (e.g., Swiss LV95, UTM zones), you can convert between WGS84 (lat/lon) and local coordinates:
from hydws.coordinates import CoordinateTransformer
# Initialize with target CRS (e.g., Swiss LV95)
transformer = CoordinateTransformer('EPSG:2056')
# Convert WGS84 to local coordinates
easting, northing, altitude = transformer.to_local_coords(lon=8.5, lat=47.4)
# Convert back to WGS84
lon, lat, alt = transformer.from_local_coords(easting, northing)
Raw Data Parser
The RawHydraulicsParser transforms raw sensor data from pandas DataFrames into HYDWS JSON format. It's useful for ingesting data from various monitoring systems and mapping sensor columns to the correct borehole sections.
from hydws.parser.rawparser import RawHydraulicsParser
import json
# Load borehole metadata (from HYDWS API or file)
with open('boreholes.json') as f:
boreholes_metadata = json.load(f)
# Initialize parser with config and metadata
parser = RawHydraulicsParser('config.json', boreholes_metadata)
# Parse a DataFrame with sensor data (datetime index required)
hydws_data = parser.parse(dataframe, format='json') # Returns list of borehole dicts
Config File Structure
The config file is a JSON array defining how to map DataFrame columns to HYDWS fields:
[
{
"columnNames": ["sensor_pressure_1", "sensor_pressure_2"],
"fieldName": "toppressure",
"unitConversion": ["mul", 1000000],
"sensorPosition": "downhole",
"assignTo": "sectionID",
"section": "ST1_section_01"
}
]
Config fields:
| Field | Required | Description |
|---|---|---|
columnNames |
Yes | Input DataFrame columns (summed if multiple) |
fieldName |
Yes | Target field: toppressure, topflow, toptemperature, bottompressure, bottomflow, bottomtemperature, fluiddensity, fluidviscosity, fluidph |
assignTo |
Yes | "sectionID" (direct) or "plan" (time-based CSV) |
section |
Yes | Section name or path to plan CSV |
unitConversion |
No | [operation, value] e.g., ["mul", 1e6], ["div", 60] |
sensorPosition |
No | "surface" or "downhole" (affects pressure correction) |
conditions |
No | Conditional column selection rules |
Time-Based Assignment (Plan Mode)
For scenarios where data should be routed to different sections based on time (e.g., injection experiments), use a plan CSV:
{
"columnNames": ["injection_pressure"],
"fieldName": "toppressure",
"assignTo": "plan",
"section": "injection_plan.csv"
}
Plan CSV format:
interval, date_from, date_until
ST1_section_11, 2024/04/15T00:00:00, 2024/04/20T23:59:59
ST1_section_12, 2024/04/21T00:00:00, 2024/04/25T23:59:59
Surface Pressure Correction
When sensorPosition: "surface" is set for pressure fields, the parser automatically adds hydrostatic pressure to correct for the water column height between the surface and the section depth.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file hydws_client-1.3.3.tar.gz.
File metadata
- Download URL: hydws_client-1.3.3.tar.gz
- Upload date:
- Size: 71.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
480aa078b5d7ba162cddd4e9942ce0750f0b7dfb2d537dbaf5378c71dc51e222
|
|
| MD5 |
b6329f35080ea14f6f27c90b57112ed0
|
|
| BLAKE2b-256 |
fd2dcc8858eab8576d9b3c91676bbd4373449111cb3722d82cdac39e98e5df55
|
Provenance
The following attestation bundles were made for hydws_client-1.3.3.tar.gz:
Publisher:
publish.yml on swiss-seismological-service/hydws-client
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
hydws_client-1.3.3.tar.gz -
Subject digest:
480aa078b5d7ba162cddd4e9942ce0750f0b7dfb2d537dbaf5378c71dc51e222 - Sigstore transparency entry: 863011378
- Sigstore integration time:
-
Permalink:
swiss-seismological-service/hydws-client@6ff9e2880c89b8239bc606fa6701b9bec431d4d5 -
Branch / Tag:
refs/tags/v1.3.3 - Owner: https://github.com/swiss-seismological-service
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@6ff9e2880c89b8239bc606fa6701b9bec431d4d5 -
Trigger Event:
push
-
Statement type:
File details
Details for the file hydws_client-1.3.3-py3-none-any.whl.
File metadata
- Download URL: hydws_client-1.3.3-py3-none-any.whl
- Upload date:
- Size: 54.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
651b91ba58c9de8d21a90de4b895cf02be250be24c3bd8ac26c439fb92b07492
|
|
| MD5 |
cf1300619edf3118d8ff3ece9b5c634e
|
|
| BLAKE2b-256 |
f37e740d5db531c8ef56d2ba47a4fb10263f7caaf381ec86ff35860c0c245434
|
Provenance
The following attestation bundles were made for hydws_client-1.3.3-py3-none-any.whl:
Publisher:
publish.yml on swiss-seismological-service/hydws-client
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
hydws_client-1.3.3-py3-none-any.whl -
Subject digest:
651b91ba58c9de8d21a90de4b895cf02be250be24c3bd8ac26c439fb92b07492 - Sigstore transparency entry: 863011416
- Sigstore integration time:
-
Permalink:
swiss-seismological-service/hydws-client@6ff9e2880c89b8239bc606fa6701b9bec431d4d5 -
Branch / Tag:
refs/tags/v1.3.3 - Owner: https://github.com/swiss-seismological-service
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@6ff9e2880c89b8239bc606fa6701b9bec431d4d5 -
Trigger Event:
push
-
Statement type: