Skip to main content

Various methods to support event-based evaluations.

Project description

OWPHydroTools :: Events

This subpackage implements methods to partition hydrometric times series into discrete "events." See the Events Documentation for a complete list and description of the currently available methods. To report bugs or request new methods, submit an issue through the OWPHydroTools Issue Tracker on GitHub.

Installation

In accordance with the python community, we support and advise the usage of virtual environments in any workflow using python. In the following installation guide, we use python's built-in venv module to create a virtual environment in which the tool will be installed. Note this is just personal preference, any python virtual environment manager should work just fine (conda, pipenv, etc. ).

# Create and activate python environment, requires python >= 3.8
$ python3 -m venv venv
$ source venv/bin/activate
$ python3 -m pip install --upgrade pip

# Install events
$ python3 -m pip install hydrotools.events

Event Detection Usage

The following example demonstrates how one might use hydrotools.events.event_detection to extract hydrological events from a time series of real streamflow measurements. This example requires the hydrotools.nwis_client package.

Code

# Import tools to retrieve data and detect events
from hydrotools.nwis_client.iv import IVDataService
from hydrotools.events.event_detection import decomposition as ev

# Use pandas to resample the data
from pandas import Grouper

# Use a helper function to detect events at multiple sites
def list_events_helper(mi_series, level, halflife, window):
    """Reduce multi-index series before applying event detection."""
    return ev.list_events(mi_series.droplevel(level), halflife, window)

# Retrieve streamflow observations for two sites
service = IVDataService()
observations = service.get(
    sites='02146750,0214676115,09489000', 
    startDT='2019-10-01', 
    endDT='2020-09-30'
    )

# Drop extra columns to be more efficient
observations = observations[[
    'usgs_site_code', 
    'value_date', 
    'value'
    ]]

# Check for duplicate time series, keep first by default
observations = observations.drop_duplicates(
    subset=['usgs_site_code', 'value_date']
    )

# Resample to hourly, keep first measurement in each 1-hour bin
observations = observations.groupby([
    'usgs_site_code',
    Grouper(key='value_date', freq='h')
    ]).first().ffill()

# Detect events
events = observations['value'].groupby(
    level='usgs_site_code').apply(
        list_events_helper, 
        level='usgs_site_code', 
        halflife='6h', 
        window='7D'
    )

# Print event list    
print(events)

Output

                                start                 end
usgs_site_code                                           
02146750       0  2019-10-13 22:00:00 2019-10-15 04:00:00
               1  2019-10-16 09:00:00 2019-10-18 02:00:00
               2  2019-10-19 03:00:00 2019-10-23 19:00:00
               3  2019-10-27 05:00:00 2019-10-28 18:00:00
               4  2019-10-30 01:00:00 2019-11-02 18:00:00
...                               ...                 ...
09489000       34 2020-09-16 00:00:00 2020-09-16 19:00:00
               35 2020-09-17 00:00:00 2020-09-17 15:00:00
               36 2020-09-18 04:00:00 2020-09-18 10:00:00
               37 2020-09-27 01:00:00 2020-09-27 03:00:00
               38 2020-09-28 05:00:00 2020-09-28 11:00:00

Event Detection Tips

The evaluation_tools.events.event_detection.decomposition method has two main parameters: halflife and window. These parameters are passed directly to underlying filters used to remove noise and model the trend in a streamflow time series (also called baseflow). Significant contiguous deviations from this trend are flagged as "events". This method was originally conceived to detect rainfall-driven runoff events in small watersheds from records of volumetric discharge or total runoff. Before using decomposition you will want to have some idea of the event timescales you hope to detect.

In general you'll want to consider the following:

  1. Ensure your time series is monotonically increasing with a frequency significantly less than the frequency of events and fill missing data
  2. Use a pandas.Timedelta compatible str to specify halflife and window. See the pandas docs
  3. Specify a halflife larger than the expected frequency of noise, but smaller than the event timescale
  4. Specify a window larger than the event timescale, but at least 4 to 5 times smaller than the entire length of the time series
  5. Filter the final list of events to remove false positive events, particularly in noisy signals. You could filter on peak event flow, event duration, or other event characteristics

Baseflow Separation Usage

Baseflow Hydrograph

The following example demonstrates how to use hydrotools.events.baseflow to estimate baseflow from a time series of streamflow values. This example requires the hydrotools.nwis_client package.

Code

# Import tools to retrieve data and perform baseflow separation
from hydrotools.events.baseflow import eckhardt as bf
from hydrotools.nwis_client.iv import IVDataService

# Get some streamflow data
client = IVDataService()
df = client.get(
    startDT="2020-10-01",
    endDT="2021-09-30T23:55",
    sites="02146470"
)

# Important: Streamflow data should be continuous with
#   no missing data and a DateTimeIndex
TIME_SCALE = "1h"
df = df.set_index("value_time").resample(TIME_SCALE).first().ffill()

# Let's rename 'value' to 'total_flow'
df = df.rename(columns={"value": "total_flow"})

# Perform baseflow separation
#   This catchment is pretty small, so we'll use a
#   12-hour recession time-scale. The default is 1-day
#   per Eckhardt (2008).
results = bf.separate_baseflow(
    df["total_flow"],
    output_time_scale = TIME_SCALE,
    recession_time_scale = "12h"
    )

# Look at the 12-hour filter parameters
print(f"Recession constant, a: {results.recession_constant:.4f}")
print(f"Maximum baseflow index, bfi_max: {results.maximum_baseflow_index:.4f}")

# Store the resulting baseflow in the original DataFrame
df["baseflow"] = results.values
print(df.head())

Output

Recession constant, a: 0.9636
Maximum baseflow index, bfi_max: 0.3186
                    variable_name usgs_site_code measurement_unit  total_flow qualifiers series  baseflow
value_time                                                                                               
2020-09-30 23:00:00    streamflow       02146470            ft3/s        1.03      ['A']      0  1.030000
2020-10-01 00:00:00    streamflow       02146470            ft3/s        1.03      ['A']      0  0.993130
2020-10-01 01:00:00    streamflow       02146470            ft3/s        1.03      ['A']      0  0.958197
2020-10-01 02:00:00    streamflow       02146470            ft3/s        1.03      ['A']      0  0.925100
2020-10-01 03:00:00    streamflow       02146470            ft3/s        1.03      ['A']      0  0.893740

Interception

The hydrotools.events.models.interception module includes a basic method that estimates canopy interception. The method includes reasonable defaults from the literature for average evaporation rates and trunk fractions. The primary point of entry is the compute_interception method which estimates canopy intercepted rainfall in units of depth. compute_interception requires a gross rainfall amount, an average rain rate, and a canopy fraction. The defaults assume that gross rainfall is in mm, rainfall rate is in mm/h, and canopy fraction is dimensionless. While this method can be used in different contexts, expected usage is for event-scale (single storm) catchment-wide estimates of intercepted rainfall.

Example usage

# Imports
from hydrotools.events.models.interception import compute_interception

# Assumed values for a single event
storm_total_rainfall = 50.0 # mm
mean_rainfall_rate = 5.0 # mm/h
vegetative_fraction = 0.5 # 50% vegetative landcover

# Estimate interception
estimate = compute_interception(
    gross_rainfall = storm_total_rainfall,
    rainfall_rate = mean_rainfall_rate,
    canopy_fraction = vegetative_fraction
)
print(f"Storm total intercepted rainfall (mm): {estimate:.1f}")
print(f"Storm total intercepted rainfall (%): {(estimate/storm_total_rainfall)*100:.1f}")
print(f"Effective rainfall (mm): {storm_total_rainfall-estimate:.1f}")

Output

Storm total intercepted rainfall (mm): 2.5
Storm total intercepted rainfall (%): 5.0
Effective rainfall (mm): 47.5

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

hydrotools_events-2.2.0.tar.gz (28.3 kB view details)

Uploaded Source

Built Distribution

hydrotools_events-2.2.0-py3-none-any.whl (23.7 kB view details)

Uploaded Python 3

File details

Details for the file hydrotools_events-2.2.0.tar.gz.

File metadata

  • Download URL: hydrotools_events-2.2.0.tar.gz
  • Upload date:
  • Size: 28.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.2

File hashes

Hashes for hydrotools_events-2.2.0.tar.gz
Algorithm Hash digest
SHA256 7085dafe39106c8534139cae91e9eedfc3e5613aea35d0491b9c74054a37404b
MD5 0cbf04ccff0182af40a67049deaae610
BLAKE2b-256 3c028374b1175cda2d88bcd817da1b8e5e240da67c6c1dd71c9e055ef213c96b

See more details on using hashes here.

File details

Details for the file hydrotools_events-2.2.0-py3-none-any.whl.

File metadata

File hashes

Hashes for hydrotools_events-2.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 ad3bd11bedc7dfbd539dec5598c90469e686e70091981fca4a26052b36f7c8da
MD5 bf094bda1419a21e16ee49c18ce0dce4
BLAKE2b-256 c884b3e54684b7efc5dd6b65d98d90c34fc12b176d7aeb0300a8f42a22b90a46

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page