Skip to main content

Decline Curve Library

Project description

PyPi Version Build Status Documentation Status Coverage Status Open in Visual Studio Code

Empirical analysis of production data requires implementation of several decline curve models spread over years and multiple SPE publications. Additionally, comprehensive analysis requires graphical analysis among multiple diagnostics plots and their respective plotting functions. While each model’s q(t) (rate) function may be simple, the N(t) (cumulative volume) may not be. For example, the hyperbolic model has three different forms (hyperbolic, harmonic, exponential), and this is complicated by potentially multiple segments, each of which must be continuous in the rate derivatives. Or, as in the case of the Power-Law Exponential model, the N(t) function must be numerically evaluated.

This library defines a single interface to each of the implemented decline curve models. Each model has validation checks for parameter values and provides simple-to-use methods for evaluating arrays of time to obtain the desired function output.

Additionally, we also define an interface to attach a GOR/CGR yield function to any primary phase model. We can then obtain the outputs for the secondary phase as easily as the primary phase.

Analytic functions are implemented wherever possible. When not possible, numerical evaluations are performed using scipy.integrate.fixed_quad. Given that most of the functions of interest that must be numerically evaluated are monotonic, this generally works well.

Primary Phase

Transient Hyperbolic, Modified Hyperbolic, Power-Law Exponential, Stretched Exponential, Duong

Secondary Phase

Power-Law Yield

Water Phase

Power-Law Yield

The following functions are exposed for use

Base Functions

rate(t), cum(t), D(t), beta(t), b(t),

Interval Volumes

interval_vol(t), monthly_vol(t), monthly_vol_equiv(t),

Transient Hyperbolic

transient_rate(t), transient_cum(t), transient_D(t), transient_beta(t), transient_b(t)

Primary Phase

add_secondary(model), add_water(model)

Secondary Phase

gor(t), cgr(t)

Water Phase

wor(t), wgr(t)

Utility

bourdet(y, x, …), get_time(…), get_time_monthly_vol(…)

Getting Started

Install the library with pip:

pip install petbox-dca

A default time array of evenly-logspaced values over 5 log cycles is provided as a convenience.

>>> from petbox import dca
>>> t = dca.get_time()
>>> mh = dca.MH(qi=1000.0, Di=0.8, bi=1.8, Dterm=0.08)
>>> mh.rate(t)
array([986.738, 982.789, 977.692, ..., 0.000])

We can also attach secondary phase and water phase models, and evaluate the rate just as easily.

>>> mh.add_secondary(dca.PLYield(c=1200.0, m0=0.0, m=0.6, t0=180.0, min=None, max=20_000.0))
>>> mh.secondary.rate(t)
array([1184.086, 1179.346, 1173.231, ..., 0.000])

>>> mh.add_water(dca.PLYield(c=2.0, m0=0.0, m=0.1, t0=90.0, min=None, max=10.0))
>>> mh.water.rate(t)
array([1.950, 1.935, 1.917, ..., 0.000])

Once instantiated, the same functions and process for attaching a secondary phase work for any model.

>>> thm = dca.THM(qi=1000.0, Di=0.8, bi=2.0, bf=0.8, telf=30.0, bterm=0.03, tterm=10.0)
>>> thm.rate(t)
array([968.681, 959.741, 948.451, ..., 0.000])

>>> thm.add_secondary(dca.PLYield(c=1200.0, m0=0.0, m=0.6, t0=180.0, min=None, max=20_000.0))
>>> thm.secondary.rate(t)
array([1162.417, 1151.690, 1138.141, ..., 0.000])

>>> ple = dca.PLE(qi=1000.0, Di=0.1, Dinf=0.00001, n=0.5)
>>> ple.rate(t)
array([904.828, 892.092, 877.768, ..., 0.000])

>>> ple.add_secondary(dca.PLYield(c=1200.0, m0=0.0, m=0.6, t0=180.0, min=None, max=20_000.0))
>>> ple.secondary.rate(t)
array([1085.794, 1070.510, 1053.322, ..., 0.000])

Applying the above, we can easily evaluate each model against a data set.

>>> import matplotlib.pyplot as plt
>>> fig = plt.figure()
>>> ax1 = fig.add_subplot(121)
>>> ax2 = fig.add_subplot(122)

>>> ax1.plot(t_data, rate_data, 'o')
>>> ax2.plot(t_data, cum_data, 'o')

>>> ax1.plot(t, thm.rate(t))
>>> ax2.plot(t, thm.cum(t) * cum_data[-1] / thm.cum(t_data[-1]))  # normalization

>>> ax1.plot(t, ple.rate(t))
>>> ax2.plot(t, ple.cum(t) * cum_data[-1] / ple.cum(t_data[-1]))  # normalization

>>> ...

>>> plt.show()
model comparison

See the API documentation for a complete listing, detailed use examples, and model comparison.

Regression

No methods for regression are included in this library, as the models are simple enough to be implemented in any regression package. I recommend using scipy.optimize.least_squares.

For detailed derivation and argument for regression techniques, please see SPE-201404-MS – Optimization Methods for Time–Rate–Pressure Production Data Analysis using Automatic Outlier Filtering and Bayesian Derivative Calculations. Additionally, you may view my blog post on the topic. The Jupyter Notebook is available here.

The following is an example of how to use the THM model with scipy.optimize.least_squares.

from petbox import dca
import numpy as np
import scipy as sc

from scipy.optimize import least_squares

from typing import NamedTuple
from numpy.typing import NDArray


class Bounds(NamedTuple):
    qi: tuple[float, float]
    Di: tuple[float, float]
    bf: tuple[float, float]
    telf: tuple[float, float]


def load_data() -> tuple[NDArray[np.float64], NDArray[np.float64]]:
    ... # load your data here
    return rate, time


def filter_buildup(rate: NDArray[np.float64], time: NDArray[np.float64]) -> tuple[NDArray[np.float64], NDArray[np.float64]]:
    """Filter out buildup data"""
    idx = np.argmax(rate)
    return rate[idx:], time[idx:]


def jitter_rates(rate: NDArray[np.float64]) -> NDArray[np.float64]:
    """Add small jitter to rates to improve gradient descent"""
    # double-precion has at least 15 digits, so for rates in the 10_000s, this leaves a lot of room
    sd = 1e-6
    return rate * np.random.normal(1.0, sd, rate.shape)


def forecast_thm(params: NDArray[np.float64], time: NDArray[np.float64]) -> NDArray[np.float64]:
    """Forecast rates using the Transient Hyperbolic Model"""
    thm = dca.THM(
        qi=params[0],
        Di=params[1],
        bi=2.0,
        bf=params[2],
        telf=params[3],
        bterm=0.0,
        tterm=0.0
    )
    return thm.rate(time)


def log1sp(x: NDArray[np.float64]) -> NDArray[np.float64]:
    """Add small epsilon to avoid log(0) error"""
    return np.log(x + 1e-6)


def residuals(params: NDArray[np.float64], time: NDArray[np.float64], rate: NDArray[np.float64]) -> NDArray[np.float64]:
    """Residuals for scipy.optimize.least_squares"""
    forecast = forecast_thm(params, time)
    return log1sp(rate) - log1sp(forecast)


rate, time = load_data()
rate, time = filter_buildup(rate, time)  # filter out buildup data
rate = jitter_rates(rate)  # add small jitter to rates to improve gradient descent
bounds = Bounds(  # these ***are not general***, they must be calibrated to your data
    qi=   (10.0,  10000.0),
    Di=   (1e-6,      0.8),
    bf=   ( 0.5,      1.5),
    telf= ( 5.0,     50.0)
)
opt = least_squares(
    fun=lambda params, time, rate: residuals(params, time, rate),  # residuals function
    bounds=list(zip(*bounds)),  # unpack bounds into list of tuples
    x0=[np.mean(p) for p in bounds],  # initial guess, mean works well enough
    args=(time, rate),  # additional arguments to `fun`
    loss='soft_l1',  # robust loss function
    f_scale=.35  # affects outlier senstivity of the regression, larger values are more sensitive
)

# no terminal segment
# bterm = 0.0
# tterm = 0.0

# hyperbolic terminal segment
bterm = 0.3
tterm = 15.0  # years

# exponential terminal segment
# bterm = 0.06  # 6.0% secant effective decline / year
# tterm = 0.0

params = np.r_[np.insert(opt.x, 2, 2.0), bterm, tterm]  # insert bi=2.0 and terminal parameters
print(params)

Which would print something like the following:

[1177.57885, 0.793357559, 2.0, 0.666515071, 7.17744813, 0.3, 15.0]

And passed into the THM constructor as follows:

thm = dca.THM.from_params(params)

Development

petbox-dca is maintained by David S. Fulford (@dsfulf). Please post an issue or pull request in this repo for any problems or suggestions!

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

petbox_dca-1.2.0.tar.gz (847.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

petbox_dca-1.2.0-py3-none-any.whl (22.4 kB view details)

Uploaded Python 3

File details

Details for the file petbox_dca-1.2.0.tar.gz.

File metadata

  • Download URL: petbox_dca-1.2.0.tar.gz
  • Upload date:
  • Size: 847.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.10.13

File hashes

Hashes for petbox_dca-1.2.0.tar.gz
Algorithm Hash digest
SHA256 7a1ef470fa78b1ed8f4120d53f8342ece553841ce32c7b12420aeaf036d98433
MD5 58908386a801fac90a682af77161502f
BLAKE2b-256 3e29518e3b5734b95b655d0bcd2116a82bd5572cc30a1022b95897045c64680d

See more details on using hashes here.

File details

Details for the file petbox_dca-1.2.0-py3-none-any.whl.

File metadata

  • Download URL: petbox_dca-1.2.0-py3-none-any.whl
  • Upload date:
  • Size: 22.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.10.13

File hashes

Hashes for petbox_dca-1.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 31a021f565fb1c9a6530eaba9a4438edfc087d7b7910a801d81305b510cb74f0
MD5 3933ab62a703cae0d475ed31d8408d83
BLAKE2b-256 45ab94ed5396bd6d40c7aa55edabbdf037f5c4638e739391339a72ee8b625100

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page