Skip to main content

Package to prepare well log data for ML projects.

Project description

MLPet

Preprocessing tools for Petrophysics ML projects at Eureka

Installation

  • Install the package by running the following (requires python 3.8 or later)

      pip install mlpet
    

Quick start

  • Short example for pre-processing data prior to making a regression model:

      from mlpet.Datasets.shear import Sheardata
      # Instantiate an empty dataset object using the example settings and mappings provided
      ds = Sheardata(
              settings="support/settings_shear.yaml", 
              mappings="support/mappings.yaml", 
              folder_path="support/")
      # Populate the dataset with data from a file 
      # (support for multiple file formats and direct cdf data collection exists)
      ds.load_from_pickle("support/data/shear.pkl")
      # The original data will be kept in ds.df_original and will remain unchanged 
      print(ds.df_original.head())
      # Split the data into train-validation sets
      df_train_original, df_valid_original, valid_wells = ds.train_test_split(
              df=ds.df_original, 
              test_size=0.3)
      # Preprocess the data for training
      df_train, train_key_wells, feats = ds.preprocess(df_train_original)
      # Preprocecss accepts some keyword arguments related to various steps 
      # (e.g. the key wells used for normalizing curves such as GR)
      df_valid, valid_key_wells, _ = ds.preprocess(
              df_valid_original, 
              _normalize_curves={'key_wells':train_key_wells})
    

The procedure will be exactly the same for the lithology class. The only difference will be in the "settings". For a full list of possible settings keys see the documentation for the main Dataset class. Make sure that the curve names are consistent with those in the dataset. The mappings will NOT be applied during the load data step.

API Documentation

Full API documentaion of the package can be found under docs/

For developers

  • to update the API documentation, from the root directory of the project run

      pip install pdoc
      pdoc --docformat google -o docs mlpet
    

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

MLPet-0.0.7.1.tar.gz (19.7 kB view details)

Uploaded Source

Built Distribution

MLPet-0.0.7.1-py3-none-any.whl (22.3 kB view details)

Uploaded Python 3

File details

Details for the file MLPet-0.0.7.1.tar.gz.

File metadata

  • Download URL: MLPet-0.0.7.1.tar.gz
  • Upload date:
  • Size: 19.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.8.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.9.7

File hashes

Hashes for MLPet-0.0.7.1.tar.gz
Algorithm Hash digest
SHA256 1a84c8a3d299c3b377ec739a089bf748913377c101c925a72fa98c6cbbf16b33
MD5 f460c1bfe401a2c2589a783c57417641
BLAKE2b-256 889e8928ea129e8de2c8ac349881d83a67cdb64ae774d2fb9bab2db11fb30d5c

See more details on using hashes here.

File details

Details for the file MLPet-0.0.7.1-py3-none-any.whl.

File metadata

  • Download URL: MLPet-0.0.7.1-py3-none-any.whl
  • Upload date:
  • Size: 22.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.8.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.9.7

File hashes

Hashes for MLPet-0.0.7.1-py3-none-any.whl
Algorithm Hash digest
SHA256 0c3288d0aabbc1a94e6054ab763f176a0e4c4834aa1a3bac24916eb52a913980
MD5 1e1a6b6dcc52c3844a25ff711c648238
BLAKE2b-256 459cbce89f15e8149d17182d8498bea39c5e60f3bce3cd1a5de0dbd371819a74

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page