Skip to main content

normalising Flow exoPlanet Parameter Inference Toolkyt

Project description

FlopPITy

normalizing Flow exoplanet Parameter Inference Toolkyt

FlopPITy allows the user to easily perform atmospheric retrievals using SNPE-C (citation) and neural spline flows (citation).

Installation guide

Currently FlopPITy doesn't work with python 3.13

$ conda create -n floppity_env python==3.12.9
$ conda activate floppity_env
$ pip install floppity

Basic usage:

  • First, import FlopPITy:
from floppity import Retrieval
from floppity.simulators import read_ARCiS_input, ARCiS
  • Now you can initialize the retrieval class with a simulator. A python wrapper for ARCiS comes built-in (you need to install ARCiS on your own tho):
R = Retrieval(ARCiS)
  • Read in observations and define parameters to retrieve:
R.get_obs(['path/to/obs_0', 'path/to/obs_1',..., 'path/to/obs_n'])
    
R.add_parameter(par_0, min, max)
R.add_parameter(par_1, min, max)
...
R.add_parameter(par_m, min, max)
  • For ARCiS, the observations and parameters can be read from the ARCiS input file:
pars, obs_list = read_ARCiS_input('path/to/ARCiS/input')
R.get_obs(obs_list)
R.parameters=pars
  • For retrievals using ARCiS, the input file and output directory need to be passed in a dictionary:
ARCiS_kwargs= dict(
                    ARCiS_dir = "/path/to/ARCiS/executable", #only needs to be set if ARCiS is not on the default path
                    input_file = arcis_input,
                    output_dir = 'path/to/output',
                  )
  • You can now run the retrieval, indicating the number of rounds and samples per round:
R.run_retrieval(n_rounds=10, n_samples=1000, simulator_kwargs=ARCiS_kwargs)
  • Great! You can now inspect your posterior:
fig = R.plot_corner()

Writing a simulator

Writing a simulator to work for FlopPITy is relatively straightforward. All that's needed is a function that takes in observations and parameters and returns spectra. The spectra need to be returned in a dictionary where each key represents each of the observations simulated (e.g. simulated[0] contains PRISM spectra and simulated[1] contains MIRI/LRS spectra):

def simulator(obs, parameters, **kwargs):
    wvl_0 = obs[0][:,0]
    wvl_1 = obs[1][:,0]
    ...
    wvl_n = obs[n][:,0]

    spectra={}
    spectra[0] = # array of shape (ndims, len(wvl_0))
    spectra[1] = # array of shape (ndims, len(wvl_1))
    ...
    spectra[n] = # array of shape (ndims, len(wvl_n))

    return spectra

Advanced options:

  • Additional post processing parameters (currently RV, vrot, offset and scaling) can be added, for example:
R.add_parameter('RV', -100, 100, post_process=True) # km/s

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

floppity-0.0.7.4.tar.gz (2.5 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

floppity-0.0.7.4-py3-none-any.whl (69.3 kB view details)

Uploaded Python 3

File details

Details for the file floppity-0.0.7.4.tar.gz.

File metadata

  • Download URL: floppity-0.0.7.4.tar.gz
  • Upload date:
  • Size: 2.5 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.7

File hashes

Hashes for floppity-0.0.7.4.tar.gz
Algorithm Hash digest
SHA256 f21c73a0a109a4b0f8bd8c4e2357df906e50751d458907a9a978d51466adfe43
MD5 222c9a58f507e457a24fb016e27e0fc1
BLAKE2b-256 b0ae911d328daa27fa562cba3c7dd164323d570f44b5eb5b21c2c04064f1b753

See more details on using hashes here.

File details

Details for the file floppity-0.0.7.4-py3-none-any.whl.

File metadata

  • Download URL: floppity-0.0.7.4-py3-none-any.whl
  • Upload date:
  • Size: 69.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.7

File hashes

Hashes for floppity-0.0.7.4-py3-none-any.whl
Algorithm Hash digest
SHA256 d074d0e86d2ebfe7c6d0e09e0e74f1f9fb2dba231334b287f156e42f8ef29d58
MD5 880e14e48dd11fb5da0d74a69f68631f
BLAKE2b-256 a50ae59d1bbfb47c97e8d5180b0c4d1d212d9a5461b64cdf2c04509d1fd7bab0

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page