Skip to main content

Plan observations with the Zwicky Transient Facility

Project description

planobs

Toolset for planning and triggering observations with ZTF. GCN parsing is currently only implemented for IceCube alerts.

It checks if the object is observable with a maximum airmass on a given date, plots the airmass vs. time, computes two optimal (minimal airmass at night) observations of 300s in g- and r and generate the ZTF field plots for all fields having a reference. There is also the option to create a longer (multiday) observation plan.

CI Coverage Status PyPI version DOI

Requirements

ztfquery for checking if fields have a reference.

planobs requires Python 3.10.

Installation

Using Pip: pip install planobs.

Otherwise, you can clone the repository: git clone https://github.com/simeonreusch/planobs, followed by poetry install This also gives you access to the Slackbot.

Note for ARM-based macs: The install of fiona might fail if you do not have gdal installed. In that case, consider using a conda and running conda install -c conda-forge gdal before running poetry install.

General usage

from planobs.plan import PlanObservation

name = "testalert" # Name of the alert object
date = "2020-05-05" #This is optional, defaults to today
ra = 133.7
dec = 13.37

plan = PlanObservation(name=name, date=date, ra=ra, dec=dec)
plan.plot_target() # Plots the observing conditions
plan.request_ztf_fields() # Checks in which ZTF fields this 
# object is observable and generates plots for them.

The observation plot and the ZTF field plots will be located in the current directory/[name]

Note: Checking if fields have references requires ztfquery, which needs IPAC credentials.

Usage for IceCube alerts

from planobs.plan import PlanObservation

name = "IC201007A" # Name of the alert object
date = "2020-10-08" #This is optional, defaults to today

# No RA and Dec values are given, because we set alertsource to icecube, which leads to automatic GCN parsing.

plan = PlanObservation(name=name, date=date, alertsource="icecube")
plan.plot_target() # Plots the observing conditions.
plan.request_ztf_fields() # Checks which ZTF fields cover the target (and have references).
print(plan.recommended_field) # This give you the field with the most overlap.

Triggering ZTF

planobs can be used to schedule ToO observations with ZTF. This is done through API calls to the Kowalski system, managed by the Kowalski Python API penquins.

To use this functionality, you must first configure the connection details. You need both an API token, and to know the address of the Kowalski host address. You can then set these as environment variables:

export KOWALSKI_HOST=something
export KOWALSKI_API_TOKEN=somethingelse

You can then import the Queue class for querying, submitting and deleting ToO triggers:

Querying

from planobs.api import Queue

q = Queue(user="yourname")

existing_too_requests = get_too_queues(names_only=True)
print(existing_too_requests)

Submitting

from planobs.api import Queue, get_too_queues
from planobs.models import TooTarget

trigger_name = "ToO_IC220513A_test"

# Instantiate the API connection
q = Queue(user="yourname")

# Add a trigger to the internal submission queue. Filter ID is 1 for r-, 2 for g- and 3 for i-band. Exposure time is given in seconds.
q.add_trigger_to_queue(
    trigger_name=trigger_name,
    validity_window_start_mjd=59719.309333333334,
    targets=[
        TooTarget(
            field_id=427,
            filter_id=1,
            exposure_time=300,
        )
    ],
)

q.submit_queue()

# Now we verify that our trigger has been successfully submitted
existing_too_requests = get_too_queues(names_only=True)
print(existing_too_requests)
assert trigger_name in existing_too_requests

Deleting

from planobs.api import Queue

q = Queue(user="yourname")

trigger_name = "ToO_IC220513A_test"

res = q.delete_trigger(trigger_name=trigger_name)

Citing the code

If you use this code, please cite it! A DOI is provided by Zenodo, which can reference both the code repository and specific releases:

DOI

Contributors

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

planobs-0.7.2.tar.gz (17.9 MB view details)

Uploaded Source

Built Distribution

planobs-0.7.2-py3-none-any.whl (18.5 MB view details)

Uploaded Python 3

File details

Details for the file planobs-0.7.2.tar.gz.

File metadata

  • Download URL: planobs-0.7.2.tar.gz
  • Upload date:
  • Size: 17.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.5.1 CPython/3.11.3 Linux/5.15.0-1038-azure

File hashes

Hashes for planobs-0.7.2.tar.gz
Algorithm Hash digest
SHA256 9ece2ddac0d1a26269f4f8688b4ddd7b101b2eceadc1f09883c859dc9af789a4
MD5 629e3b5c4901b95d1438379e83d30a40
BLAKE2b-256 332a36d6b09c1a6a13804aa87f7dfd4b6c876e5de43c01c0683afd0dd520cc15

See more details on using hashes here.

File details

Details for the file planobs-0.7.2-py3-none-any.whl.

File metadata

  • Download URL: planobs-0.7.2-py3-none-any.whl
  • Upload date:
  • Size: 18.5 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.5.1 CPython/3.11.3 Linux/5.15.0-1038-azure

File hashes

Hashes for planobs-0.7.2-py3-none-any.whl
Algorithm Hash digest
SHA256 bc2685a55e168fcec5f6f837f06c069417313a236644f4182c1df18ee6594ace
MD5 3bcbfe7fa819f89bfcbf2ac29fba35c0
BLAKE2b-256 8a1d1d8907b0dd71232e4dfa4afca80d1fc3c87707140054194f372dec856170

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page