Skip to main content

A package to extract data from intervals.icu to influxDB

Project description

Intervals.icu to InfluxDB

Intervalsicu-to-influxdb is a personal project to extract data from Intervals.icu to InfluxDB (oh, really?). But if Intervals.icu already shows a lot of graphics, statistics and more, why I need to extract it?

Full documentation can be found in here

Why

Well, as a sportsman and techie, it's just a personal project, but the main reason is because I want to create my own dashboards (using Grafana in this case).

So, for example, I can combine activity data with sleep time or quality, compare the evolution between pace/bpm for the similar activities or whatever.

Grafana Dashboard example Grafana Dashboard example2

How it works

This project exports some data from intervals.icu to influxDB. To retrieve the information the official intervals.icu API is used.

Exported data

Not all information is exported. This project has been created to extract data from activities and wellness. Besides, information about data account (like email, location, preferences, etc.), calendar or workouts are not retrieved neither (for now).

Currently the following data is exported:

  • Wellness*: this data contains information like sleep time and quality, atl/ctl or VO2Max
  • Activities*: general information about every activity, like elapsed time, time in zones (hr or pace), distance, average pace/hr, etc.
  • Streams**: streams contains detailed information about activities, like hr/pace for every second.

* There are som extra fields generated, just to facilitate the use for the dashboards (see Entities)

** Currently working on it.

How to use

There are 2 ways (3 if you count 'from source code') to use the project: with Docker or directly with Python (or from source code), but in both cases you need to create a .env file to save your credentials for Intervals.icu and InfluxDB as follow:

INFLUXDB_TOKEN=
INFLUXDB_ORG=
INFLUXDB_URL=
INFLUXDB_BUCKET=
INFLUXDB_TIMEOUT=10000
INTERVALS_ATHLETE_ID=
INTERVALS_API_KEY=

With Python

If you want to run it directly with Python, first install the dependency:

pip install intervalsicu-to-influxdb

Then, the minimum code to run it is (remember to put the .env file on the same folder):

from intervalsicu_to_influxdb.extractor import IntervalsToInflux

extractor = IntervalsToInflux()
extractor.all_data()

To run it, just save as app.py and run it:

python app.py

Arguments

All the arguments are optional, but take in consideration the following variations when run it:

  • No arguments: retrieve the wellness and activities data for today (this is the basic use to run with a cronjob)
  • date: retrieve data for the defined date (in format YYYY-MM-DD)
  • start-date: retrieve data from the starting date (in format YYYY-MM-DD) until today
  • end-date: retrieve data until specified date (in format YYYY-MM-DD). Use it with start-date
  • streams: retrieve the streams for the activities
  • reset: delete the current bucket and recreate again

NOTE: on the first run, the bucket is created automatically if not exists on InfluxDB

extractor = IntervalsToInflux(start_date="2023-01-01")
extractor = IntervalsToInflux(streams=True)
extractor = IntervalsToInflux(start_date="2023-01-01", end_date="2023-05-01")

Dynamic script

If you want to create a more dynamic script, here is a more complete example:

import argparse

from intervalsicu_to_influxdb.extractor import IntervalsToInflux

parser = argparse.ArgumentParser()

parser.add_argument("--date", type=str, help="Date in format YYYY-MM-DD")
parser.add_argument("--start-date", type=str, help="Start date in format YYYY-MM-DD")
parser.add_argument("--end-date", type=str, help="End date in format YYYY-MM-DD")
parser.add_argument(
    "--streams",
    action="store_true",
    help="Export streams for the activities",
)
parser.add_argument(
    "--reset", action="store_true", help="Reset influx bucket (delete and create)"
)

args = parser.parse_args()

if args.date:
    date = args.date
else:
    date = None
if args.start_date:
    start_date = args.start_date
else:
    start_date = None
if args.end_date:
    end_date = args.end_date
else:
    end_date = None
if args.streams:
    streams = True
else:
    streams = False
if args.reset:
    reset = True
else:
    reset = False

extractor = IntervalsToInflux(date, start_date, end_date, reset, streams)
extractor.all_data()

Then, just run the script as before, but you will can use arguments (same as the Docker section):

python app.py [-h] [--date DATE] [--start-date START_DATE] [--end-date END_DATE] [--streams] [--reset]

From source code

If you want to run it from source code, just clone the project, and the follow the next steps (remember to create the .env file):

Run with Docker

First, compile the image

docker build --tag intervals-to-influxdb .

And then, just run it like the Docker section above (but with the image name)

docker run --env-file PATH/TO/FILE -it --rm intervals-to-influxdb app.py [-h] [--date DATE] [--start-date START_DATE] [--end-date END_DATE] [--streams] [--reset]

Run with Python

First, install dependencies from source

pip install .

And then, run the script

python app.py [-h] [--date DATE] [--start-date START_DATE] [--end-date END_DATE] [--streams] [--reset]

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

intervalsicu_to_influxdb-0.2.5.tar.gz (24.5 kB view details)

Uploaded Source

Built Distribution

intervalsicu_to_influxdb-0.2.5-py3-none-any.whl (24.4 kB view details)

Uploaded Python 3

File details

Details for the file intervalsicu_to_influxdb-0.2.5.tar.gz.

File metadata

File hashes

Hashes for intervalsicu_to_influxdb-0.2.5.tar.gz
Algorithm Hash digest
SHA256 cced3889c740d9f0d8f55de040962d264e44f7390f0b165f36bb52fcbd6037c6
MD5 37d335e920d073d920d1b6e848977aac
BLAKE2b-256 86c3dc20db4dc2cabc0094939092d525e3749dab4c481e993b57f7f09e05566b

See more details on using hashes here.

Provenance

File details

Details for the file intervalsicu_to_influxdb-0.2.5-py3-none-any.whl.

File metadata

File hashes

Hashes for intervalsicu_to_influxdb-0.2.5-py3-none-any.whl
Algorithm Hash digest
SHA256 eb54d3d9add7c21d92ed2100ddd7d01b513f40a003653465498273a1d3efb4fd
MD5 c2c1ba828e833d28f12d184a27053796
BLAKE2b-256 f83f90f3edb412d5c52221d8d2bb3d6f6c8b444b717696fbd84a10a5a54a4dd7

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page