For reading feeds of sports and activity data published in the OpenActive format
Project description
This is a Python package for reading feeds of sports and activity data published in the OpenActive format. Note that this is an experimental package, and although care has been taken to cater for many typical OpenActive situations, there may still be those that don't work as expected. It is therefore recommended to not directly use this package for critical pipelines or automations, but to treat it as a toolset for exploration and education. The underlying code is relatively short and can be consulted for informing your own approach if needed.
Installation
It is recommended to first set up a virtual environment for your openactive
project, to ensure that it's isolated from your base environment and runs as intended. The only thing that must be installed in your base environment is some package for generating virtual environments, such as virtualenv
:
$ pip install virtualenv
Then, in a new project folder, create and initialise a new virtual environment as follows (virt
and venv
are typical names, but you can choose something else if needed):
$ virtualenv virt
$ source virt/bin/activate
(virt) $
Now install the openactive
package in the virtual environment, and you're good to go:
(virt) $ pip install openactive
When you're done working in the virtual environment, deactivate it by:
(virt) $ deactivate
Usage
In a Python session running in an environment with the openactive
package installed, let's import the package under the proxy name oa
:
>>> import openactive as oa
In order to effectively use the package, we must first understand the OpenActive ecosystem. OpenActive data is decentralised, so there is no single owner or location that stores and serves the data. Instead, data is released by multiple data publishers as separate Realtime Paged Data Exchange (RPDE) feeds, which are described in more detail later. There can also be multiple feeds from a given publisher, and in fact we often have complimentary pairs of feeds, such as having one for super-event data (e.g. various series of fitness classes) and one for sub-event data (e.g. various sessions in the various series). In such cases, both feeds must be read in order to get a complete picture, and items in one feed will reference items in the other feed. The alternative to this would be to copy each super-event data item into each associated sub-event data item, resulting in one feed with a lot of duplication.
A group of feeds from a data publisher is bundled together in a "dataset", a group of datasets from different data publishers is bundled together in a "catalogue", and a group of catalogues is bundled together in a "collection". There is only one collection, which is therefore the starting point for everything else. Given a list of all feed information, you will not often want to see the exact path by which the information was gathered, but there are functions in the openactive
package that break down the journey from the source collection if needed. So let's just start at the very beginning to be clear on how things work. First, let's define a printer function to give us a clear indented output display for what follows:
>>> import json
>>> def printer(arg):
... print(json.dumps(arg,indent=4))
Get feeds
Now let's get the catalogue URLs in the collection, which should take about a second:
>>> catalogue_urls = oa.get_catalogue_urls()
>>> printer(catalogue_urls)
{
"https://openactive.io/data-catalogs/data-catalog-collection.jsonld": [
"https://opendata.leisurecloud.live/api/datacatalog",
"https://openactivedatacatalog.legendonlineservices.co.uk/api/DataCatalog",
"https://openactive.io/data-catalogs/singular.jsonld",
"https://app.bookteq.com/api/openactive/catalogue"
]
}
We see that this returns a dictionary with a single key, the collection URL, which has a value that is the list of catalogue URLs. Unless otherwise stated, all data gathering functions have two optional boolean keywords which weren't used above. The first keyword is flat
, which causes a function to return a flat list structure rather than a dictionary, so losing the key-level information. The second keyword is verbose
, which causes a function to print its name and the URLs that it calls during execution. Note that, regardless of the verbose
keyword, warning and error messages are printed as standard, and we typically have such messages when a page to be read is unavailable or not set up correctly. Let's run the above again with both keywords set to True
:
>>> catalogue_urls = oa.get_catalogue_urls(flat=True, verbose=True)
get_catalogue_urls
CALLING: https://openactive.io/data-catalogs/data-catalog-collection.jsonld
>>> printer(catalogue_urls)
[
"https://opendata.leisurecloud.live/api/datacatalog",
"https://openactivedatacatalog.legendonlineservices.co.uk/api/DataCatalog",
"https://openactive.io/data-catalogs/singular.jsonld",
"https://app.bookteq.com/api/openactive/catalogue"
]
Now for each of these catalogue URLs let's get the dataset URLs they contain, which should take a few seconds. Note that get_dataset_urls
calls get_catalogue_urls
internally, the above was just for illustration of the process:
Note: Output is truncated at 'etc.'
>>> dataset_urls = oa.get_dataset_urls()
>>> printer(dataset_urls)
{
"https://opendata.leisurecloud.live/api/datacatalog": [
"https://activeleeds-oa.leisurecloud.net/OpenActive/",
etc.
],
"https://openactivedatacatalog.legendonlineservices.co.uk/api/DataCatalog": [
"https://halo-openactive.legendonlineservices.co.uk/OpenActive",
etc.
],
"https://openactive.io/data-catalogs/singular.jsonld": [
"http://data.better.org.uk/",
etc.
],
"https://app.bookteq.com/api/openactive/catalogue": [
"https://actihire.bookteq.com/api/open-active",
etc.
]
}
We again see an output dictionary, with keys that are catalogue URLs and values that are lists of dataset URLs. The above output display is truncated, and you will see many more dataset URLs if you run the command yourself.
Now for each of these dataset URLs let's get the feed information they contain, which should take a couple of minutes. Note that get_feeds
calls get_dataset_urls
internally, the above was just for illustration of the process:
Note: Output is truncated at 'etc.'
>>> feeds = oa.get_feeds()
WARNING: Retrying (1/9): https://gll-openactive.legendonlineservices.co.uk/OpenActive
WARNING: Retrying (1/9): https://sllandinspireall-openactive.legendonlineservices.co.uk/OpenActive
WARNING: Retrying (1/9): https://data.bookwhen.com/
WARNING: Retrying (2/9): https://data.bookwhen.com/
WARNING: Retrying (3/9): https://data.bookwhen.com/
WARNING: Retrying (4/9): https://data.bookwhen.com/
WARNING: Retrying (5/9): https://data.bookwhen.com/
WARNING: Retrying (6/9): https://data.bookwhen.com/
WARNING: Retrying (7/9): https://data.bookwhen.com/
WARNING: Retrying (8/9): https://data.bookwhen.com/
WARNING: Retrying (9/9): https://data.bookwhen.com/
WARNING: Max. tries (10) reached for: https://data.bookwhen.com/
ERROR: Can't get dataset: https://data.bookwhen.com/
ERROR: Can't get dataset: https://www.participant.co.uk/participant/openactive/
>>> printer(feeds)
{
"https://activeleeds-oa.leisurecloud.net/OpenActive/": [
{
"name": "Active Leeds Sessions and Facilities",
"type": "CourseInstance",
"url": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-course-instance",
"datasetUrl": "https://activeleeds-oa.leisurecloud.net/OpenActive/",
"discussionUrl": "",
"licenseUrl": "https://creativecommons.org/licenses/by/4.0/",
"publisherName": "Active Leeds"
},
etc.
],
"https://brimhamsactive.gs-signature.cloud/OpenActive/": [
{
"name": "Brimhams Active Sessions and Facilities",
"type": "CourseInstance",
"url": "https://opendata.leisurecloud.live/api/feeds/BrimhamsActive-live-course-instance",
"datasetUrl": "https://brimhamsactive.gs-signature.cloud/OpenActive/",
"discussionUrl": "",
"licenseUrl": "https://creativecommons.org/licenses/by/4.0/",
"publisherName": "Brimhams Active"
},
etc.
],
etc.
}
Once again we see an output dictionary, with keys that are dataset URLs and values that are lists of feed information dictionaries. The above output display is truncated, and you will see many more feed information dictionaries if you run the command yourself. Also note the warning messages in the above, which, as mentioned previously, occur if a URL cannot be correctly accessed due to some issue. The openactive
package will automatically retry a problematic URL up to 10 times with a 1 second wait time between tries, and these defaults can be overridden via the num_tries_max
and seconds_wait_retry
keywords, respectively. Additionally there is a default wait time of 0.2 seconds between subsequent URLs being called in a list even if no issue occurs, to ensure that servers aren't overburdened with an unrestricted call rate, and this default can be overridden via the seconds_wait_next
keyword. Any of these keywords can be given to any data gathering function and they will be passed through to other data gathering functions they call internally. So, for example, the get_feeds
function could be called as follows, which will adjust the settings not only for the get_feeds
function itself but also for the get_dataset_urls
function and the get_catalogue_urls
function which are internally called in a chain:
>>> feeds = oa.get_feeds(seconds_wait_next=0.4, seconds_wait_retry=2, num_tries_max=5)
The list of all feed information is usually where you'll want to start your project work, but it's useful to be aware of the full journey illustrated above in getting to this point.
Get opportunities
What we ultimately want to work with is the activity "opportunity" data served via a given feed starting URL, which is the entry point for data transferred via Realtime Paged Data Exchange (RPDE). In essence, this is just like what we have returned from a search engine, which breaks result items over a chain of pages rather than showing them all on a single page. As described earlier, feeds often come in pairs for efficiency of data storage and delivery, with one feed for super-event data and one feed for sub-event data. Inspecting the full output from the get_feeds
function, we have the following for the first dataset:
Note: Output is truncated at 'etc.'
>>> printer(feeds)
{
"https://activeleeds-oa.leisurecloud.net/OpenActive/": [
{
"name": "Active Leeds Sessions and Facilities",
"type": "CourseInstance",
"url": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-course-instance",
"datasetUrl": "https://activeleeds-oa.leisurecloud.net/OpenActive/",
"discussionUrl": "",
"licenseUrl": "https://creativecommons.org/licenses/by/4.0/",
"publisherName": "Active Leeds"
},
{
"name": "Active Leeds Sessions and Facilities",
"type": "SessionSeries",
"url": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series",
"datasetUrl": "https://activeleeds-oa.leisurecloud.net/OpenActive/",
"discussionUrl": "",
"licenseUrl": "https://creativecommons.org/licenses/by/4.0/",
"publisherName": "Active Leeds"
},
{
"name": "Active Leeds Sessions and Facilities",
"type": "ScheduledSession",
"url": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions",
"datasetUrl": "https://activeleeds-oa.leisurecloud.net/OpenActive/",
"discussionUrl": "",
"licenseUrl": "https://creativecommons.org/licenses/by/4.0/",
"publisherName": "Active Leeds"
},
{
"name": "Active Leeds Sessions and Facilities",
"type": "FacilityUse",
"url": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-facility-uses",
"datasetUrl": "https://activeleeds-oa.leisurecloud.net/OpenActive/",
"discussionUrl": "",
"licenseUrl": "https://creativecommons.org/licenses/by/4.0/",
"publisherName": "Active Leeds"
},
{
"name": "Active Leeds Sessions and Facilities",
"type": "Slot",
"url": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-slots",
"datasetUrl": "https://activeleeds-oa.leisurecloud.net/OpenActive/",
"discussionUrl": "",
"licenseUrl": "https://creativecommons.org/licenses/by/4.0/",
"publisherName": "Active Leeds"
}
],
etc.
}
Extracting the feed starting URLs for this dataset, we have:
>>> feed_urls = [feed['url'] for feed in feeds['https://activeleeds-oa.leisurecloud.net/OpenActive/']]
>>> printer(feed_urls)
[
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-course-instance",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-facility-uses",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-slots"
]
To help automate workflows, there is a function called get_partner_feed_url
that takes a single feed starting URL to find a partner for, and a list of feed starting URLs in which there may be a partner. It's a simple search-and-replace function using typical URL parts and their variants, such as swapping session-series
or sessionseries
for scheduled-sessions
or scheduledsessions
, until a match is found. If no match is found, then None
is returned instead. Looping through the list of feed starting URLs for the above dataset, and using the list itself as the set of matching options, we have:
>>> for feed_url in feed_urls:
... partner_feed_url = oa.get_partner_feed_url(feed_url, feed_urls)
... print('Feed-1 URL:', feed_url)
... print('Feed-2 URL:', partner_feed_url, '\n')
...
Feed-1 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-course-instance
Feed-2 URL: None
Feed-1 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series
Feed-2 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions
Feed-1 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions
Feed-2 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series
Feed-1 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-facility-uses
Feed-2 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-slots
Feed-1 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-slots
Feed-2 URL: https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-facility-uses
The course-instance
feed is the only standalone feed in this case, with the remaining four feeds being two pairs, namely the session-series
super-event feed with the scheduled-sessions
sub-event feed, and the facility-uses
super-event feed with the slots
sub-event feed. Note that it doesn't matter if the single feed starting URL provided to get_partner_feed_url
is for a super-event feed or a sub-event feed, it will find whatever partner URL matches.
Let's look at the paired session-series
and scheduled-sessions
feeds from the above dataset. As mentioned, a feed consists of a list of opportunity items split over a number of pages. Some items are in fact not live and are marked as deleted, and some items may supersede other items with more up-to-date values. To get all of the data for a given feed, we must visit each page one-by-one and retain only the most up-to-date live items, which is done by the get_opportunities
function. Note that the number of pages in a given feed is not known in advance, and so the time required to read all associated pages can vary greatly between one feed and another, from a number of seconds to a number of minutes. The verbose
keyword may be particularly useful here to monitor progress. The feeds in this example should only take a few seconds each:
>>> superevent_opportunities = oa.get_opportunities('https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series')
>>> subevent_opportunities = oa.get_opportunities('https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions')
For the super-event data we have:
Note: Output is truncated at 'etc.'
>>> printer(superevent_opportunities)
{
"items": {
"HO1ONDL23501021": {
"id": "HO1ONDL23501021",
"modified": 14554552,
"kind": "SessionSeries",
"state": "updated",
"data": {
"@context": [
"https://openactive.io/",
"https://openactive.io/ns-beta"
],
"@type": "SessionSeries",
"@id": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/session-series/HO1ONDL23501021",
etc.
}
},
etc.
},
"urls": [
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series?afterTimestamp=24571209&afterId=SH5CLPI13300124"
],
"firstUrlOrigin": "https://opendata.leisurecloud.live",
"nextUrl": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-session-series?afterTimestamp=26002956&afterId=KL2CLPL11001121"
}
For the sub-event data we have:
Note: Output is truncated at 'etc.'
>>> printer(subevent_opportunities)
{
"items": {
"00000000000170005743": {
"id": "00000000000170005743",
"modified": 34213402,
"kind": "ScheduledSession",
"state": "updated",
"data": {
"@context": [
"https://openactive.io/",
"https://openactive.io/ns-beta"
],
"@type": "ScheduledSession",
"@id": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/scheduled-sessions/170005743",
etc.
}
},
etc.
}
"urls": [
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions?afterTimestamp=34383479&afterId=00000000000060031844",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions?afterTimestamp=34385051&afterId=00000000000130035754",
"https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions?afterTimestamp=34386561&afterId=00000000000010053619"
],
"firstUrlOrigin": "https://opendata.leisurecloud.live",
"nextUrl": "https://opendata.leisurecloud.live/api/feeds/ActiveLeeds-live-scheduled-sessions?afterTimestamp=34387745&afterId=00000000000080031177"
}
The returned outputs are, once again, in dictionary form, and the main content of interest is found under the items
key. The above output display is truncated, and you will see many more items if you run the command yourself. This output cannot be flattened via the flat
keyword, as its structure is essential to maintain for potential later use. All URLs that were visited in the feed chain are also returned in the output, as well as the first URL "origin" component, and the next URL to be visited when the feed is updated by the publisher, in order to continue the read at this point in the feed chain at a later time. To do this, which can also be done if we encounter an issue and only receive output from a partial read of the feed chain, we give the output dictionary to the function as argument rather than the feed starting URL. Using the super-event feed to illustrate, we would do:
>>> superevent_opportunities_new = oa.get_opportunities(superevent_opportunities)
Assess opportunities
After obtaining a set of opportunity items, we can scan through all of them and count the various "kind" and "type" values that appear, using the functions get_item_kinds
and get_item_data_types
. Usually there is only one kind and one type for a given feed, and usually these are the same as each other too, but there can be differences and this is useful to keep in mind. Let's take a look for the opportunities obtained above.
For the super-event data we have:
>>> len(superevent_opportunities['items'].keys())
916
>>> printer(oa.get_item_kinds(superevent_opportunities))
{
"SessionSeries": 916
}
>>> printer(oa.get_item_data_types(superevent_opportunities))
{
"SessionSeries": 916
}
For the sub-event data we have:
>>> len(subevent_opportunities['items'].keys())
1476
>>> printer(oa.get_item_kinds(subevent_opportunities))
{
"ScheduledSession": 1476
}
>>> printer(oa.get_item_data_types(subevent_opportunities))
{
"ScheduledSession": 1476
}
In this case, all 916 super-event items have a kind and a type of SessionSeries
, and all 1476 sub-event items have a kind and a type of ScheduledSession
. So it's safe to say that these are pure feeds of only one of the OpenActive data variants each, and we can treat them as such in further analysis. Note that you will likely see different values to the above if you're following through, as OpenActive feeds are dynamically changing.
Even though we have commented that the SessionSeries
feed is super-event data and the ScheduledSession
feed is sub-event data, these judgements were made simply by observation and knowledge of the OpenActive data model. In order to automate an analysis that relies on knowing whether we're dealing with super-event data or sub-event data, the get_event_type
function is useful. This takes an item kind or type label and returns 'superevent'
, 'subevent'
or None
, accordingly:
>>> oa.get_event_type('SessionSeries')
'superevent'
>>> oa.get_event_type('ScheduledSession')
'subevent'
>>> oa.get_event_type('blah')
>>> oa.get_event_type('blah') is None
True
Finally, let's look at the relationship between the items in a pair of super-event and sub-event feeds. For any single sub-event item, there will be a single related super-event item i.e. a single class belongs to a single series of classes. For any single super-event item, there will be a number of related sub-event items i.e. a single series of classes has a number of classes. The functions get_superevents
and get_subevents
help with these two situations, respectively. Let's take a look at each in turn. First, let's observe a single sub-event item from the full list of sub-event items:
>>> printer(list(subevent_opportunities['items'].values())[0])
{
"id": "00000000000170005743",
"modified": 34213402,
"kind": "ScheduledSession",
"state": "updated",
"data": {
"@context": [
"https://openactive.io/",
"https://openactive.io/ns-beta"
],
"@type": "ScheduledSession",
"@id": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/scheduled-sessions/170005743",
"startDate": "2024-02-23T10:00:00+00:00",
"identifier": 170005743,
"endDate": "2024-02-23T11:00:00+00:00",
"superEvent": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/session-series/WE5CLKM10000723",
"duration": "PT1H",
"maximumAttendeeCapacity": 20,
"remainingAttendeeCapacity": 0,
"beta:sportsActivityLocation": [
{
"@type": "SportsActivityLocation",
"name": "Activity Room"
}
]
}
}
Note that there is a field called "superEvent" in the above, the value of which contains what looks like a unique ID, namely "WE5CLKM10000723". We can see that this exists within the full list of super-event items:
>>> 'WE5CLKM10000723' in superevent_opportunities['items'].keys()
True
Now let's select that super-event item itself, by giving the sub-event item and the full list of super-event items to get_superevents
:
Note: Output is truncated at 'etc.'
>>> superevent_opportunities_selection = oa.get_superevents(
... list(subevent_opportunities['items'].values())[0],
... superevent_opportunities
... )
>>> printer(superevent_opportunities_selection)
[
{
"id": "WE5CLKM10000723",
"modified": 34053122,
"kind": "SessionSeries",
"state": "updated",
"data": {
"@context": [
"https://openactive.io/",
"https://openactive.io/ns-beta"
],
"@type": "SessionSeries",
"@id": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/session-series/WE5CLKM10000723",
"eventSchedule": [
{
"@type": "PartialSchedule",
"byDay": [
"https://schema.org/Friday"
],
"duration": "PT1H",
"endTime": "11:00",
"startDate": "2023-07-21",
"endDate": "2025-02-14",
"startTime": "10:00"
}
],
"identifier": "WE5CLKM10000723",
"name": "Keep Moving",
"attendeeInstructions": "Low level, low impact exercise for people with health conditions that want to become more active and improve their wellbeing",
etc.
}
}
]
Note that the super-event ID we identified in the sub-event item appears in both the "id" and "data:@id" fields of the super-event item itself. The unique part of "id" and "data:@id" should be the same for a given super-event item, and all super-event items should be different in this respect, so there should be only one super-event for a given sub-event, as mentioned. However, hypothetically this may not always be so in spurious cases, and the output from get_superevents
is a list which could contain more than one item if multiple matches are indeed found. It is worth checking that there is only one item in the list before any further analysis is done, and, if not, then establishing exactly why not.
Now let's go the other way around, from a given super-event item to a group of sub-event items. We can use the single super-event item as found above:
>>> subevent_opportunities_selection = oa.get_subevents(
... superevent_opportunities_selection[0],
... subevent_opportunities
... )
>>> printer(subevent_opportunities_selection)
[
{
"id": "00000000000170005743",
"modified": 34213402,
"kind": "ScheduledSession",
"state": "updated",
"data": {
"@context": [
"https://openactive.io/",
"https://openactive.io/ns-beta"
],
"@type": "ScheduledSession",
"@id": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/scheduled-sessions/170005743",
"startDate": "2024-02-23T10:00:00+00:00",
"identifier": 170005743,
"endDate": "2024-02-23T11:00:00+00:00",
"superEvent": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/session-series/WE5CLKM10000723",
"duration": "PT1H",
"maximumAttendeeCapacity": 20,
"remainingAttendeeCapacity": 0,
"beta:sportsActivityLocation": [
{
"@type": "SportsActivityLocation",
"name": "Activity Room"
}
]
}
},
{
"id": "00000000000170005744",
"modified": 34365346,
"kind": "ScheduledSession",
"state": "updated",
"data": {
"@context": [
"https://openactive.io/",
"https://openactive.io/ns-beta"
],
"@type": "ScheduledSession",
"@id": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/scheduled-sessions/170005744",
"startDate": "2024-03-01T10:00:00+00:00",
"identifier": 170005744,
"endDate": "2024-03-01T11:00:00+00:00",
"superEvent": "https://activeleeds-oa.leisurecloud.net/OpenActive/api/session-series/WE5CLKM10000723",
"duration": "PT1H",
"maximumAttendeeCapacity": 20,
"remainingAttendeeCapacity": 10,
"beta:sportsActivityLocation": [
{
"@type": "SportsActivityLocation",
"name": "Activity Room"
}
]
}
}
]
We have a list of two sub-event items, one of which was the original input to get_superevents
, as expected, and the other is the only other sub-event in the same super-event series. Such a list could contain many more items, but the above provides a complete and compact full example.
The functions for assessing opportunities described in this section are a good starting point for further analysis, providing the user with a set of tools to find basic but important information about feed content. The user can then more confidently work with that content, including automations that may depend on knowing the item kind or type, and whether or not those labels indicate a pure super-event feed or sub-event feed.
Function summary
The following table summarises the inputs and outputs of all functions described above:
Function | Input | Output (not using flat ) |
---|---|---|
get_catalogue_urls |
- | dict: catalogue URLs in the collection |
get_dataset_urls |
- | dict: dataset URLs for each catalogue |
get_feeds |
- | dict: feed info for each dataset |
get_partner_feed_url |
str: feed1_url and [str]: [ feed2_url_options ] |
str: feed2_url that best partners with feed1_url |
get_opportunities |
str: feed_url or dict: opportunities |
dict: opportunities info for a given feed |
get_item_kinds |
dict: opportunities |
dict: Item kinds and counts |
get_item_data_types |
dict: opportunities |
dict: Item data types and counts |
get_event_type |
str: item kind or str: item data type |
str: "superevent" or "subevent" or None |
get_superevents |
dict: sub-event item and dict: super-event opportunities |
[dict]: [super-event items] |
get_subevents |
dict: super-event item and dict: sub-event opportunities |
[dict]: [sub-event items] |
Additionally, the data gathering functions accept the following keywords:
Function | flat bool |
verbose bool |
seconds_wait_next num |
seconds_wait_retry num |
num_tries_max int |
---|---|---|---|---|---|
get_catalogue_urls |
✓ | ✓ | ✗ | ✓ | ✓ |
get_dataset_urls |
✓ | ✓ | ✓ | ✓ | ✓ |
get_feeds |
✓ | ✓ | ✓ | ✓ | ✓ |
get_opportunities |
✗ | ✓ | ✓ | ✓ | ✓ |
References
The main locations:
The complete set of OpenActive specifications:
- Realtime Paged Data Exchange (RPDE) data transfer protocol
- Opportunity data primer
- Opportunity data model
- Dataset model
- Route model
- Booking system model
Tools:
- Feed status
- Data visualiser - for those curious about the data and for data publishers checking their feed quality
- Data validator - a more involved tool for drilling into feed details and checking content
Community and communications:
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file openactive-2.0.0.tar.gz
.
File metadata
- Download URL: openactive-2.0.0.tar.gz
- Upload date:
- Size: 30.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.8.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 57ec1248a6b21a4663f00f9c5f6a4279d028426b34732d13b84be57db63196fd |
|
MD5 | 88507c4c349bd8cecb45cca768867632 |
|
BLAKE2b-256 | afc27352d5dc8ccecd2f4cc5c8d1132aebad83de6cee3482d75d4cf9b39943ac |
File details
Details for the file openactive-2.0.0-py3-none-any.whl
.
File metadata
- Download URL: openactive-2.0.0-py3-none-any.whl
- Upload date:
- Size: 15.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.8.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 63ca1e8e9b23bef40ed40031c9d09b376665177427e27e29ca54815b4d3db79d |
|
MD5 | e3da74b6b4798abb0e5f1add2c8fc591 |
|
BLAKE2b-256 | f42ffb95885eb0be580e2d4348804118ed98e63d2f29da31d79e8517ad8a49ce |