Skip to main content

Modern Python interface to the exact online API

Project description

ExactPy

A modern, highly configurable Python interface to the Exact Online API based on Pydantic and httpx and integrates with sparkdantic to make it easy to convert your data to (py)spark dataframes.

For now, this package does not provide any controller methods POST and PUT calls; it's basically read-only. This will be added later.

Installing

You can install directly from PyPI using e.g.

uv add exactpy

or

pip install exactpy

or you can install locally from file using:

uv pip install -e <location_on_disk>

Developing

This repo provides a devcontainer to easily develop the package. It assumes you have uv and fish installed on your host system. This cannot be made optional because these assumptions involve bind mounts.

Install the package locally (symlinked) within the venv created by uv using:

uv pip install -e .

Pre-commit

We use pre-commit to run a few hooks to ensure code quality and stuff like that. Please use the devcontainer, all this is taken care for you if you do so.

Running tests

We use tox for bothing running the tests as well as running pre-commit. Please run tox to run your tests.

Contributing

Please see CONTRIBUTING.md.

Issues

If you discover bugs or other issues, please create an issue with a stack trace and code to reproduce. We have no predefined format for issues, just make sure there is enough info to reproduce.

Why?

Currently available packages aren't configurable in such a way that they are usable to me. This package attempts to fix that.

Good to know

Spark and context aware serialization

This package provides support for conversion to Spark dataframes indirectly, by using SparkModel instead of BaseModel, Spark schemas can be generated from models (see also: https://github.com/mitchelllisle/sparkdantic) by callign e.g. exactpy.models.financial.GLAccountModel.model_spark_schema().

The types used in OData are a bit unusual in some cases. The timestamp type is a string with a timestamp in millisecond precision embedded into it, for instance.

For this reason, serialization output can be changed to be (among other things) (py)spark and pandas compatible. This can be done by passing a context to the serialization methods. One example:

gl_account_instance.model_dump(by_alias=False, context={"output": "spark"})

For now, this does not change much (only timestamps are serialized as datetime.datetime instead of timestamp strings), but this might change in the future. Just pass context={"output": "spark"} in your serialization methods, and you're good.

Model field naming

This package uses Pydantic's BaseModel (well, actually sparkdantic.SparkModel, but that is derived from BaseModel) as model base classes. Field names do not correspond exactly to Exact Online API field names. ExactOnline uses a special type of pascal case. This means that all words are capitalized, like in regular pascal case. It differs in the way it handles acryonyms. For example, user_id (snake case) would normally become UserId in pascal case. In Exact Online syntax, this becomes UserID.

Similarly, for longer acronyms such as the one found in oid_connect, this will become OIDConnect in Exact Online syntax.

Because Pydantic is very pythonic, it's customary to use snake case for property (field) names. Here's is the issue with that: when converting pascal case from the Exact Online API responses to snake case, information is lost. For example: UserID will be come user_id when using e.g. Pydantic's to_pascal validation. To solve this, a special type of snake case is used in the exactpy's model definition. Every acronym is suffixed with a double underscore __, which tells the validator to capitalizer every single letter before the underscore, until it hits the beginning of the string or another underscore.

As an example: Exact Online's field with name GLAccountID is defined in our models as gl__acount_id__.

One exception is the id field in every model, which is always completely capitalized. The reason for this, is that it's so common, it would be a waste of time to double underscore it in every single model.

Sample usage

Initial oauth token request

This initial token request theoretically needs to be done once. An access token and requets token will be available after this. If the access token is refreshed in time (refresh token is valid for 30 days max), no log in has to be performed.

from exactpy.client import Client

client_id = "***"
client_secret = "***"

# This can be anything, but it needs to satisfy two conditions:
# 1. It needs to be a valid public website
# 2. It needs to be website protected by an SSL certificate (https)
# It needs to match exactly what you entered in your app registration
redirect_url = "https://some_site"

client = Client(
    client_id=client_id, client_secret=client_secret, redirect_url=redirect_url
)

# This will print a link that you can click
# You'll have to log to the Exact Online portal first
# You'll get redirected to `redirect_url`, and you'll need to copy that url
print(client.auth_client.get_authorization_url())

# Enter the url that you've copied earlier and hit enter
resp_url = input()

# If everything went well, you should now be able to acquire
# a token and refresh token with the line below
# This will also automatically cache your credentials
# to file, if caching is enabled (which it is, by default)
client.auth_client.acquire_token(resp_url)

If your cache callable was not set to None, your credentials should now have been saved. In the default case, they're saved in creds.json (plain text).

Client usage

Now you can use the client like below:

from exactpy.client import Client
from pprint import pprint

client = Client(
    client_id="xxx",
    client_Secret="xxx,
    verbose=True,
)

# (Almost) every request requires a division to be set
# This division number is set as a client property
# and set with every request, as part of the url
# You can set the division equal to the current divison
# (current as in server side) using:
client.set_initial_division()

# Print the division using:
pprint(client.get_current_division().model_dump(by_alias=True))

# Alternatively, you may pick a different division by listing them
# and selecting a specific one:
divisions = client.system.divisions.all()

# Select the "nth" division and set it as current division
# n = 0...len(divisions)-1
client.division = divisions[n].code

# Every request you do, will trigger a token
# check as well. In the default case, auto-caching is done
# This means that the current access token and refresh token
# (after a optional refresh) are cached. This is only true
# if cache_callable was not set to None.
# To disable this behaviour, either set cache_callable to None
# or set autocache_enabled to False.

# Note that endpoint controllers are namespaced by service,
# as shown below (general ledger accounts are part of the
# financial service) and above (divisions are part of the)
# system service
gl_accounts = client.financial.gl_accounts.all()

Detailed usage

Every OData query arg is supported, except $skip; this is because Exact Online does not support this. See also: https://support.exactonline.com/community/s/knowledge-base#All-All-DNO-Simulation-query-string-options

Okay fine, on some older endpoints it might actually be supported, but I'm not going to find out which ones those are.

Filters

All of the odata filter operators are supported. The full list can be found it the enum exactpy.types.FilterOperatorEnum.

Filters are passed a list to the controller and should contain dicts containing key, val, op and optionally, meta key value pairs.

The key should refer to the name of the field, the val to the filter value, the op to the filter operator. For one filter type, substring, a second parameter length is needed. This can be passed by passing an explicit meta dict, with the structure meta={"length": <some_int>}.

Example filters param:

[{"key": "reporting_year", "val": 2025, "op": "eq"}]

You can also use an explicit enum property for the op.:

from exactpy.types import FilterOperatorEnum

[{"key": "reporting_year", "val": 2025, "op": FilterOperatorEnum.EQUALS}]

Filter field names should use the model field names, not the Exact Online API field names. These field names are automatically translated into their Exact Online name variant. Multiple filters are combined using the binary and or the binary or operator, which can be set using the FilterCombinationOperatorEnum enum type. Example:

from exactpy.types import FilterCombinationOperatorEnum

# OR
reporting_balances = client.financial.reporting_balances.all(
    filters=[{"key": "reporting_year", "val": 2025, "op": "eq"}, {"key": "reporting_year", "val": 2024, "op": "eq"}],
    filter_combination_operator=FilterCombinationOperatorEnum.OR,
)

# AND
reporting_balances = client.financial.reporting_balances.all(
    filters=[{"key": "reporting_year", "val": 2025, "op": "eq"}],
    filter_combination_operator=FilterCombinationOperatorEnum.AND,
)

Obviously, because they're enums, you can also use the enum values themselves. Example:

reporting_balances = client.financial.reporting_balances.all(
    filters=[{"key": "reporting_year", "val": 2025, "op": "eq"}],
    filter_combination_operator="and", # use "and" or "or" (enum values)
)

Note that the or filter combination operator can only be used if the keys of multiple filters are all identical (e.g. all filters filter on the same field). I haven't verified this, but the Exact Online docs says so.

Top n results

Use the top argument to select the first top results:

# This should give maximum (for two reasons, more on this later) 5 results
reporting_balances = client.financial.reporting_balances.all(top=5)

Expand

Embedded models are not expanded by default. Use the expand arg to tell the API to expand them. Again, using model field names, not Exact Online API field names. Example:

accounts = client.crm.accounts.all(expand=["bank_accounts"])

Select

Use the select arg to only retrieve a select number of columns. Again, using model field names, not Exact Online API field names. Example:

gl_accounts = client.financial.gl_accounts.all(select=["reporting_year"])

Order by

You can order by a certain field name by setting the order_by arg of the controller (all or all_paged methods) to a dict of the following structure:

from types import OrderByDirectionEnum

# Ascending
gl_accounts = client.financial.gl_accounts.all(order_by={"key": "reporting_year", "direction": OrderByDirectionEnum.ASC })

# Descending
gl_accounts = client.financial.gl_accounts.all(order_by={"key": "reporting_year", "direction": OrderByDirectionEnum.DESC })

Again, as can be seen, using model field names, not Exact Online API field names.

Regular strings desc and asc are also supported:

from types import OrderByDirectionEnum

# Ascending
gl_accounts = client.financial.gl_accounts.all(order_by={"key": "reporting_year", "direction": "asc" })
gl_accounts = client.financial.gl_accounts.all(order_by={"key": "reporting_year", "direction": "desc" })

Inline count

Setting inline_count to True will retrieve the count of all records. In the client this will result in a tuple as return type, instead of a simple list of model instances:

gl_accounts, count = client.financial.gl_accounts.all(select=["reporting_year"], inline_count=True)

Count only

OData also implements a $count query arg to only retrieve a count of all records. This is implemented as the count() method on the controllers in this package. No options can be set, except for include_division, which will tell the client to include the division in the API url or not. Example:

count = client.financial.gl_accounts.count()

Paged (generator) vs all results

The underlying way for retrieving records is using a generator that yields every retrieved page as a results list. This is implemented as the all_paged method on the controllers. Example usage:

for page in client.financial.gl_accounts.all_paged(select=["reporting_year"]):
    for gl_account in page:
        print(page.model_dump(by_alias=True))

And using inline count:

for page, count in client.financial.gl_accounts.all_paged(select=["reporting_year"], inline_count=True):
    print(f"Total count: {count}")
    for gl_account in page:
        print(page.model_dump(by_alias=True))

Note that the count given every yield by the generator is the total count, and is the same every time. It is not the count per page.

Max pages

Setting the max_pages arg will result in only retrieving max_pages from the API. This competes with the top argument. If either one exhausts the number results first, the other isn't effective, obviously.

Note that when you set inline_count=True, the API responses are no longer paged and this argument will not do anything at all.

Skipping invalid

By default, invalid records are skipped. This is not the default pydantic behaviour. Lists of input is usually parsed using for example:

list_adapter = TypeAdapter(SomeModel)
list_adapter.validate_python([{"field1": "someval", ...}, ...])

This will raise a pydantic.ValidationError when it encounters invalid input (ie input that cannot be validated using that specific model).

This is not really wanted behavior when using the Exact Online API in combination with the strict Pydantic models in this package, as it's very clear that Exact Online API output is known not to adhere to field types described in the API (notoriously, types will sometimes have undocumented values).

Setting skip_invalid=True, will skip over these records and log these events (if verbose=True on the client). Setting skip_invalid=False will result in normal pydantic behavior.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

exactpy-0.0.22.tar.gz (39.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

exactpy-0.0.22-py3-none-any.whl (54.0 kB view details)

Uploaded Python 3

File details

Details for the file exactpy-0.0.22.tar.gz.

File metadata

  • Download URL: exactpy-0.0.22.tar.gz
  • Upload date:
  • Size: 39.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for exactpy-0.0.22.tar.gz
Algorithm Hash digest
SHA256 b3112ae3ea9af9defc27faaf0a090df9e9e815e4704943e1fe17fc489e2358ac
MD5 8f31cd16a950c2bd6535190bc6de909c
BLAKE2b-256 588f77c363e02dadeb7904f23497ed5e58a3df69117a91a07b9e9e3cf49475cf

See more details on using hashes here.

Provenance

The following attestation bundles were made for exactpy-0.0.22.tar.gz:

Publisher: publish.yml on riccardo92/exactpy

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file exactpy-0.0.22-py3-none-any.whl.

File metadata

  • Download URL: exactpy-0.0.22-py3-none-any.whl
  • Upload date:
  • Size: 54.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for exactpy-0.0.22-py3-none-any.whl
Algorithm Hash digest
SHA256 1242d039b630ee65cb1e7cf87c17e68ee08b3fcb51f06a8038fe7ab6bb49cf44
MD5 8a8b2fb88db6c875067b43cd4803b199
BLAKE2b-256 153bc5f3c4d06c2abc8057f9df08e8c47ad0160543d6c05cde0b3873e2cc24c1

See more details on using hashes here.

Provenance

The following attestation bundles were made for exactpy-0.0.22-py3-none-any.whl:

Publisher: publish.yml on riccardo92/exactpy

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page