Touca SDK for Python
Touca SDK For Python
Touca helps you understand the true impact of your day to day code changes on the behavior and performance of your overall software, as you write code.
Touca SDKs let you describe the behavior and performance of your code by capturing values of interesting variables and runtime of important functions. We remotely compare your description against a trusted version of your software, visualize all differences, and report them in near real-time.
You can install Touca with pip:
pip install touca
We formally support Python v3.6 and newer on Linux, macOS, and Windows platforms.
👀 Sneak Peak
For a more thorough guide of how to use Touca SDK for Python, check out the
examplesdirectory or visit our documentation website at docs.touca.io.
Let us imagine that we want to test a software workflow that reports whether a given number is prime.
def is_prime(number: int):
We can use unit testing in which we hard-code a set of input numbers and list our expected return value for each input. In this example, the input and output of our code under test are a number and a boolean. If we were testing a video compression algorithm, they may have been video files. In that case:
- Describing the expected output for a given video file would be difficult.
- When we make changes to our compression algorithm, accurately reflecting those changes in our expected values would be time-consuming.
- We would need a large number of input video files to gain confidence that our algorithm works correctly.
Touca makes it easier to continuously test workflows of any complexity and with any number of test cases.
import touca from code_under_test import is_prime @touca.Workflow def test_is_prime(testcase: str): touca.check("is_prime", is_prime(int(testcase))) if __name__ == "__main__": touca.run()
Touca tests have two main differences compared to typical unit tests:
- We have fully decoupled our test inputs from our test logic. We refer to these inputs as "test cases". The SDK retrieves the test cases from the command line, or a file, or a remote Touca server and feeds them one by one to our code under test.
- We have removed the concept of expected values. With Touca, we only describe the actual behavior and performance of our code under test by capturing values of interesting variables and runtime of important functions, anywhere within our code. For each test case, the SDK submits this description to a remote server which compares it against the description for a trusted version of our code. The server visualizes any differences and reports them in near real-time.
We can run Touca tests with any number of inputs from the command line:
python3 test_prime_app.py --api-key <TOUCA_API_KEY> --api-url <TOUCA_API_URL> --revision v1.0 --testcase 13 17 51
TOUCA_API_URL can be obtained from the Touca server
at app.touca.io. This command produces the following
Touca Test Framework Suite: is_prime_test Revision: v1.0 ( 1 of 3 ) 13 (pass, 127 ms) ( 2 of 3 ) 17 (pass, 123 ms) ( 3 of 3 ) 51 (pass, 159 ms) Processed 3 of 3 testcases Test completed in 565 ms
Touca is very effective in addressing common problems in the following situations:
- When we need to test our workflow with a large number of inputs.
- When the output of our workflow is too complex, or too difficult to describe in our unit tests.
- When interesting information to check for regression is not exposed through the interface of our workflow.
The fundamental design features of Touca that we highlighted earlier can help us test these workflows at any scale.
- Decoupling our test input from our test logic, can help us manage our long list of inputs without modifying the test logic. Managing that list on a remote server accessible to all members of our team, can help us add notes to each test case, explain why they are needed and track how their performance changes over time.
- Submitting our test results to a remote server, instead of storing them in files, can help us avoid the mundane tasks of managing and processing of those results. The Touca server retains test results and makes them accessible to all members of the team. It compares test results using their original data types and reports discovered differences in real-time to all interested members of our team. It allows us to audit how our software evolves over time and provides high-level information about our tests.
- If you are new to Touca, the best place to start is the Quickstart Guide on our documentation website.
- For information on how to use this SDK, see our Python SDK Documentation.
- If you cannot wait to start writing your first test with Touca, see our Python API Reference.
🙋 Ask for Help
🚀 Next Steps
Touca client libraries are free and open-source. Our cloud-hosted Touca server at app.touca.io has a free plan. You can create an account and explore Touca on your own. We are also happy to chat 1:1 to help you get on-boarded and answer any questions you may have in the process. We'd love to learn more about you, understand your software and its requirements, and help you decide if Touca would be useful to you and your team.
This repository is released under the Apache-2.0 License. See
Release history Release notifications | RSS feed
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
|Filename, size||File type||Python version||Upload date||Hashes|
|Filename, size touca-1.5.0.tar.gz (34.9 kB)||File type Source||Python version None||Upload date||Hashes View|
|Filename, size touca-1.5.0-py3-none-any.whl (40.5 kB)||File type Wheel||Python version py3||Upload date||Hashes View|