Skip to main content

A pytest-plugin for updating doctest outputs

Project description

pytest-accept

GitHub Workflow CI Status PyPI Version GitHub License

pytest-accept is a pytest plugin for automatically updating doctest outputs. It runs doctests, observes the generated outputs, and writes them to the doctests' documented outputs.

It's designed for a couple of use cases:

  • People who work with doctests and don't enjoy manually copying generated outputs from the pytest error log and pasting them into their doctests' documented outputs. pytest-accept does the copying & pasting for you.
  • People who generally find writing tests a bit annoying, and prefer to develop by "running the code and seeing whether it works". This library aims to make testing a joyful part of that development loop.

pytest-accept is decoupled from the doctests it works with — it can be used with existing doctests, and the doctests it edits are no different from normal doctests.

Jesse, what the?

Here's an example of pytest-accept does: given a file like add.py containing an incorrect documented output:

def add(x, y):
    """
    Adds two values.

    >>> add(1, 1)
    3

    >>> add("ab", "c")
    'bac'
    """

    return x + y

...running doctests using pytest and passing --accept replaces the existing incorrect values with correct values:

pytest --doctest-modules examples/add.py --accept
diff --git a/examples/add.py b/examples/add.py
index 10a71fd..c2c945f 100644
--- a/examples/add.py
+++ b/examples/add.py
@@ -3,10 +3,10 @@ def add(x, y):
     Adds two values.
 
     >>> add(1, 1)
-    3
+    2
 
     >>> add("ab", "c")
-    'bac'
+    'abc'
     """
 
     return x + y

This style of testing is fairly well-developed in some languages, although still doesn't receive the attention I think it deserves, and historically hasn't had good support in python.

Confusingly, it's referred to "snapshot testing" or "regression testing" or "expect testing" or "literate testing" or "acceptance testing". The best explanation I've seen on this testing style is from Ron Minsky in a Jane Street Blogpost. @matklad also has an excellent summary in his blog post How to Test.

Installation

pip install pytest-accept

What about normal tests?

A previous effort in assert_plugin.py attempted to do this for assert statements, and the file contains some notes on the effort. The biggest problem is pytest stops on the first assert failure in each test, which is very limiting. (Whereas pytest can be configured to continue on doctest failures, which this library takes advantage of.)

It's probably possible to change pytest's behavior here, but it's a significant effort on the pytest codebase.

Some alternatives:

  • Use an existing library like pytest-regtest, which offers file snapshot testing (i.e. not inline).
  • We could write a specific function / fixture, like accept(result, "abc"), similar to frameworks like rust's excellent insta (which I developed some features for), or ocaml's ppx_expect.
    • But this has the disadvantage of coupling the test to the plugin: it's not possible to run tests independently of the plugin, or use the plugin on general assert tests. And one of the great elegances of pytest is its deferral to a normal assert statement.
  • Some of this testing feels like writing a notebook and testing that. pytest-notebook fully implements this.

Anything else?

Not really! Some things to watch out for:

  • It attempts to confirm the file hasn't changed between the start and end of the test and won't overwrite the file in those cases. This can be helpful for workflows where the tests run repeatedly in the background (using something like watchexec while someone is working on the file, on when the tests take a long time, maybe because of --pdb To be doubly careful, passing --accept-copy will cause the plugin to instead create a file named {file}.py.new.
    • It will overwrite the existing documented values, though these aren't generally useful per se — they're designed to match the generated of the code. The only time they could be useful is if there's manual curation (e.g. removing volatile outputs like hashes), and in those cases ideally they can be restored from version control, or pass --accept-copy to be conservative.
  • This is early, and there are probably some small bugs. Let me know anything at all and I'll attempt to fix them.
  • It currently doesn't affect the printing of test results; the doctests will still print as failures.
    • TODO: A future version could print something about them being fixed.
  • Python's doctest library is imperfect:
    • It can't handle indents, and probably other things. (We do handle blank lines though, and TODO: check whether the output we paste is valid doctest output).
    • The syntax for .* is an ellipsis ..., which is also the syntax for continuing a code line, so it can't be at the start of a line.
    • The syntax for all the directives is arguably less than aesthetically pleasing.
    • It doesn't have an option for pretty printing, so the test must pretty print, which is verbose.
    • It reports line numbers incorrectly in some cases — two docstring lines separated with continuation character \ is counted as one, meaning this library will not have access to the correct line number for doctest inputs and outputs.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pytest-accept-0.1.5.tar.gz (16.2 kB view hashes)

Uploaded Source

Built Distribution

pytest_accept-0.1.5-py3-none-any.whl (14.8 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page