EagerPy is a thin wrapper around PyTorch, TensorFlow Eager, JAX and NumPy that unifies their interface and thus allows writing code that works natively across all of them.
Project description
EagerPy: PyTorch, TensorFlow, JAX and NumPy — all of them natively using the same code
EagerPy is a Python framework that let’s you write code that automatically works natively with PyTorch, TensorFlow, JAX, and NumPy. EagerPy is also great when you work with just one framework but prefer a clean and consistent API that is fully chainable, provides extensive type annotions and let’s you write beautiful code.
🔥 Design goals
Native Performance: EagerPy operations get directly translated into the corresponding native operations.
Fully Chainable: All functionality is available as methods on the tensor objects and as EagerPy functions.
Type Checking: Catch bugs before running your code thanks to EagerPy’s extensive type annotations.
📖 Documentation
Learn more about in the documentation.
🚀 Quickstart
pip install eagerpy
🎉 Example
import torch
x = torch.tensor([1., 2., 3., 4., 5., 6.])
import tensorflow as tf
x = tf.constant([1., 2., 3., 4., 5., 6.])
import jax.numpy as np
x = np.array([1., 2., 3., 4., 5., 6.])
import numpy as np
x = np.array([1., 2., 3., 4., 5., 6.])
# No matter which framwork you use, you can use the same code
import eagerpy as ep
# Just wrap a native tensor using EagerPy
x = ep.astensor(x)
# All of EagerPy's functionality is available as methods
x = x.reshape((2, 3))
x.flatten(start=1).square().sum(axis=-1).sqrt()
# or just: x.flatten(1).norms.l2()
# and as functions (yes, we gradients are also supported!)
loss, grad = ep.value_and_grad(loss_fn, x)
ep.clip(x + eps * grad, 0, 1)
# You can even write functions that work transparently with
# Pytorch tensors, TensorFlow tensors, JAX arrays, NumPy arrays
def my_universal_function(a, b, c):
# Convert all inputs to EagerPy tensors
a, b, c = ep.astensors(a, b, c)
# performs some computations
result = (a + b * c).square()
# and return a native tensor
return result.raw
🗺 Use cases
Foolbox Native, the latest version of Foolbox, a popular adversarial attacks library, has been rewritten from scratch using EagerPy instead of NumPy to achieve native performance on models developed in PyTorch, TensorFlow and JAX, all with one code base.
🐍 Compatibility
We currently test with the following versions:
PyTorch 1.4.0
TensorFlow 2.1.0
JAX 0.1.57
NumPy 1.18.1
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file eagerpy-0.27.0.tar.gz
.
File metadata
- Download URL: eagerpy-0.27.0.tar.gz
- Upload date:
- Size: 20.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/40.6.2 requests-toolbelt/0.9.1 tqdm/4.43.0 CPython/3.6.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 68b887c615178f5f881552d008673d7bbd702cf8dcac5c38179a55825c7bd0c7 |
|
MD5 | 9e99dc598aab3c69eb77aaa00388c9e7 |
|
BLAKE2b-256 | b0d8ba2d794a02885acba16da11e2b97817b096b4c0281c65eb378aef00f67e0 |
File details
Details for the file eagerpy-0.27.0-py3-none-any.whl
.
File metadata
- Download URL: eagerpy-0.27.0-py3-none-any.whl
- Upload date:
- Size: 29.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/40.6.2 requests-toolbelt/0.9.1 tqdm/4.43.0 CPython/3.6.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | a63f1959b1d8af4f82ee195a8a5f945fe1427e60434fa8a201d42bdc058d56b8 |
|
MD5 | 46d8a73c7e3026f263df59913b3b1469 |
|
BLAKE2b-256 | 8d4c13ed2aba954c111ea0aff7e75a5e3c95b533504ba24718f7ea18b036c440 |