JAX NN library.
Project description
The โจMagicalโจ JAX NN Library.
*Serket is the goddess of magic in Egyptian mythology
Installation |Description |Quick Example |Freezing/Fine tuning |Filtering
๐ ๏ธ Installation
pip install serket
Install development version
pip install git+https://github.com/ASEM000/serket
๐ Description
-
serket
aims to be the most intuitive and easy-to-use physics-based Neural network library in JAX. -
serket
is built on top ofpytreeclass
-
serket
currently implements
๐ง Neural network package: serket.nn
๐ง
Group | Layers |
---|---|
Linear | Linear , Bilinear , Multilinear , GeneralLinear , Identity |
Densely connected | FNN (Fully connected network), PFNN (Parallel fully connected network) |
Convolution | Conv1D , Conv2D , Conv3D , Conv1DTranspose , Conv2DTranspose , Conv3DTranspose , DepthwiseConv1D , DepthwiseConv2D , DepthwiseConv3D , SeparableConv1D , SeparableConv2D , SeparableConv3D , Conv1DLocal , Conv2DLocal , Conv3DLocal |
Containers | Sequential , Lambda |
Pooling | MaxPool1D , MaxPool2D , MaxPool3D , AvgPool1D , AvgPool2D , AvgPool3D GlobalMaxPool1D , GlobalMaxPool2D , GlobalMaxPool3D , GlobalAvgPool1D , GlobalAvgPool2D , GlobalAvgPool3D LPPool1D , LPPool2D ,LPPool3D , AdaptivePool1D , AdaptivePool2D , AdaptivePool3D (kernex backend) |
Reshaping | Flatten , Unflatten , FlipLeftRight2D , FlipUpDown2D , Repeat1D , Repeat2D , Repeat3D , Resize1D , Resize2D , Resize3D , Upsample1D , Upsample2D , Upsample3D , Pad1D , Pad2D , Pad3D |
Crop | Crop1D , Crop2D , |
Normalization | LayerNorm , InstanceNorm , GroupNorm |
Blurring | AvgBlur2D , GaussianBlur2D |
Dropout | Dropout , Dropout1D , Dropout2D , Dropout3D , |
Random transforms | RandomCrop1D , RandomCrop2D , RandomApply , RandomCutout1D , RandomCutout2D , RandomZoom2D , RandomContrast2D |
Misc | HistogramEqualization2D , AdjustContrast2D , Filter2D , PixelShuffle |
Activations | AdaptiveLeakyReLU ,AdaptiveReLU ,AdaptiveSigmoid ,AdaptiveTanh ,CeLU ,ELU ,GELU ,GLU , HardSILU ,HardShrink ,HardSigmoid ,HardSwish ,HardTanh ,LeakyReLU ,LogSigmoid ,LogSoftmax ,Mish ,PReLU ,ReLU ,ReLU6 ,SILU ,SeLU ,Sigmoid ,SoftPlus ,SoftShrink ,SoftSign ,Swish ,Tanh ,TanhShrink , ThresholdedReLU , Snake |
Recurrent | SimpleRNNCell , LSTMCell , ConvLSTM1D , ConvLSTM2D , ConvLSTM3D , SeparableConvLSTM1DCell , SeparableConvLSTM2DCell , SeparableConvLSTM3DCell |
Blocks | VGG16Block , VGG19Block , UNetBlock |
โโFinite difference package: serket.fd
โโ
Group | Function/Layer |
---|---|
Finite difference layer | Difference : apply finite difference to input array to any derivative order and accuracy |
Finite difference functions | - difference : finite difference of array with any accuracy and derivative order - generate_finitediff_coeffs : generate coeffs using sample points and derivative order - fgrad : differentiate functions (similar to jax.grad ) with custom accuracy and derivative order |
Vector operator layers | Curl , Divergence , Gradient , Laplacian |
Vector operator function | curl , divergence , gradient , laplacian |
โฉ Examples:
Finite difference package examples
import jax
jax.config.update("jax_enable_x64", True)
import jax.numpy as jnp
import numpy.testing as npt
import serket as sk
# lets first define a vector valued function F: R^3 -> R^3
# F = F1, F2
# F1 = x^2 + y^3
# F2 = x^4 + y^3
# F3 = 0
# F = [x**2 + y**3, x**4 + y**3, 0]
x, y, z = [jnp.linspace(0, 1, 100)] * 3
dx, dy, dz = x[1] - x[0], y[1] - y[0], z[1] - z[0]
X, Y, Z = jnp.meshgrid(x, y, z, indexing="ij")
F1 = X**2 + Y**3
F2 = X**4 + Y**3
F3 = jnp.zeros_like(F1)
F = jnp.stack([F1, F2, F3], axis=0)
# โF1/โx : lets differentiate F1 with respect to x (i.e axis=0)
dF1dx = sk.fd.difference(F1, axis=0, step_size=dx, accuracy=6)
dF1dx_exact = 2 * X
npt.assert_allclose(dF1dx, dF1dx_exact, atol=1e-7)
# โF2/โy : let us now differentiate F2 with respect to y (i.e axis=1)
dF2dy = sk.fd.difference(F2, axis=1, step_size=dy, accuracy=6)
dF2dy_exact = 3 * Y**2
npt.assert_allclose(dF2dy, dF2dy_exact, atol=1e-7)
# โ.F : lets take the divergence of F
divF = sk.fd.divergence(F, step_size=(dx, dy, dz), keepdims=False, accuracy=6)
divF_exact = 2 * X + 3 * Y**2
npt.assert_allclose(divF, divF_exact, atol=1e-7)
# โF1 : lets take the gradient of F1
gradF1 = sk.fd.gradient(F1, step_size=(dx, dy, dz), accuracy=6)
gradF1_exact = jnp.stack([2 * X, 3 * Y**2, 0 * X], axis=0)
npt.assert_allclose(gradF1, gradF1_exact, atol=1e-7)
# ฮF1 : lets take laplacian of F1
lapF1 = sk.fd.laplacian(F1, step_size=(dx, dy, dz), accuracy=6)
lapF1_exact = 2 + 6 * Y
npt.assert_allclose(lapF1, lapF1_exact, atol=1e-7)
# # โxF : lets take the curl of F
curlF = sk.fd.curl(F, step_size=(dx, dy, dz), accuracy=6)
curlF_exact = jnp.stack([F1 * 0, F1 * 0, 4 * X**3 - 3 * Y**2], axis=0)
npt.assert_allclose(curlF, curlF_exact, atol=1e-7)
Train Bidirectional-LSTM
import jax
import jax.numpy as jnp
import jax.random as jr
import matplotlib.pyplot as plt
import optax # for gradient optimization
import serket as sk
x = jnp.linspace(0, 1, 101).reshape(-1, 1) # 101 samples of 1D data
y = jnp.sin(2 * jnp.pi * x)
y += 0.1 * jr.normal(jr.PRNGKey(0), y.shape)
# we will use 2 time steps to predict the next time step
x_batched = jnp.stack([x[:-1], x[1:]], axis=1)
x_batched = jnp.reshape(x_batched, (100, 1, 2, 1)) # 100 minibatches x 1 sample x 2 time steps x 1D data
y_batched = jnp.reshape(y[1:], (100, 1, 1)) # 100 minibatches x 1 samples x 1D data
model = sk.nn.Sequential(
[
# first cell is the forward cell, second cell is the backward cell for bidirectional RNN
# we return the full sequence of outputs for each cell by setting return_sequences=True
# we use None in place of `in_features` to infer the input shape from the input
sk.nn.ScanRNN(
sk.nn.LSTMCell(None, 64),
backward_cell=sk.nn.LSTMCell(None, 64),
return_sequences=True,
),
# here the in_features is inferred from the previous layer by setting it to None
# or simply we can set it to 64*2 (64 for each cell from previous layer)
# we set return_sequences=False to return only the last output of the sequence
sk.nn.ScanRNN(sk.nn.LSTMCell(None, 1), return_sequences=False),
]
)
@jax.value_and_grad
def loss_func(NN, batched_x, batched_y):
# use jax.vmap to apply the model to each minibatch
# in our case single x minibatch has shape (1, 2, 1)
# and single y minibatch has shape (1, 1)
# then vmap will be applied to the leading axis
batched_preds = jax.vmap(NN)(batched_x)
return jnp.mean((batched_preds - batched_y) ** 2)
@jax.jit
def batch_step(NN, batched_x, batched_y, opt_state):
loss, grads = loss_func(NN, batched_x, batched_y)
updates, optim_state = optim.update(grads, opt_state)
NN = optax.apply_updates(NN, updates)
return NN, optim_state, loss
# dry run to infer the in_features (i.e. replace None with in_features)
# if you want restrict the model to a specific input shape or to avoid
# confusion you can manually specify the in_features as a consequence
# dry run is not necessary in this case
model(x_batched[0, 0])
optim = optax.adam(1e-3)
opt_state = optim.init(model)
epochs = 100
for i in range(1, epochs + 1):
epoch_loss = []
for x_b, y_b in zip(x_batched, y_batched):
model, opt_state, loss = batch_step(model, x_b, y_b, opt_state)
epoch_loss.append(loss)
epoch_loss = jnp.mean(jnp.array(epoch_loss))
if i % 10 == 0:
print(f"Epoch {i:3d} Loss {epoch_loss:.4f}")
# Epoch 10 Loss 0.0880
# Epoch 20 Loss 0.0796
# Epoch 30 Loss 0.0620
# Epoch 40 Loss 0.0285
# Epoch 50 Loss 0.0205
# Epoch 60 Loss 0.0187
# Epoch 70 Loss 0.0182
# Epoch 80 Loss 0.0176
# Epoch 90 Loss 0.0171
# Epoch 100 Loss 0.0166
y_pred = jax.vmap(model)(x_batched.reshape(-1, 2, 1))
plt.plot(x[1:], y[1:], "--k", label="data")
plt.plot(x[1:], y_pred, label="prediction")
plt.legend()
Lazy initialization
In cases where in_features
needs to be inferred from input, use None
instead of in_features
to infer the value at runtime.
However, since the lazy module initialize it's state after the first call (i.e. mutate it's state) jax
transformation ex: vmap, grad ...
is not allowed before initialization. Using any jax
transformation before initialization will throw a ValueError
.
import serket as sk
import jax
import jax.numpy as jnp
model = sk.nn.Sequential(
[
sk.nn.Conv2D(None, 128, 3),
sk.nn.ReLU(),
sk.nn.MaxPool2D(2, 2),
sk.nn.Conv2D(128, 64, 3),
sk.nn.ReLU(),
sk.nn.MaxPool2D(2, 2),
sk.nn.Flatten(),
sk.nn.Linear(None, 128),
sk.nn.ReLU(),
sk.nn.Linear(128, 1),
]
)
# print the first `Conv2D` layer before initialization
print(model[0].__repr__())
# Conv2D(
# weight=None,
# bias=None,
# *in_features=None,
# *out_features=None,
# *kernel_size=None,
# *strides=None,
# *padding=None,
# *input_dilation=None,
# *kernel_dilation=None,
# weight_init_func=None,
# bias_init_func=None,
# *groups=None
# )
try :
jax.vmap(model)(jnp.ones((10, 1,28, 28)))
except ValueError:
print("***** Not initialized *****")
# ***** Not initialized *****
# dry run to initialize the model
model(jnp.empty([3,128,128]))
print(model[0].__repr__())
# Conv2D(
# weight=f32[128,3,3,3],
# bias=f32[128,1,1],
# *in_features=3,
# *out_features=128,
# *kernel_size=(3,3),
# *strides=(1,1),
# *padding=((1,1),(1,1)),
# *input_dilation=(1,1),
# *kernel_dilation=(1,1),
# weight_init_func=Partial(glorot_uniform(key,shape,dtype)),
# bias_init_func=Partial(zeros(key,shape,dtype)),
# *groups=1
# )
Train MNIST
We will use tensorflow
datasets for dataloading. for more on interface of jax/tensorflow dataset see here
# imports
import tensorflow as tf
# Ensure TF does not see GPU and grab all GPU memory.
tf.config.set_visible_devices([], device_type="GPU")
import tensorflow_datasets as tfds
import tensorflow.experimental.numpy as tnp
import jax
import jax.numpy as jnp
import jax.random as jr
import optax # for gradient optimization
import serket as sk
import matplotlib.pyplot as plt
import functools as ft
# Construct a tf.data.Dataset
batch_size = 128
# convert the samples from integers to floating-point numbers
# and channel first format
def preprocess_data(x):
# convert to channel first format
image = tnp.moveaxis(x["image"], -1, 0)
# normalize to [0, 1]
image = tf.cast(image, tf.float32) / 255.0
# one-hot encode the labels
label = tf.one_hot(x["label"], 10) / 1.0
return {"image": image, "label": label}
ds_train, ds_test = tfds.load("mnist", split=["train", "test"], shuffle_files=True)
# (batches, batch_size, 1, 28, 28)
ds_train = ds_train.shuffle(1024).map(preprocess_data).batch(batch_size).prefetch(tf.data.AUTOTUNE)
# (batches, 1, 28, 28)
ds_test = ds_test.map(preprocess_data).prefetch(tf.data.AUTOTUNE)
๐๏ธ Model definition
We will use jax.vmap(model)
to apply model
on batches.
@sk.treeclass
class CNN:
def __init__(self):
self.conv1 = sk.nn.Conv2D(1, 32, (3, 3), padding="valid")
self.relu1 = sk.nn.ReLU()
self.pool1 = sk.nn.MaxPool2D((2, 2), strides=(2, 2))
self.conv2 = sk.nn.Conv2D(32, 64, (3, 3), padding="valid")
self.relu2 = sk.nn.ReLU()
self.pool2 = sk.nn.MaxPool2D((2, 2), strides=(2, 2))
self.flatten = sk.nn.Flatten(start_dim=0)
self.dropout = sk.nn.Dropout(0.5)
self.linear = sk.nn.Linear(5*5*64, 10)
def __call__(self, x):
x = self.conv1(x)
x = self.relu1(x)
x = self.pool1(x)
x = self.conv2(x)
x = self.relu2(x)
x = self.pool2(x)
x = self.flatten(x)
x = self.dropout(x)
x = self.linear(x)
return x
model = CNN()
๐จ Visualize model
Model summary
print(model.summary(show_config=False, array=jnp.empty((1, 28, 28))))
โโโโโโโโโฌโโโโโโโโโโฌโโโโโโโโโโฌโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโ
โName โType โParam # โSize โInput โOutput โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โconv1 โConv2D โ320(0) โ1.25KB(0.00B) โf32[1,28,28] โf32[32,26,26]โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โrelu1 โReLU โ0(0) โ0.00B(0.00B) โf32[32,26,26]โf32[32,26,26]โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โpool1 โMaxPool2Dโ0(0) โ0.00B(0.00B) โf32[32,26,26]โf32[32,13,13]โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โconv2 โConv2D โ18,496(0)โ72.25KB(0.00B)โf32[32,13,13]โf32[64,11,11]โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โrelu2 โReLU โ0(0) โ0.00B(0.00B) โf32[64,11,11]โf32[64,11,11]โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โpool2 โMaxPool2Dโ0(0) โ0.00B(0.00B) โf32[64,11,11]โf32[64,5,5] โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โflattenโFlatten โ0(0) โ0.00B(0.00B) โf32[64,5,5] โf32[1600] โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โdropoutโDropout โ0(0) โ0.00B(0.00B) โf32[1600] โf32[1600] โ
โโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโผโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโผโโโโโโโโโโโโโโค
โlinear โLinear โ16,010(0)โ62.54KB(0.00B)โf32[1600] โf32[10] โ
โโโโโโโโโดโโโโโโโโโโดโโโโโโโโโโดโโโโโโโโโโโโโโโดโโโโโโโโโโโโโโดโโโโโโโโโโโโโโ
Total count : 34,826(0)
Dynamic count : 34,826(0)
Frozen count : 0(0)
------------------------------------------------------------------------
Total size : 136.04KB(0.00B)
Dynamic size : 136.04KB(0.00B)
Frozen size : 0.00B(0.00B)
========================================================================
tree diagram
print(model.tree_diagram())
CNN
โโโ conv1=Conv2D
โ โโโ weight=f32[32,1,3,3]
โ โโโ bias=f32[32,1,1]
โ โ*โ in_features=1
โ โ*โ out_features=32
โ โ*โ kernel_size=(3, 3)
โ โ*โ strides=(1, 1)
โ โ*โ padding=((0, 0), (0, 0))
โ โ*โ input_dilation=(1, 1)
โ โ*โ kernel_dilation=(1, 1)
โ โโโ weight_init_func=Partial(init(key,shape,dtype))
โ โโโ bias_init_func=Partial(zeros(key,shape,dtype))
โ โ*โ groups=1
โโโ relu1=ReLU
โ*โ pool1=MaxPool2D
โ โ*โ kernel_size=(2, 2)
โ โ*โ strides=(2, 2)
โ โ*โ padding='valid'
โโโ conv2=Conv2D
โ โโโ weight=f32[64,32,3,3]
โ โโโ bias=f32[64,1,1]
โ โ*โ in_features=32
โ โ*โ out_features=64
โ โ*โ kernel_size=(3, 3)
โ โ*โ strides=(1, 1)
โ โ*โ padding=((0, 0), (0, 0))
โ โ*โ input_dilation=(1, 1)
โ โ*โ kernel_dilation=(1, 1)
โ โโโ weight_init_func=Partial(init(key,shape,dtype))
โ โโโ bias_init_func=Partial(zeros(key,shape,dtype))
โ โ*โ groups=1
โโโ relu2=ReLU
โ*โ pool2=MaxPool2D
โ โ*โ kernel_size=(2, 2)
โ โ*โ strides=(2, 2)
โ โ*โ padding='valid'
โ*โ flatten=Flatten
โ โ*โ start_dim=0
โ โ*โ end_dim=-1
โโโ dropout=Dropout
โ โ*โ p=0.5
โ โโโ eval=None
โโโ linear=Linear
โโโ weight=f32[1600,10]
โโโ bias=f32[10]
โ*โ in_features=1600
โ*โ out_features=10
Plot sample predictions before training
# set all dropout off
test_model = model.at[model == "eval"].set(True, is_leaf=lambda x: x is None)
def show_images_with_predictions(model, images, one_hot_labels):
logits = jax.vmap(model)(images)
predictions = jnp.argmax(logits, axis=-1)
fig, axes = plt.subplots(5, 5, figsize=(10, 10))
for i, ax in enumerate(axes.flat):
ax.imshow(images[i].reshape(28, 28), cmap="binary")
ax.set(title=f"Prediction: {predictions[i]}\nLabel: {jnp.argmax(one_hot_labels[i], axis=-1)}")
ax.set_xticks([])
ax.set_yticks([])
plt.show()
example = ds_test.take(25).as_numpy_iterator()
example = list(example)
sample_test_images = jnp.stack([x["image"] for x in example])
sample_test_labels = jnp.stack([x["label"] for x in example])
show_images_with_predictions(test_model, sample_test_images, sample_test_labels)
๐ Train the model
@ft.partial(jax.value_and_grad, has_aux=True)
def loss_func(model, batched_images, batched_one_hot_labels):
logits = jax.vmap(model)(batched_images)
loss = jnp.mean(optax.softmax_cross_entropy(logits=logits, labels=batched_one_hot_labels))
return loss, logits
# using optax for gradient updates
optim = optax.adam(1e-3)
optim_state = optim.init(model)
@jax.jit
def batch_step(model, batched_images, batched_one_hot_labels, optim_state):
(loss, logits), grads = loss_func(model, batched_images, batched_one_hot_labels)
updates, optim_state = optim.update(grads, optim_state)
model = optax.apply_updates(model, updates)
accuracy = jnp.mean(jnp.argmax(logits, axis=-1) == jnp.argmax(batched_one_hot_labels, axis=-1))
return model, optim_state, loss, accuracy
epochs = 5
for i in range(epochs):
epoch_accuracy = []
epoch_loss = []
for example in ds_train.as_numpy_iterator():
image, label = example["image"], example["label"]
model, optim_state, loss, accuracy = batch_step(model, image, label, optim_state)
epoch_accuracy.append(accuracy)
epoch_loss.append(loss)
epoch_loss = jnp.mean(jnp.array(epoch_loss))
epoch_accuracy = jnp.mean(jnp.array(epoch_accuracy))
print(f"epoch:{i+1:00d}\tloss:{epoch_loss:.4f}\taccuracy:{epoch_accuracy:.4f}")
# epoch:1 loss:0.2706 accuracy:0.9268
# epoch:2 loss:0.0725 accuracy:0.9784
# epoch:3 loss:0.0533 accuracy:0.9836
# epoch:4 loss:0.0442 accuracy:0.9868
# epoch:5 loss:0.0368 accuracy:0.9889
๐จ Visualize After training
test_model = model.at[model == "eval"].set(True, is_leaf=lambda x: x is None)
show_images_with_predictions(test_model, sample_test_images, sample_test_labels)
๐ฅถ Freezing parameters /Fine tuning
โจSee here for more about freezingโจ
๐ Filtering by masking
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.