Skip to main content

Impelementation of Neural Networks from Scratch based on this book: https://nnfs.io/

Project description

Neural Network from Scratch Implementation

In this repository, with only numpy as only dependency, I have implemented a neural network from scratch based on this marvelous book: nnfs.io.

Data Boundaries

Installation

python3.9 -m pip install nnfs-implementation;

P.S.: This package is compatible with python3.9+.

Examples

There is a module with some examples in the repository, but since this package is use the nnfs name internally, it will have a name collision with the original nnfs package which provides the toy datasets, and you cannot run it. although a glance on it doesn't hurt. Also, There is this notebook VanillaNN.ipynb which is the full implementation will all the examples and some nice plots at the bottom. I highly recommend you to take a look at it.

Instructions

A simple multi-class classifier can be implemented with the following code:

model = Model(loss=SoftmaxLoss(), optimizer=Adam(), metric=Accuracy())
model.add(Layer(28 * 28, 512))
model.add(ReLU())
model.add(Layer(512, 10))
model.fit(train_images, train_labels, epochs=10, batch_size=512)
validation_accuracy = Accuracy.evaluate(test_labels, model.predict(test_images))
print(f"Validation Accuracy: {validation_accuracy:.2%}")

Model

For defining the main model, you should use this pattern:

from nnfs.loss import SoftmaxLoss
from nnfs.metrics import Accuracy
from nnfs.model import Model
from nnfs.optimizers import Adam

model = Model(loss=SoftmaxLoss(), optimizer=Adam(), metric=Accuracy())

You have to define the loss function, the optimizer, and the metric. For any of these components you can change the default values like this:

model = Model(loss=SoftmaxLoss(), optimizer=Adam(learning_rate=0.001), metric=Accuracy())

Loss Criterion

The loss criterion is the function that will be used to calculate the loss of the model. nnfs-implementation provides four loss criterion: (but none of them support multi-label classification):

from nnfs.loss import BinaryLoss, CategoricalLoss, MSELoss, SoftmaxLoss

BinaryLoss

Binary cross-entropy loss, you must use a single neuron with a sigmoid activation function for the last layer. (CLASSIFICATION)

CategoricalLoss

Categorical cross-entropy loss, you can use it with combination of multiple neurons with a softmax activation function for the last layer. BE AWARE that backpropagation of softmax and CategoricalLoss is inefficient and expensive, so it's HIGHLY recommended to use SoftmaxLoss instead. (CLASSIFICATION)

MSELoss

Mean squared error loss. (REGRESSION)

SoftmaxLoss

Softmax activation layer combined with categorical cross-entropy loss. The derivatives of combining these two layers is computationally a lot cheaper than processing each one separately; so we merge them into one layer. THERE IS NO NEED TO USE A SOFTMAX AT LAST LAYER ANYMORE. (CLASSIFICATION)

Optimizers

We can take advantage of two optimizers; Momentum and Adam. Adam is superior to Momentum in most cases.

from nnfs.optimizers import Adam, Momentum

adam = Adam(learning_rate=0.01, decay=1e-5, beta_1=0.9, beta_2=0.999)  # Default Values
momentum = Momentum(learning_rate=0.01, decay=1e-5, beta=0.1) # Default Values
model = Model(loss=SoftmaxLoss(), optimizer=adam, metric=Accuracy())

You must play around with the parameters of the optimizer to find the best one for your model. decay is the learning rate decay.

Metrics

We have 5 metrics which we use during training and evaluation phase.

from nnfs.metrics import Accuracy, ExplainedVariance, FScore, Precision, Recall

model = Model(loss=SoftmaxLoss(), optimizer=adam, metric=Accuracy())
validation_metric = ExplainedVariance.evaluate(y_test, model.predict(X_test))

ExplainedVariance is for regression tasks, all the others are for classifications. Precision, Recall and FScore are calculated by Macro averaging.

Layers

We provide FullyConnectedLayer and DropoutLayer.

from nnfs.layer import Dropout, Layer

model.add(Layer(64, 128)) # FullyConnectedLayer with 64 inputs and 128 outputs.
model.add(Dropout(0.2)) # DropoutLayer with 20% dropout.
model.add(Layer(2, 64, w_l2=5e-4, b_l2=5e-4)) # FullyConnectedLayer with L2 regularization. (weight and bias)
model.add(Layer(2, 64, w_l2=5e-4, b_l2=5e-4, w_l1=5e-4, b_l1=5e-4)) # FullyConnectedLayer with L1 and L2 regularization. (weight and bias)

Activation Functions

nnfs-implementation provides six activation functions:

from nnfs.activations import LeakyReLU, Linear, ReLU, Sigmoid, Softmax, TanH

model.add(ReLU())
model.add(LeakyReLU(negative_slope=0.1))

Linear should be used for regression in the last layer.

Training

model.fit(X_train, y_train, epochs=1_000, batch_size=None, shuffle=True)  # Default Values
model.fit(X_train, y_train, epochs=20, batch_size=512, shuffle=True) 

If batch_size is None, then the whole dataset will be used.

Prediction

y_pred = model.predict(X_test)
y_pred_proba = model.predict_proba(X_test)

Loading & Saving

To load a model, simply use:

from nnfs.model import Model

model = Model.load("model.pkl")

and to save the model:

model.save("model.pkl", cleanup=True)

cleanup option will remove some unnecessary attributes that leads to dramatically reduced saved model size. default value is True.

Caution

This project is only for demonstration purpose. It works on numpy arrays (CPU) so you will face a poor performance.

Contributing

Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.

Contact me: OSS@Mahyar24.com :)

License

GNU GPLv3

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

nnfs-implementation-0.0.1.tar.gz (25.4 kB view details)

Uploaded Source

Built Distribution

nnfs_implementation-0.0.1-py3-none-any.whl (25.7 kB view details)

Uploaded Python 3

File details

Details for the file nnfs-implementation-0.0.1.tar.gz.

File metadata

  • Download URL: nnfs-implementation-0.0.1.tar.gz
  • Upload date:
  • Size: 25.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.13 CPython/3.9.5 Linux/5.13.0-52-generic

File hashes

Hashes for nnfs-implementation-0.0.1.tar.gz
Algorithm Hash digest
SHA256 a1cdf27b8b5453fd61a759516a4b2ec8b6955a26fea0587c8d9219c0da44779a
MD5 a00aa335ee7d3efe4a0e94f211e03bde
BLAKE2b-256 2a0756d3c6e5b7aa607a2bc9aac4a7470d821ee2646a3df7929cd283c9bd4917

See more details on using hashes here.

File details

Details for the file nnfs_implementation-0.0.1-py3-none-any.whl.

File metadata

File hashes

Hashes for nnfs_implementation-0.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 acd5d5f44318a1450228b67fc9a6d04ac5044324b0752c760ead62eb4b5e4c00
MD5 af84ba9800026d2a978a5924cf96577c
BLAKE2b-256 c65cfccb5e59f298b36abc29b79a3ab1b855c2f1a952ae3f78919ec3deebe2ce

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page