Skip to main content

Transformers at zeta scales

Project description

Minerva: Unleashing the Secrets of advanced Mathematics 🏛️🔢

Minerva is a groundbreaking language model that pushes the boundaries of mathematical understanding and problem-solving. Designed with an advanced math theme, Minerva embodies the spirit of renowned mathematicians such as Euclid, Pythagoras, and Archimedes. By harnessing their advanced wisdom, Minerva offers unparalleled capabilities in mathematical reasoning and exploration.


GitHub issues GitHub forks GitHub stars GitHub license

Share on Twitter Share on Facebook Share on LinkedIn

Share on Reddit Share on Hacker News Share on Pinterest Share on WhatsApp


Install

pip install minerva

Usage

import torch
from minerva import Minerva, Train

# Example usage
x = torch.randint(0, 20000, (1, 1024))

Minerva(x)

# or train
Train()

Training

To train Minerva, follow these steps:

  1. Configure the training settings by setting the environment variables:

    • ENTITY_NAME: Your wandb project name
    • OUTPUT_DIR: Specify the output directory for saving the weights (e.g., ./weights)
  2. Launch the training process using Deepspeed:

Accelerate Config
Accelerate launch train_distributed_accelerate.py

Dataset Building

To build a custom dataset for Minerva, you can preprocess the data using the build_dataset.py script. This script performs tasks such as pre-tokenization, data chunking, and uploading to the Huggingface hub. Here's an example command:

Dataset Description
Mathematical Web Pages Web pages containing mathematical expressions in MathJax format, cleaned to preserve math notation
arXiv 2 million arXiv papers up to Feb 2021, in LaTeX format
General Natural Language Data Same dataset used to pretrain PaLM models

The mathematical web pages and arXiv datasets focus on technical and mathematical content. The general natural language data provides a broad coverage of general language.

The paper states the mathematical web pages and arXiv each account for 47.5% of the total data. The remaining 5% is general natural language data which is a subset of what was used for PaLM pretraining.

Roadmap 🗺️📍

  • Create a dataset of ARXVIV papers

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

minerva_torch-0.0.1.tar.gz (15.4 kB view hashes)

Uploaded Source

Built Distribution

minerva_torch-0.0.1-py3-none-any.whl (15.1 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page