Skip to main content

Time Series Foundation Model - TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting

Project description

Time Series Foundation Model - TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting

preprint huggingface License: MIT

The official code for ["TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)"].

TEMPO is one of the very first open source Time Series Foundation Models for forecasting task v1.0 version.

🚀 News

  • Oct 2024: 🚀 We've streamlined our code structure, enabling users to download the pre-trained model and perform zero-shot inference with a single line of code! Check out our demo for more details. Our model's download count on HuggingFace is now trackable!

  • Jun 2024: 🚀 We added demos for reproducing zero-shot experiments in Colab. We also added the demo of building the customer dataset and directly do the inference via our pre-trained foundation model: Colab

  • May 2024: 🚀 TEMPO has launched a GUI-based online demo, allowing users to directly interact with our foundation model!

  • May 2024: 🚀 TEMPO published the 80M pretrained foundation model in HuggingFace!

  • May 2024: 🧪 We added the code for pretraining and inference TEMPO models. You can find a pre-training script demo in this folder. We also added a script for the inference demo.

  • Mar 2024: 📈 Released TETS dataset from S&P 500 used in multimodal experiments in TEMPO.

  • Mar 2024: 🧪 TEMPO published the project code and the pre-trained checkpoint online!

  • Jan 2024: 🚀 TEMPO paper get accepted by ICLR!

  • Oct 2023: 🚀 TEMPO paper released on Arxiv!

Build the environment

conda create -n tempo python=3.8
conda activate tempo
pip install -r requirements.txt

Script Demo

A streamlining example showing how to perform forecasting using TEMPO:

# Third-party library imports
import numpy as np
import torch
from numpy.random import choice
# Local imports
from models.TEMPO import TEMPO


model = TEMPO.load_pretrained_model(
        device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu'),
        repo_id = "Melady/TEMPO",
        filename = "TEMPO-80M_v1.pth",
        cache_dir = "./checkpoints/TEMPO_checkpoints"  
)

input_data = np.random.rand(336)    # Random input data
with torch.no_grad():
        predicted_values = model.predict(input_data, pred_length=96)
print("Predicted values:")
print(predicted_values)

Demos

1. Reproducing zero-shot experiments on ETTh2:

Please try to reproduc the zero-shot experiments on ETTh2 [here on Colab].

2. Zero-shot experiments on customer dataset:

We use the following Colab page to show the demo of building the customer dataset and directly do the inference via our pre-trained foundation model: [Colab]

3. Online demo:

Please try our foundation model demo [here].

Practice on your end

We also updated our models on HuggingFace: [Melady/TEMPO].

Get Data

Download the data from [Google Drive] or [Baidu Drive], and place the downloaded data in the folder./dataset. You can also download the STL results from [Google Drive], and place the downloaded data in the folder./stl.

Run TEMPO

Pre-Training Stage

bash [ecl, etth1, etth2, ettm1, ettm2, traffic, weather].sh

Test/ Inference Stage

After training, we can test TEMPO model under the zero-shot setting:

bash [ecl, etth1, etth2, ettm1, ettm2, traffic, weather]_test.sh

Pre-trained Models

You can download the pre-trained model from [Google Drive] and then run the test script for fun.

TETS dataset

Here is the prompts use to generate the coresponding textual informaton of time series via [OPENAI ChatGPT-3.5 API]

The time series data are come from [S&P 500]. Here is the EBITDA case for one company from the dataset:

Example of generated contextual information for the Company marked above:

You can download the processed data with text embedding from GPT2 from: [TETS].

Contact

Feel free to connect DefuCao@USC.EDU / YanLiu.CS@USC.EDU if you’re interested in applying TEMPO to your real-world application.

Cite our work

@inproceedings{
cao2024tempo,
title={{TEMPO}: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting},
author={Defu Cao and Furong Jia and Sercan O Arik and Tomas Pfister and Yixiang Zheng and Wen Ye and Yan Liu},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024},
url={https://openreview.net/forum?id=YH5w12OUuU}
}
@article{
   Jia_Wang_Zheng_Cao_Liu_2024, 
   title={GPT4MTS: Prompt-based Large Language Model for Multimodal Time-series Forecasting}, 
   volume={38}, 
   url={https://ojs.aaai.org/index.php/AAAI/article/view/30383}, 
   DOI={10.1609/aaai.v38i21.30383}, 
   number={21}, 
   journal={Proceedings of the AAAI Conference on Artificial Intelligence}, 
   author={Jia, Furong and Wang, Kevin and Zheng, Yixiang and Cao, Defu and Liu, Yan}, 
   year={2024}, month={Mar.}, pages={23343-23351} 
   }

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tempodev-0.0.7.tar.gz (91.8 kB view details)

Uploaded Source

Built Distribution

tempodev-0.0.7-py3-none-any.whl (115.5 kB view details)

Uploaded Python 3

File details

Details for the file tempodev-0.0.7.tar.gz.

File metadata

  • Download URL: tempodev-0.0.7.tar.gz
  • Upload date:
  • Size: 91.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.1 CPython/3.12.7

File hashes

Hashes for tempodev-0.0.7.tar.gz
Algorithm Hash digest
SHA256 22f85a08d77793c17c52b4533bc27ef0dcb0feb499d974d0b04e52e177aebc5a
MD5 73a2278664804b5427f0943073492bcd
BLAKE2b-256 e0e4a5473c661525b10375a143ae879f91d8d67a61b86eeff0e534a8dfc8ce65

See more details on using hashes here.

Provenance

The following attestation bundles were made for tempodev-0.0.7.tar.gz:

Publisher: publish.yml on idevede/TEMPO_for_pip

Attestations:

File details

Details for the file tempodev-0.0.7-py3-none-any.whl.

File metadata

  • Download URL: tempodev-0.0.7-py3-none-any.whl
  • Upload date:
  • Size: 115.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.1 CPython/3.12.7

File hashes

Hashes for tempodev-0.0.7-py3-none-any.whl
Algorithm Hash digest
SHA256 f7fd63faa91a9913ae305b908beb2a088da8094ccee2d8e6021e573184a648fc
MD5 f16188d00df47be4a2741f1584dd0d6c
BLAKE2b-256 0c5eb788ccda52794c8591fdf2f86f1fde2d3f10fcdb1575560fd7c62ed007f7

See more details on using hashes here.

Provenance

The following attestation bundles were made for tempodev-0.0.7-py3-none-any.whl:

Publisher: publish.yml on idevede/TEMPO_for_pip

Attestations:

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page