Time Series Foundation Model - TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting
Project description
Time Series Foundation Model - TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting
The official code for ["TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)"].
TEMPO is one of the very first open source Time Series Foundation Models for forecasting task v1.0 version.
🚀 News
-
Oct 2024: 🚀 We've streamlined our code structure, enabling users to download the pre-trained model and perform zero-shot inference with a single line of code! Check out our demo for more details. Our model's download count on HuggingFace is now trackable!
-
Jun 2024: 🚀 We added demos for reproducing zero-shot experiments in Colab. We also added the demo of building the customer dataset and directly do the inference via our pre-trained foundation model: Colab
-
May 2024: 🚀 TEMPO has launched a GUI-based online demo, allowing users to directly interact with our foundation model!
-
May 2024: 🚀 TEMPO published the 80M pretrained foundation model in HuggingFace!
-
May 2024: 🧪 We added the code for pretraining and inference TEMPO models. You can find a pre-training script demo in this folder. We also added a script for the inference demo.
-
Mar 2024: 📈 Released TETS dataset from S&P 500 used in multimodal experiments in TEMPO.
-
Mar 2024: 🧪 TEMPO published the project code and the pre-trained checkpoint online!
-
Jan 2024: 🚀 TEMPO paper get accepted by ICLR!
-
Oct 2023: 🚀 TEMPO paper released on Arxiv!
Build the environment
conda create -n tempo python=3.8
conda activate tempo
pip install -r requirements.txt
Script Demo
A streamlining example showing how to perform forecasting using TEMPO:
# Third-party library imports
import numpy as np
import torch
from numpy.random import choice
# Local imports
from models.TEMPO import TEMPO
model = TEMPO.load_pretrained_model(
device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu'),
repo_id = "Melady/TEMPO",
filename = "TEMPO-80M_v1.pth",
cache_dir = "./checkpoints/TEMPO_checkpoints"
)
input_data = np.random.rand(336) # Random input data
with torch.no_grad():
predicted_values = model.predict(input_data, pred_length=96)
print("Predicted values:")
print(predicted_values)
Demos
1. Reproducing zero-shot experiments on ETTh2:
Please try to reproduc the zero-shot experiments on ETTh2 [here on Colab].
2. Zero-shot experiments on customer dataset:
We use the following Colab page to show the demo of building the customer dataset and directly do the inference via our pre-trained foundation model: [Colab]
3. Online demo:
Please try our foundation model demo [here].
Practice on your end
We also updated our models on HuggingFace: [Melady/TEMPO].
Get Data
Download the data from [Google Drive] or [Baidu Drive], and place the downloaded data in the folder./dataset
. You can also download the STL results from [Google Drive], and place the downloaded data in the folder./stl
.
Run TEMPO
Pre-Training Stage
bash [ecl, etth1, etth2, ettm1, ettm2, traffic, weather].sh
Test/ Inference Stage
After training, we can test TEMPO model under the zero-shot setting:
bash [ecl, etth1, etth2, ettm1, ettm2, traffic, weather]_test.sh
Pre-trained Models
You can download the pre-trained model from [Google Drive] and then run the test script for fun.
TETS dataset
Here is the prompts use to generate the coresponding textual informaton of time series via [OPENAI ChatGPT-3.5 API]
The time series data are come from [S&P 500]. Here is the EBITDA case for one company from the dataset:
Example of generated contextual information for the Company marked above:
You can download the processed data with text embedding from GPT2 from: [TETS].
Contact
Feel free to connect DefuCao@USC.EDU / YanLiu.CS@USC.EDU if you’re interested in applying TEMPO to your real-world application.
Cite our work
@inproceedings{
cao2024tempo,
title={{TEMPO}: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting},
author={Defu Cao and Furong Jia and Sercan O Arik and Tomas Pfister and Yixiang Zheng and Wen Ye and Yan Liu},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024},
url={https://openreview.net/forum?id=YH5w12OUuU}
}
@article{
Jia_Wang_Zheng_Cao_Liu_2024,
title={GPT4MTS: Prompt-based Large Language Model for Multimodal Time-series Forecasting},
volume={38},
url={https://ojs.aaai.org/index.php/AAAI/article/view/30383},
DOI={10.1609/aaai.v38i21.30383},
number={21},
journal={Proceedings of the AAAI Conference on Artificial Intelligence},
author={Jia, Furong and Wang, Kevin and Zheng, Yixiang and Cao, Defu and Liu, Yan},
year={2024}, month={Mar.}, pages={23343-23351}
}
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file tempodev-0.0.7.tar.gz
.
File metadata
- Download URL: tempodev-0.0.7.tar.gz
- Upload date:
- Size: 91.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.1 CPython/3.12.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 22f85a08d77793c17c52b4533bc27ef0dcb0feb499d974d0b04e52e177aebc5a |
|
MD5 | 73a2278664804b5427f0943073492bcd |
|
BLAKE2b-256 | e0e4a5473c661525b10375a143ae879f91d8d67a61b86eeff0e534a8dfc8ce65 |
Provenance
The following attestation bundles were made for tempodev-0.0.7.tar.gz
:
Publisher:
publish.yml
on idevede/TEMPO_for_pip
-
Statement type:
https://in-toto.io/Statement/v1
- Predicate type:
https://docs.pypi.org/attestations/publish/v1
- Subject name:
tempodev-0.0.7.tar.gz
- Subject digest:
22f85a08d77793c17c52b4533bc27ef0dcb0feb499d974d0b04e52e177aebc5a
- Sigstore transparency entry: 151173340
- Sigstore integration time:
- Predicate type:
File details
Details for the file tempodev-0.0.7-py3-none-any.whl
.
File metadata
- Download URL: tempodev-0.0.7-py3-none-any.whl
- Upload date:
- Size: 115.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.1 CPython/3.12.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f7fd63faa91a9913ae305b908beb2a088da8094ccee2d8e6021e573184a648fc |
|
MD5 | f16188d00df47be4a2741f1584dd0d6c |
|
BLAKE2b-256 | 0c5eb788ccda52794c8591fdf2f86f1fde2d3f10fcdb1575560fd7c62ed007f7 |
Provenance
The following attestation bundles were made for tempodev-0.0.7-py3-none-any.whl
:
Publisher:
publish.yml
on idevede/TEMPO_for_pip
-
Statement type:
https://in-toto.io/Statement/v1
- Predicate type:
https://docs.pypi.org/attestations/publish/v1
- Subject name:
tempodev-0.0.7-py3-none-any.whl
- Subject digest:
f7fd63faa91a9913ae305b908beb2a088da8094ccee2d8e6021e573184a648fc
- Sigstore transparency entry: 151173341
- Sigstore integration time:
- Predicate type: