Skip to main content

Download and analyze historical futures data from TradeStation API

Project description

TradeStation Historical Data Downloader

Automated download of 1-minute futures data from TradeStation API with incremental updates, rate limiting, and Parquet storage.

Features

  • OAuth2 Authentication - Automatic token refresh
  • Incremental Updates - Only downloads new bars since last run
  • Parallel Downloads - Up to 4 concurrent symbol downloads (configurable)
  • Rate Limiting - Respects API limits with configurable delays
  • Parquet Storage - Fast, compressed columnar format
  • Resume Capability - Interrupted downloads can be resumed
  • All US Futures - Pre-configured list of ~50 continuous contracts

Quick Start

1. Install

# From PyPI (recommended)
pip install tradestation-downloader

# Or from source using uv
pip install uv
uv sync

# Or standard pip from source
pip install -e .

# For development (includes pytest, ruff)
pip install -e ".[dev]"

2. Get TradeStation API Credentials

  1. Go to TradeStation Developer Portal
  2. Create an application
  3. Get your client_id, client_secret, and refresh_token

3. Configure

Option A - Interactive setup (recommended):

tradestation-auth

Option B - Manual setup:

cp config.yaml.template config.yaml
# Edit config.yaml with your credentials

4. Run

After install, CLI commands are available:

# Download all configured symbols (incremental)
tradestation-download

# Download specific symbols only
tradestation-download -s "@ES" "@NQ" "@CL"

# Full download (ignore existing data)
tradestation-download --full

# Use daily or monthly partitioned storage
tradestation-download --storage-format daily
tradestation-download --storage-format monthly

# Use different compression (default: zstd)
tradestation-download --compression snappy

# List all default symbols
tradestation-download --list-symbols

# List symbol categories
tradestation-download --list-categories

# Download specific category
tradestation-download --category index

# Save without datetime index (raw format)
tradestation-download -s @ES --no-datetime-index

# Use more parallel workers for faster downloads (default: 4)
tradestation-download -w 8

# Sequential download (no parallelism)
tradestation-download -w 1

Note: On Windows, the @ symbol has special meaning in CMD/PowerShell. Always quote symbols: "@ES" instead of @ES.

Or run directly with Python:

python tradestation_downloader.py
python tradestation_downloader.py -s "@ES" "@NQ" "@CL"

Configuration

Edit config.yaml:

tradestation:
  client_id: "YOUR_CLIENT_ID"
  client_secret: "YOUR_CLIENT_SECRET"
  refresh_token: "YOUR_REFRESH_TOKEN"

data_dir: "./data"
start_date: "2007-01-01"
interval: 1
unit: "Minute"
storage_format: "single"  # single, daily, or monthly
compression: "zstd"       # zstd, snappy, gzip, lz4, or none
max_workers: 4            # parallel download workers (1 = sequential)

symbols:
  - "@ES"    # E-mini S&P 500
  - "@NQ"    # E-mini Nasdaq 100
  # ... add more symbols

Output Format

Data is saved as Parquet files in the data_dir. Structure depends on storage format:

Single file (with datetime index):

data/
├── @ES_index_1_1min.parquet
├── @NQ_index_1_1min.parquet
└── @CL_index_1_1min.parquet

Monthly partitioned (default with datetime index):

data/
├── @ES_index_1/
│   ├── year_month=2024-01/data-0.parquet
│   ├── year_month=2024-02/data-0.parquet
│   └── ...
└── @NQ_index_1/
    └── ...

Each file contains:

Column Type Description
datetime datetime Bar timestamp (UTC)
open float Opening price
high float High price
low float Low price
close float Closing price
volume int Total volume

DateTime Index Mode

By default, data is saved with datetime as the DataFrame index, enabling pandas time-series operations like resample() and time-based slicing. This adds an _index_1 suffix to folder names.

Default (recommended):

tradestation-download -s @ES

→ Creates @ES_index_1/ folder with datetime as index

Raw format (datetime as column):

tradestation-download -s @ES --no-datetime-index

→ Creates @ES/ folder with datetime as regular column

Why use datetime index?

  • df.resample('5min') works directly
  • Time slicing: df['2024-01':'2024-06']
  • Compatible with pandas time-series functions
  • Matches common conventions for OHLCV data

Folder structure with index (default):

data/
├── @ES_index_1/
│   ├── year_month=2024-01/data-0.parquet
│   └── year_month=2024-02/data-0.parquet
└── @NQ_index_1/
    └── ...

Loading Data

import pandas as pd

# Load single symbol (monthly partitioned with datetime index)
df = pd.read_parquet("data/@ES_index_1")
print(df.head())

# Resample to 5-minute bars (works because datetime is the index)
df_5min = df.resample('5min').agg({
    'open': 'first', 'high': 'max', 'low': 'min', 'close': 'last', 'volume': 'sum'
}).dropna()

# Time-based slicing
df_2024 = df['2024-01':'2024-12']

# Load single file format
df = pd.read_parquet("data/ES_1min.parquet")

Python API

pip install tradestation-downloader

Use programmatically in your project:

from tradestation import TradeStationDownloader, DownloadConfig, StorageFormat, Compression

config = DownloadConfig(
    client_id="your_client_id",
    client_secret="your_client_secret",
    refresh_token="your_refresh_token",
    symbols=["@ES"],
    data_dir="./data",
    start_date="2020-01-01",
    storage_format=StorageFormat.SINGLE,
    compression=Compression.ZSTD,  # or SNAPPY, GZIP, LZ4, NONE
    max_workers=4,  # parallel downloads (1 = sequential)
)

downloader = TradeStationDownloader(config)
data = downloader.download_all()

# Access download statistics
stats = downloader.stats
print(f"Downloaded {stats.bars_downloaded} bars")
print(f"Processed {stats.symbols_processed} symbols")
print(f"Errors: {stats.errors}")

# Use the data
es_df = data["@ES"]
print(es_df.head())

Or load from config file:

from tradestation import TradeStationDownloader, load_config

config = load_config("config.yaml")
config.symbols = ["@ES", "@NQ"]  # Override symbols

downloader = TradeStationDownloader(config)
downloader.download_all()

Scheduling Daily Updates

Linux/Mac (cron)

# Edit crontab
crontab -e

# Add line to run daily at 6 AM
0 6 * * * cd /path/to/tradestation_downloader && python tradestation_downloader.py >> logs/download.log 2>&1

Windows (Task Scheduler)

  1. Open Task Scheduler
  2. Create Basic Task
  3. Set trigger: Daily at 6:00 AM
  4. Action: Start a program
  5. Program: python
  6. Arguments: C:\path\to\tradestation_downloader.py
  7. Start in: C:\path\to\tradestation_downloader

Data Validation

import pandas as pd
from pathlib import Path

data_dir = Path("./data")

for f in data_dir.glob("*_1min.parquet"):
    df = pd.read_parquet(f)
    symbol = f.stem.replace("_1min", "")
    
    print(f"{symbol}:")
    print(f"  Date range: {df['datetime'].min()} to {df['datetime'].max()}")
    print(f"  Total bars: {len(df):,}")
    print(f"  Missing dates: {df['datetime'].diff().gt(pd.Timedelta(minutes=2)).sum()}")
    print()

Troubleshooting

"401 Unauthorized" Error

Your refresh token may have expired. Run tradestation-auth to get a new one.

"429 Rate Limited" Error

The script handles this automatically, but if persistent:

  • Reduce max_workers (e.g., -w 2)
  • Increase rate_limit_delay in config
  • Reduce number of symbols per run

TradeStation allows 500 requests per 5-minute window for bar chart data. The default settings (4 workers, 0.2s delay) stay well within these limits.

Missing Data

Some symbols may not have data going back to 2007. Check TradeStation's data availability for specific contracts.

Large Download Size

1-minute data from 2007 is ~500MB-1GB per symbol. Total for all US futures: ~30-50GB.

Default Symbols

Run tradestation-download --list-symbols to see all default symbols:

  • Equity Index: @ES, @NQ, @YM, @RTY, @MES, @MNQ, etc.
  • Energy: @CL, @NG, @RB, @HO, @BRN
  • Metals: @GC, @SI, @HG, @PL, @PA
  • Treasury: @US, @TY, @FV, @TU, @UB, @TEN, @TWE
  • Grains: @C, @S, @W, @KW, @BO, @SM
  • Softs: @KC, @SB, @CT, @CC, @OJ, @LBR
  • Meats: @LC, @LH, @FC
  • Currency: @EC, @JY, @BP, @AD, @CD, @SF, @DX
  • Volatility: @VX
  • Crypto: @BTC, @ETH, @MBT, @MET

License

MIT License - Free to use and modify.

Support

For TradeStation API issues: TradeStation Developer Support

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tradestation_downloader-1.0.6.tar.gz (98.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tradestation_downloader-1.0.6-py3-none-any.whl (23.4 kB view details)

Uploaded Python 3

File details

Details for the file tradestation_downloader-1.0.6.tar.gz.

File metadata

  • Download URL: tradestation_downloader-1.0.6.tar.gz
  • Upload date:
  • Size: 98.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.14

File hashes

Hashes for tradestation_downloader-1.0.6.tar.gz
Algorithm Hash digest
SHA256 4db5dbddf7cda8cf1fca49a6fd57d918f80dc6b52173d58ac9428edcebb9aeeb
MD5 604d112864cec351d32ee170b6c6c7ba
BLAKE2b-256 0dd027550203473b7d1aa62149c33c570c8dcd2f33dce9ea9971edef22ebf919

See more details on using hashes here.

File details

Details for the file tradestation_downloader-1.0.6-py3-none-any.whl.

File metadata

File hashes

Hashes for tradestation_downloader-1.0.6-py3-none-any.whl
Algorithm Hash digest
SHA256 9ca42e17f4c6b6c421d039968163289379b907bb971371c4a3d861d326d3fc98
MD5 79c1febb2db344d8369321a1e18aeee9
BLAKE2b-256 61410dc87eb37076d00085379e471272fffecedd634482b8ff34513a2fe697fd

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page