Skip to main content

Lookahead Keys Attention

Project description

Lookahead Keys Attention (wip)

Causal Attention with Lookahead Keys

Installation

pip install lookahead-keys-attention

Usage

import torch
from lookahead_keys_attention import Castle

# Initialize the Castle attention module
model = Castle(
    dim=512,           # input dimension
    heads=8,           # number of attention heads
    dim_head=64,       # dimension per head
    use_triton=None    # auto-detect CUDA for Triton optimization
)

# Example with CUDA sequence
batch_size = 2
seq_len = 128
dim = 512

# Move to CUDA if available
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
model = model.to(device)

# Input sequence
x = torch.randn(batch_size, seq_len, dim).to(device)

# Forward pass
output = model(x)  # Shape: [batch_size, seq_len, dim]

# For inference with caching (single token generation)
cache = None
for i in range(seq_len):
    token = x[:, i:i+1, :]  # Single token
    output, cache = model(token, cache=cache, return_next_cache=True)

Citations

@inproceedings{Song2025CausalAW,
    title   = {Causal Attention with Lookahead Keys},
    author  = {Zhuoqing Song and Peng Sun and Huizhuo Yuan and Quanquan Gu},
    year    = {2025},
    url     = {https://api.semanticscholar.org/CorpusID:281218151}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lookahead_keys_attention-0.0.6.tar.gz (154.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lookahead_keys_attention-0.0.6-py3-none-any.whl (11.2 kB view details)

Uploaded Python 3

File details

Details for the file lookahead_keys_attention-0.0.6.tar.gz.

File metadata

  • Download URL: lookahead_keys_attention-0.0.6.tar.gz
  • Upload date:
  • Size: 154.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.23

File hashes

Hashes for lookahead_keys_attention-0.0.6.tar.gz
Algorithm Hash digest
SHA256 6cc2dceaefad48c9f3395851a8670f89f70abe2a2309eaa29f093e1f4e065c87
MD5 4ac3a8f94510aaefa83ad776f9173da7
BLAKE2b-256 ac1860a7ee00a44ddff8a50f25c7851529f3b111892c4c65e93aa5f16be0dd7b

See more details on using hashes here.

File details

Details for the file lookahead_keys_attention-0.0.6-py3-none-any.whl.

File metadata

File hashes

Hashes for lookahead_keys_attention-0.0.6-py3-none-any.whl
Algorithm Hash digest
SHA256 b2bc82f9ab9e8ad259d1ed49753df45c4a8b491000d542e3e2ad4b68f03990d3
MD5 8a4c2f5439b957a89c7e075d41f9b933
BLAKE2b-256 b527fa01b96fffd315bc5eebcc75549e025f44f95ce3516eb49f16696f10d64d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page