PoPE
Project description
PoPE-pytorch
Efficient implementation (and explorations) into polar coordinate positional embedding (PoPE) - from Gopalakrishnan et al. under Schmidhuber
Install
$ pip install PoPE-pytorch
Usage
import torch
from PoPE_pytorch import PoPE
# define pope
pope = PoPE(64, heads = 8)
# pass in sequence length
pos_embed = pope(1024)
# queries and keys in attention
q = torch.randn(1, 8, 1024, 64)
k = torch.randn(1, 8, 1024, 64)
# training
rotated_q, rotated_k = pope.apply_pope_to_qk(pos_embed, q, k)
# inference
rotated_q, rotated_k = pope.apply_pope_to_qk(pos_embed, q[..., -1:, :], k)
Fused Attention Similarity
import torch
from PoPE_pytorch import PoPE, compute_attn_similarity
# define pope
pope = PoPE(dim = 64, heads = 8).cuda()
# get rotations
pos_emb = pope(1024)
# queries and keys
q = torch.randn(1, 8, 1024, 64).cuda()
k = torch.randn(1, 8, 1024, 64).cuda()
# fused attention similarity, avoiding expanding 64 to 128
sim = compute_attn_similarity(q, k, pos_emb) # (1, 8, 1024, 1024)
attn = sim.softmax(dim = -1) # the usual in attention..
Fused Flash Attention
import torch
from PoPE_pytorch import PoPE, flash_attn_with_pope
# pope
pope = PoPE(dim = 32, heads = 8).cuda()
# queries, keys, values for attention
q = torch.randn(2, 8, 1024, 64).cuda()
k = torch.randn(2, 8, 1024, 64).cuda()
v = torch.randn(2, 8, 1024, 64).cuda()
pos_emb = pope(1024)
mask = torch.ones((2, 1024)).bool().cuda()
out = flash_attn_with_pope(q, k, v, pope = pos_emb, causal = True, mask = mask)
assert out.shape == (2, 8, 1024, 64)
Citations
@misc{gopalakrishnan2025decouplingwhatwherepolar,
title = {Decoupling the "What" and "Where" With Polar Coordinate Positional Embeddings},
author = {Anand Gopalakrishnan and Robert Csordás and Jürgen Schmidhuber and Michael C. Mozer},
year = {2025},
eprint = {2509.10534},
archivePrefix = {arXiv},
primaryClass = {cs.LG},
url = {https://arxiv.org/abs/2509.10534},
}
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
pope_pytorch-0.0.14.tar.gz
(213.5 kB
view details)
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file pope_pytorch-0.0.14.tar.gz.
File metadata
- Download URL: pope_pytorch-0.0.14.tar.gz
- Upload date:
- Size: 213.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.9.25
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
e88b26c113ac6b06004313b06a9a7bbf10a4088cb58d6d08046c84264f5f9371
|
|
| MD5 |
5f8ad2a49dc48ea22ff905221da5bece
|
|
| BLAKE2b-256 |
a21d2ef0001ba47ba9d7d8c9e9fcc7aeab859e782aabd9287b80b3d51138ceb9
|
File details
Details for the file pope_pytorch-0.0.14-py3-none-any.whl.
File metadata
- Download URL: pope_pytorch-0.0.14-py3-none-any.whl
- Upload date:
- Size: 13.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.9.25
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
173f2a02c2d37f468e4ee15811c7ab18cc7956bf22a93be2cfa890a4f12ed8d7
|
|
| MD5 |
d632dbe4185914ed8abba12eeaa565a6
|
|
| BLAKE2b-256 |
8826e3e0ac9807f540baf882dbbe023c127b7e0111d0ad4945b1d9ce9d5722b9
|