Skip to main content

Pulumi components for Pinecone BYOC clusters

Project description

Pinecone BYOC

PyPI version

Deploy Pinecone in your own cloud account (AWS or GCP) with full control over your infrastructure.

Demo

Quick Start

Interactive Setup

curl -fsSL https://raw.githubusercontent.com/pinecone-io/pulumi-pinecone-byoc/main/bootstrap.sh | bash

This will:

  1. Select your cloud provider (AWS or GCP)
  2. Check that required tools are installed (Python 3.12+, uv, cloud CLI, Pulumi, kubectl)
  3. Verify your cloud credentials
  4. Run an interactive setup wizard
  5. Generate a complete Pulumi project

Then deploy:

cd pinecone-byoc
pulumi up

Provisioning takes approximately 25-30 minutes.

Prerequisites

Common Tools (Required for Both AWS and GCP)

Tool Purpose Install
Python 3.12+ Runtime python.org
uv Package manager docs.astral.sh/uv
Pulumi Infrastructure pulumi.com/docs/install
kubectl Cluster access kubernetes.io

Cloud-Specific Tools

AWS

Tool Purpose Install
AWS CLI AWS access AWS docs

GCP

Tool Purpose Install
gcloud CLI GCP access GCP docs

Architecture

┌──────────────────────┐                    ┌───────────────────────────────────────────────┐
│                      │    operations      │              Your AWS/GCP Account (VPC)       │
│  Pinecone            │───────────────────▶│                                               │
│  Control Plane       │                    │  ┌─────────────┐  ┌─────────────────────────┐ │
│                      │◀───────────────────│  │  Control    │  │                         │ │
│                      │   cluster state    │  │  Plane      │  │    Cluster Manager      │ │
└──────────────────────┘                    │  └─────────────┘  │        (EKS/GKE)        │ │
                                            │  ┌─────────────┐  └─────────────────────────┘ │
                                            │  │  Heartbeat  │                              │
                                            │  └─────────────┘                              │
┌──────────────────────┐                    │  ┌───────────────────────────────────────────┐│
│                      │◀───────────────────│  │                                           ││
│  Pinecone            │   metrics &        │  │              Data Plane                   ││
│  Observability (DD)  │   traces           │  │                                           ││
│                      │                    │  └───────────────────────────────────────────┘│
└──────────────────────┘                    │  ┌──────────┐  ┌──────────┐  ┌─────────────┐  │
                                            │  │  S3/GCS  │  |   RDS/   |  │  Route53/   │  │
        No customer data                    │  │  Buckets │  │  Alloy   │  |  Cloud DNS  |  │
        leaves the cluster                  │  └──────────┘  └──────────┘  └─────────────┘  │
                                            └───────────────────────────────────────────────┘

How It Works

Pinecone BYOC uses a pull-based model for control plane operations:

  1. Index Operations - When you create, scale, or delete indexes through the Pinecone API, these operations are queued in Pinecone's control plane
  2. Pull & Execute - Components running in your cluster continuously pull pending operations and execute them locally
  3. Heartbeat & State - Your cluster pushes health status and state back to Pinecone for monitoring
  4. Observability - Metrics and traces (not customer data) are sent to Pinecone's observability platform (Datadog) for operational insights

This architecture ensures:

  • Your data never leaves your cloud account - only operational metrics and cluster state are transmitted
  • Network security policies remain under your control
  • All communication is outbound from your cluster - Pinecone never needs inbound access

Cluster Access

After deployment, configure kubectl:

AWS:

aws eks update-kubeconfig --region <region> --name <cluster-name>

GCP:

gcloud container clusters get-credentials <cluster-name> --region <region> --project <project-id>

The exact command is output after pulumi up completes.

Upgrades

Pinecone manages upgrades automatically in the background. If you need to trigger an upgrade manually:

pulumi up -c pinecone-version=<new-version>

Replace <new-version> with the target Pinecone version (e.g., main-abc1234).

Configuration

The setup wizard creates a Pulumi stack with these configurable options:

AWS Configuration Options:

Option Description Default
pinecone-version Pinecone release version (required)
region AWS region us-east-1
availability_zones AZs for high availability ["us-east-1a", "us-east-1b"]
vpc_cidr VPC IP range 10.0.0.0/16
deletion_protection Protect RDS/S3 from accidental deletion true
public_access_enabled Enable public endpoint (false = PrivateLink only) true
tags Custom tags to apply to all resources {}

GCP Configuration Options:

Option Description Default
pinecone-version Pinecone release version (required)
gcp_project GCP project ID (required)
region GCP region us-central1
availability_zones Zones for high availability ["us-central1-a", "us-central1-b"]
vpc_cidr VPC IP range 10.112.0.0/12
deletion_protection Protect AlloyDB/GCS from accidental deletion true
public_access_enabled Enable public endpoint (false = Private Service Connect only) true
labels Custom labels to apply to all resources {}

Edit Pulumi.<stack>.yaml to modify these values.

Programmatic Usage

For advanced users who want to integrate into existing infrastructure:

import pulumi
from pulumi_pinecone_byoc.aws import PineconeAWSCluster, PineconeAWSClusterArgs

config = pulumi.Config()

cluster = PineconeAWSCluster(
    "pinecone-aws-cluster",
    PineconeAWSClusterArgs(
        pinecone_api_key=config.require_secret("pinecone_api_key"),
        pinecone_version=config.require("pinecone_version"),
        region=config.require("region"),
        availability_zones=config.require_object("availability_zones"),
        vpc_cidr=config.get("vpc_cidr") or "10.0.0.0/16",
        deletion_protection=config.get_bool("deletion_protection") if config.get_bool("deletion_protection") is not None else True,
        public_access_enabled=config.get_bool("public_access_enabled") if config.get_bool("public_access_enabled") is not None else True,
        tags=config.get_object("tags") or {},
    ),
)

# Export useful values
pulumi.export("environment", cluster.environment.env_name)
pulumi.export("cluster_name", cluster.cell_name)
pulumi.export("kubeconfig", cluster.eks.kubeconfig)

Installation

Install from PyPI with cloud-specific dependencies:

# For AWS
uv add 'pulumi-pinecone-byoc[aws]'

# For GCP
uv add 'pulumi-pinecone-byoc[gcp]'

Troubleshooting

Preflight check failures

The setup wizard runs preflight checks for cloud quotas. If these fail:

AWS:

  1. VPC Quota - Request a limit increase via AWS Service Quotas
  2. Elastic IPs - Release unused EIPs or request a limit increase
  3. NAT Gateways - Request a limit increase
  4. EKS Clusters - Request a limit increase

GCP:

  1. APIs - Enable required APIs (compute, container, alloydb, storage, dns)
  2. Compute Quotas - Request CPU/disk quota increases via GCP Console
  3. GKE Clusters - Request a limit increase if at quota
  4. IP Addresses - Release unused static IPs or request more

Deployment failures

If pulumi up fails partway through:

pulumi refresh  # Sync state with actual resources
pulumi up       # Retry deployment

Cluster access issues

Ensure your cloud credentials match the account where the cluster is deployed:

# AWS
aws sts get-caller-identity

# GCP
gcloud auth list
gcloud config get-value project

Cleanup

To destroy all resources:

pulumi destroy

Note: If deletion_protection is enabled (default), you'll need to disable it first or manually delete protected resources.

Support

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pulumi_pinecone_byoc-0.2.2.tar.gz (52.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

pulumi_pinecone_byoc-0.2.2-py3-none-any.whl (70.8 kB view details)

Uploaded Python 3

File details

Details for the file pulumi_pinecone_byoc-0.2.2.tar.gz.

File metadata

  • Download URL: pulumi_pinecone_byoc-0.2.2.tar.gz
  • Upload date:
  • Size: 52.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for pulumi_pinecone_byoc-0.2.2.tar.gz
Algorithm Hash digest
SHA256 305c9718f208fed6236843a7c1b43082c38775227c56220dc4461b85a4929e54
MD5 aedc9fb0e62c5b7d8fce277ae35e856a
BLAKE2b-256 b460c8c713b59edd1d0341a0af58fae665197bf22e3fe42d291bba1a592d7991

See more details on using hashes here.

Provenance

The following attestation bundles were made for pulumi_pinecone_byoc-0.2.2.tar.gz:

Publisher: release.yaml on pinecone-io/pulumi-pinecone-byoc

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file pulumi_pinecone_byoc-0.2.2-py3-none-any.whl.

File metadata

File hashes

Hashes for pulumi_pinecone_byoc-0.2.2-py3-none-any.whl
Algorithm Hash digest
SHA256 44a19b18ec4f20f9870fe77bbe9c450ef8306b33b8e0844e2dfa180487115750
MD5 6988590b0546a1160cd9b6991e009cfb
BLAKE2b-256 41f1e4a8b57ad214bf47d615a967b2302dc890ceb8546cae1e8d5a3277836194

See more details on using hashes here.

Provenance

The following attestation bundles were made for pulumi_pinecone_byoc-0.2.2-py3-none-any.whl:

Publisher: release.yaml on pinecone-io/pulumi-pinecone-byoc

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page