Skip to main content

A helper library for exploring and fetching data from the U.S. Census Bureau API.

Project description

cendat: A Python Helper for the Census API

Introduction

cendat is a Python library designed to simplify the process of exploring and retrieving data from the U.S. Census Bureau’s API. It provides a high-level, intuitive workflow for discovering available datasets, filtering geographies and variables, and fetching data concurrently.

The library handles the complexities of the Census API’s structure, such as geographic hierarchies and inconsistent product naming, allowing you to focus on getting the data you need.

Workflow

The library is designed around a simple, four-step “List -> Set -> Get -> Convert” workflow:

  1. List: Use the list_* methods (list_products, list_geos, list_variables) with patterns to explore what’s available and filter down to what you need.
  2. Set: Use the set_* methods (set_products, set_geos, set_variables) to lock in your selections. You can call these methods without arguments to use the results from your last “List” call.
  3. Get: Call the get_data() method to build and execute all the necessary API calls. This method handles complex geographic requirements automatically and utilizes thread pooling for speed.
  4. Convert & Analyze: Use the to_polars() or to_pandas() methods on the response object to get your data in a ready-to-use DataFrame format. The response object also includes a powerful tabulate() method for quick, Stata-like frequency tables.

Installation

You can install cendat using pip.

pip install cendat

The library has optional dependencies for converting the response data into pandas or polars DataFrames. You can install the support you need:

Install with pandas support

pip install cendat[pandas]

Install with polars support

pip install cendat[polars]

Install with both

pip install cendat[all]

API Reference

CenDatHelper Class

This is the main class for building and executing queries.

__init__(self, years=None, key=None)

Initializes the helper object.

  • years (int | list[int], optional): The year or years of interest. Can be a single integer or a list of integers. Defaults to None.
  • key (str, optional): Your Census API key. Providing a key is recommended to avoid strict rate limits. Defaults to None.

set_years(self, years)

Sets the primary year or years for data queries.

  • years (int | list[int]): The year or years to set.

load_key(self, key=None)

Loads a Census API key for authenticated requests.

  • key (str, optional): The API key to load.

list_products(self, years=None, patterns=None, to_dicts=True, logic=all, match_in='title')

Lists available data products, filtered by year and search patterns.

  • years (int | list[int], optional): Filters products available for the specified year(s). Defaults to the years set on the object.
  • patterns (str | list[str], optional): Regex pattern(s) to search for within the product metadata.
  • to_dicts (bool): If True (default), returns a list of dictionaries with full product details. If False, returns a list of product titles.
  • logic (callable): The logic to use when multiple patterns are provided. Can be all (default) or any.
  • match_in (str): The field to match patterns against. Can be 'title' (default) or 'desc'.

set_products(self, titles=None)

Sets the active data products for the session.

  • titles (str | list[str], optional): The title or list of titles of the products to set. If None, it sets all products from the last list_products() call.

list_geos(self, to_dicts=False, patterns=None, logic=all)

Lists available geographies for the currently set products.

  • to_dicts (bool): If True, returns a list of dictionaries with full geography details. If False (default), returns a list of unique summary level (sumlev) strings.
  • patterns (str | list[str], optional): Regex pattern(s) to search for within the geography description.
  • logic (callable): The logic to use when multiple patterns are provided. Can be all (default) or any.

set_geos(self, values=None, by='sumlev')

Sets the active geographies for the session.

  • values (str | list[str], optional): The geography values to set. If None, sets all geos from the last list_geos() call.
  • by (str): The key to use for matching values. Must be either 'sumlev' (default) or 'desc'.

list_variables(self, to_dicts=True, patterns=None, logic=all, match_in='label')

Lists available variables for the currently set products.

  • to_dicts (bool): If True (default), returns a list of dictionaries with full variable details. If False, returns a list of unique variable names.
  • patterns (str | list[str], optional): Regex pattern(s) to search for within the variable metadata.
  • logic (callable): The logic to use when multiple patterns are provided. Can be all (default) or any.
  • match_in (str): The field to match patterns against. Can be 'label' (default), 'name' or 'concept'.

set_variables(self, names=None)

Sets the active variables for the session.

  • names (str | list[str], optional): The name or list of names of the variables to set. If None, sets all variables from the last list_variables() call.

get_data(self, within='us', max_workers=100, timeout=30, preview_only=False)

Executes the API calls based on the set parameters and retrieves the data.

  • within (str | dict | list[dict], optional): Defines the geographic scope of the query.
    • For aggregate data, this can be a dictionary filtering parent geographies (e.g., {'state': '06'} for California). A list of dictionaries can be provided to query multiple scopes.
    • For microdata, this must be a dictionary specifying the target geography and its codes (e.g., {'public use microdata area': ['7701', '7702']}).
    • Defaults to 'us' for nationwide data where applicable.
  • max_workers (int, optional): The maximum number of concurrent threads to use for making API calls. For requests generating thousands of calls, it's wise to keep this value lower (e.g., < 100) to avoid server-side connection issues. Defaults to 100.
  • timeout (int, optional): Request timeout in seconds for each API call. Defaults to 30.
  • preview_only (bool, optional): If True, builds the list of API calls but does not execute them. Useful for debugging. Defaults to False.

CenDatResponse Class

A container for the data returned by CenDatHelper.get_data().

to_polars(self, schema_overrides=None, concat=False, destring=False)

Converts the raw response data into a list of Polars DataFrames.

  • schema_overrides (dict, optional): A dictionary mapping column names to Polars data types to override the inferred schema. Example: {'POP': pl.Int64}.
  • concat (bool): If True, concatenates all resulting DataFrames into a single DataFrame. Defaults to False.
  • destring (bool): If True, attempts to convert string representations of numbers into native numeric types. Defaults to False.

to_pandas(self, dtypes=None, concat=False, destring=False)

Converts the raw response data into a list of Pandas DataFrames.

  • dtypes (dict, optional): A dictionary mapping column names to Pandas data types, which is passed to the .astype() method. Example: {'POP': 'int64'}.
  • concat (bool): If True, concatenates all resulting DataFrames into a single DataFrame. Defaults to False.
  • destring (bool): If True, attempts to convert string representations of numbers into native numeric types. Defaults to False.

tabulate(self, *variables, strat_by=None, weight_var=None, weight_div=None, where=None, logic=all, digits=1)

Generates and prints a frequency table.

  • *variables (str): One or more column names to include in the tabulation.
  • strat_by (str, optional): A column name to stratify the results by. Percentages and cumulative stats will be calculated within each stratum. Defaults to None.
  • weight_var (str, optional): The name of the column to use for weighting. If None, each row has a weight of 1. Defaults to None.
  • weight_div (int, optional): A positive integer to divide the weight by, useful for pooled tabulations across multiple product vintages. weight_var must be provided if this is used. Defaults to None.
  • where (str | list[str], optional): A string or list of strings representing conditions to filter the data before tabulation. Each condition should be in a format like "variable operator value" (e.g., "AGE > 30"). Defaults to None.
  • logic (callable): The function to apply when multiple where conditions are provided. Use all for AND logic (default) or any for OR logic.
  • digits (int): The number of decimal places to display for floating-point numbers in the output table. Defaults to 1.

Usage Example: Stratified Tabulation

This example demonstrates how to retrieve ACS PUMS data and then use the tabulate method to create a stratified frequency table.

import os
from cendat import CenDatHelper
from dotenv import load_dotenv

load_dotenv()

# 1. Initialize and set up the query
cdh = CenDatHelper(years=[2022], key=os.getenv("CENSUS_API_KEY"))
cdh.list_products(patterns=r"acs/acs1/pums\b")
cdh.set_products()
cdh.set_geos(values="state", by="desc")
cdh.set_variables(names=["SEX", "AGEP", "ST", "PWGTP"])

# 2. Get data for two states
response = cdh.get_data(
    within={"state": ["06", "48"]}, # California and Texas
)

# 3. Create a stratified tabulation
# This shows the adult age distribution (AGEP) for each sex (SEX),
# stratified by state (ST). The results are weighted by PWGTP.
print("Age Distribution by Sex, Stratified by State")
response.tabulate(
    "SEX", "AGEP",
    strat_by="ST",
    weight_var="PWGTP",
    where="AGEP > 17" # Filter for adults
)

# 4. Convert to DataFrame for further analysis
# The `destring=True` argument allows Polars to infer the schema
# for requested variables. It can also be controlled precisely via
# `schema_overrides`.
df = response.to_polars(concat=True, destring=True)
print(df.head())

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cendat-0.4.4.tar.gz (6.4 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

cendat-0.4.4-py3-none-any.whl (19.6 kB view details)

Uploaded Python 3

File details

Details for the file cendat-0.4.4.tar.gz.

File metadata

  • Download URL: cendat-0.4.4.tar.gz
  • Upload date:
  • Size: 6.4 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.2

File hashes

Hashes for cendat-0.4.4.tar.gz
Algorithm Hash digest
SHA256 6066acf2e414a6a3504b0a85a9deb5626ff4aa06509c8a3396cad6c92a88a1f8
MD5 4e7503417d2695395dd4d097cc042c9a
BLAKE2b-256 d507afbba5c6778015bddd16124e0646e4dccdb8304676e95c2359455a9fff91

See more details on using hashes here.

File details

Details for the file cendat-0.4.4-py3-none-any.whl.

File metadata

  • Download URL: cendat-0.4.4-py3-none-any.whl
  • Upload date:
  • Size: 19.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.2

File hashes

Hashes for cendat-0.4.4-py3-none-any.whl
Algorithm Hash digest
SHA256 4f4e68c78c5eb479f15b7c7f069eb3ba3629dedd97529c46ff39ec428897b379
MD5 ad77a89624eb88149f0e46029a79330a
BLAKE2b-256 2687ca728cabe6072933d411fa8e3dbefab0f98997c65eead899b5cfd889161b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page