Nearmap AI Python Library for extracting AI features from aerial imagery
Project description
nmaipy - Nearmap AI Python Library
Extract building footprints, vegetation, damage assessments, and other AI features from Nearmap's aerial imagery using simple Python code.
What is nmaipy?
nmaipy (pronounced "en-my-pie") is a Python library that makes it easy for data scientists to access Nearmap's AI-powered geospatial data. Whether you're analyzing a few properties or processing millions of buildings across entire cities, nmaipy handles the complexity so you can focus on your analysis.
Supported countries: au (Australia), us (United States), nz (New Zealand), ca (Canada)
Quick Start for Data Scientists
1. Install
Option A: Install from PyPI
pip install nmaipy
Option B: Install from source (for development)
git clone https://github.com/nearmap/nmaipy.git
cd nmaipy
pip install -e .
Option C: Using conda
Minimal installation (core features only):
conda env create -f environment-minimal.yaml
conda activate nmaipy
Full installation (includes development and notebook tools):
conda env create -f environment.yaml
conda activate nmaipy
Option D: Install into existing conda environment
conda install -c conda-forge geopandas pandas numpy pyarrow psutil pyproj python-dotenv requests rtree shapely stringcase tqdm fsspec s3fs
pip install -e .
Additional options
For running notebooks with pip:
pip install -e ".[notebooks]"
For development with pip:
pip install -e ".[dev]"
2. Set your API key
export API_KEY=your_api_key_here
3. Run your first extraction
from nmaipy.exporter import NearmapAIExporter
# Extract building and vegetation data
exporter = NearmapAIExporter(
aoi_file='my_parcels.geojson', # Your areas of interest
output_dir='results', # Where to save outputs
country='au', # au, us, nz, or ca
packs=['building', 'vegetation'], # What features to extract
processes=4 # Parallel processing
)
exporter.run()
That's it! Your results will be saved as CSV or Parquet files in the output directory.
Note:
AOIExporteris available as a backward-compatible alias forNearmapAIExporter.
Common Use Cases
Urban Planning
Extract comprehensive data about buildings, vegetation coverage, and surface materials:
exporter = NearmapAIExporter(
aoi_file='city_blocks.geojson',
output_dir='urban_analysis',
country='au',
packs=['building', 'vegetation', 'surfaces', 'solar'],
save_features=True, # Get individual features, not just summaries
include_parcel_geometry=True # Keep boundaries for GIS analysis
)
Disaster Response
Assess damage after natural disasters like hurricanes or floods:
exporter = NearmapAIExporter(
aoi_file='affected_areas.geojson',
output_dir='damage_assessment',
country='us',
packs=['damage'],
since='2024-07-08', # Date range of the event
until='2024-07-11',
rapid=True, # Use rapid post-catastrophe imagery
save_features=True
)
Environmental Analysis
Study vegetation coverage and tree canopy:
exporter = NearmapAIExporter(
aoi_file='study_area.geojson',
output_dir='vegetation_study',
country='au',
packs=['vegetation'],
save_features=True # Get individual tree polygons
)
Market Research
Find properties with pools or solar panels:
exporter = NearmapAIExporter(
aoi_file='suburbs.geojson',
output_dir='market_analysis',
country='au',
packs=['pools', 'solar'],
include_parcel_geometry=True
)
Roof Age Analysis (US Only)
Predict roof installation dates using AI analysis of historical imagery.
Unified approach (recommended) - combines Feature API and Roof Age in one export:
exporter = NearmapAIExporter(
aoi_file='properties.geojson',
output_dir='unified_results',
country='us',
packs=['building'],
roof_age=True, # Include Roof Age API data
save_features=True
)
exporter.run()
Standalone approach - for roof age data only:
from nmaipy.roof_age_exporter import RoofAgeExporter
exporter = RoofAgeExporter(
aoi_file='properties.geojson',
output_dir='roof_age_results',
country='us',
threads=10,
output_format='both' # Generate both GeoParquet and CSV
)
exporter.run()
The Roof Age API uses machine learning to analyze multiple imagery captures over time, combined with building permit data and climate information, to predict when roofs were last installed or significantly renovated. Each roof feature includes:
- Predicted installation date
- Confidence score (trust score)
- Evidence type and number of captures analyzed
- Timeline of all imagery used in analysis
This is valuable for:
- Insurance underwriting and risk assessment
- Property valuation and market analysis
- Maintenance planning and capital budgeting
- Real estate due diligence
Available AI Features
Some of the more common AI packs are below - there are more and growing, available via API request or on the Nearmap help.nearmap.com page.
| Pack | Description | Example Use Cases |
|---|---|---|
building |
Building footprints and heights | Urban planning, property analysis |
vegetation |
Trees and vegetation coverage | Environmental studies, urban forestry |
surfaces |
Ground surface materials | Permeability studies, heat mapping |
pools |
Swimming pool detection | Compliance, market research |
solar |
Solar panel detection | Renewable energy assessment |
damage |
Post-disaster damage classification | Insurance, emergency response |
building_characteristics |
Detailed roof types, materials | Detailed property analysis |
Input Data Formats
nmaipy accepts areas of interest (AOIs) in several formats:
- GeoJSON: Standard geospatial format with polygons
- GeoPackage (GPKG): OGC standard for geospatial data
- Parquet / GeoParquet: Efficient columnar format for large datasets
- CSV: Simple format with WKT geometries (also supports TSV and PSV)
Your input file should contain polygon geometries representing the areas you want to analyze (parcels, census blocks, suburbs, etc.).
Output Data
The exporter writes results to {output_dir}/final/ with the following structure:
| File | Description |
|---|---|
rollup.csv or .parquet |
One row per AOI with summary statistics (counts, areas, confidences) |
{class}.csv |
Per-class attribute tables (e.g. roof.csv, building.csv) |
{class}_features.parquet |
Per-class GeoParquet with feature geometries (when save_features=True) |
features.parquet |
All features combined as GeoParquet (when save_features=True) |
buildings.csv or .parquet |
Per-building detail rows (when save_buildings=True) |
feature_api_errors.csv |
AOIs where the Feature API returned errors |
roof_age_errors.csv |
AOIs where the Roof Age API returned errors (US only) |
latency_stats.csv |
API timing diagnostics |
export_config.json |
Full record of export parameters and nmaipy version |
README.md |
Auto-generated data dictionary describing all output files and columns |
A {output_dir}/chunks/ directory holds intermediate per-chunk results during processing, enabling resume after interruption.
For detailed column-level documentation, refer to the auto-generated README.md inside each export's final/ directory.
S3 Output Support
nmaipy can write output directly to Amazon S3. Pass an s3:// URI as the output directory:
exporter = NearmapAIExporter(
aoi_file='properties.geojson',
output_dir='s3://my-bucket/nmaipy-results/',
country='us',
packs=['building'],
)
exporter.run()
AWS credentials are resolved automatically from environment variables (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY, AWS_DEFAULT_REGION) or ~/.aws/credentials. No additional nmaipy configuration is needed.
The cache_dir parameter also accepts S3 URIs for cloud-native workflows, though local caching is faster for iterative development.
Examples
Quick start — verify your setup with 10 US properties covering buildings, features, and roof age:
export API_KEY=your_api_key_here
python run_10_test.py
More examples — see examples.py for complete, working examples covering:
- Basic building/vegetation extraction
- Damage assessment (Hurricane Beryl)
- Urban planning (multi-pack)
- Vegetation analysis
- Pool detection
- Large area gridding
- Time series extraction
- Unified roof age + feature export
Example AOI files are provided in data/examples/:
sydney_parcels.geojson— Sydney CBD, Australiaus_parcels.geojson— Austin, Texas, USAlarge_area.geojson— 2km x 2km Melbourne area (triggers auto-gridding)
Working with Large Areas
nmaipy automatically handles large areas by:
- Splitting them into manageable grid cells
- Processing in parallel
- Combining results seamlessly
For areas larger than 1 sq km, the library will automatically use gridding:
exporter = NearmapAIExporter(
aoi_file='large_region.geojson',
output_dir='large_area_results',
country='us',
packs=['building'],
aoi_grid_inexact=True, # Allow mixing survey dates if needed
processes=16 # Use more processes for speed
)
Performance Tips
- Use parallel processing: Set
processesto the number of CPU cores available. - Tune chunk size:
chunk_sizecontrols how many AOIs are grouped into each parallel work unit (default: 500). Smaller values give finer-grained parallelism and cheaper resume after interruption; larger values reduce overhead. - Cache API responses: Use
cache_dirto persist API responses to a directory. On subsequent runs with different parameters (e.g. different packs), cached responses are reused without re-fetching. By default, cache is stored in{output_dir}/cache/. - Filter by date: Use
sinceanduntilto restrict to specific time periods, reducing data volume.
Command Line Interface
Feature API Export
python nmaipy/exporter.py \
--aoi-file "parcels.geojson" \
--output-dir "results" \
--country us \
--packs building vegetation \
--save-features \
--roof-age
Key options:
--packs: AI packs to extract (building, vegetation, surfaces, pools, solar, damage, etc.)--roof-age: Include Roof Age API data (US only)--save-features: Save per-class GeoParquet files with feature geometries--save-buildings: Save per-building detail rows--tabular-file-format: Format for tabular output files — rollup, buildings, and per-class attribute files (csvorparquet, default:csv)--cache-dir: Directory for caching API responses--no-cache: Disable caching entirely--primary-decision: Feature selection method (largest_intersection,nearest,optimal)--since/--until: Filter by survey date range--max-retries: Maximum API retry attempts (default: 10)
Run python nmaipy/exporter.py --help for all options.
Standalone Roof Age Export (US Only)
python -m nmaipy.roof_age_exporter \
--aoi-file "us_properties.geojson" \
--output-dir "roof_age_results" \
--country us \
--processes 4 \
--output-format both
Run python -m nmaipy.roof_age_exporter --help for all options.
Getting Help
- Examples: See
examples.pyfor common use cases - Installation: See
INSTALL.mdfor detailed installation options - Notebooks: Check the
notebooks/directory for Jupyter notebook tutorials - Issues: Report bugs or request features on GitHub
Requirements
- Python 3.12+
- Nearmap API key (contact Nearmap for access)
- 4GB+ RAM recommended for large extractions
- AWS credentials for S3 output (optional)
Advanced: Building a Conda Package
For system administrators who want to create a local conda package:
conda build conda.recipe
conda install --use-local nmaipy
This will create a conda package that can be shared internally or uploaded to a conda channel.
License
See LICENSE file for details.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file nmaipy-4.5.0a2.tar.gz.
File metadata
- Download URL: nmaipy-4.5.0a2.tar.gz
- Upload date:
- Size: 157.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
a8418fc01b7dbf639604592d5317d9c16a3f1714118858f22317d19c8cbb1eef
|
|
| MD5 |
f698bb98f6eac1f84304964b6f02912f
|
|
| BLAKE2b-256 |
08963a54cb7980b9203632f8da6bfa3700711960e0d95758cf0ee49ff281ee55
|
File details
Details for the file nmaipy-4.5.0a2-py3-none-any.whl.
File metadata
- Download URL: nmaipy-4.5.0a2-py3-none-any.whl
- Upload date:
- Size: 160.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5878c1712d3756c61e298f220739be8d5a7aa51dc7018e419ea3bafc91bab284
|
|
| MD5 |
f0b5eb1b70e0505e0bc423eed96916d2
|
|
| BLAKE2b-256 |
efe4478fc8a4e90354c79eff6ed3384b8dc7229fd0d0a683594b48fef817b377
|