Skip to main content

Analyze large datasets of point clouds recorded over time in an efficient way

Project description

Analyze large datasets of point clouds recorded over time in an efficient way.

test status test coverage https://github.com/virtual-vehicle/pointcloudset/actions/workflows/doc.yml/badge.svg Docker PyPi badge PyPi badge JOSS badge code style ruff

Code | Documentation

Features

  • Handles point clouds over time

  • Directly read ROS files and many pointcloud file formats.

  • Generate a dataset from multiple pointclouds. For example from thousands of .las files.

  • Building complex pipelines with a clean and maintainable code

newpointcloud = pointcloud.limit("x",-5,5).filter("quantile","reflectivity", ">",0.5)
  • Apply arbitrary functions to datasets of point clouds

def isolate_target(frame: PointCloud) -> PointCloud:
   return frame.limit("x",0,1).limit("y",0,1)

def diff_to_pointcloud(pointcloud: PointCloud, to_compare: PointCloud) -> PointCloud:
   return pointcloud.diff("pointcloud", to_compare)

result = dataset.apply(isolate_target).apply(diff_to_pointcloud, to_compare=dataset[0])
  • Includes powerful aggregation method agg similar to pandas

dataset.agg(["min","max","mean","std"])
  • Support for large files with lazy evaluation and parallel processing

https://raw.githubusercontent.com/virtual-vehicle/pointcloudset/master/images/dask.gif
  • Support for numerical data per point (intensity, range, noise …)

  • Interactive 3D visualisation

https://raw.githubusercontent.com/virtual-vehicle/pointcloudset/master/images/tree.gif
  • High level processing based on dask, pandas, scipy, scikit-learn

  • Docker image is available

  • Optimised - but not limited to - automotive lidar

  • A command line tool to convert ROS 1 & 2 files

Use case examples

  • Post processing and analytics of a lidar dataset recorded by ROS

  • A collection of multiple lidar scans from a terrestrial laser scanner

  • Comparison of multiple point clouds to a ground truth

  • Analytics of point clouds over time

  • Developing algorithms on a single frame and then applying them to huge datasets

Installation with pip

Install python package with pip:

pip install pointcloudset

Installation with Docker

The easiest way to get started is to use the pre-build docker tgoelles/pointcloudset.

Quickstart

Reading ROS1 or ROS2 files:

import pointcloudset as pcs
from pathlib import Path
import urllib.request

urllib.request.urlretrieve(
   "https://github.com/virtual-vehicle/pointcloudset/raw/master/tests/testdata/test.bag", "test.bag"
)

dataset = pcs.Dataset.from_file(Path("test.bag"), topic="/os1_cloud_node/points", keep_zeros=False)
pointcloud = dataset[1]
pointcloud.plot("x", hover_data=True)

You can also generate a dataset from multiple pointclouds from formats like las, pcd, csv, and xyz.

PointCloud.to_file(...) currently writes csv, xyz, las, and pcd. For text formats, csv defaults to writing a header and also supports header=False; xyz defaults to headerless output and also supports header=True. When reading files, PointCloud.from_file(...) supports normalize_xyz (default False). If a file uses uppercase coordinate headers X, Y, Z, reading fails unless you pass normalize_xyz=True. This makes the conversion explicit while keeping internal processing consistent with lowercase x, y, z.

import pointcloudset as pcs
from pathlib import Path
import urllib.request

urllib.request.urlretrieve(
   "https://github.com/virtual-vehicle/pointcloudset/raw/master/tests/testdata/las_files/test_tree.las",
   "test_tree.las",
)
urllib.request.urlretrieve(
   "https://github.com/virtual-vehicle/pointcloudset/raw/master/tests/testdata/pcd_files/test_tree.pcd",
   "test_tree.pcd",
)

las_pc = pcs.PointCloud.from_file(Path("test_tree.las"), normalize_xyz=True)
pcd_pc = pcs.PointCloud.from_file(Path("test_tree.pcd"))
dataset = pcs.Dataset.from_instance("pointclouds", [las_pc, pcd_pc])
pointcloud = dataset[1]

pointcloud.plot("z", hover_data=True)

CLI to convert ROS1 and ROS2 files: pointcloudset convert

The package includes a CLI to convert pointclouds in ROS1 and ROS2 files into pointcloudset directories or native file formats. It currently writes csv, xyz, las, and pcd files and handles both mcap and db3 ROS2 inputs.

pointcloudset convert test.bag --output-format las --output-dir converted_las
https://raw.githubusercontent.com/virtual-vehicle/pointcloudset/master/images/cli_demo.gif

You can view PointCloud2 messages with

pointcloudset topics test.bag

Tipp: If you have uv installed you can simply run:

uvx pointcloudset --help

Citation and contact

orcid Thomas Gölles email: thomas.goelles@v2c2.at

Please cite our JOSS paper if you use pointcloudset.

@article{Goelles2021,
  doi = {10.21105/joss.03471},
  url = {https://doi.org/10.21105/joss.03471},
  year = {2021},
  publisher = {The Open Journal},
  volume = {6},
  number = {65},
  pages = {3471},
  author = {Thomas Goelles and Birgit Schlager and Stefan Muckenhuber and Sarah Haas and Tobias Hammer},
  title = {`pointcloudset`: Efficient Analysis of Large Datasets of Point Clouds Recorded Over Time},
  journal = {Journal of Open Source Software}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pointcloudset-0.13.0.tar.gz (36.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

pointcloudset-0.13.0-py3-none-any.whl (50.8 kB view details)

Uploaded Python 3

File details

Details for the file pointcloudset-0.13.0.tar.gz.

File metadata

  • Download URL: pointcloudset-0.13.0.tar.gz
  • Upload date:
  • Size: 36.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.11.9 {"installer":{"name":"uv","version":"0.11.9","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for pointcloudset-0.13.0.tar.gz
Algorithm Hash digest
SHA256 1df5120c0c794dd23bc90475b759dac18f9a4d249c9ddc318679a5a9297917e9
MD5 8dc3fb2160598905552e925b54892549
BLAKE2b-256 e9e8cf46837702cac09abf9bab478d380f0eb07b4ffe9744cc45c7f48c19d265

See more details on using hashes here.

File details

Details for the file pointcloudset-0.13.0-py3-none-any.whl.

File metadata

  • Download URL: pointcloudset-0.13.0-py3-none-any.whl
  • Upload date:
  • Size: 50.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.11.9 {"installer":{"name":"uv","version":"0.11.9","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for pointcloudset-0.13.0-py3-none-any.whl
Algorithm Hash digest
SHA256 61392dea1eeb480fb4d320c18503a54e3203c36b1fd3c8e9e75d22fec5152133
MD5 e7f7ac9f20b50aaa52c1c0535ddfc448
BLAKE2b-256 b8535fd93a43465e2fc202ee411c50c34d7d3191a57b727859d86184c3b0afed

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page