Skip to main content

Python support for Parquet file format

Project description

fastparquet is a python implementation of the parquet format, aiming integrate into python-based big data work-flows. It is used implicitly by the projects Dask, Pandas and intake-parquet.

We offer a high degree of support for the features of the parquet format, and very competitive performance, in a small install size and codebase.

Details of this project, how to use it and comparisons to other work can be found in the documentation.


(all development is against recent versions in the default anaconda channels and/or conda-forge)


  • numpy

  • pandas

  • cython >= 0.29.23 (if building from pyx files)

  • cramjam

  • fsspec

Supported compression algorithms:

  • Available by default:

    • gzip

    • snappy

    • brotli

    • lz4

    • zstandard

  • Optionally supported


Install using conda, to get the latest compiled version:

conda install -c conda-forge fastparquet

or install from PyPI:

pip install fastparquet

You may wish to install numpy first, to help pip’s resolver. This may install an appropriate wheel, or compile from source. For the latter, you will need a suitable C compiler toolchain on your system.

You can also install latest version from github:

pip install git+

in which case you should also have cython to be able to rebuild the C files.


Please refer to the documentation.


from fastparquet import ParquetFile
pf = ParquetFile('myfile.parq')
df = pf.to_pandas()
df2 = pf.to_pandas(['col1', 'col2'], categories=['col1'])

You may specify which columns to load, which of those to keep as categoricals (if the data uses dictionary encoding). The file-path can be a single file, a metadata file pointing to other data files, or a directory (tree) containing data files. The latter is what is typically output by hive/spark.


from fastparquet import write
write('outfile.parq', df)
write('outfile2.parq', df, row_group_offsets=[0, 10000, 20000],
      compression='GZIP', file_scheme='hive')

The default is to produce a single output file with a single row-group (i.e., logical segment) and no compression. At the moment, only simple data-types and plain encoding are supported, so expect performance to be similar to numpy.savez.


This project forked in October 2016 from parquet-python, which was not designed for vectorised loading of big data or parallel access.

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fastparquet-2023.4.0.tar.gz (393.3 kB view hashes)

Uploaded source

Built Distributions

fastparquet-2023.4.0-cp311-cp311-win_amd64.whl (616.0 kB view hashes)

Uploaded cp311

fastparquet-2023.4.0-cp310-cp310-win_amd64.whl (618.0 kB view hashes)

Uploaded cp310

fastparquet-2023.4.0-cp39-cp39-win_amd64.whl (620.2 kB view hashes)

Uploaded cp39

fastparquet-2023.4.0-cp38-cp38-win_amd64.whl (630.0 kB view hashes)

Uploaded cp38

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page