Skip to main content

A quick filetree uploader for the databricks filestore, local to cloud.

Project description

Databricks FileStore Uploader

A convenience tool for uploading local directories to the DBFS FileStore


Upload Local Files

This tool uploads specified source directories to specified paths in the DBFS FileStore.

Create a file adjacent to main.py called input.yml, and copy the following:

host : <workspace-url>
token: <user-access-token>

# THIS TOOL CREATES OR OVERWRITES THE TARGET FOLDER!
# Folder paths don't start with './'
# Folder paths don't end with '/'
payloads:
  some-local-folder : some-filestore-folder
  some-other-local-folder : some-other/filestore-folder
  # ...

Replace <user-access-token> with your token in databricks, and replace the example text under payloads: one or more "source : destination" pairs.
These local source directory paths are relative to main.py.

You can then upload all files contained in the local source folders to their specified destinations with:

./main.py

Though make sure you have PyYAML installed.


Verify Uploads

In a databricks notebook, you can verify uploads with:

display(dbutils.fs.ls("/FileStore/some-filestore-folder"))

Please Note❗
This tool CREATES or OVERWRITES it's target dbfs:/FileStore path. Only target folders that are meant to be overwritten, or don't exist.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

databricks-filestore-uploader-0.1.0.tar.gz (5.2 kB view hashes)

Uploaded Source

Built Distribution

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page