Streaming read/writes to Google Storage blobs with ascynchronous buffering.
gs-chunked-io: Streams for Google Storage
gs-chunked-io provides transparently chunked io streams for google storage objects. Writable streams are managed as multipart objects, composed when the stream is closed.
IO opperations are concurrent by default. The number of concurrent threads can be adjusted using the
parameter, or disabled entirely with
import gs_chunked_io as gscio from google.cloud.storage import Client client = Client() bucket = client.bucket("my-bucket") blob = bucket.get_blob("my-key") # Readable stream: with gscio.Reader(blob) as fh: fh.read(size) # Writable stream: with gscio.Writer("my_new_key", bucket) as fh: fh.write(data) # Process blob in chunks: for chunk in gscio.for_each_chunk(blob): my_chunk_processor(chunk) # Multipart copy with processing: dst_bucket = client.bucket("my_dest_bucket") with gscio.Writer("my_dest_key", dst_bucket) as writer: for chunk in gscio.for_each_chunk(blob) process_my_chunk(chunk) writer(chunk) # Extract .tar.gz on the fly: import gzip import tarfile with gscio.Reader(blob) as fh: gzip_reader = gzip.GzipFile(fileobj=fh) tf = tarfile.TarFile(fileobj=gzip_reader) for tarinfo in tf: process_my_tarinfo(tarinfo)
pip install gs-chunked-io
Please report bugs, issues, feature requests, etc. on GitHub.
Release history Release notifications | RSS feed
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
|Filename, size||File type||Python version||Upload date||Hashes|
|Filename, size gs-chunked-io-0.5.2.tar.gz (8.1 kB)||File type Source||Python version None||Upload date||Hashes View|