Skip to main content

A wrapper for google's existing google-cloud python package that aims to make using python inside the Google Cloud framework more intuitive.

Project description

NEW IN VERSION 2.0.0:

New Features:

  • Added polars functionality
  • Added custom GCP client functionality using the gcp_client arg

QL & Fixes:

  • Fixed N rows have been added to ... message in write_bigquery()
  • Deprecated unused GCS functionality

----------------------------

LICENSE

GPL-3 Summary:

You may copy, distribute and modify the software as long as you track changes/dates in source files. Any modifications to or software including (via compiler) GPL-licensed code must also be made available under the GPL along with build & install instructions. In other words, any derivative work of this software shall be released under the same GPL license as the original software, meaning the modified code must be exactly as free and open-source as the original.

----------------------------

ABOUT

gcloudy is a wrapper for Google's GCP Python package(s) that aims to make interacting with GCP and its services more intuitive, especially for new GCP users. In doing so, it adheres to pandas-like syntax for function/method calls.

The gcloudy package is not meant to be a replacement for GCP power-users, but rather an alternative for GCP users who are interested in using Python in GCP to deploy Cloud Functions and interact with certain GCP services, especially BigQuery and Google Cloud Storage.

The gcloudy package is built on top of cononical Google Python packages(s) without any alteration to Google's base code.

----------------------------

INSTALL, IMPORT, & INITIALIZE

gcloudy is installed using pip with the terminal command:

$ pip install gcloudy

Once installed, the BigQuery class can be imported from the main GoogleCloud module with:

from gcloudy.GoogleCloud import BigQuery

Then, the bq object is initialized with the following (where "gcp-project-name" is your GCP Project ID / Name):

bq = BigQuery("gcp-project-name")

To initiate using a custom Google Cloud Platform client object, use the gcp_client arg:

bq = BigQuery("gcp-project-name", gcp_client = my_client_object)

----------------------------

METHODS

-----------

bq.read_bigquery

- Read an existing BigQuery table into a DataFrame.

read_bigquery(bq_dataset_dot_table = None, date_cols = [], preview_top = None, to_verbose = True)

  • bq_dataset_dot_table : the "dataset-name.table-name" path of the existing BigQuery table
  • date_cols : [optional] column(s) passed inside a list that should be parsed as dates
  • preview_top : [optional] only read in the top N rows
  • to_verbose : should info be printed? defaults to True
  • use_polars : [NEW IN 2.0.0] should a polars DataFrame be returned instead of a pandas DataFrame? Defaults to False

EX:

my_table = bq.read_bigquery("my_bq_dataset.my_bq_table")
my_table = bq.read_bigquery("my_bq_dataset.my_bq_table", date_cols = ['date'])
my_table = bq.read_bigquery("my_bq_dataset.my_bq_table", use_polars = True)

-----------

bq.read_custom_query

- Read in a custom BigQuery SQL query into a DataFrame.

read_custom_query(custom_query, to_verbose = True)

  • custom_query : the custom BigQuery SQL query that will produce a table to be read into a DataFrame
  • to_verbose : should info be printed? defaults to True
  • use_polars : [NEW IN 2.0.0] should a polars DataFrame be returned instead of a pandas DataFrame? Defaults to False

EX:

my_custom_table = bq.read_custom_query("""
    SELECT
        date,
        sales,
        products
    FROM
        my_bq_project_id.my_bq_dataset.my_bq_table
    WHERE
        sales_month = 'June'
""")

-----------

bq.write_bigquery

- Write a DataFrame to a BigQuery table.

write_bigquery(df, bq_dataset_dot_table = None, use_schema = None, append_to_existing = False, to_verbose = True)

  • df : the DataFrame to be written to a BigQuery table
  • bq_dataset_dot_table : the "dataset-name.table-name" path of the existing BigQuery table
  • use_schema : [optional] a custom schema for the BigQuery table. NOTE: see bq.guess_schema below
  • append_to_existing : should the DataFrame be appended to an existing BigQuery table? defaults to False (create new / overwrite)
  • to_verbose : should info be printed? defaults to True

EX:

bq.write_bigquery(my_data, "my_bq_dataset.my_data")
bq.write_bigquery(my_data, "my_bq_dataset.my_data", append_to_existing = True)

-----------

bq.guess_schema

- A helper for bq.write_bigquery, passed to its use_schema arg. Creates a custom schema based on the dtypes of a DataFrame.

guess_schema(df, bq_type_default = "STRING")

  • df : the DataFrame to be written to a BigQuery table
  • bq_type_default : default BQ type passed to dtype 'object'

EX:

bq.write_bigquery(my_data, "my_bq_dataset.my_data", use_schema = bq.guess_schema(my_data))

-----------

bq.send_query

- Send a custom SQL query to BigQuery. Process is carried out within BigQuery. Nothing is returned.

send_query(que, to_verbose = True)

  • que : the custom SQL query to be sent and carried out within BigQuery
  • to_verbose : should info be printed? defaults to True

EX:

bq.send_query("""
    CREATE TABLE my_bq_project_id.my_bq_dataset.my_new_bq_table AS 
    (
        SELECT
            date,
            sales,
            products
        FROM
            my_bq_project_id.my_bq_dataset.my_bq_table
        WHERE
            sales_month = 'June'
    )
""")

-----------

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gcloudy-2.0.1.tar.gz (17.3 kB view details)

Uploaded Source

Built Distribution

gcloudy-2.0.1-py3-none-any.whl (17.5 kB view details)

Uploaded Python 3

File details

Details for the file gcloudy-2.0.1.tar.gz.

File metadata

  • Download URL: gcloudy-2.0.1.tar.gz
  • Upload date:
  • Size: 17.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.3

File hashes

Hashes for gcloudy-2.0.1.tar.gz
Algorithm Hash digest
SHA256 c938453804d1ca1c2dfac2377296b9b59f05695c4b465fff1bd5dab075c7c009
MD5 6bdc7547f345bd63373b956bcaf78cb5
BLAKE2b-256 01842f9c8bdf3f9d15902c9ce65f13f2df029951ff7b3935e349b4eca90ed950

See more details on using hashes here.

File details

Details for the file gcloudy-2.0.1-py3-none-any.whl.

File metadata

  • Download URL: gcloudy-2.0.1-py3-none-any.whl
  • Upload date:
  • Size: 17.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.3

File hashes

Hashes for gcloudy-2.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 e22ee901243c8c48e38be7b90c60c1b5f0b489e24d6823e74ba4803e4229ebb8
MD5 ca4428764e02a57c4e357e8533ef1bed
BLAKE2b-256 50d8807ed898ccdfac354e4cae45522ae827efcf6231d9dcc56e8640b87e9024

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page