A wrapper for google's existing google-cloud python package that aims to make using python inside the Google Cloud framework more intuitive.
Project description
NEW IN VERSION 2.0.0
:
Features:
- Added
polars
functionality
QL & Fixes:
- Replaced
use_client
var withgcp_client
for user clarity - Fixed N rows have been added to ... message in
write_bigquery()
LICENSE
GPL-3 Summary:
You may copy, distribute and modify the software as long as you track changes/dates in source files. Any modifications to or software including (via compiler) GPL-licensed code must also be made available under the GPL along with build & install instructions.
In other words, any derivative work of this software shall be released under the same GPL license as the original software, meaning the modified code must be exactly as free and open-source as the original.
ABOUT
gcloudy is a wrapper for Google's GCP Python package(s) that aims to make interacting with GCP and its services more intuitive, especially for new GCP users. In doing so, it adheres to pandas-like syntax for function/method calls.
The gcloudy package is not meant to be a replacement for GCP power-users, but rather an alternative for GCP users who are interested in using Python in GCP to deploy Cloud Functions and interact with certain GCP services, especially BigQuery and Google Cloud Storage.
The gcloudy package is built on top of cononical Google Python packages(s) without any alteration to Google's base code.
INSTALL, IMPORT, & INITIALIZE
-
gcloudy is installed using pip with the terminal command:
$ pip install gcloudy
-
Once installed, the BigQuery class can be imported from the main GoogleCloud module with:
from gcloudy.GoogleCloud import BigQuery
-
Then, the
bq
object is initialized with the following (where "gcp-project-name" is your GCP Project ID / Name):
bq = BigQuery("gcp-project-name")
-
NOTE: It is important to also import the Pandas package:
import pandas as pd
METHODS
The following section contains the methods and their usage.
----------------------------
bq.read_bigquery
- Read an existing BigQuery table into a DataFrame.
read_bigquery(bq_dataset_dot_table = None, date_cols = [], preview_top = None, to_verbose = True)
- bq_dataset_dot_table : the "dataset-name.table-name" path of the existing BigQuery table
- date_cols : [optional] column(s) passed inside a list that should be parsed as dates
- preview_top : [optional] only read in the top N rows
- to_verbose : should info be printed? defaults to True
EX:
my_table = bq.read_bigquery("my_bq_dataset.my_bq_table")
my_table = bq.read_bigquery("my_bq_dataset.my_bq_table", date_cols = ['date'])
-----------
bq.write_bigquery
- Write a DataFrame to a BigQuery table.
write_bigquery(df, bq_dataset_dot_table = None, use_schema = None, append_to_existing = False, to_verbose = True)
- df : the DataFrame to be written to a BigQuery table
- bq_dataset_dot_table : the "dataset-name.table-name" path of the existing BigQuery table
- use_schema : [optional] a custom schema for the BigQuery table. NOTE: see bq.guess_schema below
- append_to_existing : should the DataFrame be appended to an existing BigQuery table? defaults to False (create new / overwrite)
- to_verbose : should info be printed? defaults to True
EX:
bq.write_bigquery(my_data, "my_bq_dataset.my_data")
bq.write_bigquery(my_data, "my_bq_dataset.my_data", append_to_existing = True)
-----------
bq.guess_schema
- A helper for bq.write_bigquery, passed to its use_schema arg. Creates a custom schema based on the dtypes of a DataFrame.
guess_schema(df, bq_type_default = "STRING")
- df : the DataFrame to be written to a BigQuery table
- bq_type_default : default BQ type passed to dtype 'object'
EX:
bq.write_bigquery(my_data, "my_bq_dataset.my_data", use_schema = bq.guess_schema(my_data))
-----------
bq.read_custom_query
- Read in a custom BigQuery SQL query into a DataFrame.
read_custom_query(custom_query, to_verbose = True)
- custom_query : the custom BigQuery SQL query that will produce a table to be read into a DataFrame
- to_verbose : should info be printed? defaults to True
EX:
my_custom_table = bq.read_custom_query("""
SELECT
date,
sales,
products
FROM
my_bq_project_id.my_bq_dataset.my_bq_table
WHERE
sales_month = 'June'
""")
-----------
bq.send_query
- Send a custom SQL query to BigQuery. Note, does not return anything as the process is carried out within BigQuery.
send_query(que, to_verbose = True)
- que : the custom SQL query to be sent and carried out within BigQuery
- to_verbose : should info be printed? defaults to True
EX:
bq.send_query("""
CREATE TABLE my_bq_project_id.my_bq_dataset.my_new_bq_table AS
(
SELECT
date,
sales,
products
FROM
my_bq_project_id.my_bq_dataset.my_bq_table
WHERE
sales_month = 'June'
)
""")
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file gcloudy-2.0.0.tar.gz
.
File metadata
- Download URL: gcloudy-2.0.0.tar.gz
- Upload date:
- Size: 17.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.0 CPython/3.12.3
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5a33c1b029b441305f7d60d13d049e5ecea78c2c351514d5d24fa2623d169d91 |
|
MD5 | c53b331caa912a72834034227d9fdea8 |
|
BLAKE2b-256 | 6853fe9270285de9212b22308d14a52fc1f20d2add3655a4b983a370f1a6d377 |
File details
Details for the file gcloudy-2.0.0-py3-none-any.whl
.
File metadata
- Download URL: gcloudy-2.0.0-py3-none-any.whl
- Upload date:
- Size: 17.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.0 CPython/3.12.3
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | deb4ec66a2523d3180303017f80d329cb78ca293f81c309039593fb9b78fcf9e |
|
MD5 | 69141845e15e96971f9dfa00c3076c7a |
|
BLAKE2b-256 | 28329ecfc159c9335daa08bf881860ec3722cb66788c971e9d320527e870909e |