Library for fetching reports from Google Ads API and saving them locally / BigQuery.
Project description
Google Ads API Report Fetcher (gaarf)
Python version of Google Ads API Report Fetcher tool a.k.a. gaarf
.
Please see the full documentation in the root README.
Getting started
Prerequisites
- Python 3.9+
- pip installed
- Google Ads API enabled
google-ads.yaml
file. Learn how to create one here.
Installation and running
- create virtual environment and install the tool
python3 -m venv gaarf
source gaarf/bin/activate
pip install google-ads-api-report-fetcher
install the latest development version with
pip install -e git+https://github.com/google/ads-api-report-fetcher.git#egg=google-ads-api-report-fetcher\&subdirectory=py
Versions of the library
google-ads-api-report-fetcher[sqlalchemy]
- version with SQLalchemy supportgoogle-ads-api-report-fetcher[simulator]
- version with support for simulating query results instead of calling Google Ads API.google-ads-api-report-fetcher[full]
- full version
- Run the tool with
gaarf
command:
gaarf <queries> [options]
Documentation on available options see in the root README.md.
Using as a library
Once google-ads-api-report-fetcher
is installed you can use it as a library.
Initialize GoogleAdsApiClient
to connect to Google Ads API
GoogleAdsApiClient
is responsible for connecting to Google Ads API and provides several method for authentication.
from gaarf.api_clients import GoogleAdsApiClient
# initialize from local file
client = GoogleAdsApiClient(path_to_config="google-ads.yaml", version="v12")
# initialize from remote file
client = GoogleAdsApiClient(path_to_config="gs://<PROJECT-ID>/google-ads.yaml", version="v12")
# initialize from dictionary
google_ads_config_dict = {
"developer_token": "",
"client_id": "",
"client_secret": "",
"refresh_token": "",
"client_customer_id": "",
"use_proto_plus": True
}
client = GoogleAdsApiClient(config_dict=google_ads_config_dict, version="v12")
initialize AdsReportFetcher
to get reports
from gaarf.query_executor import AdsReportFetcher, AdsQueryExecutor
report_fetcher = AdsReportFetcher(client)
# create query text
query_text = "SELECT campaign.id AS campaign_id FROM campaign"
# Execute query and store `campaigns` variable
# specify customer_ids explicitly
customer_ids = ['1', '2']
# or perform mcc expansion for mcc 1234567890
customer_ids = report_fetcher.expand_mcc('1234567890')
campaigns = report_fetcher.fetch(query_text, customer_ids)
# perform mcc expansion when calling `fetch` method
campaigns = report_fetcher.fetch(query_text, '1234567890', auto_expand=True)
Use macros in your queries
parametrized_query_text = """
SELECT
campaign.id AS campaign_id
FROM campaign
WHERE campaign.status = '{status}'
"""
active_campaigns = report_fetcher.fetch(parametrized_query_text, customer_ids,
{"status": "ENABLED"})
Define queries
There are three ways how you can define a query:
- in a variable
- in a file
- in a class (useful when you have complex parametrization and validation)
from gaarf.base_query import BaseQuery
from gaarf.io import reader
# 1. define query as a string an save in a variable
query_string = "SELECT campaign.id FROM campaign"
# 2. define path to a query file and read from it
# path can be local
query_path = "path/to/query.sql"
# or remote
query_path = "gs://PROJECT_ID/path/to/query.sql"
# Instantiate reader
reader_client = reader.FileReader()
# And read from the path
query = reader_client.read(query_path)
# 3. define query as a class
class Campaigns(BaseQuery):
def __init__(self, status: str = "ENABLED"):
self.query_text = f"""
SELECT
campaign.id
FROM campaign
WHERE campaign.status = {status}
"""
active_campaigns = report_fetcher.fetch(Campaigns())
inactive_campaigns = report_fetcher.fetch(Campaigns("INACTIVE"))
Iteration and slicing
AdsReportFetcher.fetch
method returns an instance of GaarfReport
object which you can use to perform simple iteration.
query_text = """
SELECT
campaign.id AS campaign_id,
campaign.name AS campaign_name,
metrics.clicks AS clicks
FROM campaign
WHERE segments.date DURING LAST_7_DAYS
"""
campaigns = report_fetcher.fetch(query_text, '1234567890', auto_expand=True)
# iterate over each row of `campaigns` report
for row in campaigns:
# Get element as an attribute
print(row.campaign_id)
# Get element as a slice
print(row["campaign_name"])
# Get element as an index (will print number of clicks)
print(row[2])
# Create new column
row["new_campaign_id"] = row["campaign_id"] + 1
You can easily slice the report
# Create new reports by selecting one or more columns
campaign_only_report = campaigns["campaign_name"]
campaign_name_clicks_report = campaigns[["campaign_name", "clicks"]]
# Get subset of the report
# Get first row only
first_campaign_row = campaigns[0]
# Get first ten rows from the report
first_10_rows_from_campaigns = campaigns[0:10]
Convert report
GaarfReport
can be easily converted to common data structures:
# convert `campaigns` to list
campaigns_list = campaigns.to_list()
# convert `campaigns` to pandas DataFrame
campaigns_df = campaigns.to_pandas()
Save report
GaarfReport
can be easily saved to local or remote storage:
from gaarf.io import writer
# initialize CSV writer
csv_writer = writer.CsvWriter(destination_folder="/tmp")
# initialize BigQuery writer
bq_writer = writer.BigQueryWriter(project="", dataset="", location="")
# initialize SQLAlchemy writer
sqlalchemy_writer = writer.SqlAlchemyWriter(connection_string="")
# initialize Console writer
console_writer = writer.Console(page_size=10)
# save report using one of the writers
csv_writer.write(campaigns, destination="my_file_name")
bq_writer.write(campaigns, destination="my_table_name")
sqlalchemy_writer.write(campaigns, destination="my_table_name")
Combine fetching and saving with AdsQueryExecutor
If your job is to execute query and write it to local/remote storage you can use AdsQueryExecutor
to do it easily.
When reading query from file
AdsQueryExecutor
will use query file name as a name for output file/table.
from gaarf.io import reader, writer
from gaarf.query_executor import AdsQueryExecutor
# initialize query_executor to fetch report and store them in local/remote storage
query_executor = AdsQueryExecutor(client)
# initialize writer
csv_writer = writer.CsvWriter(destination_folder="/tmp")
reader_client = reader.FileReader()
query_text = """
SELECT
campaign.id AS campaign_id,
campaign.name AS campaign_name,
metrics.clicks AS clicks
FROM campaign
WHERE segments.date DURING LAST_7_DAYS
"""
# execute query and save results to `/tmp/campaign.csv`
query_executor.execute(
query_text=query_text,
query_name="campaign",
customer_ids=customer_ids,
write_client=csv_writer)
# execute query from file and save to results to `/tmp/query.csv`
query_path="path/to/query.sql"
query_executor.execute(
query_text=reader_client.read(query_path),
query_name=query_path,
customer_ids=customer_ids,
write_client=csv_writer)
Python specific command line flags
--optimize-performance
- accepts one of the following values:NONE
- no optimizations are donePROTOBUF
- convert Google Ads API response to protobuf before parsing (speeds up query execution 5x times but forces conversion of ENUMs to integers instead of strings)BATCH
- converts all response of Ads API to a list and then parses its content in parallelBATCH_PROTOBUF
- combinesBATCH
andPROTOBUF
approaches.
Disclaimer
This is not an officially supported Google product.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file google-ads-api-report-fetcher-1.12.1.tar.gz
.
File metadata
- Download URL: google-ads-api-report-fetcher-1.12.1.tar.gz
- Upload date:
- Size: 40.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 381c836f497794cd7e6dc9f9e3eeb18d9b7040425fa4824a0a485387c1430a69 |
|
MD5 | 9b245ee706c07276b16111e3a7ee1522 |
|
BLAKE2b-256 | 224972c05da945038dcc29a0dff6531efb9ea816bb2f8bc58b47f25cae219af5 |