Google Ads API Report Fetcher (gaarf)
Python version of Google Ads API Report Fetcher tool a.k.a. gaarf
.
Please see the full documentation in the root README.
Getting started
Prerequisites
- Python 3.8+
- pip installed
- Google Ads API enabled
google-ads.yaml
file. Learn how to create one here.
Installation and running
- create virtual environment and install the tool
python3 -m venv gaarf
source gaarf/bin/activate
pip install google-ads-api-report-fetcher
install the latest development version with pip install -e git+https://github.com/google/ads-api-report-fetcher.git#egg=google-ads-api-report-fetcher\&subdirectory=py
Versions of the library
google-ads-api-report-fetcher[bq]
- version with BigQuery supportgoogle-ads-api-report-fetcher[pandas]
- version with Pandas supportgoogle-ads-api-report-fetcher[sqlalchemy]
- version with SQLalchemy supportgoogle-ads-api-report-fetcher[simulator]
- version with support for simulating
query results instead of calling Google Ads API.google-ads-api-report-fetcher[sheets]
- version with Google Sheets supportgoogle-ads-api-report-fetcher[full]
- full version
- Run the tool with
gaarf
command:
gaarf <queries> [options]
Documentation on available options see in the root README.md.
Using as a library
Once google-ads-api-report-fetcher
is installed you can use it as a library.
Initialize GoogleAdsApiClient
to connect to Google Ads API
GoogleAdsApiClient
is responsible for connecting to Google Ads API and provides several methods for authentication.
from gaarf import GoogleAdsApiClient
client = GoogleAdsApiClient(path_to_config="google-ads.yaml")
client = GoogleAdsApiClient(path_to_config="gs://<PROJECT-ID>/google-ads.yaml")
google_ads_config_dict = {
"developer_token": "",
"client_id": "",
"client_secret": "",
"refresh_token": "",
"client_customer_id": "",
"use_proto_plus": True
}
client = GoogleAdsApiClient(config_dict=google_ads_config_dict)
initialize AdsReportFetcher
to get reports
from gaarf.report_fetcher import AdsReportFetcher
report_fetcher = AdsReportFetcher(client)
query_text = "SELECT campaign.id AS campaign_id FROM campaign"
customer_ids = ['1', '2']
customer_ids = report_fetcher.expand_mcc('1234567890')
campaigns = report_fetcher.fetch(query_text, customer_ids)
campaigns = report_fetcher.fetch(query_text, '1234567890', auto_expand=True)
Use macros in your queries
parametrized_query_text = """
SELECT
campaign.id AS campaign_id
FROM campaign
WHERE campaign.status = '{status}'
"""
active_campaigns = report_fetcher.fetch(parametrized_query_text, customer_ids,
{"macro": {
"status": "ENABLED"
}})
Define queries
There are three ways how you can define a query:
- in a variable
- in a file
- in a class (useful when you have complex parametrization and validation)
from gaarf.base_query import BaseQuery
from gaarf.io import reader
query_string = "SELECT campaign.id FROM campaign"
query_path = "path/to/query.sql"
query_path = "gs://PROJECT_ID/path/to/query.sql"
reader_client = reader.FileReader()
query = reader_client.read(query_path)
class Campaigns(BaseQuery):
query_text = """
SELECT
campaign.id
FROM campaign
WHERE campaign.status = {status}
"""
def __init__(self, status: str = "ENABLED") -> None:
self.status = status
from dataclasses import dataclass
@dataclass
class Campaigns(BaseQuery):
query_text = """
SELECT
campaign.id
FROM campaign
WHERE campaign.status = {status}
"""
status: str = "ENABLED"
class Campaigns(BaseQuery):
def __init__(self, status: str = "ENABLED"):
self.query_text = f"""
SELECT
campaign.id
FROM campaign
WHERE campaign.status = {status}
"""
active_campaigns = report_fetcher.fetch(Campaigns())
inactive_campaigns = report_fetcher.fetch(Campaigns("INACTIVE"))
Iteration and slicing
AdsReportFetcher.fetch
method returns an instance of GaarfReport
object which you can use to perform simple iteration.
query_text = """
SELECT
campaign.id AS campaign_id,
campaign.name AS campaign_name,
metrics.clicks AS clicks
FROM campaign
WHERE segments.date DURING LAST_7_DAYS
"""
campaigns = report_fetcher.fetch(query_text, '1234567890', auto_expand=True)
for row in campaigns:
print(row.campaign_id)
print(row["campaign_name"])
print(row[2])
row["new_campaign_id"] = row["campaign_id"] + 1
You can easily slice the report
campaign_only_report = campaigns["campaign_name"]
campaign_name_clicks_report = campaigns[["campaign_name", "clicks"]]
first_campaign_row = campaigns[0]
first_10_rows_from_campaigns = campaigns[0:10]
Convert report
GaarfReport
can be easily converted to common data structures:
campaigns_list = campaigns.to_list()
campaigns_list = campaigns.to_list(row_type="scalar")
campaigns_list = campaigns["campaign_id"].to_list()
campaigns_list = campaigns["campaign_id"].to_list(distinct=True)
campaigns_list = campaigns.to_list(row_type="dict")
campaigns_df = campaigns.to_pandas()
campaigns_df = campaigns.to_dict(
key_column="campaign_id",
value_column="campaign_name",
value_column_output="scalar",
)
campaigns_df = campaigns.to_dict(
key_column="campaign_id",
value_column="campaign_name",
value_column_output="list",
)
Build report
GaarfReport
can be easily built from pandas data frame:
import pandas as pd
df = pd.DataFrame(data=[[1]], columns=["one"])
report = GaarfReport.from_pandas(df)
Save report
GaarfReport
can be easily saved to local or remote storage:
from gaarf.io import writers
csv_writer = writers.csv_writer.CsvWriter(destination_folder="/tmp")
bq_writer = writers.bigquery_writer.BigQueryWriter(
project="", dataset="", location="")
sqlalchemy_writer = writers.sqlalchemy_writer.SqlAlchemyWriter(
connection_string="")
console_writer = writers.console_writer.ConsoleWriter(page_size=10)
json_writer = writers.json_writer.JsonWriter(destination_folder="/tmp")
sheet_writer = writers.sheets_writer.SheetWriter(
share_with="you@email.com",
credential_files="path/to/credentials.json"
)
csv_writer.write(campaigns, destination="my_file_name")
bq_writer.write(campaigns, destination="my_table_name")
sqlalchemy_writer.write(campaigns, destination="my_table_name")
json_writer.write(campaigns, destination="my_table_name")
sheet_writer.write(campaigns, destination="my_table_name")
Combine fetching and saving with AdsQueryExecutor
If your job is to execute query and write it to local/remote storage you can use AdsQueryExecutor
to do it easily.
When reading query from file AdsQueryExecutor
will use query file name as a name for output file/table.
from gaarf.io import reader, writers
from gaarf.executors import AdsQueryExecutor
query_executor = AdsQueryExecutor(client)
csv_writer = writers.csv_writer.CsvWriter(destination_folder="/tmp")
reader_client = reader.FileReader()
query_text = """
SELECT
campaign.id AS campaign_id,
campaign.name AS campaign_name,
metrics.clicks AS clicks
FROM campaign
WHERE segments.date DURING LAST_7_DAYS
"""
query_executor.execute(
query_text=query_text,
query_name="campaign",
customer_ids=customer_ids,
write_client=csv_writer)
query_path="path/to/query.sql"
query_executor.execute(
query_text=reader_client.read(query_path),
query_name=query_path,
customer_ids=customer_ids,
write_client=csv_writer)
Python specific command line flags
--optimize-performance
- accepts one of the following values:
NONE
- no optimizations are donePROTOBUF
- convert Google Ads API response to protobuf before parsing
(speeds up query execution 5x times but forces conversion of ENUMs to integers instead of strings)BATCH
- converts all response of Ads API to a list and then parses its content in parallelBATCH_PROTOBUF
- combines BATCH
and PROTOBUF
approaches.
Disclaimer
This is not an officially supported Google product.