Ready to Run!

  1. Github Actions Run Terraform ⚠️ for initial set-up only!!

  • Go to Github Actions in your repo

  • Go to Teraform GCP Deployment

  • Click Run workflow -> Run workflow

  • You can then click on the event and the run to look at details

  • Once it completes ✅ Go to Google cloud and check:

    1. Bucket name of your BUCKET secret in Github

    2. BigQuery database named petfinderapi / petfinder_data


  1. Github Actions Daily Python Run ⚠️ The cron job is commented out in the yml file. If you would like it to run daily, go to `` and uncomment the cron line

  • Go to Github Actions in your repo

  • Go to Fetch and Upload PetFinder Data

  • Click Run workflow -> Run workflow

  • You can then click on the event and the run to look at details

  • Once it completes ✅ Go to Google cloud and check:

    1. A csv with today's date was added to your bucket

    2. raw_petfinder data in your BigQuery for today's petfinder values!


  1. Github Actions Daily DBT ⚠️ The cron job is commented out in the yml file. If you would like it to run daily, go to `` and uncomment the cron line

  • Go to Github Actions in your repo

  • Go to Run dbt Transformations

  • Click Run workflow -> Run workflow

  • You can then click on the event and the run to look at details

  • Once it completes ✅ Go to Google cloud and check in BigQuery:

    1. Created 'active_pets', 'stg_petfinder', and partitioned and clustered 'transformed_petfinder'

See plots from my API Looker here:

https://lookerstudio.google.com/s/r_UWAUGaVC8

Last updated