We have a lightweight ETL pipeline setup using CloudRun that pulls data from a 3rd party SaaS software via APIs and inserts them into BQ tables using GCS as the temporary storage. We are facing a challenge in updating BQ tables daily via costly BQ inserts. Is there a better way to automate data refresh cycles meaning any data that changes in the 3rd party software needs to be updated in BQ to keep BQ tables in sync (upsert operation)? This operation needs to be run at a a tunable frequency (once per day, every 3 hours, once a week etc) via a cloud scheduler job.
Solved! Go to Solution.
Hi @dheerajpanyam,
Welcome back to Google Cloud Community!
Some suggestions to improve your data refresh cycles while keeping your BQ tables in sync with the 3rd party SaaS:
Remember: Choose the solution that best fits your specific needs and SaaS capabilities.
Hi @dheerajpanyam,
Welcome back to Google Cloud Community!
Some suggestions to improve your data refresh cycles while keeping your BQ tables in sync with the 3rd party SaaS:
Remember: Choose the solution that best fits your specific needs and SaaS capabilities.
Thanks so much @christianpaula 🙏