Many BigQuery users ask for database triggers—a way to run some procedural code in response to events on a particular BigQuery table, model, or dataset. Maybe you want to run an ELT job whenever a new table partition is created, or maybe you want to retrain your ML model whenever new rows are inserted into the table. In the general category of “Cloud gets easier”, this article will show how to quite simply and cleanly tie together BigQuery and Cloud Run.
In Part 1 we learned how to set up our Xplenty pipeline to work with Chartio and prepared the data source. In Part 2, we will focus on using the data Xplenty provides in the Chartio platform. If you're new to Chartio, you can read through their QuickStart docs (shouldn't take more than 5-10 minutes) to gain some familiarity.
Snowflake met with Jan Doumen, Head of Expertise for Allianz Benelux, and Naveed Memon, Program Director, Data and Analytics for Emirates, at Data Cloud Summit 2020. Read excerpts from the conversation to learn how capturing data insights in the Data Cloud brings value to their businesses. Data’s value in the 21st century is often compared to oil’s value in the 18th century. It can transform organizations, opening doors to unprecedented opportunities.