Easy Monitoring of DBT Cloud jobs with Streamlit

This post also appeared on the Cazoo Tech Blog, read more here.

DBT is a core tool here in the Cazoo data team for taking raw data from our S3 datalake and transforming it into a form that is much more easily consumed by our reporting & analytics tools (Looker and AWS Athena to name just two).  We run multiple jobs throughout the day to serve data that is somewhat near real time (updated hourly) to keep our internal customers informed on how the business is performing.

We use DBT cloud to schedule our jobs, which unfortunately comes with one reasonably frustrating issue - detailed information on the status of jobs (including any reasons for failure) is not available for standard users, only Admins get access. Ideally we want to expose this information to a wider audience without needing to grant admin access, both because that would increase costs, but also, most people don't need admin level access, a read only view on these jobs is more than sufficient.  Fortunately DBT Cloud does provide an API, that allows us to pull some fairly detailed data from our jobs which we can use to, we just need some suitable way to display this.

Below we'll get into how we solved this, but if you want to jump right into the source code, here it is.

How Streamlit fits in

This is where Streamlit comes in. Streamlit is a fantastic way to build shareable web applications without any knowledge of front end technologies (which is rather handy for me, as whilst I can spell HTML and CSS that’s about as far as my knowledge goes). I’ve tried similar libraries (such as Dash) in the past but Streamlit has blown us all away with how easy it is to use, and how nice the resulting apps look.  Streamlit is open source, so you can deploy apps locally, or on your own infrastructure if you so wish.  

At Cazoo, we have a broad strategy of making use of external expertise for running tools to lessen the overhead for our engineering teams, and so instead of deploying Streamlit on some Fargate instances at AWS we’ve been enthusiastic participants in the ongoing Streamlit for Teams beta.  This allows us to deploy our apps at the click of a button, and let someone else (the nice folks at Streamlit) worry about keeping the lights on.  

As well as using Streamlit for the more obvious case of Analytics & Data Science apps - here in Data Engineering we’re also making use of it for apps where we want to expose operational information such as Dead Letter Queues, or in this case what’s happening with our DBT jobs.

The DBT Cloud + Streamlit app

Full disclosure, I didn’t build this app, one of my colleagues Oliver Morgans did. I’m just helping out on this one.

We’ve made the app available for anyone who wants to use it, and we’d certainly welcome any questions, comments or pull requests for ways to make this better. I’m sure as our analysts start using this they’ll be making suggestions for things they want to see also. You can find the source code at here.

Detailed instructions for getting up and running are in the repo README.  You’ll need to add a few things like your DBT Cloud account_id and API key into the streamlit secrets.toml, and you’ll probably want to swap out the logo we’ve stored in images too (unless you’re particularly enamoured with the Cazoo logo of course).

We have included basic username/password auth in the code, though if you’re using Streamlit for Teams I’d recommend you use the built in SSO as that’s a much nicer experience than typing in a long password each time.

The app is currently composed of a few pieces:

You can select to see all runs, successful runs or failed runs. And you can filter down by project (this project list is driven by a set of project ids you can add in secrets.toml)

RAG Status

We’re pulling the last 500 job runs, and this table will let you sort by the columns to get at what you’re after.

Failed Steps

Of particular interest to our analysts, is the step (or steps) that a DBT job failed on, as most often that represents a unit test that has failed. So you have the option to select a specific run, and to see any failed steps in that run.

That blue text is indeed a link which will take you to GitHub and the specific file that threw the error. You can configure the base URL for your repo as part of the secrets.toml.

Historical Runs

Picking a job name also exposes the last ten historical runs, which will also allow you to inspect them for failure steps if needed.

Wrapping up

I am a huge fan of Streamlit, and the fact we’re able to easily use it not just for Data Science but for other operational things in engineering is making it a really key part of our toolbox.  I’d welcome any comments below on what you would add to this app, or feel free to tell us if we’re doing it all wrong - because every day is a school day!


This is a companion discussion topic for the original entry at https://blog.streamlit.io/dbt-cloud-jobs-with-streamlit/
2 Likes