How I Broke Our SLA and Delighted Our Customer

Failing the SLA was the price we paid for trust. And it was worth every second.

I broke one of our most critical SLAs just last week, and it was the best thing that could have happened.

It was shaping up to be a major embarrassment.ย  One of our key data warehouse refreshes had failed. No new data. No dashboard updates. The refresh was long past its deadline, the projectโ€™s key data engineer was on vacation, and I was playing backup. Where was I?ย  At the moment, I was flying home from a data quality conference. This was not good.

Naturally, the airline’s Wi-Fi was terrible, but I managed to get through to a colleague, who knew little about the project. And then we started troubleshootingโ€”one Slack message at a time. I slacked him SQL. He slacked back data. Back and forth we went. An embedded test had failed. Where had our process gone wrong? Everything I could see looked fine. And I was tempted, so tempted, as the clock kept ticking, to disable the test and let it go.ย 

Then it dawned on me that this test wasnโ€™t even ours.ย  It was one of a series of tests that had been added by one of our principal stakeholders in collaboration with his analytics team โ€“ esoteric referential integrity checks that had been placed at key points in the build, validating consistency between multiple data sources that fed hourly into our reporting layer.

These tests werenโ€™t easy to define or implement. They came from an understanding of the business rules that we, as data engineers, lacked. And they caught a major problem: the new records we received from one source were completely out of sync with our other data.

Something had gone wrong upstreamโ€”very wrong. And because our process was designed to fail fast when integrity was violated, it refused to publish anything at all.

From an SLA standpoint, we failed. The data didnโ€™t arrive on time.

But from a trust standpoint? We won.

We trusted stakeholders to define critical business rules that would test for major problems. They trusted us to implement their logic with every refresh. We trusted each other not to blame and shame, but to steadily build out better and better testing to benefit everyone.

Now, together, we had protected leadership from a serious integrity issue. We had prevented a flood of inaccurate reports from reaching decision makers. And when we explained what happenedโ€”why the data was late and what we had caughtโ€”they werenโ€™t frustrated. They were grateful.

The value of data quality is often invisible. When things go right, no one notices. But this was a moment where things went wrong exactly the way we hoped they would: the system raised a red flag, we caught the issue before it caused damage, and we preserved the integrity of our analytics.

Failing the SLA was the price we paid for trust. And it was worth every second.

The takeaway?ย  If youโ€™re not building comprehensive, automated data quality checks into your production pipelines, youโ€™re one bad refresh away from losing stakeholder confidence and risking devastating business outcomes.ย 

Your data processes shouldnโ€™t just be fastโ€”they should be right. And sometimes, the best outcome is one where the system refuses to run.

Sign-Up for our Newsletter

Get the latest straight into your inbox

DataOps Data Quality TestGen:

Simple, Fast, Generative Data Quality Testing, Execution, and Scoring.

[Open Source, Enterprise]

DataOps Observability:

Monitor every date pipeline, from source to customer value, & find problems fast

[Open Source, Enterprise]

DataOps Automation:

Orchestrate and automate your data toolchain with few errors and a high rate of change.

[Enterprise]

recipes for dataops success

DataKitchen Consulting Services


DataOps Assessments

Identify obstacles to remove and opportunities to grow

DataOps Consulting, Coaching, and Transformation

Deliver faster and eliminate errors

DataOps Training

Educate, align, and mobilize

Commercial Data & Analytics Platform for Pharma

Get trusted data and fast changes to create a single source of truth

 

dataops-cookbook-download

DataOps Learning and Background Resources


DataOps Journey FAQ
DataOps Observability basics
Data Journey Manifesto
Why it matters!
DataOps FAQ
All the basics of DataOps
DataOps 101 Training
Get certified in DataOps
Maturity Model Assessment
Assess your DataOps Readiness
DataOps Manifesto
Thirty thousand signatures can't be wrong!

 

DataKitchen Basics


About DataKitchen

All the basics on DataKitchen

DataKitchen Team

Who we are; Why we are the DataOps experts

Careers

Come join us!

Contact

How to connect with DataKitchen

 

DataKitchen News


Newsroom

Hear the latest from DataKitchen

Events

See DataKitchen live!

Partners

See how partners are using our Products

 

Monitor every Data Journey in an enterprise, from source to customer value, in development and production.

Simple, Fast Data Quality Test Generation and Execution. Your Data Journey starts with verifying that you can trust your data.

Orchestrate and automate your data toolchain to deliver insight with few errors and a high rate of change.