Is Your Team in Denial about Data Quality? Here’s How to Tell

Data quality problems are ignored, rationalized, or swept aside. How do you tell if your team is in denial about data quality?   Uncle Chip has provided a bingo card ... play with your data team! 

Is Your Team in Denial about Data Quality?

Here’s How to Tell

In many organizations, data quality problems fester in the shadows—ignored, rationalized, or swept aside with confident-sounding statements that mask a deeper dysfunction. While the cost of poor data quality is well-documented—wasted time, lost trust, and flawed decisions—teams often operate in collective denial. The problem isn’t that they don’t care. They’ve built a culture of rationalization, where signs of trouble are misread as proof of success.

Take a closer look at the illustration Uncle Chip created above. It’s a snapshot of common justifications that data teams use to avoid grappling with the real condition of their data. These aren’t caricatures; they’re alarmingly familiar echoes from meetings, Slack threads, and status updates across the industry. A pipeline ran “all green”? That doesn’t mean the data inside was correct. QA passed the code? Great, but that doesn’t mean the business logic downstream was valid. “We passed the data exactly as we received it”? That’s not a defense—that’s a surrender of responsibility.

The woman on the phone saying, “The users know about that—they have a workaround!” might be trying to help, but she’s normalizing failure. Workarounds don’t fix broken data; they institutionalize it. When someone claims, “I haven’t heard a single complaint,” it usually means users have stopped expecting anything better or have worked around the data team entirely. And the engineer proudly pointing to COUNT(*) as validation? That’s not testing; that’s wishful thinking.

One of the most telling lines comes from the self-assured analyst who declares that metrics must be accurate because he calculated them in Excel. Confidence is not a substitute for verification. Manual processes, however well-intentioned, are rarely repeatable, scalable, or trustworthy. That’s not data quality; that’s data folklore.

Denial about data quality doesn’t always look like laziness or malice. More often, it’s baked into team culture, shaped by tools that don’t make quality visible, roles that lack ownership, and a lack of shared language to even talk about what’s broken. When success is measured by the absence of red flags instead of the presence of verified data quality, teams fool themselves into thinking everything is fine—until the report goes live. Someone important finds a problem in the data, and you sit in a ‘war room‘ all night looking at log files and writing random queries.

So, how do you break the cycle?

Start by playing Data Quality Denial Bingo with your team. Please print out the image above or share it during a team meeting. Ask each team member to check off every statement they’ve heard—or said themselves—in the past month. If someone gets five in a row, they win… or the whole team loses. Use the laughter and mild embarrassment that follows as a gateway to having a serious conversation about your data quality testing practices, data assessment, and data observability. Ask your team what it means to take ownership of data quality.

Because the first step to fixing the problem is admitting you have one. Is your team ready?


We built DataOps Data Quality TestGen as the first tool for data teams looking to break the cycle of surrender, failure normalization, wishful thinking, and unverified confidence. Those teams need an open-source, intelligent data quality tool.  DataKitchen’s TestGen lets you start data testing, profiling, and assessing thousands of tables immediately. It’s AI-driven, so you can get valuable results in just a few button clicks.

It profiles your data, generates a catalog, instantly generates hundreds of valuable data quality checks, and allows you to build data quality assessment dashboards.  Interested?  Install here:  https://info.datakitchen.io/testgen

Sign-Up for our Newsletter

Get the latest straight into your inbox

DataOps Data Quality TestGen:

Simple, Fast, Generative Data Quality Testing, Execution, and Scoring.

[Open Source, Enterprise]

DataOps Observability:

Monitor every date pipeline, from source to customer value, & find problems fast

[Open Source, Enterprise]

DataOps Automation:

Orchestrate and automate your data toolchain with few errors and a high rate of change.

[Enterprise]

recipes for dataops success

DataKitchen Consulting Services


DataOps Assessments

Identify obstacles to remove and opportunities to grow

DataOps Consulting, Coaching, and Transformation

Deliver faster and eliminate errors

DataOps Training

Educate, align, and mobilize

Commercial Data & Analytics Platform for Pharma

Get trusted data and fast changes to create a single source of truth

 

dataops-cookbook-download

DataOps Learning and Background Resources


DataOps Journey FAQ
DataOps Observability basics
Data Journey Manifesto
Why it matters!
DataOps FAQ
All the basics of DataOps
DataOps 101 Training
Get certified in DataOps
Maturity Model Assessment
Assess your DataOps Readiness
DataOps Manifesto
Thirty thousand signatures can't be wrong!

 

DataKitchen Basics


About DataKitchen

All the basics on DataKitchen

DataKitchen Team

Who we are; Why we are the DataOps experts

Careers

Come join us!

Contact

How to connect with DataKitchen

 

DataKitchen News


Newsroom

Hear the latest from DataKitchen

Events

See DataKitchen live!

Partners

See how partners are using our Products

 

Monitor every Data Journey in an enterprise, from source to customer value, in development and production.

Simple, Fast Data Quality Test Generation and Execution. Your Data Journey starts with verifying that you can trust your data.

Orchestrate and automate your data toolchain to deliver insight with few errors and a high rate of change.