Over the past year, I’ve worked with dozens of distributors on AI adoption.
Same conversation every time.
“We need to clean our data first.” “Our ERP is a mess.” “We’ve got duplicate customers across three systems.” “Half our product descriptions are garbage.”
Then they wait.
And wait.
And nothing happens.
I understand the instinct. For decades, we’ve been taught that data quality determines system performance. Garbage in, garbage out. It’s been true for every major technology implementation—ERP, CRM, business intelligence platforms. Clean data was table stakes.
But here’s what I’ve learned watching distributors navigate AI adoption: The belief that you need perfect data before starting with AI is costing you years of progress.
Not because the belief is entirely wrong.
Because it misses a critical distinction that changes everything.
The Two Categories That Matter
Some AI applications do require pristine data.
Predictive analytics for demand forecasting need clean historical sales data, accurate inventory records, and consistent product hierarchies. Automated pricing engines need structured competitor data and reliable cost information. Machine learning models trained on your specific data need that data to be accurate, complete, and properly labeled.
These are real requirements. If you’re building custom models or deploying AI that learns from your historical patterns, data quality absolutely matters.
But here’s what most distributors miss: Many of the most valuable AI tools available today work brilliantly with messy data.
Large language models don’t care if your product descriptions are inconsistent. They don’t need your customer records deduplicated across systems. They can read your badly formatted PDFs, interpret your inconsistent naming conventions, and extract meaning from documents that have been sitting in network drives for fifteen years.
This isn’t theoretical. I’ve watched these tools:
- Parse vendor catalogs with wildly different formats and extract comparable product specifications
- Read decades of email correspondence and identify recurring customer issues
- Analyze RFQ documents where every customer uses different terminology for the same products
- Review contract language across hundreds of agreements with zero standardization
The technology finds patterns in chaos. That’s what makes it different from everything that came before.
Why the Waiting Strategy Fails
When distributors tell me they’re waiting to clean their data first, I ask a simple question: “When does that project finish?”
The answer is usually uncomfortable silence.
Because data cleanup isn’t a project with a finish line. It’s an ongoing operational challenge that never ends. Your data is being created and modified every day—by sales reps, customer service teams, warehouse staff, and automated systems. Perfect data is a moving target you’ll never hit.
Meanwhile, the distributors who started experimenting with AI tools six months ago—messy data and all—are already seeing results.
They’re using AI to:
- Draft customer correspondence faster
- Research technical specifications across product lines
- Analyze bid opportunities more thoroughly
- Train new employees on complex product knowledge
None of these applications required a data warehouse project. None waited for the ERP cleanup initiative. They started where they were.
The Real Question You Should Be Asking
Stop asking: “Is our data clean enough for AI?”
Start asking: “Which AI applications require clean data, and which ones work despite our mess?”
This reframes the entire conversation.
Instead of a binary choice between “wait until data is perfect” or “proceed recklessly,” you can make strategic decisions about where to experiment now and where to invest in data quality for specific, high-value use cases later.
The distributors making real progress aren’t waiting for permission from their IT roadmap. They’re running small experiments with tools that meet them where they are—imperfect data and all.
Then something interesting happens.
As they work with these tools, they start understanding which data actually matters for their highest-value applications. They’re making smarter, more targeted decisions about data quality investments because they’re learning from real use cases instead of theoretical requirements.
What This Means for Your Operation
Your data will never be perfect. That’s not pessimism—it’s operational reality.
But you can start gaining value from AI today with the data you have.
Here’s how to think about it:
For exploratory AI work—using ChatGPT, Claude, or similar tools to help your team work faster—your current data is fine. These tools are designed to work with messy, unstructured information. Start here. Build familiarity. Learn what’s possible.
For analytical AI applications—demand forecasting, pricing optimization, custom models—invest in the specific data quality improvements those applications require. But make these investments after you’ve proven the use case matters, not before.
For everything in between—run small pilots. Test assumptions. See what works with your actual data before committing to major cleanup initiatives.
The pattern I’ve seen consistently: Distributors who experiment first and clean data strategically later move faster and make better investment decisions than those who try to perfect everything upfront.
The Competitive Reality
Here’s the uncomfortable truth: While you’re waiting for perfect data, your competitors are learning how to work with AI using imperfect data.
They’re not smarter. They’re not better funded. They don’t have cleaner systems.
They just started.
And every month they’re building organizational muscle memory around AI tools—understanding what works, training their teams, identifying high-value applications—while the “wait for perfect data” companies are still in planning mode.
This isn’t about being reckless. It’s about being strategic.
Know which applications need clean data. Invest accordingly.
But don’t let the pursuit of perfect data prevent you from capturing value that’s available right now, with the systems you already have.
What Changes Monday Morning
Look at your current AI initiatives—or the ones you’ve been delaying.
Ask yourself honestly: Are you waiting because the data genuinely isn’t sufficient for the specific application you have in mind?
Or are you waiting because you believe all AI requires perfect data?
If it’s the latter, you’re solving the wrong problem.
The technology that can help you today doesn’t need you to spend two years on a data warehouse project first. It needs you to start asking better questions of the messy data you already have.
Ready to move from waiting to learning?
Join us at Applied AI for Distributors in Chicago, June 23-25, 2026. Three days focused on practical AI implementation—not theory, not vendor pitches, just distributors learning from distributors who’ve already started.
You’ll see real use cases. Meet peers solving the same problems. Leave with applications you can test Monday morning.
Learn more and register at distributionstrategy.com/
Because perfect data isn’t coming. But better results are available right now.
Share this article:


Leave a Reply