Executive Summary
Why this matters: AI projects fail not because of technology, but because of data. Most organisations overestimate their data readiness.
What to know: Data readiness means knowing what you have, where it lives, who maintains it, and whether it’s usable. Most Nigerian organisations have significant gaps in all four areas.
What to do: Conduct a data audit for one use case. Document what exists, where, and who owns it. Accept that you will find problems — that’s the point.
Watch out for: Confusing ‘having data’ with ‘having usable data.’ They are not the same thing.
Why AI Projects Fail Before They Start: The Data Reality Check
The Situation
Every AI vendor will tell you the same thing: AI needs data. The more data, the better. Feed the machine, and magic happens.
This creates a comforting assumption. Your organisation has data. Lots of it. Customer records, transaction histories, employee files, operational reports. Years of accumulated information sitting in systems, spreadsheets, and shared drives.
So when someone proposes an AI project, the natural response is: we have the data, let’s do it.
This is where most AI projects begin to fail. Not because the technology is wrong, but because the assumption about data is wrong.
Having data is not the same as having usable data. And the gap between those two things is where AI initiatives go to die.

What data readiness actually means
When AI experts talk about ‘data readiness,’ they mean something specific. It is not about volume. It is not about having a database. It is about four things:
Knowing what data exists. This sounds obvious, but most organisations cannot answer the question: What data do we have? Information is scattered across systems, departments, personal drives, and email inboxes. No one has a complete picture. When asked about customer data, different departments give different answers — because they are looking at different systems that have never been reconciled.
Knowing where data lives. Even when organisations know what data exists in theory, they often do not know where it physically resides. Is the customer database on the local server or in the cloud? Who has access? When was it last backed up? These questions reveal uncomfortable gaps.
Understanding what data represents. Data without context is noise. A column labelled ‘status’ means nothing until you know what statuses are possible, what they indicate, and how they were assigned. This contextual knowledge often lives in people’s heads, not in documentation. When those people leave, the knowledge leaves with them.
Trusting data accuracy. Data degrades. Customers move and addresses become outdated. Employees leave and records are not updated. Workarounds become permanent. Over time, the gap between what the data says and what is actually true grows wider. AI systems trained on inaccurate data produce inaccurate outputs — confidently.
The Nigerian context
These challenges are universal, but they have particular dimensions in Nigerian organisations.
Many Nigerian businesses were built on relationships, not systems. The most important information lives in the heads of key people — the accountant who knows which clients actually pay, the operations manager who knows which suppliers are reliable, the secretary who knows how to navigate the bureaucracy. This is not a flaw; it is how business works in relationship-driven economies.
But AI does not understand relationships. It understands data. And when the data is incomplete or misleading, AI amplifies the problem rather than solving it.
There is also the infrastructure reality. Power outages interrupt data collection. Internet connectivity affects cloud systems. Hardware failures cause data loss. These are not excuses; they are operating conditions that shape what data exists and how reliable it is.
Add to this the compliance dimension. The Nigeria Data Protection Act 2023 has specific requirements for how personal data is collected, stored, and processed. Many organisations have never audited their data against these requirements. They do not know what personal data they hold, where it came from, or whether they have proper consent to use it. This is a legal risk that AI adoption makes more urgent, not less.
Why this matters for AI
AI systems are pattern-recognition machines. They find patterns in data and use those patterns to make predictions or generate outputs. This works brilliantly when the data is clean, comprehensive, and representative.
When the data is messy, incomplete, or biased, the AI still finds patterns — but they are the wrong patterns. It learns to replicate errors, amplify biases, and make confident predictions based on flawed foundations.
This is why ‘garbage in, garbage out’ is not just a cliché. It is the single most common reason AI projects fail to deliver value. The technology works exactly as designed. The problem is what it was fed.
And here is the uncomfortable truth: you cannot fix data problems by spending money on AI tools. You fix data problems by doing the unglamorous work of understanding, cleaning, and documenting your data. There are no shortcuts.
Practical takeaway
Before investing in any AI project, conduct a data reality check. Start small — one use case, one data set — and answer these questions honestly:
- What data do we need for this use case? Be specific. Not ‘customer data’ but ‘customer purchase history for the past 24 months, including product, price, date, and channel.’
- Where does this data live? Which systems? Who controls access? How is it updated? If you cannot answer this, you have already found a problem.
- What does this data actually represent? How was it collected? What are the definitions? What are the known limitations? Document this, even if it feels tedious.
- How accurate is this data? When was it last validated? What error rate is acceptable? How would you know if it was wrong?
- Who is the data owner? Not the IT team — the business owner who understands what the data means and is accountable for its quality.
This audit will surface problems. That is the point. Better to discover them now than after you have spent money on AI tools that cannot work with what you have.
Risks or limitations
There is a risk that data readiness becomes an excuse for permanent inaction. The data will never be perfect. Waiting for perfect data means waiting forever.
The goal is not perfection. It is honesty. Know what you have. Know its limitations. Make informed decisions about what is good enough for your specific use case.
Some AI applications require high data quality. Others are more forgiving. The key is matching ambition to reality, not pretending the reality is different from what it is.
The organisations that succeed with AI are not the ones with the best data. They are the ones that understand their data honestly — and design AI initiatives accordingly.
But knowing your data situation is only part of the picture. The next question is: who decides what to do with it? That’s where role clarity comes in. Part 2: Who Decides What Happens to the Data?.


