AI is forcing a rethink of data privacy - governance is the missing piece
- Summary:
- AI is changing data privacy faster than most organizations can govern it. Confluent's Peter Pugh-Jones explains why governance is the missing piece.
When organizations talk about data privacy, the focus is still largely on security breaches and leaks. But heading further into 2026, the bigger risk lies in how AI is increasingly being used to automate processes and tasks.
Coupled with the growing push for AI to make decisions, it changes the privacy equation entirely. Data is no longer just being stored or shared, it is being interpreted and acted on, often at speed and often in ways that are difficult to fully explain.
Even where organizations put lots of safeguards in place, these alone are no guarantee of safety. As AI becomes more and more embedded in everyday operations, the ability to explain how decisions were made — and what data they relied on — becomes essential.
In this context, governance is the topic leaders need to be discussing. Not as a compliance exercise, but as the practical foundation for using AI responsibly. Without that clarity, organizations risk moving fast while unknowingly accumulating serious privacy and regulatory exposure.
AI privacy now hinges on explanation
As AI-driven systems take on more responsibility, privacy risk is defined by whether organizations can clearly explain what their systems are doing.
That challenge is becoming sharper as regulation evolves. New ‘AI Acts’ are emerging across regions, such as South Korea, Europe and the UK, creating extra complications. Businesses are already struggling to demonstrate that automated decisions are lawful, proportionate, or appropriate, particularly when they affect different groups of users.
For many organizations, this exposes a practical gap. It’s one thing to deploy AI, it’s another to know exactly how its output was created and where it came from.This is where privacy has to be operational. If organizations cannot explain how an AI-driven outcome was produced, they may be forced to pause, rework, or even withdraw systems already in use. In some cases, they might have to pull products they’ve already released into the marketplace — a scenario that would be disastrous for any company.
A privacy challenge that’s often overlooked
One of the most underestimated privacy challenges emerging with AI is segmentation — understanding who systems are interacting with and what is appropriate for them.
Organizations have long segmented users by geography or interests. But AI introduces a more complex requirement in the ability to recognize when different rules should apply to different people. As regulations evolve, particularly around age and vulnerability, organizations are being pushed to understand how to segment their customer bases in entirely new ways.
This becomes especially difficult when AI systems are trained broadly. If a model is built on the whole internet, it can be hard to determine where information came from or whether it should be used in a particular context at all. Content, recommendations, or automated messages that may be acceptable for one group can quickly become inappropriate for another.
Without this level of segmentation, organizations risk losing control over how AI behaves in real-world interactions. The ability or inability to distinguish between users and apply the right safeguards accordingly creates a new and largely untested privacy fault line.
Governing data late is no longer an option
One of the most common mistakes organizations make with AI and data platforms is treating governance as something that can be added later. The technology goes live, use cases multiply, and only then does the question arise - how well is any of this actually governed?
My experience suggests this approach doesn’t hold up. The reality is that the best place to govern data is as soon as it arrives in your organisation. Once data is already moving through systems, feeding models, and generating outcomes, retrofitting controls becomes complex, expensive, and risky.
This problem is particularly acute in real-time and streaming environments, where data is high-volume, high-speed, and often reused across multiple applications. Governance that isn’t designed in from the start quickly becomes a bottleneck.
Organizations may initially be excited by what their platforms can do, only to realise months later that critical controls are missing. At that point, they are forced to go back and rework foundations that should have been in place from day one, slowing innovation rather than enabling it.
Someone else has it covered. Or do they?
There is also a persistent assumption at senior levels that responsibility for AI risk sits neatly with someone else. A specialist team, a data science function, or a single executive role. The thinking goes - the right people have been hired, so the problem is being handled.
In practice, that assumption rarely holds. AI systems today are fundamentally different from those built even a decade ago. They are trained on vast, often opaque datasets, and their behavior can be difficult to predict without close scrutiny. Treating governance as a delegated task, rather than a shared responsibility, creates blind spots at precisely the moment clarity is most needed.
When accountability is unclear, so is ownership of outcomes. Decisions informed by AI still belong to the organisation and ultimately to its leaders. Without senior oversight and an understanding of how systems are built and governed, privacy risks will reach a stage where they’ll be impossible to ignore.
What organizations need to ask in 2026
Too often, the conversation about data privacy is framed around awareness. But in 2026, awareness alone is no longer enough.
The reality is that organizations can already build powerful AI-driven capabilities. The harder task is ensuring they are governed in a way that stands up to scrutiny. That means understanding data provenance, segmenting users responsibly, and putting controls in place before automation scales.
Above all, it means recognising that privacy is no longer a downstream concern. In an AI-driven world, it is inseparable from how systems are designed, deployed, and owned.
Businesses need to move beyond good intentions and focus instead on whether organizations can truly stand behind the decisions their technology is making.