Data as infrastructure

Photo by Franki Chamaki on Unsplash

We’ve explored how modern software-intensive companies must move beyond projects and Agile rituals toward continuous value delivery, superset platforms, data- and AI-driven learning loops and autonomous cross-functional teams. These aren’t isolated concepts, but they form a coherent operating model that I’ve coined Radical. But beneath all of them lies a deeper, more fundamental layer: data as infrastructure. Continuous value delivery, learning loops, team autonomy and platform thinking only scale when they’re built on a shared, reliable and evolving data substrate. Without that foundation, these concepts remain aspirations rather than operational realities.

Everything we do is driven by data. What we choose to measure, what we optimize, where we apply AI, how we experiment and what we learn from those experiments are all constrained by the data we collect and trust. Continuous value delivery is impossible without continuous, high-quality data on customer behavior, system performance, cost, risk and outcomes. Without data, value becomes a matter of opinion; with data, it becomes something we can sense, quantify and improve over time.

This is one of the clearest fault lines between Agile and Radical thinking. Agile largely treats data as feedback. Not unimportant, but secondary. Metrics are often team-local, sprint-focused and operational in nature. Learning happens, but it’s fragmented, slow and weakly connected to strategic decision-making. Data is consumed, not architected.

Radical treats data as infrastructure. Learning isn’t incidental; it’s designed. Data isn’t a by-product; it’s a strategic asset. Measurement, experimentation and AI aren’t add-ons to delivery but rather embedded into the operating model itself. In Radical organizations, data isn’t something teams ‘use’; it’s what the organization is structurally built around.

This shift changes how organizations organize themselves. Instead of organizing primarily around functions, projects or even products, leading organizations increasingly organize around their data. Data cuts across development, operations, sales, customer support and strategy. Treating data as infrastructure means acknowledging that it’s the shared foundation upon which all coordination, learning and scaling depend.

This is also what makes superset platforms possible. A superset platform is not just a technical platform; rather, it’s an organizational integration mechanism. Data is its connective tissue. Shared data models, shared telemetry and shared experimentation results allow teams to optimize locally while the organization optimizes globally. Without a strong data infrastructure, superset platforms collapse into loosely coupled silos with incompatible views of reality.

The same applies to data- and AI-driven learning loops. Learning loops only function when data is collected strategically, managed for quality and made broadly accessible. Experiments that can’t be trusted or reproduced don’t create learning; they create noise. Treating data as infrastructure means designing learning loops deliberately, ensuring that signals are reliable, comparable over time and usable by both humans and AI systems.

Finally, it reframes autonomous cross-functional teams. True autonomy is not independence from the organization; it’s independence in execution combined with alignment through shared data and performance metrics. When teams operate on a common data substrate, they can move fast locally without fragmenting the system globally. Autonomy without shared data leads to divergence; autonomy on top of shared data infrastructure creates speed with coherence.

To enable this, data must be treated as a first-class architectural asset. Like code, it must be governed, versioned, tested and evolved. Data schemas, pipelines and contracts must be explicit. Changes must be observable. Ownership must be clear, but access must be broad. Rather than viewing it as bureaucracy, it’s engineering discipline applied to learning and decision-making.

Many organizations invest heavily in cloud platforms, CI/CD pipelines and software architecture, while leaving data as an unmanaged side-effect. The result is fragile analytics, brittle AI models and endless internal debates about “whose numbers are right.” In contrast, organizations that embrace data as infrastructure apply the same rigor to data that they do to software. This unlocks faster experimentation, more effective AI, better decisions and sustained value creation.

In an AI-driven world, data is no longer something you analyze after delivery. It’s the infrastructure that makes continuous value delivery, superset platforms, learning loops and autonomous teams possible. If Agile optimized how teams build software, Radical is about how organizations build learning systems. And data, treated explicitly as infrastructure, is the foundation of that system. To end with William Edwards Deming: “Without data, you’re just another person with an opinion.”

Want to read more like this? Sign up for my newsletter at jan@janbosch.com or follow me on janbosch.com/blog, LinkedIn (linkedin.com/in/janbosch) or X (@JanBosch).