2025 in Review: The Year AI Stopped Being Experimental
Looking back at 12 months of AI development — the models, the infrastructure, the real deployments, and the lessons learned. What actually changed in 2025, and what it means for 2026.
2025 in Review: The Year AI Stopped Being Experimental
At the start of 2025, most organisations were still asking "should we be doing AI?" By the end, the question had become "how do we scale what we've started?"
That shift — from experimentation to operationalisation — is the defining narrative of the year. It didn't happen uniformly or without friction, but it happened. And the gap between organisations that made this transition and those that didn't grew substantially over the course of 12 months.
Here's how I'd characterise the year, month by month and theme by theme.
The Model Landscape Transformed
Open source caught up
The year opened with DeepSeek R1 in January — a signal that efficient, high-quality open-source models were no longer a step behind frontier proprietary models. Llama 4 Scout and Maverick in April confirmed the trend. By the end of 2025, the question for enterprise AI builders had genuinely shifted from "open vs. closed" to "which deployment model fits our requirements?"
For organisations with data sovereignty requirements, high-volume economics that favour self-hosting, or specific domain fine-tuning needs — open models became a viable primary option, not a compromise.
Reasoning became standard
Early 2025 saw reasoning models (o3, R1, Claude 3.7 with extended thinking) as a distinct premium tier. By the second half of the year, reasoning capabilities had diffused across the entire model landscape — even smaller, cheaper models incorporated structured chain-of-thought inference. This raised the floor of what "standard" AI capability looks like significantly.
The multimodal baseline shifted
2025 was the year when "it handles images too" stopped being a feature and became table stakes. By Q3, every frontier model was natively multimodal. More importantly, reasoning across modalities — connecting what's in an image with what's in the accompanying text, or reasoning about a chart in the context of a financial document — reached practical quality thresholds for production use.
The Infrastructure Matured
Agent infrastructure became enterprise-ready
The biggest infrastructure shift of 2025 was the maturation of agentic AI infrastructure. Microsoft's Azure AI Agent Service reached GA, Anthropic released frameworks for multi-agent coordination, and LangChain/LangGraph reached production stability.
More significantly: organisations stopped asking "can we build agents?" and started asking "what's the right agent architecture for this workflow?" That's a sign of real maturity — when the conversation moves from capability to implementation trade-offs.
Observability and evaluation became first-class concerns
One of the quieter but most important developments of 2025 was the growth of AI observability tooling. LangSmith, Azure Monitor AI integration, Helicone, and similar tools matured to the point where production AI systems could be monitored, debugged, and optimised with the same rigour as any other software system.
This matters because you can't improve what you can't measure. Organisations that invested in observability in 2025 have a systematic understanding of where their AI systems add value and where they fail. Those that didn't are operating on intuition and hope.
RAG pipelines stabilised
Retrieval-Augmented Generation — the pattern of grounding AI responses in retrieved context rather than relying on parametric model knowledge — is now mature. The tooling is stable, the patterns are well-documented, and the failure modes are understood. Building a production RAG system in 2025 is a week-long project rather than a research exercise.
What Actually Changed in Organisations
The experimentation tax was paid
2025 was the year most organisations finished paying the experimentation tax — the POCs that didn't pan out, the internal chatbots that nobody used, the AI projects that delivered less than the business case promised.
This sounds negative, but it's actually healthy. Learning what doesn't work, and why, is essential to building what does. The organisations that have been experimenting since 2023 now have a much clearer picture of where AI creates value in their specific context.
AI literacy spread beyond technical teams
Twelve months ago, AI projects were predominantly driven by IT and data teams. By the end of 2025, business stakeholders — heads of operations, finance directors, customer experience teams — had become active participants in AI project definition. This is essential: the best AI projects come from people who deeply understand the business problem, not just the technology.
The talent picture changed
The talent shortage for AI implementation didn't resolve, but it evolved. The acute shortage of "build it from scratch" engineers became less acute as platforms matured and lowered the implementation barrier. The growing shortage is in AI product thinking — the ability to identify which business problems are good AI candidates, define success metrics, and iterate on deployed systems.
Three Lessons for 2026
Looking ahead, the three themes I expect to define 2026 are already visible in the trends of 2025:
1. AI governance becomes non-optional. The window of regulatory ambiguity is closing. Organisations without documented AI governance practices will face increasing friction — from regulators, from enterprise customers, from their own workforces. Building governance infrastructure now is table stakes for 2026.
2. Multi-agent systems go mainstream. Single-agent deployments are well-understood. The next wave of value comes from orchestrated networks of specialised agents — and the infrastructure to run them reliably is now in place. 2026 will be the year multi-agent systems move from advanced projects to standard deployments.
3. The productivity compounding begins. Organisations that deployed AI systematically in 2025 will see their second-order benefits in 2026: the internal knowledge that accumulated, the feedback loops that improved model performance, the processes that were redesigned around AI capabilities. Those that started later will find the gap harder to close.
2025 was the year AI became real in the enterprise. 2026 is when the advantage starts to compound.
Whatever your current starting point, the best time to build is now.