|8 min read

2023: The Year LLMs Ate the World

Looking back on a year that transformed AI from a research curiosity into a force reshaping every industry

As 2023 comes to a close, I find myself looking back on twelve months that changed technology more profoundly than any single year I can remember. More than the iPhone launch. More than the cloud computing revolution. More than the container and microservices transformation. This was the year that large language models went from impressive research demos to tools that hundreds of millions of people use daily, and the ripple effects are only beginning.

The Timeline

The speed of this year's developments is staggering when you lay them out chronologically.

January brought Microsoft's multi-billion dollar investment in OpenAI, signaling that the largest technology companies saw generative AI as a platform-level shift worth enormous bets.

February saw Google scramble to announce Bard, revealing that even the company that invented the transformer architecture felt threatened by OpenAI's momentum. The demo stumble foreshadowed the challenges of deploying AI systems under competitive pressure.

March delivered GPT-4, a model that passed professional exams, processed images, and demonstrated reasoning capabilities that made the previous generation look primitive. The gap between GPT-3.5 and GPT-4 was so large that it forced everyone to recalibrate their expectations about the pace of AI progress.

April introduced autonomous agent frameworks like AutoGPT, showing that language models could be orchestrated into systems that plan, execute, and iterate. The implementations were crude, but the concept was powerful enough to capture the imagination of developers worldwide.

The middle of the year brought an explosion of tooling. LangChain, LlamaIndex, and dozens of other frameworks made it dramatically easier to build applications with LLMs. Vector databases proliferated. The concept of retrieval-augmented generation went from a research technique to an industry standard pattern.

July brought Llama 2 from Meta, the first genuinely capable open source LLM with a permissive commercial license. This single release expanded the universe of who could build with frontier AI capabilities.

The fall intensified everything. OpenAI's DevDay unveiled GPTs and the Assistants API. The board crisis that followed exposed the deep tensions between commercial ambition and safety concerns. AWS's re:Invent put generative AI at the center of Amazon's cloud strategy with Bedrock and related services.

December capped the year with Google's Gemini launch and Mixtral's demonstration that architectural innovation could challenge the scaling orthodoxy. The competitive landscape at year's end looks nothing like it did in January.

What Changed for Me

On a personal level, 2023 was the year I pivoted from cloud infrastructure architect to someone who builds with AI every day. That pivot was not a single decision; it was a series of steps, each building on the last.

I started the year as a curious observer. I ended it with working prototypes deployed to internal teams, a systematic understanding of transformer architectures and training methodologies, hands-on experience with multiple model providers and frameworks, and a growing conviction that the intersection of infrastructure engineering and AI is where the most important work of the next decade will happen.

The skills I built over years of Linux administration, cloud architecture, container orchestration, and infrastructure automation did not become irrelevant. They became the foundation for understanding how AI systems operate in production. The problems of model serving, data pipeline management, evaluation at scale, and system reliability are fundamentally infrastructure problems. My background gives me a perspective that many people entering AI from the machine learning side do not have.

What I Got Right

Looking back at my assessments throughout the year, some held up well.

I predicted that the AI arms race would intensify, and it did. Microsoft, Google, Amazon, Meta, and Anthropic are all investing billions of dollars and competing aggressively for talent, compute, and market position.

I argued that multi-provider strategy would become essential for enterprise AI, and the OpenAI board crisis validated that dramatically. Any organization that had built its entire AI strategy on a single provider had a very uncomfortable week in November.

I believed that open source models would close the gap with proprietary offerings faster than expected, and Llama 2 and Mixtral proved that right. The open source AI movement is more vibrant and capable than most people anticipated.

I predicted that agent architectures would become the next major application pattern, and while autonomous agents remain unreliable, the ecosystem has matured enormously. OpenAI's Assistants API, Amazon's Agents for Bedrock, and countless open source frameworks are all building toward a future of AI agents that can take actions, not just generate text.

What Surprised Me

Several developments caught me off guard.

The speed of consumer adoption was faster than I expected. ChatGPT did not gradually grow; it erupted. The rate at which non-technical people integrated AI into their daily workflows surprised me. Teachers, lawyers, marketers, writers, and people in every profession found ways to use these tools within weeks of discovering them.

The quality gap between models narrowed faster than I anticipated. In January, GPT-3.5 was clearly ahead of everything else. By December, there are multiple models from multiple providers that are competitive across most tasks. The idea that any single company would maintain a durable capability advantage seems increasingly unlikely.

The corporate drama at OpenAI was something nobody predicted. The firing and rehiring of Sam Altman was a reminder that the development of the most powerful technology in the world is subject to the same human dynamics, politics, ambition, disagreement, and ego, that affect every other organization.

What Concerns Me

Not everything about this year's developments is unambiguously positive.

The concentration of AI capability in a handful of large companies raises important questions about power, access, and accountability. Training frontier models costs hundreds of millions of dollars, creating a natural barrier to entry that limits who can participate in pushing the boundaries.

The safety and alignment challenges are real and growing. Models that are more capable are also capable of more harm. The governance frameworks, evaluation methodologies, and safety techniques needed to deploy AI responsibly are not maturing as fast as the capabilities themselves.

The impact on employment is beginning to manifest. Not in the dramatic "AI replaces all jobs" narrative, but in subtler ways: writing tasks that used to take hours take minutes, code generation tools change how software teams are structured, customer service automation reduces headcount in call centers. The full economic impact of these changes will unfold over years, but the direction is clear.

The environmental cost of training and serving large models is significant and often overlooked. The energy consumption of GPU clusters running at scale is substantial, and as AI usage grows, so does the carbon footprint.

What I Am Watching for 2024

Several threads from 2023 will continue into the new year:

Model capability: Will GPT-5 or its competitors deliver another generational leap? The scaling laws suggest it is possible, but there are real questions about whether the current paradigm of more data and more compute can continue to deliver proportional improvements.

Agent maturity: The agent architecture is the right concept but the current implementations are unreliable. Improvements in model reasoning, tool use, and planning could make agents viable for production use cases in 2024.

Open source momentum: The trajectory of open source models suggests that 2024 could bring models competitive with GPT-4 on many practical tasks. This would further democratize AI capabilities and shift the value from model providers to application builders.

Enterprise deployment: 2023 was the year of AI experimentation in enterprises. 2024 should be the year of AI production deployment. The organizations that have been building expertise and prototyping will start to deploy at scale, and the real challenges of operating AI systems in production will become apparent.

Regulation: Governments around the world are developing AI regulations. The EU AI Act, executive orders in the US, and various national frameworks will shape how AI can be developed and deployed. The regulatory landscape will become a significant factor in enterprise AI strategy.

The Year in Perspective

2023 was the year that artificial intelligence transitioned from a specialized technical domain to a general-purpose technology that touches every industry and every role. It was messy, chaotic, exciting, and sometimes concerning. The technology advanced faster than the institutions, regulations, and social norms needed to manage it.

For me personally, it was the most intellectually stimulating year of my career. Every week brought something new to learn, something to build, something to rethink. The combination of deep technical challenges and broad societal implications makes AI the most compelling area of technology I have encountered.

I do not know what 2024 will bring. Nobody does. But I know I will be building in this space, learning as fast as I can, and documenting the journey along the way.

What a year.

Share: