Blogs

GPT-5: OpenAI just rewrote the AI playbook — again

On August 8th, OpenAI released its latest flagship model, GPT-5 — just three days after unveiling its open-weight models, gpt-oss. With foundational models now shaping everything from customer service to product design and enterprise workflows, this release deserves closer attention. I’m sharing a few reflections on what GPT-5 signals — not just about model capabilities, but about where AI is headed. 

You can watch the full launch stream here.

A new age of intelligence

Model consolidation is moving faster than we anticipated — and GPT-5 is the clearest signal yet. For everyday users, the distinction between “thinking” and “non-thinking” modes is now invisible. GPT-5 introduces a real-time model router that automatically decides when to reason, when to use tools, and when to act — without any manual switching. That’s a win for many (myself included) who often forget to toggle to o3 for complex tasks. But for power users who value explicit control, it may feel like the model is making too many decisions on their behalf. (Though to be fair, paid users still have access to the dedicated Thinking mode)

This shift feels like a step closer to AGI. Not because it is AGI — but because it increasingly behaves like a system that understands what we want, even before we do. In a world where intelligence is often judged by perception, that alone may be enough to change how we interact with machines.

A model that understands users, not just prompts

OpenAI seems to have found its sweet spot in coding, math, voice, and health—key verticals where AI usability and business value are both high. Others are keeping pace: Claude is very competitive in coding with its Claude Code CLI approach while Perplexity is carving a niche in finance. In tech circles today, the hot question is whether you're building a real product—or just a wrapper.  

But here’s the truth: even the standout names are wrappers. Cursor, Eleven Labs, ChatPRD, Lovable, Bolt—they all build on top of foundation models. Their success doesn’t refute the dominance of base models; it reinforces a more important point: distribution and experience still wins. Most users don’t care if your app runs on GPT-4, GPT-5, or something else entirely. They care if it helps them get tasks done faster and better.

That window is quickly getting smaller. As general-purpose models like GPT-5 become more capable, many wrappers will struggle to differentiate. OpenAI, for its part, isn’t just a model lab—it’s the most successful AI product company in the world. One of the best at designing product experiences around evolving model capabilities. But even so, during the GPT-5 demo, they showcased Cursor (a company it once tried to acquire)—a reminder that distribution, UX, and well-crafted workflows still matter, even when the foundation model is equally capable.

One model for everything (?)

GPT-5 is reaching the point of one-shotting complex tasks such as coding a computer game or a live dashboard. It now tops SWE benchmarks and delivers state-of-the-art performance and reliability with far less scaffolding. In the live demo, a four-line prompt generated a production-ready, gamified French language learning app. Lovable does this well too—but if one subscription can help me ship an app on the weekend and nail a client proposal by Monday, I’ll happily spend the extra $25 for 100 credits somewhere else.

The writing edge (still has edges)

Writing remains the biggest use case. GPT-5 is said to be the most capable writing collaborator yet, though some still prefer GPT-4.5 or Claude depending on style. GPT-5’s advantage lies in instruction-following and personalisation. Its ability to adopt distinct voices—like Cynic, Listener, Robot, Nerd—hints at a future of creative and technical writing that’s as expressive as it is assistive.

Coding as the killer app

In the enterprise, coding still holds the strongest case for adoption. GPT-5 is now trained to be a better collaborator—what the industry calls pair programming—across both vibe-coding and serious production work. It understands user intent with much more nuance, breaks down complex prompts into execution plans, and handles longer sessions with ease. It can retrace its steps, walk users through its reasoning, and even backtrack on changes it made earlier.

It also comes with a stronger sense of “engineering taste”, better UI preferences, development frameworks and best practices, built-in steerability, and more advanced agentic capabilities—like reasoning, tool use, and parallel tool calling. GPT-5 isn’t just generating code anymore; it’s working alongside you to ship. With out-of-the-box support for free-form function calling, it can take a high-level user instruction and spin up a team of sub-agents to get the job done. Together, these push us to rethink how we build software—and how much of it we still need to build ourselves.

Rewriting the enterprise stack

GPT-5 is positioned to expand OpenAI’s lead in enterprise use, especially in large production-level deployments. In the demo, it built a production-grade finance dashboard in under 5 minutes. The implication is clear: internal tools, workflows, and clunky enterprise systems are now vulnerable to reinvention—quickly, cheaply, and with better UX.

The downstream opportunity is massive:

  • Rebuilding legacy internal apps
  • Accelerating developer productivity
  • Improving internal user experience at scale

The stack beneath the stack

Model capabilities will keep getting better. But it’s the infrastructure and compute layer that’s heating up the fastest. Demand for compute, training, and infrastructure has become the new economic and geopolitical fault line—driving policy shifts, data sovereignty debates, and energy planning. Countries are competing not just on model performance, but on who controls the stack that powers intelligence.  

In the enterprise world, the conversation has moved from building AI features to securing the capability to run them—at scale, with reliability, and under cost pressure. Model performance is no longer limited by innovation alone; it’s gated by access to GPUs, memory throughput, and the energy footprint of inferencing at scale.

That’s why Nvidia’s recent achievement of $4 trillion in market cap isn’t just hype—it’s a signal of where the real power lies in the AI economy. But Nvidia isn’t alone. AMD is taking a different route to power the stack—betting on modularity, open ecosystem partnerships, and price-performance optimization to carve out space in hyperscaler and enterprise deployments. While Nvidia leans into an integrated, end-to-end AI platform, AMD is positioning itself as the more flexible, cost-efficient alternative for infrastructure builders who want optionality without sacrificing performance.

The pseudo-integration race has begun

As new models drop, the pressure to integrate—at least visibly—is ramping up. Companies are already refreshing their libraries and deployment stacks to accommodate GPT-5 in production. OpenAI has released pricing for three variants of the GPT-5 family via API, and coding-centric platforms like Microsoft (naturally), Cursor, GitHub, and Windsurf have jumped in with early access.

But beneath the flurry of integrations lies a deeper shift: the stack is moving—and no one wants to be caught flat-footed. Even if what’s being integrated is largely a wrapper for now, the signalling value is real. Enterprise buyers are watching. Developers are experimenting. And the next wave of adoption may be shaped less by what the model can do, and more by who made it easiest to plug in.

A smarter but safer intelligence, really?

Developers managing production deployments will find fewer safety concerns with GPT-5. Its responses are significantly more accurate and reliable, especially on complex, open-ended questions. Benchmarks like LongFact show that GPT-5 is six times less likely to hallucinate compared to o3. This improvement comes from a new training approach—introducing evaluations designed to probe the model’s reasoning on impossible, underspecified, or resource-scarce tasks, while also lowering its susceptibility to deception.

Its system card—unlike the conspicuous absence in Grok 4—makes this clear: GPT-5 is less agreeable, less fluffy, and far less sycophantic. Fewer emojis. Fewer overconfident wrong answers. A model that's finally learning when to say “I don’t know.” and gives users a reason why.

Now in everyone’s hands

GPT-5 is now available to all Plus, Pro, Team, and Free users, with enterprise and education access rolling out next week. Personally, I’m looking forward to seeing what it unlocks—both at work and in the messier parts of everyday life. Let’s just hope I don’t hit the usage limits too soon.

Enjoyed the read? Let’s take it further.


Connect to unlock exclusive insights, smart AI tools, and real connections that spark action.

Schedule a chat to unlock the full experience