Scaling LLM Products for the Future
A few months ago, I came across a fascinating article that argued two divergent possibilities about the future of large language models (LLMs). Unfortunately, I neglected to bookmark it, and now I can’t find it anywhere. Since I brought it up again in a recent conversation with a friend, I realized I might as well outline these ideas myself rather than letting them go un-cited altogether.
In this post, I’d like to lay out the scenario described in that (now lost) article and reflect on how we, as software engineers or product builders, might navigate the rapidly shifting terrain of AI capabilities.
Scenario
In short products today fall into two camps of startups. Imagine that if you’re an ambitious software engineer or founder and want to build a new product that leverages the power of LLMs. Before you write your first line of code or wire up your favorite AI API, you need to make a critical decision:
- Bet that your AI will remain on par with today's most popular LLMs
In other words, assume that any improvements in LLM technology will be incremental, and the performance we see now is roughly the “ceiling” for the product you plan to build. - Bet that LLMs will continue to improve at the exponential pace of recent years
This means designing with the expectation that future LLMs might be orders of magnitude more powerful, flexible, and capable than what we have at our disposal today.
Why Option 1 Appears Tempting at First
- Simplicity: You don’t need to constantly re-architect your system to accommodate new model sizes, advanced features, or more complex integrations.
- Predictable Costs: You can optimize for a known set of capabilities and computational requirements, helping keep infrastructure and engineering budgets stable.
- Faster Time-to-Market: By locking in a stable, well-known model, you can focus on shipping quickly, especially if your product satisfies a straightforward need.
For some early-stage teams, these can be compelling reasons to keep things simple. However, these advantages may be short-lived if the state of the art in LLMs continues to advance rapidly.
Risks of Betting on Today's LLMS
Rapid Technological Advancements
- Historical Evidence: Over just a few years, LLMs have evolved from simple text predictors to sophisticated, multimodal reasoning systems (ChatGPT, GPT-4, etc.).
- Exponential Growth: Many experts in AI believe the pace of improvement will continue—or even accelerate—as new architectures, larger training datasets, and refined techniques are developed.
Implication: By building around today’s models, you risk being blindsided by breakthrough capabilities that arrive sooner and more powerfully than anticipated. Your product could quickly look outdated compared to newer competitors leveraging next-generation models.
Limited Ability to Innovate
- Restricted Feature Set: If you’re building around a certain set of LLM capabilities, you may close yourself off to new features that could emerge from more advanced models (such as real-time learning, complex reasoning on large knowledge graphs, or sophisticated context management).
- Lack of Flexibility: A system architected for one stable AI model can be a nightmare to retrofit for bigger and more complex models down the road.
Implication: You might miss out on entirely new opportunities—like offering specialized domain expertise or advanced analytics—because you designed everything around a stagnant AI core.
Mitigating the Risk
Betting wholeheartedly on “today’s ceiling” can be a perilous path. Instead, you could:
- Build Modular Architectures
Design a product that can easily plug in new models or integrate with emerging AI services. Even if you start small, keep the door open to future upgrades. - Adopt an “MVP + Research” Approach
Launch a minimal viable product with current LLMs for market validation, but allocate resources to exploring how next-gen models could integrate down the road. - Monitor AI Advances
Keep an eye on major AI research (e.g., from OpenAI, Google DeepMind, Anthropic). Maintaining a light R&D function or partnership ensures you won’t miss critical breakthroughs.
Bringing it Together
While it’s tempting to assume we’ve reached a plateau in LLM performance—or that the cost or complexity of chasing the AI frontier is too high—history suggests this is a risky bet. If AI continues to advance as swiftly as it has in recent years, products that settle for today’s models could be outpaced before they ever achieve significant market traction.
Bottom Line: Designing with the expectation of growth can save you from future headaches, enable you to stay ahead of the competition, and ensure that your product scales in tandem with AI capabilities. In a field evolving this quickly, planning for tomorrow isn’t a luxury—it’s a necessity.