· Nolwen Brosson · Blog · 7 min read
Test-Time Compute vs Pre-Training: Is This the End of Giant Models?
For the past two years, one idea has kept coming up in AI research: what if the real driver of progress is no longer just training bigger and bigger models, but giving them more compute at the moment they answer? This approach has a slightly technical name, but it matters: test-time compute, also called inference-time scaling.
The question behind this debate is simple: should we still invest heavily in pre-training giant models, or can we get part of the gains by letting the model “think” longer at the moment of use?
TLDR: pre-training is still necessary. But yes, the center of gravity has started to shift. And that shift is already changing how AI products are designed, how infrastructure costs are managed, and even how competition between players is evolving.
Pre-training: the historical engine behind large models
For a long time, the dominant recipe was always the same: use more data, more parameters, more training compute, and watch performance improve steadily. The famous scaling laws largely confirmed this intuition. DeepMind’s Chinchilla paper showed in particular that, at a fixed compute budget, many models were mainly undertrained, and that model size and data volume had to be balanced more effectively.
In concrete terms, pre-training means exposing a model to a massive amount of text, code, or other data so it can learn the regularities of the world. This is the phase that builds its knowledge base, its statistical intuitions, its internal grammar, and part of its general abilities.
Without that foundation, a model cannot reason properly about concepts it has never learned to represent. Pre-training therefore remains the base layer.
Test-time compute: making the model work while it answers
Test-time compute is built on a different idea: instead of only improving the model’s brain before deployment, you give it more resources while it is solving a problem.
This can take several forms:
- generating several possible answers instead of just one,
- breaking a problem into intermediate steps,
- checking candidate solutions,
- backtracking when a path looks wrong,
- allocating more tokens or more reasoning effort to certain queries.
In other words, the model is no longer asked only to answer quickly. In some cases, it is asked to search, test, compare, and correct before producing a final answer.
This is exactly what made the new so-called “reasoning” models visible, such as OpenAI o1, introduced in September 2024 as a family of models designed to “spend more time thinking before they respond.” DeepSeek also helped popularize this direction with DeepSeek-R1, announced in January 2025 and then improved in May 2025.
Why this approach is so appealing
The main reason is economic as much as technical.
Pre-training a very large model is expensive. Once a player has already reached a certain scale, every marginal gain often requires heavier and heavier compute and data budgets. Test-time compute suggests a different curve: perform better at usage time, query by query, without necessarily rebuilding a giant model from scratch.
Work published in 2024 also showed that, under a certain inference budget, increasing compute intelligently at execution time can, on some difficult tasks, be more effective than increasing model size. This matters because it does not mean large models are useless. It means the optimal trade-off between training and inference is starting to change.
For companies, the promise is strong: instead of relying on one huge model that is expensive for everyone, you can imagine a system that saves its maximum effort only for complex queries.
What the general public should take away
The difference can be summed up like this:
Pre-training teaches the model about the world.
Test-time compute helps it organize itself better to solve a specific problem.
The first builds general capabilities. The second improves how those capabilities are used in demanding cases.
A simple analogy: pre-training is the years spent studying. Test-time compute is the time you are given during the exam, with the option to draft, check, and start over. A well-trained student still has an advantage. But when levels are close, the one who has more time to think can produce a much better paper.
Is this the end of giant models?
The headline is tempting, but the answer is no.
First, because test-time compute does not replace the knowledge learned during training. A weak model does not become excellent just because you give it more time. It may explore more, but it still explores with the tools it already has.
Second, because many test-time compute techniques work better when the base model is already strong. Inference gains often depend on the ability to generate useful intermediate steps, self-evaluate, or produce several plausible solutions.
And finally, because the best recent results rarely come from a single lever. In practice, progress often combines:
- strong pre-training,
- post-training,
- reinforcement learning,
- and smarter compute allocation at inference time.
So the real change is not the disappearance of giant models. It is the end of a simpler belief: “bigger at training time” is no longer enough on its own to explain the most visible progress.
The real question: where should the compute budget go?
The central issue is becoming almost a question of industrial strategy.
Is it better to spend the budget:
- on pre-training, to raise the baseline quality of the whole model,
- or on inference, to make certain queries much more capable?
The answer depends a lot on the product.
For a general-purpose assistant aimed at the public, latency and cost per query matter a lot. You cannot make every request “think” for thirty seconds. For rarer but critical use cases, such as code, math, document analysis, or some business workflows, accepting more inference compute can make sense if it increases reliability significantly.
In other words, the market may become more segmented:
- fast, low-cost models for everyday tasks,
- slower, more expensive reasoning models for complex tasks.
An important consequence for AI products
In the past, many AI products focused on choosing the “best model.” Tomorrow, the competitive advantage will also come from orchestration:
- when to trigger a more expensive reasoning mode,
- how many attempts to generate,
- how to verify an answer,
- when to use a smaller model,
- when to escalate to a stronger one.
The limits of test-time compute
Test-time compute still comes with several very real costs:
- more latency,
- higher cost per query,
- more orchestration complexity,
- and sometimes more instability, because generating several paths can also multiply errors when verification is weak.
It is not useful everywhere either. The gains are often especially visible on tasks that require explicit reasoning, such as mathematics, coding, or certain structured problems. On simple tasks, very common tasks, or tasks where speed matters most, the benefit is smaller.
Conclusion
This is probably the end of an era when AI progress could be summed up in one sentence: “we trained a bigger model.” Giant models still matter, but they are no longer enough. What matters now is the balance between what the model learned beforehand and how much compute we are willing to give it while it works.
For companies as well as product teams, the real question is no longer “big model or not.” The question is: where should compute be allocated intelligently to get the best quality, at the right cost, with the right user experience?
