next BIG future

next BIG future

Test Time Training Will Take LLM AI to the Next Level

By Brian Wang

NextBigFuture's avatar
NextBigFuture
Nov 16, 2024
∙ Paid
Share

MIT researchers achieved 61.9% on ARC tasks by updating model parameters during inference.

Is this key to AGI?

We might reach the 85% AGI doorstep by scaling and integrating it with COT (Chain of thought) next year.

Test-time training (TTT) for large language models typically requires additional compute resources during inference compared to standard inference. The amount of extra compute needed can vary depending on the specific implementation and approach used. Here are some key points about the inference compute requirements for test-time training:

Compute Requirements

Increased Computation: TTT generally requires more computation than standard inference, as it involves adapting the model parameters for each test input or small batch of inputs

Variability: The exact amount of additional compute can vary significantly based on factors like the complexity of the task, the size of the model, and the specific TTT strategy employed

Comparison to Best-of-N: In some implementations, TTT can be more efficient than traditional best-of-N sampling approaches. For example, one study showed that a compute-optimal TTT strategy achieved better performance while using only about 25% of the computation required by best-of-N sampling

Factors Affecting Compute Requirements

Several factors influence the amount of inference compute needed for test-time training:

Task Difficulty: The complexity of the task or question being addressed affects the compute requirements. Easier tasks may require less additional compute, while more challenging problems might necessitate more intensive computation

Model Size: The base size of the language model impacts the overall compute needs. Smaller models adapted with TTT might require less total compute than much larger pre-trained models for certain tasks2

TTT Strategy: Different TTT approaches have varying compute requirements. For instance, strategies that involve multiple iterations of revision or complex search algorithms may require more computation than simpler methods1

Adaptive Allocation: Some advanced TTT implementations use adaptive strategies that allocate compute resources based on the perceived difficulty of the input. This can lead to more efficient use of compute, applying more resources only when necessary

Efficiency Considerations

While TTT does require additional compute during inference, it can potentially offer efficiency benefits:

Keep reading with a 7-day free trial

Subscribe to next BIG future to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Nextbigfuture
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture