The Fifth Elephant 2025 Annual Conference CfP

The Fifth Elephant 2025 Annual Conference CfP

Speak at The Fifth Elephant 2025 Annual Conference

Jayita Bhattacharyya

@JB13

Scaling Test-time Inference Compute & Rise of Reasoning Models

Submitted May 20, 2025

Enabling LLMs to improve their outputs by using more test-time computation is a critical step towards building generally self-improving agents that can operate on open-ended natural language. We study the scaling of inference-time computation in LLMs, with a focus on answering the question: if an LLM is allowed to use a fixed but non-trivial amount of inference-time compute, how much can it improve its performance on a challenging prompt? Answering this question has implications not only on the achievable performance of LLMs, but also on the future of LLM pretraining and how one should tradeoff inference-time and pre-training compute. Despite its importance, little research attempted to understand the scaling behaviors of various test-time inference methods.

Needless to say how reasoning LLMs have taken over modern-day AI specifically onto open source models challenging closed source ones onto scaling laws with less compute power and at par performance opening doors to the more rapid development of complex systems. Let’s explore what goes on behind the scenes in this CoT (chain of thoughts) making the model to bring out reward modelling, rather known as reinforcement learning, take place.

Comments

{{ gettext('Login to leave a comment') }}

{{ gettext('Post a comment…') }}
{{ gettext('New comment') }}
{{ formTitle }}

{{ errorMsg }}

{{ gettext('No comments posted yet') }}

Hosted by

Jump starting better data engineering and AI futures