Papers
arxiv:2604.09482

Process Reward Agents for Steering Knowledge-Intensive Reasoning

Published on Apr 10
· Submitted by
Michael Moor
on Apr 13
Authors:
,
,
,
,

Abstract

Process Reward Agents provide domain-grounded, online step-wise rewards for frozen policies in knowledge-intensive reasoning, enabling improved search-based decoding and generalizing across different model sizes without retraining.

AI-generated summary

Reasoning in knowledge-intensive domains remains challenging as intermediate steps are often not locally verifiable: unlike math or code, evaluating step correctness may require synthesizing clues across large external knowledge sources. As a result, subtle errors can propagate through reasoning traces, potentially never to be detected. Prior work has proposed process reward models (PRMs), including retrieval-augmented variants, but these methods operate post hoc, scoring completed trajectories, which prevents their integration into dynamic inference procedures. Here, we introduce Process Reward Agents (PRA), a test-time method for providing domain-grounded, online, step-wise rewards to a frozen policy. In contrast to prior retrieval-augmented PRMs, PRA enables search-based decoding to rank and prune candidate trajectories at every generation step. Experiments on multiple medical reasoning benchmarks demonstrate that PRA consistently outperforms strong baselines, achieving 80.8% accuracy on MedQA with Qwen3-4B, a new state of the art at the 4B scale. Importantly, PRA generalizes to unseen frozen policy models ranging from 0.5B to 8B parameters, improving their accuracy by up to 25.7% without any policy model updates. More broadly, PRA suggests a paradigm in which frozen reasoners are decoupled from domain-specific reward modules, allowing the deployment of new backbones in complex domains without retraining.

Community

Paper submitter

Process reward agents (PRA) is new a framework for disentangling reasoning (i.e. frozen LRM) from the domain knowledge (eg medical guidelines), here operated by a process reward agent that can search, reward - to then steer a frozen reasoning policy and improve overall and step correctness.

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2604.09482
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2604.09482 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2604.09482 in a dataset README.md to link it from this page.

Spaces citing this paper 1

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.