A curated list of awesome resources about reward construction for AI agents. This repository covers cutting-edge research, and practical guides on defining and collecting rewards to build more intelligent and aligned AI agents.
- Introduction
- Synthesizing Verifiable Task
- Real-World Task Reward Construction
- Unsupervised Reward Construction
- Reward Model
- Evaluation and Benchmarks
- Contributing
What is Reward Construction?
Reward construction is the process of designing and collecting reward signals that guide AI agents toward desired behaviors and outcomes.
Why is Reward Construction Important?
Background: The Second Half & Era of Experience
The Second Half: Transitioning from creating new methods and models to defining new tasks
- First Half Focus: Exam-like tasks with universal methods (next token prediction, RL) and architectures (Transformer, GPT)
- Turning Point: Organic combination of universal methods and architectures, where RL on large models achieves generalization
- Second Half Focus: Project-based scenarios with multi-turn interactions and temporal learning
Era of Experience: Large Models + Reinforcement Learning = General Superhuman Agents
- Previous Era: Human Data Era with limitations of human-generated data and capabilities
- Current Opportunity: Combining self-discovery capabilities with task generality from the human data era
- Key Components: Environmental rewards, autonomous interaction, continuous experience streams, non-human planning and reasoning
In conclusion, reward construction provides interactive environments and learning signals. It becomes crucial for AI agent to get experience for new project. We divided Reward Construction research into 5 categories, including Synthesizing Verifiable Task, Real-World Task Reward Construction, Unsupervised Reward Construction, Reward Model and Evaluation and Benchmarks.
Scaling task quantities through constructing new verifiable task gyms, such as puzzles, games. Training agents to solve these task can enhance model general reasoning capabilities. We divided Reward Construction research to 4 types, including Multi-Modal Reasoning, Text-Based Puzzle Solving, Zero-Sum Games, Converting Open-Domain Tasks to Verifiable Tasks and Curriculum Learning.
-
(Arxiv'25) [2505]
Code2Logic: Game-Code-Driven Data Synthesis for Enhancing VLMs General Reasoning (Fudan University, Douyin Co., Ltd., Shanghai Innovation Institute, Shanghai Key Laboratory of Intelligent Information Processing, Institute of Modern Languages, Fudan University, Southern University of Science and Technology) - Using game code to synthesize verifiable multi-modal reasoning data for improving VLM general reasoning through RL
-
(Arxiv'25) [2506]
Play to Generalize: Learning to Reason Through Game Play (Rice University, Johns Hopkins University, NVIDIA) - Training expert models on Snake game to provide ground truth for multi-modal reasoning rewards, improving general reasoning capabilities through RL
-
(Arxiv'25) [2508]
InternBootcamp Technical Report: Boosting LLM Reasoning with Verifiable Task Scaling (Shanghai AI Laboratory, Fudan University) - Advanced reasoning through structured thinking processes
-
(Arxiv'25) [2505]
SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond (The Hong Kong University of Science and Technology, MiniMax, The City University of Hong Kong) - Systematic approach to generating logical reasoning data
-
(Arxiv'25) [2505]
Enigmata: Scaling Logical Reasoning in Large Language Models with Synthetic Verifiable Puzzles (ByteDance Seed, Fudan University, Institute for AI Industry Research (AIR), Tsinghua University, Nanjing University, Shanghai Jiao Tong University, SIA-Lab of Tsinghua AIR and ByteDance Seed) - Creating synthetic puzzles to enhance logical reasoning capabilities
- (Arxiv'25) [2506]
SPIRAL: Self-Play on Zero-Sum Games Incentivizes Reasoning via Multi-Agent Multi-Turn Reinforcement Learning (National University of Singapore, Centre for Frontier AI Research (CFAR), A*STAR, Northeastern University, Sea AI Lab, Plastic Labs, University of Washington) - Using chess games' natural verifiable environments for self-play with win/loss rewards to enhance general reasoning
Transforming general tasks which usually trained through Pretraining and SFT, into RL-compatible formats.
-
(Arxiv'25) [2506] Reinforcement Pre-training (Microsoft Research, Peking University, Tsinghua University) - Converting next token prediction tasks into verifiable rewards where the next token serves as the verification signal
-
(COLM'25) [2503]
Learning to Reason for Long-Form Story Generation (University of Edinburgh) - Constructing next chapter prediction tasks for story generation, using subsequent chapters as ground truth for reward construction
-
(Arxiv'25) [2506]
RLPR: Extrapolating RLVR to General Domains without Verifiers (Tsinghua University, National University of Singapore, Shanghai Qi Zhi Institute, Harbin Institute of Technology, Beijing University of Posts and Telecommunications, University of Illinois Urbana-Champaign) - Extending reinforcement learning from verifiable reasoning to general domains
-
(Arxiv'25) [2505]
Reinforcing General Reasoning without Verifiers (Sea AI Lab, Singapore, University of Chinese Academy of Sciences, Institute of Automation, Chinese Academy of Sciences, National University of Singapore, University of Oxford, Renmin University of China) - General reasoning enhancement without explicit verification mechanisms
Scaling difficuity of task through curriculum learning, converting sparse reward to dense reward.
-
(Arxiv'25) [2507] Blending Supervised and Reinforcement Fine-Tuning with Prefix Sampling (University of Edinburgh, Fudan University, Alibaba Group, University of Amsterdam, Stepfun) - The paper introduces Prefix-RFT, a hybrid approach that combines Supervised Fine-Tuning and Reinforcement Fine-Tuning to enhance AI agent performance by synergizing learning from demonstration and exploration, particularly effective in mathematical reasoning tasks.
-
(ICML'24) [2402]
Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning (Fudan University, ByteDance Research) - The paper introduces R³, a method that uses reverse curriculum reinforcement learning to transform sparse outcome supervision into step-level supervisory signals, enhancing reasoning in large language models without requiring extensive manual annotations.
-
(ICLR'24) [2405]
Reverse Forward Curriculum Learning for Extreme Sample and Demonstration Efficiency in Reinforcement Learning (University of California, San Diego) - The paper introduces the Reverse Forward Curriculum Learning (RFCL) algorithm, which enhances reward construction for AI agents by leveraging a reverse curriculum to train an initial policy on a narrow state distribution and a forward curriculum to generalize it, significantly improving demonstration and sample efficiency without relying on engineered dense reward signals.
-
(Arxiv'25) [2505]
UFT: Unifying Supervised and Reinforcement Fine-Tuning (Massachusetts Institute of Technology) - The paper introduces Unified Fine-Tuning (UFT), which integrates supervision from SFT and reward from RFT into a single training paradigm, blending reward optimization with log-likelihood maximization on hints to enable simultaneous exploration and learning, addressing the trade-off between memorization and generalization.
Design reward function and synthesis data to scale up the quantities of the real-world reward. We divided Real-World Task Reward Construction research into 4 types, including Web Search, GUI, VLA and World Model.
-
(Arxiv'25) [2505]
WebDancer: Towards Autonomous Information Seeking Agency (Tongyi Lab, Alibaba Group) - Synthesizing agent action trajectories for search tasks, constructing verifiable QA pairs for RL training on information seeking
-
(Arxiv'25) [2507]
WebSailor: Navigating Super-human Reasoning for Web Agent (Tongyi Lab, Alibaba Group) - The paper's main contribution is the development of WebSailor, a post-training methodology that enhances AI agents' reward construction by generating high-uncertainty tasks and using an efficient RL algorithm, Duplicating Sampling Policy Optimization (DUPO), to match proprietary agents' performance in complex information-seeking tasks.
-
(Arxiv'25) [2507]
WebShaper: Agentically Data Synthesizing via Information-Seeking Formalization (Tongyi Lab, Alibaba Group) - Formalization-driven data synthesis framework for information-seeking tasks.
-
(Arxiv'25) [2508]
WebWatcher: Breaking New Frontiers of Vision-Language Deep Research Agent (Tongyi Lab, Alibaba Group) - The paper's main contribution regarding reward construction for AI agents is the development of a reinforcement learning approach using Group-Relative Policy Optimization (GRPO) with a reward function that balances format conformance and semantic accuracy, enhancing decision-making for multimodal deep research tasks.
-
(COLM'25) [2503]
Search-R1: Training LLMs to Reason and Leverage Search Engines with Reinforcement Learning (University of Illinois at Urbana-Champaign, University of Massachusetts Amherst, Google Cloud AI Research) - Search-R1 introduces a reinforcement learning approach that enables large language models to autonomously generate search queries during reasoning, using a simple outcome-based reward function to optimize multi-turn search interactions.
-
(ICLR'25) [2408]
Agent Q: Advanced Reasoning and Learning for Autonomous AI Agents (The AGI Company (MultiOn), Stanford University) - Combination of DPO, MCTS, and process supervision for web navigation tasks with multi-modal reasoning rewards
-
(ICLR'25) [2411]
WebRL: Training LLM Web Agents via Self-Evolving Online Curriculum Reinforcement Learning (Tsinghua University, Zhipu AI) - The paper introduces a robust outcome-supervised reward model (ORM) within the WEB RL framework to evaluate task success and provide binary reward signals, addressing the sparsity and cost of feedback signals in training LLM-based web agents.
-
(Arxiv'25) [2508]
Beyond Ten Turns: Unlocking Long-Horizon Agentic Search with Large-Scale Asynchronous RL (IIIS, Tsinghua University, Ant Research, RL Lab, University of Washington) - The paper introduces ASearcher, which employs a sparse-reward setting where rewards are computed at trajectory completion, using a combination of format reward and F1 score for base LLMs and LLM-as-Judge for fine-tuning large reasoning models, to enhance search intelligence in AI agents through reinforcement learning.
-
(Arxiv'25) [2504]
WebThinker: Empowering Large Reasoning Models with Deep Research Capability (Renmin University of China, BAAI, Huawei Poisson Lab) - The paper introduces an RL-based training strategy using iterative online Direct Preference Optimization (DPO) to enhance large reasoning models' ability to effectively utilize research tools by constructing preference pairs based on accuracy, tool usage, and output quality from sampled reasoning trajectories.
-
(Arxiv'25) [2506]
AgentCPM-GUI: Building Mobile-Use Agents with Reinforcement Fine-Tuning (Tsinghua University, Renmin University of China, ModelBest Inc.) - The paper's main contribution is the development of a reinforcement fine-tuning stage using Group Relative Policy Optimization (GRPO) to enhance the reasoning and decision-making capabilities of AgentCPM-GUI, enabling robust performance on complex, long-horizon GUI tasks.
-
(Arxiv'25) [2505]
ARPO:End-to-End Policy Optimization for GUI Agents with Experience Replay (The Chinese University of Hong Kong, SmartMore, Hong Kong University of Science and Technology) - The paper's main contribution is the introduction of Agentic Replay Policy Optimization (ARPO), which enhances reward construction for AI agents by integrating a replay buffer with Group Relative Policy Optimization to reuse successful experiences and stabilize training in sparse-reward GUI environments.
-
(Arxiv'25) [2504]
GUI-R1: A Generalist R1-Style Vision-Language Action Model For GUI Agents (Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, University of Chinese Academy of Sciences, National University of Singapore) - The paper introduces GUI-R1, which designs a rule-based unified action space reward function to evaluate and enhance the accuracy of predicted GUI actions across different platforms and task granularities, improving the performance of large vision-language models with significantly less data.
-
(ACL'25) [2412]
OS-Genesis: Automating GUI Agent Trajectory Construction via Reverse Task Synthesis (Shanghai AI Laboratory, The University of Hong Kong, Johns Hopkins University, Shanghai Jiao Tong University, University of Oxford, Hong Kong University of Science and Technology) - The paper introduces a Trajectory Reward Model (TRM) in the OS-Genesis pipeline, which uses a graded evaluation with scores from 1 to 5 to assess trajectory quality based on completion and coherence, enhancing the training of GUI agents by effectively utilizing both complete and incomplete trajectories.
-
(Arxiv'25) [2506]
AgentSynth: Scalable Task Generation for Generalist Computer-Use Agents (UC Berkeley) - AgentSynth's main contribution is a scalable and cost-efficient pipeline that automatically synthesizes diverse and realistic tasks for generalist computer-use agents by leveraging information asymmetry to create challenging long-horizon tasks from simple subtasks.
-
(Arxiv'25) [2505]
VLA-RL: Towards Masterful and General Robotic Manipulation with Scalable Reinforcement Learning (Tsinghua Shenzhen International Graduate School, Tsinghua University, School of Electrical and Electronic Engineering, Nanyang Technological University) - Reward design in embodied AI systems
-
(Arxiv'25) [2501] Improving vision-language-action model with online reinforcement learning (Institute for Interdisciplinary Information Sciences, Tsinghua University, Shanghai Qi Zhi Institute, University of California, Berkeley) - The paper introduces the iRe-VLA framework, which enhances reward construction for AI agents by iterating between reinforcement learning and supervised learning to stabilize training and improve performance in vision-language-action models for robotic control.
-
(Arxiv'25) [2505]
What Can RL Bring to VLA Generalization? (Tsinghua University, Shanghai Qi Zhi Institute, Beijing Zhongguancun Academy) - The paper's main contribution is developing an efficient PPO-based RL fine-tuning recipe that significantly enhances VLA generalization in semantic understanding and execution robustness compared to SFT, while maintaining comparable visual robustness.
-
(Arxiv'25) [2505]
Interactive Post-Training for Vision-Language-Action Models (UT Austin, Nankai University) - RIPT-VLA's main contribution is a reinforcement learning paradigm that fine-tunes pretrained Vision-Language-Action models using only sparse binary success rewards, enabling stable and data-efficient optimization without the need for shaped rewards, value functions, or reward modeling.
Towards future: Using world models and real-world interactions for reward construction.
-
[2508] Genie 3: A new frontier for world models (Google DeepMind) - Genie 3 introduces a general purpose world model that enhances AI agent training by generating consistent, interactive, and promptable environments, allowing for longer sequences of actions and more complex goal achievement through real-time navigation and text-based world event modifications.
-
(Arxiv'25) [2508]
Matrix-Game 2.0: An Open-Source, Real-Time, andStreaming Interactive World Model (Skywork AI) - Matrix-Game 2.0 for real-time interactive video generation
-
(ACL'24) [2406]
Can Language Models Serve as Text-Based World Simulators? (University of Arizona, New York University, Johns Hopkins University, Microsoft Research Montréal, Allen Institute for AI) - Exploring LLMs as world simulators for reward construction
Finding reward signals from model internals. Model generates data to train itself. We divided Unsupervised Reward Construction into 2 types, including Proposer and Solver and the discussion of can Large Reasoning Models Self-Train.
Models simultaneously act as problem proposers and solution generators, creating new training data.
-
(Arxiv'25) [2505] Absolute Zero: Reinforced Self-play Reasoning with Zero Data (Tsinghua University, Beijing Institute for General Artificial Intelligence, Pennsylvania State University) - Models serving dual roles as problem proposers and solvers, using compiler verification for self-improvement without external data
-
(Arxiv'25) [2508] R-Zero: Self-Evolving Reasoning LLM from Zero Data (Tencent AI Seattle Lab, Washington University in St. Louis, University of Maryland, College Park, The University of Texas at Dallas) - R-Zero's main contribution is introducing a fully autonomous framework where a Challenger and a Solver LLM co-evolve through interaction, with rewards based on proposing and solving increasingly challenging tasks, respectively, to improve reasoning capabilities without relying on pre-existing human-curated data.
-
(COLM'25) [2508] Self-Questioning Language Models (Carnegie Mellon University) - The paper introduces Self-Questioning Language Models (SQLM), an asymmetric self-play framework where AI agents improve reasoning skills by generating and solving their own questions, using reinforcement learning with rewards based on problem difficulty and majority voting or unit tests for verification, without needing curated training data.
Extracting learning signals from model internals, confidence scores, and consistant behaviors without external validation or verification mechanisms.
-
(Arxiv'25) [2505]
The Entropy Mechanism of Reinforcement Learning for Reasoning Language Models (Shanghai AI Laboratory, Tsinghua University, UIUC, Peking University, Nanjing University, CUHK) - Understanding how entropy affects reinforcement learning in reasoning models, it indicated that the improvement of self-train may cause by the entropy mechanism.
-
(Arxiv'25) [2506]
Spurious Rewards: Rethinking Training Signals in RLVR (University of Washington, Allen Institute for Artificial Intelligence, University of California, Berkeley) - Can random reward improve model performance?
-
(Arxiv'25) [2505]
Can Large Reasoning Models Self-Train (Ubiquant) - The paper's main contribution is demonstrating that entropy minimization with a single unlabeled data point and 10 optimization steps can achieve performance improvements comparable to or better than traditional methods using thousands of data points and carefully designed rewards in rule-based reinforcement learning for AI agents.
-
(Arxiv'25) [2504]
TTRL: Test-Time Reinforcement Learning (Tsinghua University, Shanghai AI Lab) - The paper introduces Test-Time Reinforcement Learning (TTRL), a novel method that uses majority voting from Test-Time Scaling to construct effective rewards for training Large Language Models on unlabeled data, enabling significant performance improvements.
-
(Arxiv'25) [2505]
Enhancing Reasoning Ability through RL without Labels (Sea AI Lab, Singapore, University of Chinese Academy of Sciences, Institute of Automation, Chinese Academy of Sciences, National University of Singapore, University of Oxford, Renmin University of China) - Completely unsupervised methods for reasoning enhancement
-
(Arxiv'25) [2505]
Maximizing Confidence Alone Improves Reasoning (Carnegie Mellon University) - The paper introduces RENT, a fully unsupervised reinforcement learning method that uses a model's entropy as an intrinsic reward to improve reasoning by reinforcing chains of thought with high model confidence, eliminating the need for external rewards or ground-truth answers.
-
(Arxiv'25) [2505]
VeriFree: Reinforcing General Reasoning without Verifiers (Sea AI Lab, Singapore, University of Chinese Academy of Sciences, Institute of Automation, Chinese Academy of Sciences, National University of Singapore, University of Oxford, Renmin University of China)
-
(COLM'25) [2504]
Right Question is Already Half the Answer: Fully Unsupervised LLM Reasoning Incentivization (Tianjin University, Tencent AI Lab, NUS) - The paper introduces Entropy Minimized Policy Optimization, a fully unsupervised method that enhances large language model reasoning by minimizing predictive entropy on unlabeled questions, achieving competitive performance without supervised signals.
-
(Arxiv'25) [2505]
Learning to Reason without External Rewards (UC Berkeley, Yale University) - The paper introduces Intuitor, a method that uses a model's self-certainty as an intrinsic reward signal for unsupervised reinforcement learning, enabling effective reasoning and generalization without external rewards.
Scaling preference data for reward models training to enable policy learning on general tasks.
-
(Arxiv'25) [2504] Inference-Time Scaling for Generalist Reward Modeling (DeepSeek-AI, Dept. of Computer Sci. & Tech., Tsinghua University, Institute for AI Industry Research (AIR), Tsinghua University) - The paper's main contribution is the development of Self-Principled Critique Tuning (SPCT) to enhance the inference-time scalability and quality of generalist reward modeling for AI agents using pointwise generative reward modeling and online reinforcement learning.
-
(ACL'25) [2410] Generative Reward Models (SynthLabs, Stanford University) - The paper's main contribution is the introduction of GenRM, an iterative algorithm that trains LLMs on self-generated reasoning traces to create synthetic preference labels that align with human judgments, enhancing reward construction for AI agents by improving both in-distribution accuracy and out-of-distribution generalization compared to traditional methods.
-
(ICLR'25) [2408] Generative Verifiers: Reward Modeling as Next-Token Prediction (Google DeepMind, University of Toronto, Mila, UCLA, Carnegie Mellon University) - The paper introduces Generative Verifiers (GenRM), which reformulate reward modeling as next-token prediction, leveraging LLMs' text generation capabilities to enhance verification performance through chain-of-thought reasoning and majority voting, outperforming traditional discriminative verifiers.
-
(Arxiv'25) [2505]
WorldPM: Scaling Human Preference Modeling (Qwen Team, Alibaba Group, Fudan University) - Exploring scalability of preference modeling using 15M human forum data, showing clear scaling laws in objective tasks like mathematics
-
(Arxiv'25) [2507]
POLAR:Pre-Trained Policy Discriminators are General Reward Models (Shanghai AI Laboratory, Fudan University) - The paper introduces POLAR, a novel criterion-agnostic pre-training paradigm for reward modeling that uses policy discrimination to enhance scalability and generalization, significantly outperforming traditional methods in reward signal accuracy and policy performance.
-
(Arxiv'25) [2505]
WavReward: Spoken Dialogue Models With Generalist Reward Evaluators (Zhejiang University, Alibaba Group) - Reward evaluation for spoken dialogue systems
-
(NeurIPS'23) [2312] Vision-Language Models as a Source of Rewards (Google DeepMind) - The paper demonstrates that off-the-shelf vision-language models, specifically from the CLIP family, can be used as a source of rewards to train reinforcement learning agents capable of achieving a variety of language-specified visual goals.
-
(ICML'24) [2402] Code as Reward: Empowering Reinforcement Learning with VLMs (Mila, McGill University, Google DeepMind, University of California, Berkeley) - The paper introduces the Code as Reward (VLM-CaR) framework, which generates dense and accurate reward functions for reinforcement learning agents by leveraging Vision-Language Models through code generation, reducing computational costs compared to direct VLM querying.
-
(ICML'24) [2402]
RL-VLM-F: Reinforcement Learning from Vision Language Foundation Model Feedback (Robotics Institute, Carnegie Mellon University, Department of Computer Science, University of Southern California) - RL-VLM-F introduces a method that automatically generates reward functions for AI agents by leveraging feedback from vision language foundation models, using task goal descriptions and visual observations to learn rewards from preference labels rather than raw scores.
-
(ACL'25) [2507] Dynamic and Generalizable Process Reward Modeling (College of Computer Science and Artificial Intelligence, Fudan University, The University of Hong Kong) - The paper introduces Dynamic and Generalizable Process Reward Modeling (DG-PRM), which uses a reward tree and Pareto dominance estimation to dynamically construct and allocate fine-grained, multi-dimensional process rewards, enhancing AI agent performance and generalizability across diverse tasks.
-
(ACL'25) [2501] The Lessons of Developing Process Reward Models in Mathematical Reasoning (Alibaba Group) - Supervising intermediate reasoning steps rather than just final outcomes
-
(ACL'24) [2312]
Math-Shepherd: Verify and Reinforce LLMs Step-by-step without Human Annotations (National Key Laboratory for Multimedia Information Processing, Peking University, DeepSeek-AI, The University of Hong Kong, Tsinghua University, The Ohio State University) - The paper introduces Math-Shepherd, a process-oriented reward model that assigns scores to each step of math problem solutions using automatically constructed supervision data, enhancing the performance of LLMs in verification and reinforcement learning without relying on human annotations.
-
(ICLR'24) [2305]
Let’s Verify Step by Step (OpenAI) - The paper demonstrates that process supervision, which provides feedback for each intermediate reasoning step, significantly outperforms outcome supervision in training reliable reward models for AI agents solving mathematical problems, achieving a 78% solve rate on a subset of the MATH test set.
Providing benchmarks or gyms to evaluate the model proformance. We divided Evaluation and Benchmarks into 4 types: Reward Model Benchmarks, Game Gym, Web Search, Computer Use and New Evaluation Dimension.
-
(NAACL'25) [2403]
RewardBench (Allen Institute for Artificial Intelligence, University of Washington, Berkman Klein Center, Harvard Law) - Comprehensive benchmark for reward model evaluation
-
(CVPR'25) [2411]
VL-RewardBench (HKU, SCUT, SJTU, PKU, UW, Allen AI) - Vision-language reward model benchmarking
-
(ACL'25) [2501]
PRMBench: A Fine-grained and Challenging Benchmark for Process-Level Reward Models (Fudan University, Shanghai AI Laboratory, Soochow University, Stony Brook University, The Chinese University of Hong Kong) - The paper introduces PRMBENCH, a comprehensive benchmark with 6,216 samples and 83,456 step-level labels, to evaluate the fine-grained error detection capabilities of Process-level Reward Models (PRMs) across dimensions like simplicity, soundness, and sensitivity, revealing significant weaknesses in current models and guiding future research.
-
(ACL'25) [2412]
PROCESSBENCH (Alibaba Inc.) - The paper introduces PROCESSBENCH, a benchmark for evaluating the ability of process reward models (PRMs) and critic models to identify errors in mathematical reasoning, revealing that existing PRMs struggle to generalize to challenging problems and underperform compared to critic models and a newly fine-tuned PRM on the PRM800K dataset.
-
(NeurIPS'24) [2402]
GTBench: Uncovering the strategic reasoning limitations of llms via game-theoretic evaluations (Drexel University, Boston University, LLNL, Lehigh University, UNC Chapel Hill, MIT, Harvard University) - The paper introduces GTBench, a language-driven environment with 10 game-theoretic tasks, to evaluate and characterize the strategic reasoning limitations of LLMs in competitive scenarios, highlighting their performance across different game types and the impact of code-pretraining on strategic reasoning.
-
(ICLR'25) [2412]
GameArena: Evaluating LLM Reasoning through Live Computer Games (University of California, San Diego, University of California, Berkeley) - The paper does not discuss reward construction for AI agents, so no summary can be provided for this specific aspect.
-
(Arxiv'24) [2403]
How Far Are We on the Decision-Making of LLMs? Evaluating LLMs’ Gaming Ability in Multi-Agent Environments (The Chinese University of Hong Kong, Tencent AI Lab, The Chinese University of Hong Kong, Shenzhen, Tsinghua University) - The paper introduces GAMA(
$\gamma$ )-Bench, a new framework with a dynamic scoring scheme to evaluate LLMs' gaming ability in multi-agent environments, enhancing the assessment of robustness and generalizability. -
(ICLR'25) [2411]
Balrog: Benchmarking agentic llm and vlm reasoning on games (AI Centre, University College London, IDEAS NCBR, University of Warsaw, University of Oxford, New York University, Anthropic, Institute of Mathematics, Polish Academy of Sciences) - A benchmark for evaluating the agentic capabilities of LLMs and VLMs in game environments.
-
(ICML'25) [2505]
ScienceBoard: Evaluating Multimodal Autonomous Agents in Realistic Scientific Workflows (The University of Hong Kong, Shanghai AI Laboratory, Fudan University, Peking University, Nanjing University, East China Normal University, Yale University) - Evaluating multimodal autonomous agents in scientific workflows through the SCIENCE BOARD environment and benchmark.
-
(NeurIPS'24) [2406]
GameBench: Evaluating strategic reasoning abilities of llm agents (Olin College of Engineering, University of Texas at Dallas, University of Oxford, University College London, ShanghaiTech University, Columbia University, University of Manchester) - The paper introduces GameBench, a cross-domain benchmark to evaluate the strategic reasoning abilities of LLM agents across 9 game environments, focusing on reasoning skills not heavily represented in pretraining data, and tests enhancements like Chain-of-Thought and Reasoning Via Planning.
-
(ICLR'24) [2311] GAIA: a benchmark for General AI Assistants (FAIR, Meta, HuggingFace, AutoGPT, GenAI, Meta) - The paper introduces GAIA, a benchmark designed to evaluate AI assistants on real-world questions requiring fundamental abilities like reasoning and tool-use, highlighting the gap between human and AI performance to guide the development of more robust AI systems.
-
(Arxiv'25) [2504]
BrowseComp: A Simple Yet Challenging Benchmark for Browsing Agents (OpenAI) - The paper introduces BrowseComp, a benchmark designed to evaluate the persistence and creativity of browsing agents in navigating the web to find hard-to-find information, analogous to programming competitions for coding agents.
-
(Arxiv'25) [2504]
BrowseComp-ZH: Benchmarking Web Browsing Ability of Large Language Models in Chinese (Hong Kong University of Science and Technology (Guangzhou), Peking University, Mindverse AI, Alibaba Group, Zhejiang University, Zhejiang University of Technology, MBZUAI, NIO, HSBC, Harvard T.H. Chan School of Public Health) - The paper introduces BrowseComp-ZH, a benchmark to evaluate the web browsing ability of large language models in Chinese, highlighting the challenges in retrieval and reasoning for these models in a linguistically and infrastructurally complex environment.
-
(Arxiv'24) [2406]
τ-bench: A Benchmark for Tool-Agent-User Interaction in Real-World Domains (Mila – Quebec AI Research Institute, Polytechnique Montréal) - The paper introduces a reward system for AI agents in τ-bench that evaluates success based on whether the final database state matches the unique ground truth outcome and whether the agent's responses contain all necessary information, ensuring faithful and fast evaluation.
-
(NeurIPS'24) [2404]
OSWORLD: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments (The University of Hong Kong, CMU, Salesforce Research, University of Waterloo) - The paper introduces an execution-based reward function in the OSWORLD environment that awards a value of 1 or a positive decimal under 1 at the final step if the state transitions meet the task objective, or 0 otherwise, to evaluate AI agents' performance on real computer tasks.
-
(Arxiv'24) [2412]
TheAgentCompany: Benchmarking LLM Agents on Consequential Real World Tasks (Carnegie Mellon University, Duke University) - A benchmark for evaluating LLM agents on real-world professional tasks.
-
(ICML'24) [2403]
WorkArena: How Capable Are Web Agents at Solving Common Knowledge Work Tasks? (ServiceNow Research, Mila – Quebec AI Research Institute, Polytechnique Montréal, McGill University, Université de Montréal) - The paper introduces WorkArena and BrowserGym to evaluate and design large language model-based agents for web-based knowledge work tasks, highlighting their current limitations and performance disparities between open and closed-source models.
-
(Arxiv'25) [2504]
BrowseComp: A Simple Yet Challenging Benchmark for Browsing Agents (OpenAI) - The paper introduces BrowseComp, a benchmark designed to evaluate the persistence and creativity of browsing agents in navigating the web to find hard-to-find information, without focusing on reward construction.
- Game data isn't utilized thoroughly. Games are shown effective to enhance model's general abilities, but open-source models rarely include games as training data.
- Using World Model to construct rewards. World models are capable of generating an endless variety of action-controllable, playable 3D environments for training and evaluating embodied agents.
- Evaluating interactable environments. A method to evaluate interactable environments, find high quality ones and choose environments fitting policy model's ability level would boost training a lot.
We welcome contributions to this repository! Please feel free to:
- Submit pull requests to add new papers
- Improve paper categorization and descriptions
- Add implementation details or code repositories
- Suggest new categories or reorganization
When adding papers, please include:
- Paper title and authors
- Brief description of the reward construction method
- Key contributions and results
- Links to paper and code (if available)
If you find this repository useful, please consider citing:
@misc{awesome-agent-reward,
title={Awesome Agent Reward: Reward Construction for AI Agents},
author={[Jingqi Tong, Yurong Mou, Jun Zhao, Hangcheng Li, Yongzhuo Yang, Mingzhe Li, Zhangye Yin]},
year={2025},
url={https://github.com/tongjingqi/Awesome-Agent-Reward}
}Note: This is a living document that will be continuously updated as the field of agent reward construction evolves. Stay tuned for the latest developments!
