#Reward Modeling
Latest news articles tagged with "Reward Modeling". Follow the timeline of events, related topics, and entities.
Articles (5)
-
🇺🇸 CDRRM: Contrast-Driven Rubric Generation for Reliable and Interpretable Reward Modeling
[USA]
arXiv:2603.08035v1 Announce Type: new Abstract: Reward modeling is essential for aligning Large Language Models(LLMs) with human preferences, yet conventional reward models suffer from poor interpret...
Related: #AI Alignment -
🇺🇸 RM-R1: Reward Modeling as Reasoning
[USA]
arXiv:2505.02387v4 Announce Type: replace-cross Abstract: Reward modeling is essential for aligning large language models with human preferences through reinforcement learning. To provide accurate re...
Related: #AI Alignment -
🇺🇸 MARS: Margin-Aware Reward-Modeling with Self-Refinement
[USA]
arXiv:2602.17658v1 Announce Type: cross Abstract: Reward modeling is a core component of modern alignment pipelines including RLHF and RLAIF, underpinning policy optimization methods including PPO an...
Related: #Human Preference Data, #Data Augmentation, #Margin‑Aware Techniques, #Self‑Refinement -
🇺🇸 Capturing Individual Human Preferences with Reward Features
[USA]
arXiv:2503.17338v2 Announce Type: replace Abstract: Reinforcement learning from human feedback usually models preferences using a reward function that does not distinguish between people. We argue th...
Related: #Artificial Intelligence, #Reinforcement Learning from Human Feedback, #Personalization, #Large Language Models -
🇺🇸 Automatically Finding Reward Model Biases
[USA]
arXiv:2602.15222v1 Announce Type: cross Abstract: Reward models are central to large language model (LLM) post-training. However, past work has shown that they can reward spurious or undesirable attr...
Related: #Large Language Models, #Bias Detection, #AI Safety, #Iterative Machine Learning
Key Entities (1)
- AI alignment (2 news)
About the topic: Reward Modeling
The topic "Reward Modeling" aggregates 5+ news articles from various countries.