Reinforcement learning from human feedback
Machine learning technique
📊 Rating
4 news mentions · 👍 0 likes · 👎 0 dislikes
📌 Topics
- AI Alignment (2)
- Machine Learning (2)
- Causal Inference (1)
- Reinforcement Learning (1)
- AI Efficiency (1)
- AI Technology (1)
- Generative Models (1)
🏷️ Keywords
RLHF (3) · AI alignment (2) · computational efficiency (2) · CausalRM (1) · reward modeling (1) · observational feedback (1) · causal inference (1) · user feedback (1) · AdaBoN (1) · adaptive alignment (1) · best-of-N (1) · human preferences (1) · model uncertainty (1) · Partial Policy Gradients (1) · reinforcement learning (1) · large language models (1) · fine-tuning (1) · model alignment (1) · Curriculum-DPO (1) · Text-to-image generation (1)
📖 Key Information
📰 Related News (4)
-
🇺🇸 CausalRM: Causal-Theoretic Reward Modeling for RLHF from Observational User Feedbacks
arXiv:2603.18736v1 Announce Type: cross Abstract: Despite the success of reinforcement learning from human feedback (RLHF) in aligning language model...
-
🇺🇸 AdaBoN: Adaptive Best-of-N Alignment
arXiv:2505.12050v3 Announce Type: replace-cross Abstract: Recent advances in test-time alignment methods, such as Best-of-N sampling, offer a simple ...
-
🇺🇸 Partial Policy Gradients for RL in LLMs
arXiv:2603.06138v1 Announce Type: cross Abstract: Reinforcement learning is a framework for learning to act sequentially in an unknown environment. W...
-
🇺🇸 Curriculum-DPO++: Direct Preference Optimization via Data and Model Curricula for Text-to-Image Generation
arXiv:2602.13055v1 Announce Type: cross Abstract: Direct Preference Optimization (DPO) has been proposed as an effective and efficient alternative to...
🔗 Entity Intersection Graph
People and organizations frequently mentioned alongside Reinforcement learning from human feedback:
-
🌐
AI alignment · 2 shared articles
-
Generative artificial intelligence · 1 shared articles