SP
BravenNow
Reinforcement learning from human feedback
🌐 Entity

Reinforcement learning from human feedback

Machine learning technique

📊 Rating

4 news mentions · 👍 0 likes · 👎 0 dislikes

📌 Topics

  • AI Alignment (2)
  • Machine Learning (2)
  • Causal Inference (1)
  • Reinforcement Learning (1)
  • AI Efficiency (1)
  • AI Technology (1)
  • Generative Models (1)

🏷️ Keywords

RLHF (3) · AI alignment (2) · computational efficiency (2) · CausalRM (1) · reward modeling (1) · observational feedback (1) · causal inference (1) · user feedback (1) · AdaBoN (1) · adaptive alignment (1) · best-of-N (1) · human preferences (1) · model uncertainty (1) · Partial Policy Gradients (1) · reinforcement learning (1) · large language models (1) · fine-tuning (1) · model alignment (1) · Curriculum-DPO (1) · Text-to-image generation (1)

📖 Key Information

In machine learning, reinforcement learning from human feedback (RLHF) is a technique to align an intelligent agent with human preferences. It involves training a reward model to represent preferences, which can then be used to train other models through reinforcement learning. In classical reinforcement learning, an intelligent agent's goal is to learn a function that guides its behavior, called a policy.

📰 Related News (4)

🔗 Entity Intersection Graph

AI alignment(2)Generative artificial intelligence(1)Reinforcement learning from human feedback

People and organizations frequently mentioned alongside Reinforcement learning from human feedback:

🔗 External Links