Glossary
Reinforcement Learning from Human Feedback — a training technique that aligns LLM outputs with human preferences.