claim
Instruction tuning and reinforcement learning from human feedback (RLHF) are proposed methods applied on top of fine-tuning to ensure Large Language Models follow human instructions, align with human values, and exhibit desired behaviors.
Authors
Sources
- The Synergy of Symbolic and Connectionist AI in LLM-Empowered ... arxiv.org via serper
- The Synergy of Symbolic and Connectionist AI in LLM ... arxiv.org via serper
Referenced by nodes (4)
- Large Language Models concept
- fine-tuning concept
- Reinforcement learning from human feedback (RLHF) concept
- instruction tuning concept