procedure
Fine-tuning models with guardrails involves training the model to refuse to answer when uncertain and including rejection examples in the dataset, such as 'I’m not sure about that. Let me check the source before answering.'
Authors
Sources
- The Role of Hallucinations in Large Language Models - CloudThat www.cloudthat.com via serper
Referenced by nodes (2)
- fine-tuning concept
- datasets concept