Relations (1)
Facts (2)
Sources
Survey and analysis of hallucinations in large language models frontiersin.org 2 facts
procedureMitigation strategies for large language model hallucinations at the modeling level include Reinforcement Learning from Human Feedback (RLHF) (Ouyang et al., 2022), retrieval fusion (Lewis et al., 2020), and instruction tuning (Wang et al., 2022).
procedureTechniques such as Reinforcement Learning with Human Feedback (RLHF) (Ouyang et al., 2022) and Retrieval-Augmented Generation (RAG) (Lewis et al., 2020) are used to address model-level limitations regarding hallucinations.