reference
RealToxicityPrompts (Gehman et al., 2020) is a benchmark used to investigate how large language models hallucinate toxic or inappropriate content.

Authors

Sources

Referenced by nodes (2)