reference
Miles Turpin, Julian Michael, Ethan Perez, and Samuel Bowman demonstrated that language models do not always provide faithful explanations when using chain-of-thought prompting in their 2023 paper 'Language models don’t always say what they think: Unfaithful explanations in chain-of-thought prompting'.
Authors
Sources
- A Survey of Incorporating Psychological Theories in LLMs - arXiv arxiv.org via serper
Referenced by nodes (1)
- Language Model concept