reference
Yang Liu, Yuanshun Yao, Jean-Francois Ton, Xiaoying Zhang, Ruocheng Guo, Hao Cheng, Yegor Klochkov, Muhammad Faaiz Taufiq, and Hang Li authored 'Trustworthy llms: a survey and guideline for evaluating large language models’ alignment', published as an arXiv preprint in 2023.
Authors
Sources
- A Survey of Incorporating Psychological Theories in LLMs - arXiv arxiv.org via serper
Referenced by nodes (2)
- Large Language Models concept
- AI alignment concept