claim
Existing approaches for auditing large language models (LLMs) often focus on isolated aspects of model behavior, such as detecting specific biases or evaluating fairness, rather than understanding how outputs depend on each input token.

Authors

Sources

Referenced by nodes (1)