reference
The research paper 'What features in prompts jailbreak LLMs? Investigating the mechanisms behind attacks' explores the mechanisms behind adversarial attacks on large language models.

Authors

Sources

Referenced by nodes (1)