claim
The AI assistant Claude, developed by Anthropic, utilizes sixteen rules to filter unsafe queries, such as threatening statements, gender-specific responses, and financial advice.
Authors
Sources
- Building Trustworthy NeuroSymbolic AI Systems - arXiv arxiv.org via serper