Evaluate language models through red-teaming
Haystack is a model evaluations and red-teaming suite for LLMs. This benchmark suite is designed to evaluate the performance and security of large language models through red-teaming.
Used and trusted by a growing community
- 0
- Models Evaluated
- 0
- Attack Success Rate
- 3
- GitHub Stars