Evaluate language models through red-teaming

Haystack is a model evaluations and red-teaming suite for LLMs. This benchmark suite is designed to evaluate the performance and security of large language models through red-teaming.

Used and trusted by a growing community

  • 0
    Models Evaluated
  • 0
    Attack Success Rate
  • 3
    GitHub Stars