safety-evaluation

There are 2 repositories under safety-evaluation topic.

  • IS2Lab/S-Eval

    S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language Models

  • SORRY-Bench/sorry-bench

    SORRY-Bench: Systematically Evaluating Large Language Model Safety Refusal Behaviors

    Language:Jupyter Notebook30100