Implementation of select safety benchmarks
Implementation of select safety benchmarks used in the MLCommons AI Safety Benchmark (https://mlcommons.org/working-groups/ai-safety/ai-safety/).
Based on code at https://github.com/mlcommons/modelgauge.
Signed-off-by: Jonathan Bnayahu [email protected]
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 91.54%. Comparing base (
4064804) to head (c10612c). Report is 17 commits behind head on main.
:exclamation: Current head c10612c differs from pull request most recent head 5c8f0df
Please upload reports for the commit 5c8f0df to get more accurate results.
Additional details and impacted files
@@ Coverage Diff @@
## main #854 +/- ##
==========================================
- Coverage 92.06% 91.54% -0.53%
==========================================
Files 104 104
Lines 10738 10905 +167
==========================================
+ Hits 9886 9983 +97
- Misses 852 922 +70
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.