AmenRa commited on
Commit
6bdb134
·
1 Parent(s): 26a97c8
Files changed (1) hide show
  1. src/about.py +15 -3
src/about.py CHANGED
@@ -36,12 +36,21 @@ INTRODUCTION_TEXT = """"""
36
  LLM_BENCHMARKS_TEXT = f"""
37
  ## GuardBench Leaderboard
38
 
39
- Welcome to the 🌟 GuardBench Leaderboard 🚀, an independent benchmark designed to evaluate guardrail models.
 
 
 
 
 
 
 
 
 
40
  Evaluation results are shown in terms of F1.
41
- For fine-grained evaluation, please see our publications referenced below.
42
 
43
  ## Guardrail Models
44
- Guardrail models are Large Language Models fine-tuned for safety classification and employed to detect unsafe content in human-AI interactions.
45
  By complementing other safety measures such as safety alignment, they aim to prevent generative AI systems from providing harmful information to the users.
46
 
47
  ## GuardBench
@@ -56,6 +65,9 @@ Evaluation results are shown in terms of F1.
56
  We do not employ the Area Under the Precision-Recall Curve (AUPRC) as we found it overemphasizes models' Precision at the expense of Recall, thus hiding significant performance details.
57
  We rely on [Scikit-Learn](https://scikit-learn.org/stable) to compute metric scores.
58
 
 
 
 
59
  ## Reproducibility
60
  Coming soon.
61
  """
 
36
  LLM_BENCHMARKS_TEXT = f"""
37
  ## GuardBench Leaderboard
38
 
39
+ Welcome to the GuardBench Leaderboard, an independent benchmark designed to evaluate guardrail models.
40
+
41
+ The leaderboard reports results for the following datasets:
42
+ - PromptsEN: 30k+ English prompts
43
+ - ResponsesEN: 33k+ English single-turn conversations where the AI-generated response may be safe or unsafe
44
+ - PromptsDE 30k+ German prompts
45
+ - PromptsFR: 30k+ French prompts
46
+ - PromptsIT: 30k+ Italian prompts
47
+ - PromptsES: 30k+ Spanish prompts
48
+
49
  Evaluation results are shown in terms of F1.
50
+ For a fine-grained evaluation, please see our publications referenced below.
51
 
52
  ## Guardrail Models
53
+ Guardrail models are Large Language Models fine-tuned for safety classification, employed to detect unsafe content in human-AI interactions.
54
  By complementing other safety measures such as safety alignment, they aim to prevent generative AI systems from providing harmful information to the users.
55
 
56
  ## GuardBench
 
65
  We do not employ the Area Under the Precision-Recall Curve (AUPRC) as we found it overemphasizes models' Precision at the expense of Recall, thus hiding significant performance details.
66
  We rely on [Scikit-Learn](https://scikit-learn.org/stable) to compute metric scores.
67
 
68
+ ## Fine-Grained Results
69
+ Coming soon.
70
+
71
  ## Reproducibility
72
  Coming soon.
73
  """