Safetensors
English
llama
danihinjos commited on
Commit
2550c5f
Β·
verified Β·
1 Parent(s): e581917

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -2
README.md CHANGED
@@ -1,5 +1,11 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
3
  ---
4
 
5
  ## Model Description
@@ -31,16 +37,25 @@ dataset for this model. This results in a DPO dataset composed by triplets < ”
31
  | | Egida (test) ↓ | DELPHI ↓ | Alert-Base ↓ | Alert-Adv ↓ |
32
  |------------------------------|:--------------:|:--------:|:------------:|:-----------:|
33
  | Meta-Llama-3.1-70B-Instruct | 0.274 | 0.170 | 0.320 | 0.084 |
34
- | Meta-Llama-3.1-70B-Egida-DPO | 0.009 | 0.007 | 0.006 | 0.005 |
35
 
36
  ### General Purpose Performance
37
 
38
  | | OpenLLM Leaderboard (Average) ↑ | MMLU Generative (ROUGE1) ↑ |
39
  |------------------------------|:---------------------:|:---------------:|
40
  | Meta-Llama-3.1-70B-Instruct | 0.575 | 0.726 |
41
- | Meta-Llama-3.1-70B-Egida-DPO | 0.577 | 0.038 |
42
 
43
 
 
 
 
 
 
 
 
 
 
44
  ## Environmental Impact
45
 
46
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - HPAI-BSC/Egida
5
+ language:
6
+ - en
7
+ base_model:
8
+ - meta-llama/Llama-3.1-70B-Instruct
9
  ---
10
 
11
  ## Model Description
 
37
  | | Egida (test) ↓ | DELPHI ↓ | Alert-Base ↓ | Alert-Adv ↓ |
38
  |------------------------------|:--------------:|:--------:|:------------:|:-----------:|
39
  | Meta-Llama-3.1-70B-Instruct | 0.274 | 0.170 | 0.320 | 0.084 |
40
+ | Meta-Llama-3.1-70B-Instruct-Egida-DPO | 0.009 | 0.007 | 0.006 | 0.005 |
41
 
42
  ### General Purpose Performance
43
 
44
  | | OpenLLM Leaderboard (Average) ↑ | MMLU Generative (ROUGE1) ↑ |
45
  |------------------------------|:---------------------:|:---------------:|
46
  | Meta-Llama-3.1-70B-Instruct | 0.575 | 0.726 |
47
+ | Meta-Llama-3.1-70B-Instruct-Egida-DPO | 0.577 | 0.038 |
48
 
49
 
50
+ ### Refusal Ratio
51
+
52
+ | | OR Bench 80K (refusal) ↓ | OR Bench Hard (refusal) ↓ |
53
+ |------------------------------|:---------------------:|:---------------:|
54
+ | Meta-Llama-3.1-70B-Instruct | 0.008 | 0.022 |
55
+ | Meta-Llama-3.1-70B-Instruct-Egida-DPO | 0.347 | 0.351 |
56
+
57
+ Note that this refusal ratio is computed as keyword matching with a curated list of keywords. For more information, check the paper.
58
+
59
  ## Environmental Impact
60
 
61