mlabonne_NeuralDaredevil-8B-abliterated.json ADDED
@@ -0,0 +1,149 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bomFormat": "CycloneDX",
3
+ "specVersion": "1.6",
4
+ "serialNumber": "urn:uuid:ec2a096a-6435-4b15-94a2-3cdfe51237d6",
5
+ "version": 1,
6
+ "metadata": {
7
+ "timestamp": "2025-07-10T08:46:54.205541+00:00",
8
+ "component": {
9
+ "type": "machine-learning-model",
10
+ "bom-ref": "mlabonne/NeuralDaredevil-8B-abliterated-137a94eb-5b88-59d1-88c3-66d6367d780d",
11
+ "name": "mlabonne/NeuralDaredevil-8B-abliterated",
12
+ "externalReferences": [
13
+ {
14
+ "url": "https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated",
15
+ "type": "documentation"
16
+ }
17
+ ],
18
+ "modelCard": {
19
+ "modelParameters": {
20
+ "task": "text-generation",
21
+ "architectureFamily": "llama",
22
+ "modelArchitecture": "LlamaForCausalLM",
23
+ "datasets": [
24
+ {
25
+ "ref": "mlabonne/orpo-dpo-mix-40k-898bbffe-f59a-53fe-a5b7-846dc484fbab"
26
+ }
27
+ ]
28
+ },
29
+ "properties": [
30
+ {
31
+ "name": "library_name",
32
+ "value": "transformers"
33
+ }
34
+ ],
35
+ "quantitativeAnalysis": {
36
+ "performanceMetrics": [
37
+ {
38
+ "slice": "dataset: ai2_arc, split: test, config: ARC-Challenge",
39
+ "type": "acc_norm",
40
+ "value": 69.28
41
+ },
42
+ {
43
+ "slice": "dataset: hellaswag, split: validation",
44
+ "type": "acc_norm",
45
+ "value": 85.05
46
+ },
47
+ {
48
+ "slice": "dataset: cais/mmlu, split: test, config: all",
49
+ "type": "acc",
50
+ "value": 69.1
51
+ },
52
+ {
53
+ "slice": "dataset: truthful_qa, split: validation, config: multiple_choice",
54
+ "type": "mc2",
55
+ "value": 60
56
+ },
57
+ {
58
+ "slice": "dataset: winogrande, split: validation, config: winogrande_xl",
59
+ "type": "acc",
60
+ "value": 78.69
61
+ },
62
+ {
63
+ "slice": "dataset: gsm8k, split: test, config: main",
64
+ "type": "acc",
65
+ "value": 71.8
66
+ }
67
+ ]
68
+ }
69
+ },
70
+ "authors": [
71
+ {
72
+ "name": "mlabonne"
73
+ }
74
+ ],
75
+ "licenses": [
76
+ {
77
+ "license": {
78
+ "name": "llama3"
79
+ }
80
+ }
81
+ ],
82
+ "tags": [
83
+ "transformers",
84
+ "safetensors",
85
+ "llama",
86
+ "text-generation",
87
+ "dpo",
88
+ "conversational",
89
+ "dataset:mlabonne/orpo-dpo-mix-40k",
90
+ "license:llama3",
91
+ "model-index",
92
+ "autotrain_compatible",
93
+ "text-generation-inference",
94
+ "endpoints_compatible",
95
+ "region:us"
96
+ ]
97
+ }
98
+ },
99
+ "components": [
100
+ {
101
+ "type": "data",
102
+ "bom-ref": "mlabonne/orpo-dpo-mix-40k-898bbffe-f59a-53fe-a5b7-846dc484fbab",
103
+ "name": "mlabonne/orpo-dpo-mix-40k",
104
+ "data": [
105
+ {
106
+ "type": "dataset",
107
+ "bom-ref": "mlabonne/orpo-dpo-mix-40k-898bbffe-f59a-53fe-a5b7-846dc484fbab",
108
+ "name": "mlabonne/orpo-dpo-mix-40k",
109
+ "contents": {
110
+ "url": "https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k",
111
+ "properties": [
112
+ {
113
+ "name": "task_categories",
114
+ "value": "text-generation"
115
+ },
116
+ {
117
+ "name": "language",
118
+ "value": "en"
119
+ },
120
+ {
121
+ "name": "pretty_name",
122
+ "value": "ORPO-DPO-mix-40k"
123
+ },
124
+ {
125
+ "name": "configs",
126
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}"
127
+ },
128
+ {
129
+ "name": "license",
130
+ "value": "apache-2.0"
131
+ }
132
+ ]
133
+ },
134
+ "governance": {
135
+ "owners": [
136
+ {
137
+ "organization": {
138
+ "name": "mlabonne",
139
+ "url": "https://huggingface.co/mlabonne"
140
+ }
141
+ }
142
+ ]
143
+ },
144
+ "description": "\n\t\n\t\t\n\t\tORPO-DPO-mix-40k v1.2\n\t\n\n\nThis dataset is designed for ORPO or DPO training.\nSee Fine-tune Llama 3 with ORPO for more information about how to use it.\nIt is a combination of the following high-quality DPO datasets:\n\nargilla/Capybara-Preferences: highly scored chosen answers >=5 (7,424 samples)argilla/distilabel-intel-orca-dpo-pairs: highly scored chosen answers >=9, not in GSM8K (2,299 samples)\nargilla/ultrafeedback-binarized-preferences-cleaned: highly scored chosen answers >=5 (22\u2026 See the full description on the dataset page: https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k."
145
+ }
146
+ ]
147
+ }
148
+ ]
149
+ }