bjoernp commited on
Commit
a433937
1 Parent(s): ab25ecc

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +169 -0
README.md ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - Open-Orca/SlimOrca-Dedup
4
+ - teknium/openhermes
5
+ - meta-math/MetaMathQA
6
+ - migtissera/Synthia-v1.3
7
+ - THUDM/AgentInstruct
8
+ - LeoLM/German_Songs
9
+ - LeoLM/German_Poems
10
+ - LeoLM/OpenSchnabeltier
11
+ - bjoernp/ultrachat_de
12
+ language:
13
+ - en
14
+ - de
15
+ library_name: transformers
16
+ pipeline_tag: text-generation
17
+ license: llama2
18
+ model_creator: DiscoResearch
19
+ model_type: llama
20
+ tags:
21
+ - goliath
22
+ - deutsch
23
+ - llama2
24
+ - discoresearch
25
+ ---
26
+
27
+
28
+ ![EM Logo](https://raw.githubusercontent.com/jphme/jpdus.github.io/master/images/discoresearch.webp)
29
+
30
+ # DiscoLM 70b
31
+
32
+ **DiscoLM 70b** is a 70b model based on [Laion's LeoLM 70b](https://huggingface.co/LeoLM/leo-hessianai-70b) which underwent additional continued pretraing for 65b tokens of German
33
+ text, strengthening it's multilingual capabilities while retaining (or partially improving) English capabilities.
34
+ This was then further finetuned on a combination of some the most popular open-source instruction sets.
35
+ DiscoLM 70b is a [DiscoResearch](https://huggingface.co/DiscoResearch) project and was trained by [Björn Plüster](https://huggingface.co/bjoernp).
36
+
37
+ The model was trained with compute provided by [HessianAI](https://hessian.ai/) - we are very grateful for their support; please check out their wesbite and projects!
38
+
39
+ <img src="https://hessian.ai/wp-content/themes/hessianai/img/hessian-ai-logo.svg" width="120">
40
+
41
+ ## Table of Contents
42
+
43
+ 1. [Download](#download)
44
+ 2. [Benchmarks](#benchmarks)
45
+ 3. [Prompt Format](#prompt-format)
46
+ 4. [Dataset](#dataset)
47
+ 5. [Acknowledgements](#acknowledgements)
48
+ 6. [Contact](#contact)
49
+ 7. [About DiscoResearch](#about-discoresearch)
50
+ 8. [Disclaimer](#disclaimer)
51
+
52
+ ## Download
53
+
54
+ | Huggingface | GPTQ | GGUF | AWQ | *Base Model* |
55
+ |-------|-------|-------|-------|-------|
56
+ | [Link](https://huggingface.co/DiscoResearch/DiscoLM-70b) | soon | soon | soon | [LeoLM 70b](https://huggingface.co/LeoLM/leo-hessianai-70b) |
57
+
58
+ ## Benchmarks
59
+
60
+ ### Hugginface Leaderboard
61
+
62
+ This models is still an early Alpha and we can't guarantee that there isn't any contamination.
63
+ However, the average of **71.24** would earn the #2 spot on the HF leaderboard at the time of writing and the highest score for a >70b model yet.
64
+
65
+ | Metric | Value |
66
+ |-----------------------|-------|
67
+ | ARC (25-shot) | 68.77 |
68
+ | HellaSwag (10-shot) | 85.41 |
69
+ | MMLU (5-shot) | 68.64 |
70
+ | TruthfulQA (0-shot) | 57.69 |
71
+ | Winogrande (5-shot) | 83.27 |
72
+ | GSM8k (5-shot) | 63.68 |
73
+ | **Avg.** | **71.24** |
74
+
75
+ We use [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard.
76
+
77
+ <!--
78
+ ### FastEval
79
+
80
+ | Metric | Value |
81
+ |-----------------------|-------|
82
+ | GSM8K | 81.2 |
83
+ | Math | 22.3 |
84
+ | BBH | 72.9 |
85
+ | MMLU | 67.9 |
86
+ | **Avg.** | **53.3** |
87
+
88
+ ### MTBench
89
+
90
+ ```json
91
+ {
92
+ "first_turn": 8.45,
93
+ "second_turn": 7.45,
94
+ "categories": {
95
+ "writing": 9.4,
96
+ "roleplay": 8.65,
97
+ "reasoning": 6.85,
98
+ "math": 5.55,
99
+ "coding": 4.95,
100
+ "extraction": 9.15,
101
+ "stem": 9.225,
102
+ "humanities": 9.825
103
+ },
104
+ "average": 7.95
105
+ }
106
+ ```
107
+ -->
108
+ ## Prompt Format
109
+
110
+ This model follows the ChatML format:
111
+
112
+ ```
113
+ <|im_start|>system
114
+ You are DiscoLM, a helpful assistant.
115
+ <|im_end|>
116
+ <|im_start|>user
117
+ Please tell me possible reasons to call a research collective "Disco Research"<|im_end|>
118
+ <|im_start|>assistant
119
+ ```
120
+
121
+ This formatting is also available via a pre-defined Transformers chat template, which means that lists of messages can be formatted for you with the apply_chat_template() method:
122
+
123
+ ```python
124
+ chat = [
125
+ {"role": "system", "content": "You are DiscoLM, a helpful assistant."},
126
+ {"role": "user", "content": "Please tell me possible reasons to call a research collective Disco Research"}
127
+ ]
128
+ tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True)
129
+ ```
130
+
131
+ If you use `tokenize=True` and `return_tensors="pt"` instead, then you will get a tokenized and formatted conversation ready to pass to `model.generate()`.
132
+
133
+ ## Dataset
134
+
135
+ The dataset curation for DiscoLM 120b followed a "brute force"/"PoC" approach, as one goal was to see whether a 120b model can "absorb" more instruction data than a 70b model.
136
+
137
+ The following datasets were used for training DiscoLM 120b:
138
+
139
+ * [SlimOrca-Dedup](https://huggingface.co/datasets/Open-Orca/SlimOrca-Dedup)
140
+ * [OpenPlatypus](https://huggingface.co/datasets/garage-bAInd/Open-Platypus)
141
+ * [OpenHermes](https://huggingface.co/datasets/teknium/openhermes)
142
+ * [MetaMathQA](https://huggingface.co/datasets/meta-math/MetaMathQA)
143
+ * [UltraChat](https://huggingface.co/datasets/HuggingFaceH4/ultrachat_200k)
144
+ * [Synthia v.1.3](https://huggingface.co/datasets/migtissera/Synthia-v1.3)
145
+ * [AgentInstruct](https://huggingface.co/datasets/THUDM/AgentInstruct)
146
+
147
+ Many thanks for all dataset providers/curators!
148
+
149
+ ## Contact
150
+
151
+ Best way to reach us is on our [Discord](https://discord.gg/4pAqJP7W).
152
+
153
+ ## About DiscoResearch
154
+
155
+ DiscoResearch is an aspiring open research community. Disco should be a place where researchers from many communities can come together to combine their expertise and create innovative and groundbreaking LLMs. Come join our Discord, share your opinions and ideas, and advance open LLM research with us!
156
+
157
+ ## Acknowledgements
158
+
159
+ Disco 120b is a [DiscoResearch](https://huggingface.co/DiscoResearch) project and was trained by [Björn Plüster](https://huggingface.co/bjoernp). [Jan Harries](https://huggingface.co/jphme) helped with technical adivce, logistics and the Model Card and [AutoMeta](https://huggingface.co/Alignment-Lab-AI) also provided helpful technical adivce.
160
+ The model was trained with compute provided by [HessianAI](https://hessian.ai/) - many thanks in particular to [Patrick Schramowski](https://huggingface.co/PSaiml) for his support.
161
+
162
+ We are standing on the shoulders of giants; many thanks in no particular order to [alpindale](https://huggingface.co/alpindale) for Goliath 120b (with important contributions by [Charles Goddard](https://huggingface.co/chargoddard) and [Undi95](https://huggingface.co/Undi95)), [TheBloke](https://huggingface.co/TheBloke) for providing quantized versions, [winglian](https://huggingface.co/winglian) for Axolotl which was used to train the model and the SlimOrca dataset, [garage-bAInd](https://huggingface.co/garage-bAInd), [Teknium](https://huggingface.co/teknium), [Migel Tissera](https://huggingface.co/migtissera), [MetaMath](https://huggingface.co/meta-math) for their great datasets (please contact us if we forgot to mention you here!).
163
+
164
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
165
+
166
+ ## Disclaimer
167
+
168
+ The license on this model does not constitute legal advice. We are not responsible for the actions of third parties who use this model.
169
+ This model should only be used for research purposes. The original Llama2 license and all restrictions of datasets used to train this model apply.