wasertech commited on
Commit
1466f0c
·
1 Parent(s): 99b57b7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -237
README.md CHANGED
@@ -7,13 +7,13 @@ language:
7
  - fr
8
  pipeline_tag: text-generation
9
  widget:
10
- - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n<s>[INST] Introduce yourself to the HuggingFace community. [/INST] "
11
  example_title: "Introduction"
12
- - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n<s>[INST] Describe your model. [/INST] "
13
  example_title: "Model Description"
14
- - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n<s>[INST] What the meaning of life? [/INST] "
15
  example_title: "Life's Meaning"
16
- - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n<s>[INST] What does the future reserve us? [/INST] "
17
  example_title: "What's next?"
18
  ---
19
 
@@ -27,236 +27,3 @@ It is also now supported by continuous batching server vLLM, allowing use of Lla
27
 
28
  As of September 25th 2023, preliminary Llama-only AWQ support has also been added to Huggingface Text Generation Inference (TGI).
29
 
30
- # Table of Contents
31
-
32
- - [Model Card for Assistant Llama 2 Chat AWQ](#model-card-for--model_id-)
33
- - [Table of Contents](#table-of-contents)
34
- - [Table of Contents](#table-of-contents-1)
35
- - [Model Details](#model-details)
36
- - [Model Description](#model-description)
37
- - [Uses](#uses)
38
- - [Direct Use](#direct-use)
39
- - [Downstream Use [Optional]](#downstream-use-optional)
40
- - [Out-of-Scope Use](#out-of-scope-use)
41
- - [Bias, Risks, and Limitations](#bias-risks-and-limitations)
42
- - [Recommendations](#recommendations)
43
- - [Training Details](#training-details)
44
- - [Training Data](#training-data)
45
- - [Training Procedure](#training-procedure)
46
- - [Preprocessing](#preprocessing)
47
- - [Speeds, Sizes, Times](#speeds-sizes-times)
48
- - [Evaluation](#evaluation)
49
- - [Testing Data, Factors & Metrics](#testing-data-factors--metrics)
50
- - [Testing Data](#testing-data)
51
- - [Factors](#factors)
52
- - [Metrics](#metrics)
53
- - [Results](#results)
54
- - [Model Examination](#model-examination)
55
- - [Environmental Impact](#environmental-impact)
56
- - [Technical Specifications [optional]](#technical-specifications-optional)
57
- - [Model Architecture and Objective](#model-architecture-and-objective)
58
- - [Compute Infrastructure](#compute-infrastructure)
59
- - [Hardware](#hardware)
60
- - [Software](#software)
61
- - [Citation](#citation)
62
- - [Glossary [optional]](#glossary-optional)
63
- - [More Information [optional]](#more-information-optional)
64
- - [Model Card Authors [optional]](#model-card-authors-optional)
65
- - [Model Card Contact](#model-card-contact)
66
- - [How to Get Started with the Model](#how-to-get-started-with-the-model)
67
-
68
-
69
- # Model Details
70
-
71
- ## Model Description
72
-
73
- <!-- Provide a longer summary of what this model is/does. -->
74
- This model is a quantitized export of [wasertech/assistant-llama2-7b-chat](https://huggingface.co/wasertech/assistant-llama2-7b-chat) using AWQ.
75
-
76
- AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference.
77
-
78
- It is also now supported by continuous batching server vLLM, allowing use of Llama AWQ models for high-throughput concurrent inference in multi-user server scenarios.
79
-
80
- As of September 25th 2023, preliminary Llama-only AWQ support has also been added to Huggingface Text Generation Inference (TGI).
81
-
82
- - **Developed by:** More information needed
83
- - **Shared by [Optional]:** More information needed
84
- - **Model type:** Language model
85
- - **Language(s) (NLP):** en, fr
86
- - **License:** llama2
87
- - **Parent Model:** More information needed
88
- - **Resources for more information:** More information needed
89
-
90
-
91
-
92
- # Uses
93
-
94
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
95
-
96
- ## Direct Use
97
-
98
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
99
- <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
100
-
101
-
102
-
103
-
104
- ## Downstream Use [Optional]
105
-
106
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
107
- <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
108
-
109
-
110
-
111
-
112
- ## Out-of-Scope Use
113
-
114
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
115
- <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
116
-
117
-
118
-
119
-
120
- # Bias, Risks, and Limitations
121
-
122
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
123
-
124
- Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
125
-
126
-
127
- ## Recommendations
128
-
129
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
130
-
131
-
132
-
133
-
134
-
135
- # Training Details
136
-
137
- ## Training Data
138
-
139
- <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
140
-
141
- More information on training data needed
142
-
143
-
144
- ## Training Procedure
145
-
146
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
147
-
148
- ### Preprocessing
149
-
150
- More information needed
151
-
152
- ### Speeds, Sizes, Times
153
-
154
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
155
-
156
- More information needed
157
-
158
- # Evaluation
159
-
160
- <!-- This section describes the evaluation protocols and provides the results. -->
161
-
162
- ## Testing Data, Factors & Metrics
163
-
164
- ### Testing Data
165
-
166
- <!-- This should link to a Data Card if possible. -->
167
-
168
- More information needed
169
-
170
-
171
- ### Factors
172
-
173
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
174
-
175
- More information needed
176
-
177
- ### Metrics
178
-
179
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
180
-
181
- More information needed
182
-
183
- ## Results
184
-
185
- More information needed
186
-
187
- # Model Examination
188
-
189
- More information needed
190
-
191
- # Environmental Impact
192
-
193
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
194
-
195
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
196
-
197
- - **Hardware Type:** More information needed
198
- - **Hours used:** More information needed
199
- - **Cloud Provider:** More information needed
200
- - **Compute Region:** More information needed
201
- - **Carbon Emitted:** More information needed
202
-
203
- # Technical Specifications [optional]
204
-
205
- ## Model Architecture and Objective
206
-
207
- More information needed
208
-
209
- ## Compute Infrastructure
210
-
211
- More information needed
212
-
213
- ### Hardware
214
-
215
- More information needed
216
-
217
- ### Software
218
-
219
- More information needed
220
-
221
- # Citation
222
-
223
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
224
-
225
- **BibTeX:**
226
-
227
- More information needed
228
-
229
- **APA:**
230
-
231
- More information needed
232
-
233
- # Glossary [optional]
234
-
235
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
236
-
237
- More information needed
238
-
239
- # More Information [optional]
240
-
241
- More information needed
242
-
243
- # Model Card Authors [optional]
244
-
245
- <!-- This section provides another layer of transparency and accountability. Whose views is this model card representing? How many voices were included in its construction? Etc. -->
246
-
247
- More information needed
248
-
249
- # Model Card Contact
250
-
251
- More information needed
252
-
253
- # How to Get Started with the Model
254
-
255
- Use the code below to get started with the model.
256
-
257
- <details>
258
- <summary> Click to expand </summary>
259
-
260
- More information needed
261
-
262
- </details>
 
7
  - fr
8
  pipeline_tag: text-generation
9
  widget:
10
+ - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n\n<s>[INST] Introduce yourself to the HuggingFace community. [/INST] "
11
  example_title: "Introduction"
12
+ - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n\n<s>[INST] Describe your model. [/INST] "
13
  example_title: "Model Description"
14
+ - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n\n<s>[INST] What the meaning of life? [/INST] "
15
  example_title: "Life's Meaning"
16
+ - text: "<<SYS>>\nYou are Assistant, a sentient AI.\n<</SYS>>\n\n<s>[INST] What does the future reserve us? [/INST] "
17
  example_title: "What's next?"
18
  ---
19
 
 
27
 
28
  As of September 25th 2023, preliminary Llama-only AWQ support has also been added to Huggingface Text Generation Inference (TGI).
29