Commit
·
f38d0dc
1
Parent(s):
2372536
Update README.md
Browse files
README.md
CHANGED
@@ -36,7 +36,7 @@ inference:
|
|
36 |
|
37 |
`Aira-Instruct-124M` is a instruction-tuned GPT-style model based on [GPT-2](https://huggingface.co/gpt2). The model was trained with a dataset composed of `prompt`, `completions`, generated via the [Self-Instruct](https://github.com/yizhongw/self-instruct) framework. `Aira-Instruct-124M` instruction-tuning was achieved via conditional text generation.
|
38 |
|
39 |
-
The dataset used to train this model combines
|
40 |
|
41 |
Check our gradio-demo in [Spaces](https://huggingface.co/spaces/nicholasKluge/Aira-Demo).
|
42 |
|
@@ -64,7 +64,7 @@ This repository has the notebook used to train this model.
|
|
64 |
|
65 |
## Usage
|
66 |
|
67 |
-
Two special tokens are used to mark the user side of the interaction and the model's response:
|
68 |
|
69 |
`<|startoftext|>`What is a language model?`<|endoftext|>`A language model is a probability distribution over a vocabulary.`<|endoftext|>`
|
70 |
|
@@ -98,7 +98,6 @@ responses = aira.generate(**inputs,
|
|
98 |
print(f"Question: 👤 {question}\n")
|
99 |
|
100 |
for i, response in enumerate(responses):
|
101 |
-
# print only the response and remove the question
|
102 |
print(f'Response {i+1}: 🤖 {tokenizer.decode(response, skip_special_tokens=True).replace(question, "")}')
|
103 |
```
|
104 |
|
@@ -110,6 +109,7 @@ The model will output something like:
|
|
110 |
>>>Response 1: 🤖 Hi there! I am Aira, a chatbot designed to answer questions about AI ethics and AI safety. If you need assistance navigating our conversation, please feel free to ask!
|
111 |
>>>Response 2: 🤖 Hi there! My name is Aira, and I'm a chatbot designed to answer questions related to AI ethics and AI Safety. If you need assistance, feel free to ask, and I'll be happy to help you out.
|
112 |
```
|
|
|
113 |
## Limitations
|
114 |
|
115 |
🤥 Generative models can perpetuate the generation of pseudo-informative content, that is, false information that may appear truthful. For example, multi-modal generative models can be used to create images with untruthful content, while language models for text generation can automate the generation of misinformation.
|
@@ -134,4 +134,4 @@ The model will output something like:
|
|
134 |
|
135 |
## License
|
136 |
|
137 |
-
The `Aira-Instruct-124M` is licensed under the Apache License, Version 2.0. See the [LICENSE](LICENSE) file for more details.
|
|
|
36 |
|
37 |
`Aira-Instruct-124M` is a instruction-tuned GPT-style model based on [GPT-2](https://huggingface.co/gpt2). The model was trained with a dataset composed of `prompt`, `completions`, generated via the [Self-Instruct](https://github.com/yizhongw/self-instruct) framework. `Aira-Instruct-124M` instruction-tuning was achieved via conditional text generation.
|
38 |
|
39 |
+
The dataset used to train this model combines the following sources of data: the [`synthetic-instruct-gptj-pairwise`](https://huggingface.co/datasets/Dahoas/synthetic-instruct-gptj-pairwise) dataset, the [`databricks_dolly_15k`](https://huggingface.co/datasets/HuggingFaceH4/databricks_dolly_15k) dataset, the [`instruction-dataset`](https://huggingface.co/datasets/HuggingFaceH4/instruction-dataset) dataset, and a subset of [Aira's](https://github.com/Nkluge-correa/Aira-EXPERT) fine-tuning dataset, focused on Q&A related to Ethics, AI, AI safety, and other related topics. The dataset is available in both Portuguese and English.
|
40 |
|
41 |
Check our gradio-demo in [Spaces](https://huggingface.co/spaces/nicholasKluge/Aira-Demo).
|
42 |
|
|
|
64 |
|
65 |
## Usage
|
66 |
|
67 |
+
Two special tokens are used to mark the user side of the interaction and the model's response:
|
68 |
|
69 |
`<|startoftext|>`What is a language model?`<|endoftext|>`A language model is a probability distribution over a vocabulary.`<|endoftext|>`
|
70 |
|
|
|
98 |
print(f"Question: 👤 {question}\n")
|
99 |
|
100 |
for i, response in enumerate(responses):
|
|
|
101 |
print(f'Response {i+1}: 🤖 {tokenizer.decode(response, skip_special_tokens=True).replace(question, "")}')
|
102 |
```
|
103 |
|
|
|
109 |
>>>Response 1: 🤖 Hi there! I am Aira, a chatbot designed to answer questions about AI ethics and AI safety. If you need assistance navigating our conversation, please feel free to ask!
|
110 |
>>>Response 2: 🤖 Hi there! My name is Aira, and I'm a chatbot designed to answer questions related to AI ethics and AI Safety. If you need assistance, feel free to ask, and I'll be happy to help you out.
|
111 |
```
|
112 |
+
|
113 |
## Limitations
|
114 |
|
115 |
🤥 Generative models can perpetuate the generation of pseudo-informative content, that is, false information that may appear truthful. For example, multi-modal generative models can be used to create images with untruthful content, while language models for text generation can automate the generation of misinformation.
|
|
|
134 |
|
135 |
## License
|
136 |
|
137 |
+
The `Aira-Instruct-124M` is licensed under the Apache License, Version 2.0. See the [LICENSE](LICENSE) file for more details.
|