Gustavo de Rosa
commited on
Commit
·
57faa53
1
Parent(s):
34d6e8c
chore(root): Adds top_k information even if 50 is already the default.
Browse files- README.md +2 -1
- generation_config.json +1 -0
README.md
CHANGED
@@ -56,7 +56,7 @@ library_name: transformers
|
|
56 |
## Usage
|
57 |
|
58 |
> [!IMPORTANT]
|
59 |
-
> To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
|
60 |
|
61 |
### Input Formats
|
62 |
|
@@ -88,6 +88,7 @@ outputs = model.generate(
|
|
88 |
inputs.to(model.device),
|
89 |
max_new_tokens=4096,
|
90 |
temperature=0.8,
|
|
|
91 |
top_p=0.95,
|
92 |
do_sample=True,
|
93 |
)
|
|
|
56 |
## Usage
|
57 |
|
58 |
> [!IMPORTANT]
|
59 |
+
> To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_k=50`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
|
60 |
|
61 |
### Input Formats
|
62 |
|
|
|
88 |
inputs.to(model.device),
|
89 |
max_new_tokens=4096,
|
90 |
temperature=0.8,
|
91 |
+
top_k=50,
|
92 |
top_p=0.95,
|
93 |
do_sample=True,
|
94 |
)
|
generation_config.json
CHANGED
@@ -5,6 +5,7 @@
|
|
5 |
"eos_token_id": 100265,
|
6 |
"pad_token_id": 100349,
|
7 |
"temperature": 0.8,
|
|
|
8 |
"top_p": 0.95,
|
9 |
"transformers_version": "4.51.1"
|
10 |
}
|
|
|
5 |
"eos_token_id": 100265,
|
6 |
"pad_token_id": 100349,
|
7 |
"temperature": 0.8,
|
8 |
+
"top_k": 50,
|
9 |
"top_p": 0.95,
|
10 |
"transformers_version": "4.51.1"
|
11 |
}
|