Gustavo de Rosa commited on
Commit
57faa53
·
1 Parent(s): 34d6e8c

chore(root): Adds top_k information even if 50 is already the default.

Browse files
Files changed (2) hide show
  1. README.md +2 -1
  2. generation_config.json +1 -0
README.md CHANGED
@@ -56,7 +56,7 @@ library_name: transformers
56
  ## Usage
57
 
58
  > [!IMPORTANT]
59
- > To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
60
 
61
  ### Input Formats
62
 
@@ -88,6 +88,7 @@ outputs = model.generate(
88
  inputs.to(model.device),
89
  max_new_tokens=4096,
90
  temperature=0.8,
 
91
  top_p=0.95,
92
  do_sample=True,
93
  )
 
56
  ## Usage
57
 
58
  > [!IMPORTANT]
59
+ > To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_k=50`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
60
 
61
  ### Input Formats
62
 
 
88
  inputs.to(model.device),
89
  max_new_tokens=4096,
90
  temperature=0.8,
91
+ top_k=50,
92
  top_p=0.95,
93
  do_sample=True,
94
  )
generation_config.json CHANGED
@@ -5,6 +5,7 @@
5
  "eos_token_id": 100265,
6
  "pad_token_id": 100349,
7
  "temperature": 0.8,
 
8
  "top_p": 0.95,
9
  "transformers_version": "4.51.1"
10
  }
 
5
  "eos_token_id": 100265,
6
  "pad_token_id": 100349,
7
  "temperature": 0.8,
8
+ "top_k": 50,
9
  "top_p": 0.95,
10
  "transformers_version": "4.51.1"
11
  }