barius commited on
Commit
2ba33ce
·
verified ·
1 Parent(s): a411e68

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -106,6 +106,11 @@ print (f"model answer: {answer_content}")
106
  ```
107
  > Note: We have included the system prompt in the tokenizer configuration, as it was used during both the SFT and RL stages. To ensure consistent output quality, we recommend including the same system prompt during actual usage; otherwise, the model's responses may be significantly affected.
108
 
 
 
 
 
 
109
  ## 🔧 Post-training pipeline
110
 
111
  To achieve its strong reasoning ability, AM‑Thinking‑v1 goes through a carefully designed post-training pipeline.
 
106
  ```
107
  > Note: We have included the system prompt in the tokenizer configuration, as it was used during both the SFT and RL stages. To ensure consistent output quality, we recommend including the same system prompt during actual usage; otherwise, the model's responses may be significantly affected.
108
 
109
+ ### Quantized versions for compact devices
110
+ A series of quantized versions for [AM-Thinking-v1](https://huggingface.co/a-m-team/AM-Thinking-v1-gguf) model.
111
+ For use with [llama.cpp](https://github.com/ggml-org/llama.cpp) and [Ollama](https://github.com/ollama/ollama) is available at (AM-Thinking-v1-gguf)[https://huggingface.co/a-m-team/AM-Thinking-v1-gguf].
112
+
113
+
114
  ## 🔧 Post-training pipeline
115
 
116
  To achieve its strong reasoning ability, AM‑Thinking‑v1 goes through a carefully designed post-training pipeline.