Update README.md
Browse files
README.md
CHANGED
@@ -106,6 +106,11 @@ print (f"model answer: {answer_content}")
|
|
106 |
```
|
107 |
> Note: We have included the system prompt in the tokenizer configuration, as it was used during both the SFT and RL stages. To ensure consistent output quality, we recommend including the same system prompt during actual usage; otherwise, the model's responses may be significantly affected.
|
108 |
|
|
|
|
|
|
|
|
|
|
|
109 |
## 🔧 Post-training pipeline
|
110 |
|
111 |
To achieve its strong reasoning ability, AM‑Thinking‑v1 goes through a carefully designed post-training pipeline.
|
|
|
106 |
```
|
107 |
> Note: We have included the system prompt in the tokenizer configuration, as it was used during both the SFT and RL stages. To ensure consistent output quality, we recommend including the same system prompt during actual usage; otherwise, the model's responses may be significantly affected.
|
108 |
|
109 |
+
### Quantized versions for compact devices
|
110 |
+
A series of quantized versions for [AM-Thinking-v1](https://huggingface.co/a-m-team/AM-Thinking-v1-gguf) model.
|
111 |
+
For use with [llama.cpp](https://github.com/ggml-org/llama.cpp) and [Ollama](https://github.com/ollama/ollama) is available at (AM-Thinking-v1-gguf)[https://huggingface.co/a-m-team/AM-Thinking-v1-gguf].
|
112 |
+
|
113 |
+
|
114 |
## 🔧 Post-training pipeline
|
115 |
|
116 |
To achieve its strong reasoning ability, AM‑Thinking‑v1 goes through a carefully designed post-training pipeline.
|