MISHANM commited on
Commit
779413c
·
verified ·
1 Parent(s): cb8a335

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Qwen/QwQ-32B
4
+ library_name: transformers
5
+ ---
6
+
7
+ # MISHANM/Qwen-QwQ-32B.gguf
8
+
9
+ This model is a GGUF version of Qwen/QwQ-32B model, It is specially designed to work smoothly with the llama.cpp framework. It's built to run efficiently on CPU systems and has been tested on the AMD EPYC™ 9755 processor. The model handles various natural language processing tasks really well. It not only processes text quickly but also has strong reasoning and thinking skills, allowing it to manage difficult language-related challenges effectively.
10
+
11
+ ## Model Details
12
+ 1. Language: English
13
+ 2. Tasks: Text generation
14
+ 3. Base Model: gQwen/QwQ-32B
15
+
16
+ ## Building and Running the Model
17
+
18
+ To build and run the model using `llama.cpp`, follow these steps:
19
+
20
+ ### Model
21
+
22
+ Steps to Download the Model:
23
+ 1. Go to the "Files and Versions" section.
24
+ 2. Click on the model.
25
+ 3. Copy the download link.
26
+ 4. Create a directory (e.g., for Linux: mkdir Qwen32B).
27
+ 5. Navigate to that directory (cd Qwen32B).
28
+ 6. Download both model parts: Qwen-QwQ-32B.gguf.part_01 and Qwen-QwQ-32B.gguf.part_02 (e.g., using wget with the copied link).
29
+
30
+ After downloading the model parts, use the following command to combine them into a complete model:
31
+
32
+ ```
33
+ cat Qwen-QwQ-32B.gguf.part_01 Qwen-QwQ-32B.gguf.part_02 > Qwen-QwQ-32B.gguf
34
+ ```
35
+
36
+
37
+ ### Build llama.cpp Locally
38
+
39
+ ```bash
40
+ git clone https://github.com/ggerganov/llama.cpp
41
+ cd llama.cpp
42
+ cmake -B build
43
+ cmake --build build --config Release
44
+
45
+ ```
46
+ ## Run the Model
47
+
48
+ Navigate to the build directory and run the model with a prompt:
49
+
50
+ ```
51
+ cd llama.cpp/build/bin
52
+ ```
53
+ ## Inference with llama.cpp
54
+
55
+ ```
56
+ ./llama-cli -m /path/to/model/ -p "Your prompt here" -n 128 --ctx-size 8192 --temp 0.6 --seed 3407
57
+ ```
58
+
59
+ ## Citation Information
60
+ ```
61
+ @misc{MISHANM/Qwen-QwQ-32B.gguf,
62
+ author = {Mishan Maurya},
63
+ title = {Introducing Qwen QwQ-32B GGUF Model},
64
+ year = {2025},
65
+ publisher = {Hugging Face},
66
+ journal = {Hugging Face repository},
67
+
68
+ }
69
+ ```