Update README.md
Browse files
README.md
CHANGED
@@ -15,9 +15,9 @@ base_model:
|
|
15 |
|
16 |
This is an experimental conversion that uses AWQ to apply scaling and export in fp16, then pack it and quantize to GGUF.
|
17 |
The AWQ calibration was done only using EN<-->PL pairs, 100 calibration samples each way, 200 total.
|
18 |
-
It has not been tested nor evaluated yet. Depending on the outcomes
|
19 |
-
* try to add LoRA PEFT for the same languages to improve quality further then release other language pairs/groups
|
20 |
-
* replace AWQ with LoRA PEFT for the same languages, merge, then redo GGUF quants
|
21 |
|
22 |
## Important note on quality
|
23 |
|
|
|
15 |
|
16 |
This is an experimental conversion that uses AWQ to apply scaling and export in fp16, then pack it and quantize to GGUF.
|
17 |
The AWQ calibration was done only using EN<-->PL pairs, 100 calibration samples each way, 200 total.
|
18 |
+
It has not been tested nor evaluated yet. Depending on the outcomes:
|
19 |
+
* If AWQ turns helpful, I will try to add LoRA PEFT for the same languages to improve quality further, merge, then run AWQ, then release other language pairs/groups
|
20 |
+
* If AWQ doesn't provide improvements, or turns otherwise problematic I will replace AWQ with LoRA PEFT for the same languages, merge, then redo GGUF quants, and if that works I'll release for other language pairs/groups.
|
21 |
|
22 |
## Important note on quality
|
23 |
|