Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,9 @@ base_model:
|
|
15 |
|
16 |
This is an experimental conversion that uses AWQ to apply scaling and export in fp16, then pack it and quantize to GGUF.
|
17 |
The AWQ calibration was done only using EN<-->PL pairs, 100 calibration samples each way, 200 total.
|
18 |
-
It has not been tested nor evaluated yet.
|
|
|
|
|
19 |
|
20 |
## Important note on quality
|
21 |
|
|
|
15 |
|
16 |
This is an experimental conversion that uses AWQ to apply scaling and export in fp16, then pack it and quantize to GGUF.
|
17 |
The AWQ calibration was done only using EN<-->PL pairs, 100 calibration samples each way, 200 total.
|
18 |
+
It has not been tested nor evaluated yet. Depending on the outcomes I will either:
|
19 |
+
* try to add LoRA PEFT for the same languages to improve quality further then release other language pairs/groups
|
20 |
+
* replace AWQ with LoRA PEFT for the same languages, merge, then redo GGUF quants
|
21 |
|
22 |
## Important note on quality
|
23 |
|