Model Card for Model ID

Merged and GPTQ quantized version of rayliuca/TRagx-internlm2-7b

Note: I'm having some difficulties quantizing the models using GPTQ. Mistral and NeuralOmniBeagle's GPTQ models have significantly degraded output, while quantized TowerInstruct v0.2 was not working out right

While this quantized model for InternLM2 seems to work all right, the translation accuracy is not validated.

These AWQ quantized models are recommended:

GPTQ Dataset

Qutanized with nsamples=45 * 3 languages [ja, zh, en] from the c4 dataset

License

See the original InternLM2 repo https://huggingface.co/internlm/internlm2-7b#open-source-license

Downloads last month
0
Safetensors
Model size
1.69B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including rayliuca/TRagx-GPTQ-internlm2-7b