Rombos-LLM-V2.6-Nemotron-70b by Rombodawg


ExLlamaV2 Quantization

Quantized with ExLlamaV2 v0.2.3

2.2 Bits Per Weight

4.65 Bits Per Weight


image/jpeg

I applied the last step of my continuous finetuning method to the Nemotron-70b model from Nvidia. More details bellow:

Quants: (Coming Soon)

Open-LLM-Leaderboard scores: (Coming soon)

Downloads last month
0
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support