Model Card for Model ID

This is Meta's Llama 2 7B quantized in 3-bit using AutoGPTQ from Hugging Face Transformers.

Model Details

Model Description

Model Sources

The method and code used to quantize the model are explained here: Quantize and Fine-tune LLMs with GPTQ Using Transformers and TRL

Uses

This model is pre-trained and not fine-tuned. You may fine-tune it with PEFT using adapters.

Other versions

Model Card Contact

The Kaitchup

Downloads last month
84
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.