Update README.md
Browse files
README.md
CHANGED
|
@@ -14,7 +14,7 @@ tags:
|
|
| 14 |
|
| 15 |
# Llama-2-7b-pruned50-retrained-evolcodealpaca-quant-ds
|
| 16 |
|
| 17 |
-
This repo contains a [50% sparse Llama 2 7B](https://huggingface.co/neuralmagic/Llama-2-7b-
|
| 18 |
It was then quantized to 8-bit weights + activations and exported to deploy with [DeepSparse](https://github.com/neuralmagic/deepsparse), a CPU inference runtime for sparse models.
|
| 19 |
|
| 20 |
**Authors**: Neural Magic, Cerebras
|
|
@@ -46,9 +46,9 @@ print(outputs.generations[0].text)
|
|
| 46 |
|
| 47 |
Model evaluation metrics and results.
|
| 48 |
|
| 49 |
-
| Benchmark | Metric | Llama-2-7b-
|
| 50 |
|------------------------------------------------|---------------|-------------|-------------------------------|
|
| 51 |
-
| [HumanEval](https://arxiv.org/abs/2107.03374) | pass@1 |
|
| 52 |
|
| 53 |
## Help
|
| 54 |
|
|
|
|
| 14 |
|
| 15 |
# Llama-2-7b-pruned50-retrained-evolcodealpaca-quant-ds
|
| 16 |
|
| 17 |
+
This repo contains a [50% sparse Llama 2 7B](https://huggingface.co/neuralmagic/Llama-2-7b-pruned50-retrained-evolcodealpaca) finetuned for code generation tasks using the [Evolved CodeAlpaca](https://huggingface.co/datasets/theblackcat102/evol-codealpaca-v1) dataset.
|
| 18 |
It was then quantized to 8-bit weights + activations and exported to deploy with [DeepSparse](https://github.com/neuralmagic/deepsparse), a CPU inference runtime for sparse models.
|
| 19 |
|
| 20 |
**Authors**: Neural Magic, Cerebras
|
|
|
|
| 46 |
|
| 47 |
Model evaluation metrics and results.
|
| 48 |
|
| 49 |
+
| Benchmark | Metric | Llama-2-7b-evolcodealpaca | Llama-2-7b-pruned50-retrained-evolcodealpaca-quant-ds |
|
| 50 |
|------------------------------------------------|---------------|-------------|-------------------------------|
|
| 51 |
+
| [HumanEval](https://arxiv.org/abs/2107.03374) | pass@1 | 32.03 | 32.2 |
|
| 52 |
|
| 53 |
## Help
|
| 54 |
|