Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,7 @@ We believe the future of AI is open. That’s why we’re sharing our latest mod
|
|
22 |
- **Customize and fine-tune models for your workflows**, from experimentation to production, with tools and frameworks built to support reproducible research and enterprise AI pipelines.
|
23 |
- **Maximize inference efficiency across hardware** using production-grade compression and optimization techniques like quantization (FP8, INT8, INT4), structured/unstructured sparsity, distillation, and more, ready for cost-efficient deployments with vLLM.
|
24 |
|
25 |
-
🔗 **Explore
|
26 |
- [**vLLM**](https://github.com/vllm-project/vllm) – Serve large language models efficiently across GPUs and environments.
|
27 |
- [**LLM Compressor**](https://github.com/vllm-project/llm-compressor) – Compress and optimize your own models with SOTA quantization and sparsity techniques.
|
28 |
- [**InstructLab**](https://github.com/instructlab) – Fine-tune open models with your data using scalable, community-backed workflows.
|
|
|
22 |
- **Customize and fine-tune models for your workflows**, from experimentation to production, with tools and frameworks built to support reproducible research and enterprise AI pipelines.
|
23 |
- **Maximize inference efficiency across hardware** using production-grade compression and optimization techniques like quantization (FP8, INT8, INT4), structured/unstructured sparsity, distillation, and more, ready for cost-efficient deployments with vLLM.
|
24 |
|
25 |
+
🔗 **Explore relevant open-source tools**:
|
26 |
- [**vLLM**](https://github.com/vllm-project/vllm) – Serve large language models efficiently across GPUs and environments.
|
27 |
- [**LLM Compressor**](https://github.com/vllm-project/llm-compressor) – Compress and optimize your own models with SOTA quantization and sparsity techniques.
|
28 |
- [**InstructLab**](https://github.com/instructlab) – Fine-tune open models with your data using scalable, community-backed workflows.
|