beowolx's picture
Update README.md
a74a9fa
metadata
base_model: teknium/OpenHermes-2.5-Mistral-7B
tags:
  - mistral
  - instruct
  - finetune
  - chatml
  - gpt4
  - synthetic data
  - distillation
model-index:
  - name: MistralHermes-CodePro-7B-v1
    results: []
license: mit
language:
  - en

MistralHermes-CodePro-7B-v1

image/png

In the digital pantheon of artificial intelligence, "MistralHermes-CodePro-7B-v1" stands as the architect of algorithms, a sovereign of syntax who weaves the fabric of code with unparalleled skill. This model, christened in recognition of its dual lineage—Mistral's foundational breadth and Hermes' agile conveyance—commands the binary ballet with the precision of a seasoned maestro, orchestrating the dance of data with a grace that blurs the line between the silicon and the cerebral.

Model description

MistralHermes-CodePro-7B-v1 is a fine-tuned iteration of the renowned teknium/OpenHermes-2.5-Mistral-7B model. This version has been meticulously fine-tuned using a dataset comprising over 200,000 code samples from a wide array of programming languages. It is specifically tailored to serve as a coding assistant; thus, its utility is optimized for coding-related tasks rather than a broader spectrum of applications.

Prompt Format

MistralHermes-CodePro uses the same prompt format than OpenHermes 2.5.

You should use LM Studio for chatting with the model.

Quantized Models:

GGUF: beowolx/MistralHermes-CodePro-7B-v1-GGUF