---
license: apache-2.0
datasets:
- pankajmathur/orca_mini_v1_dataset
- pankajmathur/WizardLM_Orca
- pankajmathur/dolly-v2_orca
- pankajmathur/alpaca_orca
language:
- en
library_name: transformers
---

### Mistral-7B-model_45k6e2e4
This model is trained on Mistral-7B-v0.1

### License Disclaimer:

This model is released under Apache 2.0, and comes with no warranty or guarantees of any kind.


### Limitations & Biases:

While this model aims for accuracy, it can occasionally produce inaccurate or misleading results.

Despite diligent efforts in refining the training data, there remains a possibility for the generation of inappropriate, biased, or offensive content.

Exercise caution and cross-check information when necessary. This is an uncensored model.

### Citation:

Please kindly cite using the following BibTeX:
```
@misc{Mistral-7B-model_45k6e2e4,
author = {Pankaj Mathur},
title = {Mistral-7B-model_45k6e2e4: An Orca style Mistral-7B-v0.1 model},
year = {2023},
publisher = {HuggingFace},
journal = {HuggingFace repository},
howpublished = {\url{https://huggingface.co/pankajmathur/Mistral-7B-model_45k6e2e4},
}
```