--- license: apache-2.0 tags: - finetuned pipeline_tag: text-generation inference: true widget: - messages: - role: user content: What is your favorite condiment? extra_gated_description: If you want to learn more about how we process your personal data, please read our Privacy Policy. --- ## Use below code to download the mistral. ```py #pip install -U transformers accelerate torch import torch from transformers import pipeline, set_seed pipe = pipeline("text-generation", model=model_path, torch_dtype=torch.bfloat16, device_map="cuda") messages = [{"role": "user", "content": query}] outputs = pipe(messages, max_new_tokens=1000, do_sample=True, temperature=0.71, top_k=50, top_p=0.92,repetition_penalty=1) print(outputs[0]["generated_text"][-1]["content"]) ``` ## Limitations The Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance. It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs. ## Develop By vickys9715@gmail.com, https://www.linkedin.com/in/vicky-s-41135319a/overlay/create-post/