--- library_name: transformers datasets: - bitext/Bitext-customer-support-llm-chatbot-training-dataset --- # Fine-Tuned GEMMA Model for Chatbot This repository hosts the fine-tuned version of the GEMMA 1.1-2B model, specifically fine-tuned for a customer support chatbot use case. ## Model Description The GEMMA 1.1-2B model has been fine-tuned on the [Bitext Customer Support Dataset](https://huggingface.co/datasets/bitext/customer-support-l1m-chatbot-training-dataset) for answering customer support queries. The fine-tuning process involved adjusting the model's weights based on question and answer pairs, which should enable it to generate more accurate and contextually relevant responses in a conversational setting. ## How to Use You can use this model directly with a pipeline for text generation: ```python from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline tokenizer = AutoTokenizer.from_pretrained("your-username/your-model-name") model = AutoModelForCausalLM.from_pretrained("your-username/your-model-name") chatbot = pipeline("text-generation", model=model, tokenizer=tokenizer) response = chatbot("How can I cancel my order?") print(response[0]['generated_text']) ```