πŸš€ Fine-tuned Gemma 3 Model (4B, 4-bit) by Webkul

This repository contains a fine-tuned version of Unsloth's gemma-3-4b-it model, optimized for lightweight 4-bit inference and instruction tuning using Hugging Face's TRL and Unsloth's speed-optimized framework.


πŸ”§ Model Details


πŸ“š Fine-tuning Dataset

This model was fine-tuned on unopim dev documentation available at https://devdocs.unopim.com/, focusing on structured software documentation and developer support content.


πŸ’‘ Intended Use

  • Conversational AI assistants trained on UnoPIM developer docs
  • API documentation question answering
  • Developer tools and chatbot integrations
  • Contextual helpdesk or onboarding bots for UnoPIM products

πŸ§ͺ How to Use

You can use this model with the Hugging Face transformers library:

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "webkul/gemma-3-4b-it-unopim-docs"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

input_text = "How do I integrate the UnoPIM API for product syncing?"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=300)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

πŸ“„ License This model is licensed under the Apache License 2.0.


Downloads last month
16
Safetensors
Model size
4.3B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for webkul/unopim-devdocs

Finetuned
(596)
this model