ArunKr commited on
Commit
70fbd7c
·
verified ·
1 Parent(s): ddbe3d5

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +50 -0
README.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ language:
4
+ - en
5
+ license: apache-2.0
6
+ tags:
7
+ - text-generation
8
+ - instruct
9
+ - manim
10
+ - lora
11
+ - gguf
12
+ datasets:
13
+ - ArunKr/verified-data-manim
14
+ base_model: HuggingFaceTB/SmolLM-135M-Instruct
15
+ library_name: transformers
16
+ pipeline_tag: text-generation
17
+ ---
18
+
19
+ # gemma-3-270m-it-web-agent - Fine-tuned
20
+
21
+ This repository contains three variants of the model:
22
+
23
+ - **LoRA adapters** → [ArunKr/gemma-3-270m-it-web-agent-lora](https://huggingface.co/ArunKr/gemma-3-270m-it-web-agent-lora)
24
+ - **Merged FP16 weights** → [ArunKr/gemma-3-270m-it-web-agent-16bit](https://huggingface.co/ArunKr/gemma-3-270m-it-web-agent-16bit)
25
+ - **GGUF quantizations** → [ArunKr/gemma-3-270m-it-web-agent-gguf](https://huggingface.co/ArunKr/gemma-3-270m-it-web-agent-gguf)
26
+
27
+ ### Training
28
+ - Base model: `unsloth/gemma-3-270m-it`
29
+ - Dataset: `ArunKr/gui_grounding_dataset-100`
30
+ - Method: LoRA fine-tuning with [Unsloth](https://github.com/unslothai/unsloth)
31
+
32
+ ### Quantizations
33
+ We provide `f16`, `bf16`, `f32`, and `q8_0` GGUF files for llama.cpp / Ollama.
34
+
35
+ ### Usage Example
36
+ ```python
37
+ from transformers import AutoModelForCausalLM, AutoTokenizer
38
+
39
+ tok = AutoTokenizer.from_pretrained("ArunKr/gemma-3-270m-it-web-agent-16bit")
40
+ model = AutoModelForCausalLM.from_pretrained("ArunKr/gemma-3-270m-it-web-agent-16bit")
41
+ print(model.generate(**tok("Hello", return_tensors="pt")))
42
+ ```
43
+
44
+ ### Ollama Example
45
+
46
+ ```bash
47
+ ollama run ArunKr/SmolLM-135M-Instruct-manim-gguf:<file_name>.gguf
48
+ ```
49
+
50
+ www.ollama.com