Update README.md
Browse files
README.md
CHANGED
@@ -15,9 +15,7 @@ This is a LoRA fine-tuned version of the `Qwen/Qwen3-1.7B` model, specifically t
|
|
15 |
|
16 |
**Developed by:** Mohammed Orabi
|
17 |
|
18 |
-
**
|
19 |
-
|
20 |
-
**Shared by \[optional]:** mohammed-orabi2
|
21 |
|
22 |
**Model type:** Causal Language Model with LoRA adaptation
|
23 |
|
@@ -25,9 +23,9 @@ This is a LoRA fine-tuned version of the `Qwen/Qwen3-1.7B` model, specifically t
|
|
25 |
|
26 |
**License:** Apache 2.0 (inherits from Qwen3-1.7B)
|
27 |
|
28 |
-
**Finetuned from model
|
29 |
|
30 |
-
**Model Sources
|
31 |
**Repository:** [https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2](https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2)
|
32 |
|
33 |
|
@@ -38,7 +36,7 @@ This is a LoRA fine-tuned version of the `Qwen/Qwen3-1.7B` model, specifically t
|
|
38 |
**Direct Use:**
|
39 |
This model can be used for generating Arabic poetry in response to user queries, particularly in cultural, educational, or creative chatbot applications.
|
40 |
|
41 |
-
**Downstream Use
|
42 |
|
43 |
* Poetry recommendation systems
|
44 |
* Arabic literature generation tools
|
@@ -96,7 +94,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=True))
|
|
96 |
**Training Data:**
|
97 |
1,000 synthetic Arabic poetic dialogues (prompt + poetic response) generated programmatically.
|
98 |
|
99 |
-
**Preprocessing
|
100 |
|
101 |
* Applied Qwen chat template
|
102 |
* Tokenized using Qwen3-1.7B tokenizer with padding/truncation
|
@@ -109,7 +107,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=True))
|
|
109 |
* Learning rate: 2e-4
|
110 |
* LoRA config: r=8, alpha=16, dropout=0.05, target: \["q\_proj", "v\_proj"]
|
111 |
|
112 |
-
**Speeds, Sizes, Times
|
113 |
|
114 |
* Training time: \~24 minutes on L4 GPU
|
115 |
* Model size: LoRA adapter \~100MB
|
@@ -155,7 +153,7 @@ Output behavior consistent with training intent. Performs well within poetic use
|
|
155 |
|
156 |
---
|
157 |
|
158 |
-
## Technical Specifications
|
159 |
|
160 |
**Model Architecture and Objective:** Transformer decoder (CausalLM) + LoRA injection
|
161 |
|
@@ -171,7 +169,7 @@ Output behavior consistent with training intent. Performs well within poetic use
|
|
171 |
|
172 |
---
|
173 |
|
174 |
-
## Citation
|
175 |
|
176 |
**BibTeX:**
|
177 |
|
@@ -189,18 +187,18 @@ Mohammed Orabi. (2025). *Qwen Arabic Poetry LoRA* \[Model]. Hugging Face. [https
|
|
189 |
|
190 |
---
|
191 |
|
192 |
-
## Glossary
|
193 |
|
194 |
* **LoRA**: Low-Rank Adaptation, a method for efficient model fine-tuning
|
195 |
* **CausalLM**: Causal Language Modeling, predicts the next token in a sequence
|
196 |
|
197 |
---
|
198 |
|
199 |
-
## More Information
|
200 |
|
201 |
For support or feedback, please open an issue on the Hugging Face repo or contact via Hugging Face profile.
|
202 |
|
203 |
-
## Model Card Authors
|
204 |
|
205 |
Mohammed Orabi
|
206 |
|
|
|
15 |
|
16 |
**Developed by:** Mohammed Orabi
|
17 |
|
18 |
+
**Shared by :** mohammed-orabi2
|
|
|
|
|
19 |
|
20 |
**Model type:** Causal Language Model with LoRA adaptation
|
21 |
|
|
|
23 |
|
24 |
**License:** Apache 2.0 (inherits from Qwen3-1.7B)
|
25 |
|
26 |
+
**Finetuned from model :** Qwen/Qwen3-1.7B
|
27 |
|
28 |
+
**Model Sources **
|
29 |
**Repository:** [https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2](https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2)
|
30 |
|
31 |
|
|
|
36 |
**Direct Use:**
|
37 |
This model can be used for generating Arabic poetry in response to user queries, particularly in cultural, educational, or creative chatbot applications.
|
38 |
|
39 |
+
**Downstream Use :**
|
40 |
|
41 |
* Poetry recommendation systems
|
42 |
* Arabic literature generation tools
|
|
|
94 |
**Training Data:**
|
95 |
1,000 synthetic Arabic poetic dialogues (prompt + poetic response) generated programmatically.
|
96 |
|
97 |
+
**Preprocessing :**
|
98 |
|
99 |
* Applied Qwen chat template
|
100 |
* Tokenized using Qwen3-1.7B tokenizer with padding/truncation
|
|
|
107 |
* Learning rate: 2e-4
|
108 |
* LoRA config: r=8, alpha=16, dropout=0.05, target: \["q\_proj", "v\_proj"]
|
109 |
|
110 |
+
**Speeds, Sizes, Times :**
|
111 |
|
112 |
* Training time: \~24 minutes on L4 GPU
|
113 |
* Model size: LoRA adapter \~100MB
|
|
|
153 |
|
154 |
---
|
155 |
|
156 |
+
## Technical Specifications
|
157 |
|
158 |
**Model Architecture and Objective:** Transformer decoder (CausalLM) + LoRA injection
|
159 |
|
|
|
169 |
|
170 |
---
|
171 |
|
172 |
+
## Citation
|
173 |
|
174 |
**BibTeX:**
|
175 |
|
|
|
187 |
|
188 |
---
|
189 |
|
190 |
+
## Glossary
|
191 |
|
192 |
* **LoRA**: Low-Rank Adaptation, a method for efficient model fine-tuning
|
193 |
* **CausalLM**: Causal Language Modeling, predicts the next token in a sequence
|
194 |
|
195 |
---
|
196 |
|
197 |
+
## More Information
|
198 |
|
199 |
For support or feedback, please open an issue on the Hugging Face repo or contact via Hugging Face profile.
|
200 |
|
201 |
+
## Model Card Authors
|
202 |
|
203 |
Mohammed Orabi
|
204 |
|