mohammed-orabi2 commited on
Commit
47a0f5e
·
verified ·
1 Parent(s): 2396c37

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -13
README.md CHANGED
@@ -15,9 +15,7 @@ This is a LoRA fine-tuned version of the `Qwen/Qwen3-1.7B` model, specifically t
15
 
16
  **Developed by:** Mohammed Orabi
17
 
18
- **Funded by \[optional]:** N/A
19
-
20
- **Shared by \[optional]:** mohammed-orabi2
21
 
22
  **Model type:** Causal Language Model with LoRA adaptation
23
 
@@ -25,9 +23,9 @@ This is a LoRA fine-tuned version of the `Qwen/Qwen3-1.7B` model, specifically t
25
 
26
  **License:** Apache 2.0 (inherits from Qwen3-1.7B)
27
 
28
- **Finetuned from model \[optional]:** Qwen/Qwen3-1.7B
29
 
30
- **Model Sources \[optional]**
31
  **Repository:** [https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2](https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2)
32
 
33
 
@@ -38,7 +36,7 @@ This is a LoRA fine-tuned version of the `Qwen/Qwen3-1.7B` model, specifically t
38
  **Direct Use:**
39
  This model can be used for generating Arabic poetry in response to user queries, particularly in cultural, educational, or creative chatbot applications.
40
 
41
- **Downstream Use \[optional]:**
42
 
43
  * Poetry recommendation systems
44
  * Arabic literature generation tools
@@ -96,7 +94,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=True))
96
  **Training Data:**
97
  1,000 synthetic Arabic poetic dialogues (prompt + poetic response) generated programmatically.
98
 
99
- **Preprocessing \[optional]:**
100
 
101
  * Applied Qwen chat template
102
  * Tokenized using Qwen3-1.7B tokenizer with padding/truncation
@@ -109,7 +107,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=True))
109
  * Learning rate: 2e-4
110
  * LoRA config: r=8, alpha=16, dropout=0.05, target: \["q\_proj", "v\_proj"]
111
 
112
- **Speeds, Sizes, Times \[optional]:**
113
 
114
  * Training time: \~24 minutes on L4 GPU
115
  * Model size: LoRA adapter \~100MB
@@ -155,7 +153,7 @@ Output behavior consistent with training intent. Performs well within poetic use
155
 
156
  ---
157
 
158
- ## Technical Specifications \[optional]
159
 
160
  **Model Architecture and Objective:** Transformer decoder (CausalLM) + LoRA injection
161
 
@@ -171,7 +169,7 @@ Output behavior consistent with training intent. Performs well within poetic use
171
 
172
  ---
173
 
174
- ## Citation \[optional]
175
 
176
  **BibTeX:**
177
 
@@ -189,18 +187,18 @@ Mohammed Orabi. (2025). *Qwen Arabic Poetry LoRA* \[Model]. Hugging Face. [https
189
 
190
  ---
191
 
192
- ## Glossary \[optional]
193
 
194
  * **LoRA**: Low-Rank Adaptation, a method for efficient model fine-tuning
195
  * **CausalLM**: Causal Language Modeling, predicts the next token in a sequence
196
 
197
  ---
198
 
199
- ## More Information \[optional]
200
 
201
  For support or feedback, please open an issue on the Hugging Face repo or contact via Hugging Face profile.
202
 
203
- ## Model Card Authors \[optional]
204
 
205
  Mohammed Orabi
206
 
 
15
 
16
  **Developed by:** Mohammed Orabi
17
 
18
+ **Shared by :** mohammed-orabi2
 
 
19
 
20
  **Model type:** Causal Language Model with LoRA adaptation
21
 
 
23
 
24
  **License:** Apache 2.0 (inherits from Qwen3-1.7B)
25
 
26
+ **Finetuned from model :** Qwen/Qwen3-1.7B
27
 
28
+ **Model Sources **
29
  **Repository:** [https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2](https://huggingface.co/mohammed-orabi2/qwen-poetry-lora2)
30
 
31
 
 
36
  **Direct Use:**
37
  This model can be used for generating Arabic poetry in response to user queries, particularly in cultural, educational, or creative chatbot applications.
38
 
39
+ **Downstream Use :**
40
 
41
  * Poetry recommendation systems
42
  * Arabic literature generation tools
 
94
  **Training Data:**
95
  1,000 synthetic Arabic poetic dialogues (prompt + poetic response) generated programmatically.
96
 
97
+ **Preprocessing :**
98
 
99
  * Applied Qwen chat template
100
  * Tokenized using Qwen3-1.7B tokenizer with padding/truncation
 
107
  * Learning rate: 2e-4
108
  * LoRA config: r=8, alpha=16, dropout=0.05, target: \["q\_proj", "v\_proj"]
109
 
110
+ **Speeds, Sizes, Times :**
111
 
112
  * Training time: \~24 minutes on L4 GPU
113
  * Model size: LoRA adapter \~100MB
 
153
 
154
  ---
155
 
156
+ ## Technical Specifications
157
 
158
  **Model Architecture and Objective:** Transformer decoder (CausalLM) + LoRA injection
159
 
 
169
 
170
  ---
171
 
172
+ ## Citation
173
 
174
  **BibTeX:**
175
 
 
187
 
188
  ---
189
 
190
+ ## Glossary
191
 
192
  * **LoRA**: Low-Rank Adaptation, a method for efficient model fine-tuning
193
  * **CausalLM**: Causal Language Modeling, predicts the next token in a sequence
194
 
195
  ---
196
 
197
+ ## More Information
198
 
199
  For support or feedback, please open an issue on the Hugging Face repo or contact via Hugging Face profile.
200
 
201
+ ## Model Card Authors
202
 
203
  Mohammed Orabi
204