Melvin56 commited on
Commit
18d9451
·
verified ·
1 Parent(s): cffe76b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -73
README.md CHANGED
@@ -1,73 +1,32 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- base_model:
6
- - deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
7
- pipeline_tag: text-generation
8
- library_name: transformers
9
- tags:
10
- - text-generation-inference
11
- - R1
12
- - Qwen
13
- - Deepseek
14
- ---
15
- ![ASFSADFCS.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/U8o3r3pREezkHYVAEEcqc.png)
16
-
17
- # **Elita-0.1-Distilled-R1-Abliterated**
18
-
19
- Elita-0.1-Distilled-R1-Abliterated is based on the *Qwen [ KT ] model*, which was distilled by *DeepSeek-AI/DeepSeek-R1-Distill-Qwen-7B*. It has been fine-tuned on the long chain-of-thought reasoning model and specialized datasets, focusing on chain-of-thought (CoT) reasoning for problem-solving. This model is optimized for tasks requiring logical reasoning, detailed explanations, and multi-step problem-solving, making it ideal for applications such as instruction-following, text generation, and complex reasoning tasks.
20
-
21
- # **Quickstart with Transformers**
22
-
23
- Here provides a code snippet with `apply_chat_template` to show you how to load the tokenizer and model and how to generate contents.
24
-
25
- ```python
26
- from transformers import AutoModelForCausalLM, AutoTokenizer
27
-
28
- model_name = "prithivMLmods/Elita-0.1-Distilled-R1-Abliterated"
29
-
30
- model = AutoModelForCausalLM.from_pretrained(
31
- model_name,
32
- torch_dtype="auto",
33
- device_map="auto"
34
- )
35
- tokenizer = AutoTokenizer.from_pretrained(model_name)
36
-
37
- prompt = "Give me a short introduction to large language model."
38
- messages = [
39
- {"role": "system", "content": "You are Elita, created by DeepSeek-AI. You are a powerful reasoning assistant."},
40
- {"role": "user", "content": prompt}
41
- ]
42
- text = tokenizer.apply_chat_template(
43
- messages,
44
- tokenize=False,
45
- add_generation_prompt=True
46
- )
47
- model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
48
-
49
- generated_ids = model.generate(
50
- **model_inputs,
51
- max_new_tokens=512
52
- )
53
- generated_ids = [
54
- output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
55
- ]
56
-
57
- response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
58
- ```
59
-
60
- ### **Intended Use:**
61
- 1. **Instruction-Following:** The model excels in understanding and executing detailed instructions, making it ideal for automation systems, virtual assistants, and educational tools.
62
- 2. **Text Generation:** It can produce coherent, logically structured, and contextually relevant text for use in content creation, summarization, and report writing.
63
- 3. **Complex Reasoning Tasks:** With its fine-tuning for chain-of-thought reasoning, the model is well-suited for multi-step problem-solving, logical deduction, and question-answering tasks.
64
- 4. **Research and Development:** It can support researchers and developers in exploring advancements in logical reasoning and fine-tuning methodologies.
65
- 5. **Educational Applications:** The model can assist in teaching logical reasoning and problem-solving by generating step-by-step solutions.
66
-
67
- ### **Limitations:**
68
- 1. **Domain-Specific Knowledge:** While fine-tuned on reasoning datasets, the model may lack deep expertise in highly specialized or technical domains.
69
- 2. **Hallucination:** Like many large language models, it can generate incorrect or fabricated information, especially when reasoning beyond its training data.
70
- 3. **Bias in Training Data:** The model's outputs may reflect biases present in the datasets it was fine-tuned on, which could limit its objectivity in certain contexts.
71
- 4. **Performance on Non-Reasoning Tasks:** The model is optimized for chain-of-thought reasoning and may underperform on tasks that require simpler, less structured responses.
72
- 5. **Resource-Intensive:** Running the model efficiently requires significant computational resources, which may limit accessibility for smaller-scale deployments.
73
- 6. **Dependence on Input Quality:** The model’s performance heavily depends on the clarity and quality of the input provided. Ambiguous or poorly structured prompts may yield suboptimal results.
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - prithivMLmods/Elita-0.1-Distilled-R1-abliterated
7
+ pipeline_tag: text-generation
8
+ library_name: transformers
9
+ tags:
10
+ - R1
11
+ - Qwen
12
+ - Deepseek
13
+ ---
14
+ # Melvin56/Elita-0.1-Distilled-R1-abliterated-GGUF
15
+
16
+ Original Model : [prithivMLmods/Elita-0.1-Distilled-R1-abliterated](https://huggingface.co/prithivMLmods/Elita-0.1-Distilled-R1-abliterated)
17
+
18
+ All quants are made using the imatrix option.
19
+
20
+ | Model | Size (GB) |
21
+ |:-------------------------------------------------|:-------------:|
22
+ | Q2_K_S | 2.82 |
23
+ | Q2_K | 3.01 |
24
+ | Q3_K_M | 3.80 |
25
+ | Q3_K_L | 4.08 |
26
+ | Q4_K_S | 4.46 |
27
+ | Q4_K_M | 4.68 |
28
+ | Q5_K_S | 5.30 |
29
+ | Q5_K_M | 5.44 |
30
+ | Q6_K | 6.25 |
31
+ | Q8_0 | 8.10 |
32
+ | F16 | 15.24 |