Lab1806 commited on
Commit
0e74032
·
verified ·
1 Parent(s): e1428b5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -3
README.md CHANGED
@@ -20,7 +20,6 @@ library_name: transformers
20
 
21
  FairyR1-32B, a highly efficient large-language-model (LLM) that matches or exceeds larger models on select tasks despite using only ~5% of their parameters. Built atop the DeepSeek-R1-Distill-Qwen-32B base, FairyR1-32B leverages a novel “distill-and-merge” pipeline—combining task-focused fine-tuning with model-merging techniques to deliver competitive performance with drastically reduced size and inference cost. This project was funded by NSFC, Grant 624B2005.
22
 
23
-
24
  ## Model Details
25
 
26
  The FairyR1 model represents a further exploration of our earlier work [TinyR1](https://arxiv.org/pdf/2503.04872), retaining the core “Branch-Merge Distillation” approach while introducing refinements in data processing and model architecture.
@@ -37,7 +36,6 @@ These results indicate that, by building on the DeepSeek‑R1‑Distill‑Qwen
37
 
38
  This work demonstrates the feasibility of significantly reducing model size and potential inference cost through optimized data processing and model fusion techniques while maintaining strong task-specific performance.
39
 
40
-
41
  ## Model Description
42
 
43
  - **Developed by:** PKU-DS-LAB
@@ -64,7 +62,6 @@ This work demonstrates the feasibility of significantly reducing model size and
64
  - [LiveCodeBench (code)](https://huggingface.co/datasets/livecodebench/code_generation_lite): We evaluate 8 times and report the average accuracy. The dataset version is "release_v5" (date range: 2024-08-01 to 2025-02-01), consisting of 279 problems.<br>
65
  - [GPQA-Diamond (Sci-QA)](https://huggingface.co/datasets/Idavidrein/gpqa): We evaluate 8 times and report the average accuracy. The dataset consists of 198 problems.<br>
66
 
67
-
68
  ## FairyR1 series Team Members:
69
 
70
  Leading By:
 
20
 
21
  FairyR1-32B, a highly efficient large-language-model (LLM) that matches or exceeds larger models on select tasks despite using only ~5% of their parameters. Built atop the DeepSeek-R1-Distill-Qwen-32B base, FairyR1-32B leverages a novel “distill-and-merge” pipeline—combining task-focused fine-tuning with model-merging techniques to deliver competitive performance with drastically reduced size and inference cost. This project was funded by NSFC, Grant 624B2005.
22
 
 
23
  ## Model Details
24
 
25
  The FairyR1 model represents a further exploration of our earlier work [TinyR1](https://arxiv.org/pdf/2503.04872), retaining the core “Branch-Merge Distillation” approach while introducing refinements in data processing and model architecture.
 
36
 
37
  This work demonstrates the feasibility of significantly reducing model size and potential inference cost through optimized data processing and model fusion techniques while maintaining strong task-specific performance.
38
 
 
39
  ## Model Description
40
 
41
  - **Developed by:** PKU-DS-LAB
 
62
  - [LiveCodeBench (code)](https://huggingface.co/datasets/livecodebench/code_generation_lite): We evaluate 8 times and report the average accuracy. The dataset version is "release_v5" (date range: 2024-08-01 to 2025-02-01), consisting of 279 problems.<br>
63
  - [GPQA-Diamond (Sci-QA)](https://huggingface.co/datasets/Idavidrein/gpqa): We evaluate 8 times and report the average accuracy. The dataset consists of 198 problems.<br>
64
 
 
65
  ## FairyR1 series Team Members:
66
 
67
  Leading By: