yuyuzhang commited on
Commit
c38b2b2
·
1 Parent(s): a6ddf6f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +74 -3
README.md CHANGED
@@ -1,3 +1,74 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - ByteDance-Seed/Seed-Coder-8B-Base
5
+ ---
6
+
7
+ # Seed-Coder-8B-Reasoning
8
+
9
+ ## Introduction
10
+ **Seed-Coder-8B-Reasoning** is an 8-billion-parameter model further optimized for **code reasoning**, **problem-solving**, and **algorithmic thinking** tasks.
11
+ Built upon the strong base of Seed-Coder, it undergoes additional training in sandbox environments to significantly enhance its ability to tackle complex coding problems and competitions. It features:
12
+ - Trained on a **massively curated corpus**, filtered using an **LLM-based method** to ensure high-quality real-world code, text-code alignment, and synthetic datasets.
13
+ - **Sandbox fine-tuning** to specifically strengthen **multi-step reasoning**, **algorithm design**, and **competitive programming** capabilities.
14
+ - Maintains **long-context handling** up to 32K tokens, enabling it to reason over extended problem descriptions and large input-output examples.
15
+
16
+ ## Requirements
17
+ You will need to install the latest versions of `transformers` and `accelerate`:
18
+
19
+ ```bash
20
+ pip install -U transformers accelerate
21
+ ```
22
+
23
+ ## Quickstart
24
+
25
+ Here is a simple example demonstrating how to load the model and perform code generation using the Hugging Face `pipeline` API:
26
+
27
+ ```python
28
+ import transformers
29
+ import torch
30
+
31
+ model_id = "ByteDance-Seed/Seed-Coder-8B-Reasoning"
32
+
33
+ pipeline = transformers.pipeline(
34
+ "text-generation",
35
+ model=model_id,
36
+ model_kwargs={"torch_dtype": torch.bfloat16},
37
+ device_map="auto",
38
+ )
39
+
40
+ messages = [
41
+ {"role": "user", "content": "Solve the following problem: Given an array of integers, find two numbers such that they add up to a specific target number."},
42
+ ]
43
+
44
+ outputs = pipeline(
45
+ messages,
46
+ max_new_tokens=512,
47
+ )
48
+ print(outputs[0]["generated_text"][-1]["content"])
49
+ ```
50
+
51
+ ## Evaluation
52
+
53
+ Seed-Coder-8B-Reasoning has been evaluated extensively on reasoning-intensive code benchmarks, showing:
54
+ - Significant improvements on **competitive programming** datasets and coding challenges.
55
+ - Enhanced ability to **break down complex problems**, **design correct algorithms**, and **produce efficient implementations**.
56
+ - Strong generalization to unseen problems across multiple domains (math, strings, arrays, graphs, DP, etc.).
57
+
58
+ For detailed results, please check our [📑 paper](https://arxiv.org/pdf/xxx.xxxxx).
59
+
60
+ ## Citation
61
+
62
+ If you find our work helpful, please consider citing our work:
63
+
64
+ ```
65
+ @article{zhang2025seedcoder,
66
+ title={Seed-Coder: Let the Code Model Curate Data for Itself},
67
+ author={Xxx},
68
+ year={2025},
69
+ eprint={2504.xxxxx},
70
+ archivePrefix={arXiv},
71
+ primaryClass={cs.CL},
72
+ url={https://arxiv.org/abs/xxxx.xxxxx},
73
+ }
74
+ ```