Improve model card: Add library, paper link, project page, and usage instructions
Browse filesThis PR significantly enhances the model card for better discoverability, usability, and adherence to Hugging Face best practices.
Key changes include:
* **Metadata update**:
* Added `library_name: transformers` to enable the "how to use" widget and improve model discoverability.
* Added `project_page: https://julienp.netlify.app/posts/soar/` to the metadata for structured information about the associated blog/project.
* **Content improvements**:
* Updated the paper link in the model card badges and main paper description to the official Hugging Face Paper page: [https://huggingface.co/papers/2507.14172](https://huggingface.co/papers/2507.14172).
* Added a prominent "GitHub" badge and a direct link to the official GitHub repository ([https://github.com/flowersteam/SOAR](https://github.com/flowersteam/SOAR)) for easier access to the code.
* Included the detailed installation instructions and "Run SOAR" guidance from the GitHub README, allowing users to get started directly from the Hugging Face Hub.
* Added a BibTeX citation section for proper academic attribution.
These updates aim to provide a more complete and accessible resource for users interacting with the SOAR-ARC model on the Hugging Face Hub.
@@ -1,11 +1,11 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
2 |
license: other
|
3 |
license_name: mistral
|
4 |
license_link: LICENSE
|
5 |
-
datasets:
|
6 |
-
- julien31/soar_arc_train_5M
|
7 |
-
base_model:
|
8 |
-
- mistralai/Mistral-Large-Instruct-2407
|
9 |
pipeline_tag: text-generation
|
10 |
tags:
|
11 |
- text-generation
|
@@ -15,22 +15,26 @@ tags:
|
|
15 |
- arc
|
16 |
- arc-agi
|
17 |
- soar
|
|
|
|
|
18 |
---
|
|
|
19 |
# SOAR-ARC Models: Self-Improving Language Models for Program Synthesis
|
20 |
|
21 |
<p align="center">
|
22 |
-
🤗 <a href="https://huggingface.co/collections/julien31/soar-arc-6856d27681fce01d9af4c4a3">Hugging Face (data and model)</a>   |    📑 <a href="https://
|
23 |
</p>
|
24 |
|
25 |
This repository contains one of the models fine-tuned using the **SOAR** (**S**elf-improving **O**perators for **A**utomated program **R**efinements) framework, as presented in the paper:
|
26 |
|
27 |
-
> [**Self-Improving Language Models for Evolutionary Program Synthesis: A Case Study on ARC-AGI**](https://
|
28 |
>
|
29 |
> Julien Pourcel, Cédric Colas, Pierre-Yves Oudeyer.
|
30 |
> *Proceedings of the 42nd International Conference on Machine Learning (ICML), 2025.*
|
31 |
|
32 |
These models are specialized in solving tasks from the challenging [Abstraction and Reasoning Corpus (ARC)](https://github.com/fchollet/ARC) by synthesizing Python programs.
|
33 |
|
|
|
34 |
|
35 |
## SOAR
|
36 |
Large Language Models (LLMs) have become incredibly powerful, but they often hit a wall when faced with truly complex reasoning tasks that require discovering a solution from scratch. Simply throwing more computing power or using a bigger model often yields diminishing returns. But what if a model could learn from its own experience, getting smarter with every attempt?
|
@@ -71,4 +75,57 @@ For a complete, end-to-end example of how to format the prompt, run inference, e
|
|
71 |
* **Official SOAR GitHub Repository**: [https://github.com/flowersteam/SOAR](https://github.com/flowersteam/SOAR)
|
72 |
* **Inference & Visualization Notebook**: [https://github.com/flowersteam/SOAR/blob/main/notebook/inference_visualisation.ipynb](https://github.com/flowersteam/SOAR/blob/main/notebook/inference_visualisation.ipynb)
|
73 |
|
74 |
-
<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/made
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- mistralai/Mistral-Large-Instruct-2407
|
4 |
+
datasets:
|
5 |
+
- julien31/soar_arc_train_5M
|
6 |
license: other
|
7 |
license_name: mistral
|
8 |
license_link: LICENSE
|
|
|
|
|
|
|
|
|
9 |
pipeline_tag: text-generation
|
10 |
tags:
|
11 |
- text-generation
|
|
|
15 |
- arc
|
16 |
- arc-agi
|
17 |
- soar
|
18 |
+
library_name: transformers
|
19 |
+
project_page: https://julienp.netlify.app/posts/soar/
|
20 |
---
|
21 |
+
|
22 |
# SOAR-ARC Models: Self-Improving Language Models for Program Synthesis
|
23 |
|
24 |
<p align="center">
|
25 |
+
🤗 <a href="https://huggingface.co/collections/julien31/soar-arc-6856d27681fce01d9af4c4a3">Hugging Face (data and model)</a>   |    📑 <a href="https://huggingface.co/papers/2507.14172">Paper</a>    |    📑 <a href="https://julienp.netlify.app/posts/soar/">Blog</a>    |    🐙 <a href="https://github.com/flowersteam/SOAR">GitHub</a>
|
26 |
</p>
|
27 |
|
28 |
This repository contains one of the models fine-tuned using the **SOAR** (**S**elf-improving **O**perators for **A**utomated program **R**efinements) framework, as presented in the paper:
|
29 |
|
30 |
+
> [**Self-Improving Language Models for Evolutionary Program Synthesis: A Case Study on ARC-AGI**](https://huggingface.co/papers/2507.14172)
|
31 |
>
|
32 |
> Julien Pourcel, Cédric Colas, Pierre-Yves Oudeyer.
|
33 |
> *Proceedings of the 42nd International Conference on Machine Learning (ICML), 2025.*
|
34 |
|
35 |
These models are specialized in solving tasks from the challenging [Abstraction and Reasoning Corpus (ARC)](https://github.com/fchollet/ARC) by synthesizing Python programs.
|
36 |
|
37 |
+
**GitHub Repository**: [https://github.com/flowersteam/SOAR](https://github.com/flowersteam/SOAR)
|
38 |
|
39 |
## SOAR
|
40 |
Large Language Models (LLMs) have become incredibly powerful, but they often hit a wall when faced with truly complex reasoning tasks that require discovering a solution from scratch. Simply throwing more computing power or using a bigger model often yields diminishing returns. But what if a model could learn from its own experience, getting smarter with every attempt?
|
|
|
75 |
* **Official SOAR GitHub Repository**: [https://github.com/flowersteam/SOAR](https://github.com/flowersteam/SOAR)
|
76 |
* **Inference & Visualization Notebook**: [https://github.com/flowersteam/SOAR/blob/main/notebook/inference_visualisation.ipynb](https://github.com/flowersteam/SOAR/blob/main/notebook/inference_visualisation.ipynb)
|
77 |
|
78 |
+
<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/made%20with%20unsloth.png" width="20%" />
|
79 |
+
|
80 |
+
## Installation
|
81 |
+
|
82 |
+
### Conda Inference Environment
|
83 |
+
```
|
84 |
+
pip install --upgrade pip
|
85 |
+
|
86 |
+
git clone https://github.com/flowersteam/SOAR
|
87 |
+
cd SOAR
|
88 |
+
conda create --name sglang47 \
|
89 |
+
python=3.11 \
|
90 |
+
-y
|
91 |
+
conda activate sglang47
|
92 |
+
|
93 |
+
pip install "sglang[all]>=0.4.7"
|
94 |
+
|
95 |
+
pip install -e .
|
96 |
+
pip install -r requirements
|
97 |
+
|
98 |
+
```
|
99 |
+
|
100 |
+
### Conda Training Environment
|
101 |
+
```
|
102 |
+
conda create --name unsloth_env \
|
103 |
+
python=3.11 \
|
104 |
+
pytorch-cuda=12.1 \
|
105 |
+
pytorch cudatoolkit xformers -c pytorch -c nvidia -c xformers \
|
106 |
+
-y
|
107 |
+
conda activate unsloth_env
|
108 |
+
|
109 |
+
pip install unsloth
|
110 |
+
cd SOAR
|
111 |
+
pip install -e .
|
112 |
+
pip install -r requirements.txt
|
113 |
+
```
|
114 |
+
|
115 |
+
## Run SOAR
|
116 |
+
To run SOAR, please refer to execution instructions located in the experience folder.
|
117 |
+
|
118 |
+
For simple instructions on running sampling and refinement with SOAR, as well as exploring the dataset, please see the Jupyter notebooks provided in the `notebook` folder. These notebooks walk through the basic SOAR step, including how to generate candidate solutions, perform refinement, and analyze results. This hands-on guide will help you get started quickly and understand each step of the SOAR process.
|
119 |
+
|
120 |
+
## Citation
|
121 |
+
Please cite the following paper if you find this model useful for your research:
|
122 |
+
|
123 |
+
```bibtex
|
124 |
+
@inproceedings{pourcel2025selfimproving,
|
125 |
+
title={{Self-Improving Language Models for Evolutionary Program Synthesis: A Case Study on ARC-AGI}},
|
126 |
+
author={Pourcel, Julien and Colas, C{\'e}dric and Oudeyer, Pierre-Yves},
|
127 |
+
booktitle={Proceedings of the 42nd International Conference on Machine Learning (ICML)},
|
128 |
+
year={2025},
|
129 |
+
url={https://huggingface.co/papers/2507.14172}
|
130 |
+
}
|
131 |
+
```
|