Add link to Github repo, change pipeline tag to robotics
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,25 +1,22 @@
|
|
1 |
---
|
2 |
-
license: other
|
3 |
-
license_name: qwen-research
|
4 |
-
license_link: https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct/blob/main/LICENSE
|
5 |
-
language:
|
6 |
-
- en
|
7 |
base_model:
|
8 |
- Qwen/Qwen2.5-Coder-3B-Instruct
|
9 |
-
|
|
|
10 |
library_name: transformers
|
|
|
|
|
|
|
|
|
11 |
tags:
|
12 |
- code
|
13 |
- chat
|
14 |
- qwen
|
15 |
- qwen-coder
|
16 |
- agent
|
|
|
17 |
---
|
18 |
|
19 |
-
# Tiny-Agent-α
|
20 |
-
|
21 |
-
## Introduction
|
22 |
-
|
23 |
***Tiny-Agent-α*** is an extension of Dria-Agent-a, trained on top of the [Qwen2.5-Coder](https://huggingface.co/collections/Qwen/qwen25-coder-66eaa22e6f99801bf65b0c2f) series to be used in edge devices. These models are carefully fine tuned with quantization aware training to minimize performance degradation after quantization.
|
24 |
|
25 |
Tiny-Agent-α employs ***Pythonic function calling***, which is LLMs using blocks of Python code to interact with provided tools and output actions. This method was inspired by many previous work, including but not limited to [DynaSaur](https://arxiv.org/pdf/2411.01747), [RLEF](https://arxiv.org/pdf/2410.02089), [ADAS](https://arxiv.org/pdf/2408.08435) and [CAMEL](https://arxiv.org/pdf/2303.17760). This way of function calling has a few advantages over traditional JSON-based function calling methods:
|
@@ -237,4 +234,6 @@ Current benchmark results for various models **(strict)**:
|
|
237 |
title={Dria-Agent-a},
|
238 |
author={"andthattoo", "Atakan Tekparmak"}
|
239 |
}
|
240 |
-
```
|
|
|
|
|
|
1 |
---
|
|
|
|
|
|
|
|
|
|
|
2 |
base_model:
|
3 |
- Qwen/Qwen2.5-Coder-3B-Instruct
|
4 |
+
language:
|
5 |
+
- en
|
6 |
library_name: transformers
|
7 |
+
license: other
|
8 |
+
license_name: qwen-research
|
9 |
+
license_link: https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct/blob/main/LICENSE
|
10 |
+
pipeline_tag: robotics
|
11 |
tags:
|
12 |
- code
|
13 |
- chat
|
14 |
- qwen
|
15 |
- qwen-coder
|
16 |
- agent
|
17 |
+
- robotics
|
18 |
---
|
19 |
|
|
|
|
|
|
|
|
|
20 |
***Tiny-Agent-α*** is an extension of Dria-Agent-a, trained on top of the [Qwen2.5-Coder](https://huggingface.co/collections/Qwen/qwen25-coder-66eaa22e6f99801bf65b0c2f) series to be used in edge devices. These models are carefully fine tuned with quantization aware training to minimize performance degradation after quantization.
|
21 |
|
22 |
Tiny-Agent-α employs ***Pythonic function calling***, which is LLMs using blocks of Python code to interact with provided tools and output actions. This method was inspired by many previous work, including but not limited to [DynaSaur](https://arxiv.org/pdf/2411.01747), [RLEF](https://arxiv.org/pdf/2410.02089), [ADAS](https://arxiv.org/pdf/2408.08435) and [CAMEL](https://arxiv.org/pdf/2303.17760). This way of function calling has a few advantages over traditional JSON-based function calling methods:
|
|
|
234 |
title={Dria-Agent-a},
|
235 |
author={"andthattoo", "Atakan Tekparmak"}
|
236 |
}
|
237 |
+
```
|
238 |
+
|
239 |
+
Code: https://github.com/firstbatchxyz/dria-agent
|