nielsr HF Staff commited on
Commit
6163ee5
·
verified ·
1 Parent(s): f1b9e61

Add link to Github repo, change pipeline tag to robotics

Browse files

This PR adds a link to the Github repository to https://github.com/firstbatchxyz/dria-agent, updates the pipeline tag to `robotics` to reflect the model's application, and adds the robotics tag.

Files changed (1) hide show
  1. README.md +10 -11
README.md CHANGED
@@ -1,25 +1,22 @@
1
  ---
2
- license: other
3
- license_name: qwen-research
4
- license_link: https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct/blob/main/LICENSE
5
- language:
6
- - en
7
  base_model:
8
  - Qwen/Qwen2.5-Coder-3B-Instruct
9
- pipeline_tag: text-generation
 
10
  library_name: transformers
 
 
 
 
11
  tags:
12
  - code
13
  - chat
14
  - qwen
15
  - qwen-coder
16
  - agent
 
17
  ---
18
 
19
- # Tiny-Agent-α
20
-
21
- ## Introduction
22
-
23
  ***Tiny-Agent-α*** is an extension of Dria-Agent-a, trained on top of the [Qwen2.5-Coder](https://huggingface.co/collections/Qwen/qwen25-coder-66eaa22e6f99801bf65b0c2f) series to be used in edge devices. These models are carefully fine tuned with quantization aware training to minimize performance degradation after quantization.
24
 
25
  Tiny-Agent-α employs ***Pythonic function calling***, which is LLMs using blocks of Python code to interact with provided tools and output actions. This method was inspired by many previous work, including but not limited to [DynaSaur](https://arxiv.org/pdf/2411.01747), [RLEF](https://arxiv.org/pdf/2410.02089), [ADAS](https://arxiv.org/pdf/2408.08435) and [CAMEL](https://arxiv.org/pdf/2303.17760). This way of function calling has a few advantages over traditional JSON-based function calling methods:
@@ -237,4 +234,6 @@ Current benchmark results for various models **(strict)**:
237
  title={Dria-Agent-a},
238
  author={"andthattoo", "Atakan Tekparmak"}
239
  }
240
- ```
 
 
 
1
  ---
 
 
 
 
 
2
  base_model:
3
  - Qwen/Qwen2.5-Coder-3B-Instruct
4
+ language:
5
+ - en
6
  library_name: transformers
7
+ license: other
8
+ license_name: qwen-research
9
+ license_link: https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct/blob/main/LICENSE
10
+ pipeline_tag: robotics
11
  tags:
12
  - code
13
  - chat
14
  - qwen
15
  - qwen-coder
16
  - agent
17
+ - robotics
18
  ---
19
 
 
 
 
 
20
  ***Tiny-Agent-α*** is an extension of Dria-Agent-a, trained on top of the [Qwen2.5-Coder](https://huggingface.co/collections/Qwen/qwen25-coder-66eaa22e6f99801bf65b0c2f) series to be used in edge devices. These models are carefully fine tuned with quantization aware training to minimize performance degradation after quantization.
21
 
22
  Tiny-Agent-α employs ***Pythonic function calling***, which is LLMs using blocks of Python code to interact with provided tools and output actions. This method was inspired by many previous work, including but not limited to [DynaSaur](https://arxiv.org/pdf/2411.01747), [RLEF](https://arxiv.org/pdf/2410.02089), [ADAS](https://arxiv.org/pdf/2408.08435) and [CAMEL](https://arxiv.org/pdf/2303.17760). This way of function calling has a few advantages over traditional JSON-based function calling methods:
 
234
  title={Dria-Agent-a},
235
  author={"andthattoo", "Atakan Tekparmak"}
236
  }
237
+ ```
238
+
239
+ Code: https://github.com/firstbatchxyz/dria-agent