Update README.md
Browse files
README.md
CHANGED
@@ -2,17 +2,19 @@
|
|
2 |
license: apache-2.0
|
3 |
tags:
|
4 |
- generated_from_trainer
|
|
|
|
|
5 |
model-index:
|
6 |
- name: t5-v1_1-base-finetuned-English-to-BASH
|
7 |
results: []
|
8 |
---
|
9 |
|
10 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
11 |
-
should probably proofread and complete it, then remove this comment. -->
|
12 |
|
13 |
# t5-v1_1-base-finetuned-English-to-BASH
|
14 |
|
15 |
-
|
|
|
|
|
16 |
It achieves the following results on the evaluation set:
|
17 |
- Loss: 0.7958
|
18 |
- Nl2bash M: 0.6179
|
@@ -28,7 +30,12 @@ More information needed
|
|
28 |
|
29 |
## Training and evaluation data
|
30 |
|
31 |
-
|
|
|
|
|
|
|
|
|
|
|
32 |
|
33 |
## Training procedure
|
34 |
|
|
|
2 |
license: apache-2.0
|
3 |
tags:
|
4 |
- generated_from_trainer
|
5 |
+
metrics:
|
6 |
+
- nl2bash_m
|
7 |
model-index:
|
8 |
- name: t5-v1_1-base-finetuned-English-to-BASH
|
9 |
results: []
|
10 |
---
|
11 |
|
|
|
|
|
12 |
|
13 |
# t5-v1_1-base-finetuned-English-to-BASH
|
14 |
|
15 |
+
Created by: [Josh Shih](https://huggingface.co/Josh98), [Alex Sha](https://huggingface.co/alexsha), [Kevin Um](https://huggingface.co/kevinum) for EEP 596 - Natural Language Processing at University of Washington (Seattle).
|
16 |
+
|
17 |
+
This model is a fine-tuned version of [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) on a more balanced iteration of the [NL2BASH](https://github.com/TellinaTool/nl2bash/tree/master/data) dataset.
|
18 |
It achieves the following results on the evaluation set:
|
19 |
- Loss: 0.7958
|
20 |
- Nl2bash M: 0.6179
|
|
|
30 |
|
31 |
## Training and evaluation data
|
32 |
|
33 |
+
This model was trained and evaluated using a custom iteration of [NL2BASH](https://github.com/TellinaTool/nl2bash/tree/master/data). The original NL2BASH dataset contains a large class imbalance with too many bash commands which begin with 'find'.
|
34 |
+
|
35 |
+
A maximum threshold was set to remove text/BASH pairs which exceeded the threshold, and [GPT-3](https://openai.com/blog/gpt-3-apps/) API was used to generate text/BASH pairs for those below the threshold.
|
36 |
+
|
37 |
+
~5500 original text/BASH pairs and ~5700 generated text/BASH pairs were used, giving a total of ~11200 lines of text/BASH pairs. Shown below is the class distribution for the top-5 commands.
|
38 |
+

|
39 |
|
40 |
## Training procedure
|
41 |
|