Update README.md
Browse files
README.md
CHANGED
@@ -43,11 +43,10 @@ These were the arguments/hyperparameters, I used. I tried using higher epochs, b
|
|
43 |
## Evaluation
|
44 |
I had three benchmarks, the WikiTableQuestions dataset, the TabFact dataset, and the Synthetic Validation set. Fine-tuning did not harm the results of on the WTQ Validation Set and the TabFact Dataset, in which I got accuracies of .3405 and .5005, respectively for both the pre-trained and fine-tuned model. There were improvements in the validation and test results after training though. On Validation, there was a jump from 0.4000 to 0.4222. On the test set, there was quite a larger jump in accuracy from 0.2033 to 0.4667 after fine-tuning.
|
45 |
|
46 |
-
| Model | Benchmark 1 (WTQ Validation Set) | Benchmark 2 (TabFact) | Benchmark 3 (
|
47 |
-
|
48 |
-
| google/tapas-base-finetuned-wtq (before Fine-tuning) | 0.3405 | .5005
|
49 |
-
| google/tapas-base-finetuned-wtq (Fine-tuned) | 0.3405 | .5005
|
50 |
-
|
51 |
|
52 |
## Usage
|
53 |
The prompt for the TAPAS model should be a natural language question paired with a structured table that can be passed in in dataframe format. The prompt should look like this:
|
|
|
43 |
## Evaluation
|
44 |
I had three benchmarks, the WikiTableQuestions dataset, the TabFact dataset, and the Synthetic Validation set. Fine-tuning did not harm the results of on the WTQ Validation Set and the TabFact Dataset, in which I got accuracies of .3405 and .5005, respectively for both the pre-trained and fine-tuned model. There were improvements in the validation and test results after training though. On Validation, there was a jump from 0.4000 to 0.4222. On the test set, there was quite a larger jump in accuracy from 0.2033 to 0.4667 after fine-tuning.
|
45 |
|
46 |
+
| Model | Benchmark 1 (WTQ Validation Set) | Benchmark 2 (TabFact) | Benchmark 3 (SQA) | Validation Set of Synthetic Dataset | Test Set of Synthetic Dataset |
|
47 |
+
|------------------------------------------------------|----------------------------------|-----------------------|-------------------|-------------------------------------|-------------------------------|
|
48 |
+
| google/tapas-base-finetuned-wtq (before Fine-tuning) | 0.3405 | 0.5005 | 0.2512 | 0.4000 | 0.2933 |
|
49 |
+
| google/tapas-base-finetuned-wtq (Fine-tuned) | 0.3405 | 0.5005 | 0.2525 | 0.4222 | 0.4667 |
|
|
|
50 |
|
51 |
## Usage
|
52 |
The prompt for the TAPAS model should be a natural language question paired with a structured table that can be passed in in dataframe format. The prompt should look like this:
|