am5uc commited on
Commit
40a0da2
·
verified ·
1 Parent(s): 263c7ca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -43,11 +43,10 @@ These were the arguments/hyperparameters, I used. I tried using higher epochs, b
43
  ## Evaluation
44
  I had three benchmarks, the WikiTableQuestions dataset, the TabFact dataset, and the Synthetic Validation set. Fine-tuning did not harm the results of on the WTQ Validation Set and the TabFact Dataset, in which I got accuracies of .3405 and .5005, respectively for both the pre-trained and fine-tuned model. There were improvements in the validation and test results after training though. On Validation, there was a jump from 0.4000 to 0.4222. On the test set, there was quite a larger jump in accuracy from 0.2033 to 0.4667 after fine-tuning.
45
 
46
- | Model | Benchmark 1 (WTQ Validation Set) | Benchmark 2 (TabFact) | Benchmark 3 (Synthetic Validation Set) | Test Set of Synthetic Dataset |
47
- |------------------------------------------------------|----------------------------------|-----------------------|----------------------------------------|-------------------------------|
48
- | google/tapas-base-finetuned-wtq (before Fine-tuning) | 0.3405 | .5005 | 0.4000 | 0.2933 |
49
- | google/tapas-base-finetuned-wtq (Fine-tuned) | 0.3405 | .5005 | 0.4222 | 0.4667 |
50
-
51
 
52
  ## Usage
53
  The prompt for the TAPAS model should be a natural language question paired with a structured table that can be passed in in dataframe format. The prompt should look like this:
 
43
  ## Evaluation
44
  I had three benchmarks, the WikiTableQuestions dataset, the TabFact dataset, and the Synthetic Validation set. Fine-tuning did not harm the results of on the WTQ Validation Set and the TabFact Dataset, in which I got accuracies of .3405 and .5005, respectively for both the pre-trained and fine-tuned model. There were improvements in the validation and test results after training though. On Validation, there was a jump from 0.4000 to 0.4222. On the test set, there was quite a larger jump in accuracy from 0.2033 to 0.4667 after fine-tuning.
45
 
46
+ | Model | Benchmark 1 (WTQ Validation Set) | Benchmark 2 (TabFact) | Benchmark 3 (SQA) | Validation Set of Synthetic Dataset | Test Set of Synthetic Dataset |
47
+ |------------------------------------------------------|----------------------------------|-----------------------|-------------------|-------------------------------------|-------------------------------|
48
+ | google/tapas-base-finetuned-wtq (before Fine-tuning) | 0.3405 | 0.5005 | 0.2512 | 0.4000 | 0.2933 |
49
+ | google/tapas-base-finetuned-wtq (Fine-tuned) | 0.3405 | 0.5005 | 0.2525 | 0.4222 | 0.4667 |
 
50
 
51
  ## Usage
52
  The prompt for the TAPAS model should be a natural language question paired with a structured table that can be passed in in dataframe format. The prompt should look like this: