FiD-TQA / README.md
wyu1's picture
update readme
9611c20
---
license: cc-by-4.0
---
# FiD model trained on TQA
-- This is the model checkpoint of FiD [2], based on the T5 large (with 770M parameters) and trained on the TriviaQA dataset [1].
-- Hyperparameters: 8 x 40GB A100 GPUs; batch size 8; AdamW; LR 3e-5; 30000 steps
References:
[1] TriviaQA: A Large Scale Dataset for Reading Comprehension and Question Answering. ACL 2017
[2] Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering. EACL 2021.
## Model performance
We evaluate it on the TriviaQA dataset, the EM score is 68.5 (0.8 higher than the original performance reported in the paper).
<a href="https://huggingface.co/exbert/?model=bert-base-uncased">
<img width="300px" src="https://cdn-media.huggingface.co/exbert/button.png">
</a>
---
license: cc-by-4.0
---