AdrienB134 commited on
Commit
16a7b85
1 Parent(s): 345ca60

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -0
README.md ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - unicamp-dl/mmarco
5
+ language:
6
+ - es
7
+ ---
8
+
9
+ ## Training
10
+
11
+ #### Details
12
+
13
+ The model is initialized from the [ColBERTv1.0-bert-based-spanish-mmarcoES](https://huggingface.co/AdrienB134/ColBERTv1.0-bert-based-spanish-mmarcoES) checkpoint and trained using the ColBERTv2 style of training.
14
+ It was trained on 2 Tesla T4 GPU with 16GBs of memory each with 20k warmup steps warmup using a batch size of 64 and the AdamW optimizer with a constant learning rate of 1e-05.
15
+ Total training time was around 48 hours.
16
+
17
+ #### Data
18
+
19
+ The model is fine-tuned on the Spanish version of the [mMARCO](https://huggingface.co/datasets/unicamp-dl/mmarco) dataset, a multi-lingual machine-translated version of the MS MARCO dataset.
20
+
21
+
22
+ ## Evaluation
23
+
24
+ The model is evaluated on the smaller development set of mMARCO-es, which consists of 6,980 queries for a corpus of 8.8M candidate passages. We report the mean reciprocal rank (MRR) and recall at various cut-offs (R@k).
25
+
26
+ | model | Vocab. | #Param. | Size | MRR@10 | R@50 | R@1000 |
27
+ |:------------------------------------------------------------------------------------------------------------------------|:--------|--------:|------:|---------:|-------:|--------:|
28
+ | **ColBERTv1.0-bert-based-spanish-mmarcoES** | spanish | 110M | 440MB | 32.86 | 76.46 | 81.06 |