ar-mmarco-sample / README.md
akhooli's picture
Update README.md
6a31c33 verified
|
raw
history blame
1.56 kB
metadata
license: mit
language:
  - ar
configs:
  - config_name: pipe
    data_files: arabic-queries-no-latin.tsv
    sep: '|'

akhooli/ar-mmarco-sample

This repo has samples from the Arabic (machine translation) version of the mMARCO dataset, together with mined rankings (English, but should apply as translations are aligned across languages). The purpose is to train (using free compute, so not fully trained) an Arabic ColBERT V2 model.
The original dataset has a little over 800K queries (training set). I filtered out ones with English words, leaving around 700K, then sampled 250K along with their ranking examples (for 250K, the examples file size is a little over 8GB).

The source of this curated data is unicamp-dl/mmarco and the full examples json file (27 GB) is linked in the ColBERT V2 repo.

Following an observation of Arabic tokenization issues (ex. in BERT models) - see https://www.linkedin.com/posts/akhooli_arabic-bert-tokenizers-you-may-need-to-normalize-activity-7225747473523216384-D1oH - two new files were uploaded to this dataset (normalized queries and collection - visually they are the same). Models based on these files require normalizing the query first.

from unicodedata import normalize
normalized_text = normalize('NFKC', text)

More: https://www.linkedin.com/posts/akhooli_arabic-mmarco-sample-dataset-and-colbert-activity-7225135682044743680-35nN