BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of Deep-Research Agent
Paper
•
2508.06600
•
Published
•
36
BEIR (Benchmarking IR) consists of a homogenous benchmark for diverse sentence or passage level IR tasks. It provides a common and easy framework for the cross-domain evaluation of your retrieval models.
nthakur/swim-ir-cross-lingual
nthakur/swim-ir-monolingual
nthakur/indic-swim-ir-cross-lingual