diff --git "a/data/dataset_meta/dataset_meta_en.csv" "b/data/dataset_meta/dataset_meta_en.csv" new file mode 100644--- /dev/null +++ "b/data/dataset_meta/dataset_meta_en.csv" @@ -0,0 +1,737 @@ +๏ปฟ"id","author","created_at","lastModified","sha","downloads_30","downloads_alltime","likes","tags","tasks","description","citation","languages","language_category","size_categories","paperswithcode_id","private","gated","disabled","license","arxiv_id","url","task_ids" +"argilla/databricks-dolly-15k-curated-en","argilla","2023-05-30 09:54:44","2023-10-02 12:32:53+00:00","4dcd1dedbe148307a833c931b21ca456a1fc4281","167431","95821337","45","None","None","Guidelines In this dataset, you will find a collection of records that show a category, an instruction, a context and a response to that instruction. The aim of the project is to correct the instructi","None","en","mono","None","None","False","False","False","None","None","None","None" +"lighteval/mmlu","lighteval","2023-05-16 09:39:28","2023-06-09 16:36:19+00:00","e24764f1fb58c26b5f622157644f2e5fe77e5b01","10443","59386649","39","None","question-answering, multiple-choice-qa","This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and ","@article{hendryckstest2021, title={Measuring Massive Multitask Language Understanding}, author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt}, journal={Proceedings of the International Conference on Learning Representations (ICLR)}, year={2021} }","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","mit","2005.00700","None","multiple-choice-qa" +"cais/mmlu","cais","2022-03-02 23:29:22","2024-03-08 20:36:26+00:00","c30699e8356da336a370243923dbaf21066bb9fe","105911","36901288","371","None","question-answering, multiple-choice-qa","Dataset Card for MMLU Dataset Summary Measuring Massive Multitask Language Understanding by Dan Hendrycks, Collin Burns, Steven Basart, Andy Zou, Mantas Mazeika, Dawn Song, and Jacob Steinhardt (ICLR ","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","mit","2005.00700","None","multiple-choice-qa" +"nyu-mll/glue","nyu-mll","2022-03-02 23:29:22","2024-01-30 07:41:18+00:00","bcdcba79d07bc864c1c254ccfcedcce55bcc9a8c","320395","35945159","390","None","text-classification, acceptability-classification, natural-language-inference, semantic-similarity-scoring, sentiment-classification, text-scoring","Dataset Card for GLUE Dataset Summary GLUE, the General Language Understanding Evaluation benchmark (https://gluebenchmark.com/) is a collection of resources for training, evaluating, and analyzing na","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","other","1804.07461","None","acceptability-classification, natural-language-inference, semantic-similarity-scoring, sentiment-classification, text-scoring" +"rajpurkar/squad_v2","rajpurkar","2022-03-02 23:29:22","2024-03-04 13:55:27+00:00","3ffb306f725f7d2ce8394bc1873b24868140c412","21214","33952921","191","None","question-answering, open-domain-qa, extractive-qa","Dataset Card for SQuAD 2.0 Dataset Summary Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles,","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","cc-by-sa-4.0","1806.03822","None","open-domain-qa, extractive-qa" +"Salesforce/wikitext","Salesforce","2022-03-02 23:29:22","2024-01-04 16:49:18+00:00","b08601e04326c79dfdd32d625aee71d232d685c3","390766","18287258","391","None","text-generation, fill-mask, language-modeling, masked-language-modeling","Dataset Card for ""wikitext"" Dataset Summary The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified Good and Featured articles on Wikipedia","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","cc-by-sa-3.0","1609.07843","None","language-modeling, masked-language-modeling" +"nyu-mll/blimp","nyu-mll","2022-03-02 23:29:22","2024-01-23 09:58:08+00:00","877fba0801ffb7cbd8c39c1ff314a46f053f6036","15552","13186826","36","task_categories:text-classification, task_ids:acceptability-classification, annotations_creators:crowdsourced, language_creators:machine-generated, multilinguality:monolingual, source_datasets:original, language:en, license:cc-by-4.0, size_categories:10K|1|T","None","False","False","False","odc-by","2406.17557","None","None" +"Rowan/hellaswag","Rowan","2022-03-02 23:29:22","2023-09-28 14:49:00+00:00","6002345709e0801764318f06bf06ce1e7d1a1fe3","144626","2267737","105","None","None","HellaSwag: Can a Machine Really Finish Your Sentence? is a new dataset for commonsense NLI. A paper was published at ACL2019.","@inproceedings{zellers2019hellaswag, title={HellaSwag: Can a Machine Really Finish Your Sentence?}, author={Zellers, Rowan and Holtzman, Ari and Bisk, Yonatan and Farhadi, Ali and Choi, Yejin}, booktitle ={Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics}, year={2019} }","en","mono","None","None","False","False","False","None","1905.07830","None","None" +"wmt/wmt16","wmt","2022-03-02 23:29:22","2024-04-03 12:30:24+00:00","41d8a4013aa1489f28fea60ec0932af246086482","5755","2169229","21","None","translation","Dataset Card for ""wmt16"" Dataset Summary Warning: There are issues with the Common Crawl corpus data (training-parallel-commoncrawl.tgz): Non-English files contain many English sentences. Their ""paral","None","cs, de, en, fi, ro, ru, tr","multi","1|0|M|<|n|<|1|0|0|M","None","False","False","False","unknown","None","None","None" +"fancyzhx/ag_news","fancyzhx","2022-03-02 23:29:22","2024-03-07 12:02:37+00:00","eb185aade064a813bc0b7f42de02595523103ca4","19103","2165309","149","None","text-classification, topic-classification","Dataset Card for ""ag_news"" Dataset Summary AG is a collection of more than 1 million news articles. News articles have been gathered from more than 2000 news sources by ComeToMyHead in more than 1 yea","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","unknown","None","None","topic-classification" +"cardiffnlp/tweet_eval","cardiffnlp","2022-03-02 23:29:22","2024-01-04 16:40:33+00:00","b3a375baf0f409c77e6bc7aa35102b7b3534f8be","11115","2117074","118","None","text-classification, intent-classification, multi-class-classification, sentiment-classification","Dataset Card for tweet_eval Dataset Summary TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. The tasks include - irony, hate, offensive, stanc","None","en","mono","1|0|0|K|<|n|<|1|M|||1|0|K|<|n|<|1|0|0|K|||1|K|<|n|<|1|0|K|||n|<|1|K","None","False","False","False","unknown","None","None","intent-classification, multi-class-classification, sentiment-classification" +"bigscience/P3","bigscience","2022-03-02 23:29:22","2024-03-04 18:08:03+00:00","db485208b9f41d46c1d0975202328d08f8199046","26680","2045636","209","task_categories:other, annotations_creators:crowdsourced, annotations_creators:expert-generated, multilinguality:monolingual, language:en, license:apache-2.0, size_categories:100M|1|T","None","False","False","False","odc-by","2406.17557","None","None" +"uclanlp/wino_bias","uclanlp","2022-03-02 23:29:22","2024-01-04 16:50:33+00:00","3f31267586e4408e3b3f77ec22198fd24ea8dc1d","2471","1722477","16","None","token-classification, coreference-resolution","Dataset Card for Wino_Bias dataset Dataset Summary WinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias. The corpus contains Winograd-schema style sentences with entit","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","mit","1804.06876","None","coreference-resolution" +"allenai/openbookqa","allenai","2022-03-02 23:29:22","2024-01-04 16:09:20+00:00","388097ea7776314e93a529163e0fea805b8a6454","76510","1623970","85","None","question-answering, open-domain-qa","Dataset Card for OpenBookQA Dataset Summary OpenBookQA aims to promote research in advanced question-answering, probing a deeper understanding of both the topic (with salient facts summarized as an op","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","unknown","None","None","open-domain-qa" +"Helsinki-NLP/opus-100","Helsinki-NLP","2022-03-02 23:29:22","2024-02-28 09:17:34+00:00","805090dc28bf78897da9641cdf08b61287580df9","24897","1450598","168","None","translation","Dataset Card for OPUS-100 Dataset Summary OPUS-100 is an English-centric multilingual corpus covering 100 languages. OPUS-100 is English-centric, meaning that all training pairs include English on eit","None","af, am, an, ar, as, az, be, bg, bn, br, bs, ca, cs, cy, da, de, dz, el, en, eo, es, et, eu, fa, fi, fr, fy, ga, gd, gl, gu, ha, he, hi, hr, hu, hy, id, ig, is, it, ja, ka, kk, km, kn, ko, ku, ky, li, lt, lv, mg, mk, ml, mn, mr, ms, mt, my, nb, ne, nl, nn, no, oc, or, pa, pl, ps, pt, ro, ru, rw, se, sh, si, sk, sl, sq, sr, sv, ta, te, tg, th, tk, tr, tt, ug, uk, ur, uz, vi, wa, xh, yi, yo, zh, zu","multi","1|0|0|K|<|n|<|1|M|||1|0|K|<|n|<|1|0|0|K|||1|K|<|n|<|1|0|K|||1|M|<|n|<|1|0|M|||n|<|1|K","None","False","False","False","unknown","2004.11867","None","None" +"ehovy/race","ehovy","2022-03-02 23:29:22","2024-01-04 16:22:34+00:00","2fec9fd81f1dc971569a9b729c43f2f0e6436637","1908","1441331","54","task_categories:multiple-choice, task_ids:multiple-choice-qa, annotations_creators:expert-generated, language_creators:found, multilinguality:monolingual, source_datasets:original, language:en, license:other, size_categories:100K|1|T","None","False","False","False","odc-by","None","None","None" +"fancyzhx/dbpedia_14","fancyzhx","2022-03-02 23:29:22","2024-01-22 11:57:58+00:00","9abd46cf7fc8b4c64290f26993c540b92aa145ac","2199","513261","28","None","text-classification, topic-classification","Dataset Card for DBpedia14 Dataset Summary The DBpedia ontology classification dataset is constructed by picking 14 non-overlapping classes from DBpedia 2014. They are listed in classes.txt. From each","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","cc-by-sa-3.0","1509.01626","None","topic-classification" +"wyu1/Leopard-Instruct","wyu1","2024-10-29 20:51:58","2024-11-08 00:12:25+00:00","93317b272c5a9d9c0417fa6ea6e2be89ac9215ea","136887","503585","56","language:en, license:apache-2.0, size_categories:1M|1|T","None","False","False","False","apache-2.0","None","None","None" +"allenai/swag","allenai","2022-03-02 23:29:22","2024-06-14 10:21:05+00:00","dc48148372b3853a9c7bae7bb06c161b46d8364a","2081","341155","22","None","text-classification, natural-language-inference","Dataset Card for Situations With Adversarial Generations Dataset Summary Given a partial description like ""she opened the hood of the car,"" humans can reason about the situation and anticipate what mi","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","unknown","1808.05326","None","natural-language-inference" +"mteb/sprintduplicatequestions-pairclassification","mteb","2022-04-19 10:40:02","2022-09-27 19:15:57+00:00","5a8256d0dff9c4bd3be3ba3e67e4e70173f802ea","1643","336732","0","language:en, size_categories:n<1K, modality:text, region:us","None","None","None","en","mono","None","None","False","False","False","None","None","None","None" +"Sterzhang/PVIT-3M","Sterzhang","2024-10-07 09:28:17","2024-11-02 07:41:57+00:00","68c0ad34851b06e7e408b092c1f8ee1004f6c92b","21207","336555","17","task_categories:visual-question-answering, task_categories:image-text-to-text, language:en, license:apache-2.0, size_categories:1M|1|T","None","False","False","False","odc-by","2404.14219","None","None" +"deepmind/narrativeqa","deepmind","2022-03-02 23:29:22","2024-03-06 07:33:05+00:00","2e643e7363944af1c33a652d1c87320d0871c4e4","4770","151533","41","task_categories:text2text-generation, task_ids:abstractive-qa, annotations_creators:crowdsourced, language_creators:found, multilinguality:monolingual, source_datasets:original, language:en, license:apache-2.0, size_categories:10K|1|T","None","False","False","False","cc0-1.0","None","None","language-modeling" +"common-canvas/commoncatalog-cc-by","common-canvas","2024-04-22 18:07:35","2024-05-16 19:01:29+00:00","80f50fe4a1ca937f37a11be3f8eee5199d776ff3","8166","132149","26","task_categories:text-to-image, language:en, license:cc-by-4.0, size_categories:10M>> from datasets import load_dataset >>> dat","None","aai, aak, aau, aaz, abt, abx, aby, acf, acr, acu, adz, aer, aey, agd, agg, agm, agn, agr, agt, agu, aia, aii, aka, ake, alp, alq, als, aly, ame, amf, amk, amm, amn, amo, amp, amr, amu, amx, anh, anv, aoi, aoj, aom, aon, apb, ape, apn, apr, apu, apw, apz, arb, are, arl, arn, arp, asm, aso, ata, atb, atd, atg, att, auc, aui, auy, avt, awb, awk, awx, azb, azg, azz, bao, bba, bbb, bbr, bch, bco, bdd, bea, bef, bel, ben, beo, beu, bgs, bgt, bhg, bhl, big, bjk, bjp, bjr, bjv, bjz, bkd, bki, bkq, bkx, bla, blw, blz, bmh, bmk, bmr, bmu, bnp, boa, boj, bon, box, bpr, bps, bqc, bqp, bre, bsj, bsn, bsp, bss, buk, bus, bvd, bvr, bxh, byr, byx, bzd, bzh, bzj, caa, cab, cac, caf, cak, cao, cap, car, cav, cax, cbc, cbi, cbk, cbr, cbs, cbt, cbu, cbv, cco, ceb, cek, ces, cgc, cha, chd, chf, chk, chq, chz, cjo, cjv, ckb, cle, clu, cme, cmn, cni, cnl, cnt, cof, con, cop, cot, cpa, cpb, cpc, cpu, cpy, crn, crx, cso, csy, cta, cth, ctp, ctu, cub, cuc, cui, cuk, cut, cux, cwe, cya, daa, dad, dah, dan, ded, deu, dgc, dgr, dgz, dhg, dif, dik, dji, djk, djr, dob, dop, dov, dwr, dww, dwy, ebk, eko, emi, emp, eng, enq, epo, eri, ese, esk, etr, ewe, faa, fai, far, ffm, for, fra, fue, fuf, fuh, gah, gai, gam, gaw, gdn, gdr, geb, gfk, ghs, glk, gmv, gng, gnn, gnw, gof, grc, gub, guh, gui, guj, gul, gum, gun, guo, gup, gux, gvc, gvf, gvn, gvs, gwi, gym, gyr, hat, hau, haw, hbo, hch, heb, heg, hin, hix, hla, hlt, hmo, hns, hop, hot, hrv, hto, hub, hui, hun, hus, huu, huv, hvn, ian, ign, ikk, ikw, ilo, imo, inb, ind, ino, iou, ipi, isn, ita, iws, ixl, jac, jae, jao, jic, jid, jiv, jni, jpn, jvn, kan, kaq, kbc, kbh, kbm, kbq, kdc, kde, kdl, kek, ken, kew, kgf, kgk, kgp, khs, khz, kik, kiw, kiz, kje, kjn, kjs, kkc, kkl, klt, klv, kmg, kmh, kmk, kmo, kms, kmu, kne, knf, knj, knv, kos, kpf, kpg, kpj, kpr, kpw, kpx, kqa, kqc, kqf, kql, kqw, ksd, ksj, ksr, ktm, kto, kud, kue, kup, kvg, kvn, kwd, kwf, kwi, kwj, kyc, kyf, kyg, kyq, kyz, kze, lac, lat, lbb, lbk, lcm, leu, lex, lgl, lid, lif, lin, lit, llg, lug, luo, lww, maa, maj, mal, mam, maq, mar, mau, mav, maz, mbb, mbc, mbh, mbj, mbl, mbs, mbt, mca, mcb, mcd, mcf, mco, mcp, mcq, mcr, mdy, med, mee, mek, meq, met, meu, mgc, mgh, mgw, mhl, mib, mic, mie, mig, mih, mil, mio, mir, mit, miz, mjc, mkj, mkl, mkn, mks, mle, mlh, mlp, mmo, mmx, mna, mop, mox, mph, mpj, mpm, mpp, mps, mpt, mpx, mqb, mqj, msb, msc, msk, msm, msy, mti, mto, mux, muy, mva, mvn, mwc, mwe, mwf, mwp, mxb, mxp, mxq, mxt, mya, myk, myu, myw, myy, mzz, nab, naf, nak, nas, nay, nbq, nca, nch, ncj, ncl, ncu, ndg, ndj, nfa, ngp, ngu, nhe, nhg, nhi, nho, nhr, nhu, nhw, nhy, nif, nii, nin, nko, nld, nlg, nmw, nna, nnq, noa, nop, not, nou, npi, npl, nsn, nss, ntj, ntp, ntu, nuy, nvm, nwi, nya, nys, nyu, obo, okv, omw, ong, ons, ood, opm, ory, ote, otm, otn, otq, ots, pab, pad, pah, pan, pao, pes, pib, pio, pir, piu, pjt, pls, plu, pma, poe, poh, poi, pol, pon, por, poy, ppo, prf, pri, ptp, ptu, pwg, qub, quc, quf, quh, qul, qup, qvc, qve, qvh, qvm, qvn, qvs, qvw, qvz, qwh, qxh, qxn, qxo, rai, reg, rgu, rkb, rmc, rmy, ron, roo, rop, row, rro, ruf, rug, rus, rwo, sab, san, sbe, sbk, sbs, seh, sey, sgb, sgz, shj, shp, sim, sja, sll, smk, snc, snn, snp, snx, sny, som, soq, soy, spa, spl, spm, spp, sps, spy, sri, srm, srn, srp, srq, ssd, ssg, ssx, stp, sua, sue, sus, suz, swe, swh, swp, sxb, tac, taj, tam, tav, taw, tbc, tbf, tbg, tbl, tbo, tbz, tca, tcs, tcz, tdt, tee, tel, ter, tet, tew, tfr, tgk, tgl, tgo, tgp, tha, thd, tif, tim, tiw, tiy, tke, tku, tlf, tmd, tna, tnc, tnk, tnn, tnp, toc, tod, tof, toj, ton, too, top, tos, tpa, tpi, tpt, tpz, trc, tsw, ttc, tte, tuc, tue, tuf, tuo, tur, tvk, twi, txq, txu, tzj, tzo, ubr, ubu, udu, uig, ukr, uli, ulk, upv, ura, urb, urd, uri, urt, urw, usa, usp, uvh, uvl, vid, vie, viv, vmy, waj, wal, wap, wat, wbi, wbp, wed, wer, wim, wiu, wiv, wmt, wmw, wnc, wnu, wol, wos, wrk, wro, wrs, wsk, wuv, xav, xbi, xed, xla, xnn, xon, xsi, xtd, xtm, yaa, yad, yal, yap, yaq, yby, ycn, yka, yle, yml, yon, yor, yrb, yre, yss, yuj, yut, yuw, yva, zaa, zab, zac, zad, zai, zaj, zam, zao, zap, zar, zas, zat, zav, zaw, zca, zga, zia, ziw, zlm, zos, zpc, zpl, zpm, zpo, zpq, zpu, zpv, zpz, zsr, ztq, zty, zyp, be, br, cs, ch, zh, de, en, eo, fr, ht, he, hr, id, it, ja, la, nl, ru, sa, so, es, sr, sv, to, uk, vi","multi","1|M|<|n|<|1|0|M","None","False","False","False","cc-by-4.0","None","None","None" +"Helsinki-NLP/multiun","Helsinki-NLP","2022-03-02 23:29:22","2024-02-27 16:59:52+00:00","489448d66a808cfc421b9f82affc1efbef9d5299","2604","91255","9","None","translation","Dataset Card for OPUS MultiUN Dataset Summary The MultiUN parallel corpus is extracted from the United Nations Website , and then cleaned and converted to XML at Language Technology Lab in DFKI GmbH (","None","ar, de, en, es, fr, ru, zh","multi","1|0|0|K|<|n|<|1|M","None","False","False","False","unknown","None","None","None" +"allenai/tulu-v2-sft-mixture","allenai","2023-11-13 21:56:34","2024-05-24 21:29:24+00:00","6248b175d2ccb5ec7c4aeb22e6d8ee3b21b2c752","1221","88218","121","task_categories:question-answering, task_categories:text-generation, language:en, license:odc-by, size_categories:100K|1|T","None","False","False","False","None","None","None","None" +"mteb/stackoverflowdupquestions-reranking","mteb","2022-04-19 12:24:06","2022-09-27 19:13:01+00:00","ef807ea29a75ec4f91b50fd4191cb4ee4589a9f9","2364","81633","3","language:en, size_categories:10K|1|T","None","False","auto","False","apache-2.0","None","None","None" +"mlfoundations/MINT-1T-PDF-CC-2023-50","mlfoundations","2024-07-12 05:42:22","2024-09-19 21:06:23+00:00","b9a0d67f6048cf79615e63fa44d7ac729958fc71","2468","76558","3","None","image-to-text, text-generation","๐Ÿƒ MINT-1T:Scaling Open-Source Multimodal Data by 10x: A Multimodal Dataset with One Trillion Tokens ๐Ÿƒ MINT-1T is an open-source Multimodal INTerleaved dataset with 1 trillion text tokens and 3.4 billi","None","en","mono","1|0|0|B|<|n|<|1|T","None","False","False","False","cc-by-4.0","2406.11271","None","None" +"mteb/reddit-clustering","mteb","2022-04-07 09:12:22","2022-09-27 19:13:31+00:00","b2805658ae38990172679479369a78b86de8c390","1172","75128","0","None","None","None","None","en","mono","None","None","False","False","False","None","None","None","None" +"mteb/imdb","mteb","2022-05-26 08:50:50","2022-09-27 19:14:44+00:00","8d743909f834c38949e8323a8a6ce8721ea6c7f4","14109","74014","1","language:en, size_categories:10K>> from datasets import load_dataset >>> dat","None","aai, aak, aau, aaz, abt, abx, aby, acf, acr, acu, adz, aer, aey, agd, agg, agm, agn, agr, agt, agu, aia, aii, aka, ake, alp, alq, als, aly, ame, amf, amk, amm, amn, amo, amp, amr, amu, amx, anh, anv, aoi, aoj, aom, aon, apb, ape, apn, apr, apu, apw, apz, arb, are, arl, arn, arp, asm, aso, ata, atb, atd, atg, att, auc, aui, auy, avt, awb, awk, awx, azb, azg, azz, bao, bba, bbb, bbr, bch, bco, bdd, bea, bef, bel, ben, beo, beu, bgs, bgt, bhg, bhl, big, bjk, bjp, bjr, bjv, bjz, bkd, bki, bkq, bkx, bla, blw, blz, bmh, bmk, bmr, bmu, bnp, boa, boj, bon, box, bpr, bps, bqc, bqp, bre, bsj, bsn, bsp, bss, buk, bus, bvd, bvr, bxh, byr, byx, bzd, bzh, bzj, caa, cab, cac, caf, cak, cao, cap, car, cav, cax, cbc, cbi, cbk, cbr, cbs, cbt, cbu, cbv, cco, ceb, cek, ces, cgc, cha, chd, chf, chk, chq, chz, cjo, cjv, ckb, cle, clu, cme, cmn, cni, cnl, cnt, cof, con, cop, cot, cpa, cpb, cpc, cpu, cpy, crn, crx, cso, csy, cta, cth, ctp, ctu, cub, cuc, cui, cuk, cut, cux, cwe, cya, daa, dad, dah, dan, ded, deu, dgc, dgr, dgz, dhg, dif, dik, dji, djk, djr, dob, dop, dov, dwr, dww, dwy, ebk, eko, emi, emp, eng, enq, epo, eri, ese, esk, etr, ewe, faa, fai, far, ffm, for, fra, fue, fuf, fuh, gah, gai, gam, gaw, gdn, gdr, geb, gfk, ghs, glk, gmv, gng, gnn, gnw, gof, grc, gub, guh, gui, guj, gul, gum, gun, guo, gup, gux, gvc, gvf, gvn, gvs, gwi, gym, gyr, hat, hau, haw, hbo, hch, heb, heg, hin, hix, hla, hlt, hmo, hns, hop, hot, hrv, hto, hub, hui, hun, hus, huu, huv, hvn, ian, ign, ikk, ikw, ilo, imo, inb, ind, ino, iou, ipi, isn, ita, iws, ixl, jac, jae, jao, jic, jid, jiv, jni, jpn, jvn, kan, kaq, kbc, kbh, kbm, kbq, kdc, kde, kdl, kek, ken, kew, kgf, kgk, kgp, khs, khz, kik, kiw, kiz, kje, kjn, kjs, kkc, kkl, klt, klv, kmg, kmh, kmk, kmo, kms, kmu, kne, knf, knj, knv, kos, kpf, kpg, kpj, kpr, kpw, kpx, kqa, kqc, kqf, kql, kqw, ksd, ksj, ksr, ktm, kto, kud, kue, kup, kvg, kvn, kwd, kwf, kwi, kwj, kyc, kyf, kyg, kyq, kyz, kze, lac, lat, lbb, lbk, lcm, leu, lex, lgl, lid, lif, lin, lit, llg, lug, luo, lww, maa, maj, mal, mam, maq, mar, mau, mav, maz, mbb, mbc, mbh, mbj, mbl, mbs, mbt, mca, mcb, mcd, mcf, mco, mcp, mcq, mcr, mdy, med, mee, mek, meq, met, meu, mgc, mgh, mgw, mhl, mib, mic, mie, mig, mih, mil, mio, mir, mit, miz, mjc, mkj, mkl, mkn, mks, mle, mlh, mlp, mmo, mmx, mna, mop, mox, mph, mpj, mpm, mpp, mps, mpt, mpx, mqb, mqj, msb, msc, msk, msm, msy, mti, mto, mux, muy, mva, mvn, mwc, mwe, mwf, mwp, mxb, mxp, mxq, mxt, mya, myk, myu, myw, myy, mzz, nab, naf, nak, nas, nay, nbq, nca, nch, ncj, ncl, ncu, ndg, ndj, nfa, ngp, ngu, nhe, nhg, nhi, nho, nhr, nhu, nhw, nhy, nif, nii, nin, nko, nld, nlg, nmw, nna, nnq, noa, nop, not, nou, npi, npl, nsn, nss, ntj, ntp, ntu, nuy, nvm, nwi, nya, nys, nyu, obo, okv, omw, ong, ons, ood, opm, ory, ote, otm, otn, otq, ots, pab, pad, pah, pan, pao, pes, pib, pio, pir, piu, pjt, pls, plu, pma, poe, poh, poi, pol, pon, por, poy, ppo, prf, pri, ptp, ptu, pwg, qub, quc, quf, quh, qul, qup, qvc, qve, qvh, qvm, qvn, qvs, qvw, qvz, qwh, qxh, qxn, qxo, rai, reg, rgu, rkb, rmc, rmy, ron, roo, rop, row, rro, ruf, rug, rus, rwo, sab, san, sbe, sbk, sbs, seh, sey, sgb, sgz, shj, shp, sim, sja, sll, smk, snc, snn, snp, snx, sny, som, soq, soy, spa, spl, spm, spp, sps, spy, sri, srm, srn, srp, srq, ssd, ssg, ssx, stp, sua, sue, sus, suz, swe, swh, swp, sxb, tac, taj, tam, tav, taw, tbc, tbf, tbg, tbl, tbo, tbz, tca, tcs, tcz, tdt, tee, tel, ter, tet, tew, tfr, tgk, tgl, tgo, tgp, tha, thd, tif, tim, tiw, tiy, tke, tku, tlf, tmd, tna, tnc, tnk, tnn, tnp, toc, tod, tof, toj, ton, too, top, tos, tpa, tpi, tpt, tpz, trc, tsw, ttc, tte, tuc, tue, tuf, tuo, tur, tvk, twi, txq, txu, tzj, tzo, ubr, ubu, udu, uig, ukr, uli, ulk, upv, ura, urb, urd, uri, urt, urw, usa, usp, uvh, uvl, vid, vie, viv, vmy, waj, wal, wap, wat, wbi, wbp, wed, wer, wim, wiu, wiv, wmt, wmw, wnc, wnu, wol, wos, wrk, wro, wrs, wsk, wuv, xav, xbi, xed, xla, xnn, xon, xsi, xtd, xtm, yaa, yad, yal, yap, yaq, yby, ycn, yka, yle, yml, yon, yor, yrb, yre, yss, yuj, yut, yuw, yva, zaa, zab, zac, zad, zai, zaj, zam, zao, zap, zar, zas, zat, zav, zaw, zca, zga, zia, ziw, zlm, zos, zpc, zpl, zpm, zpo, zpq, zpu, zpv, zpz, zsr, ztq, zty, zyp, be, br, cs, ch, zh, de, en, eo, fr, ht, he, hr, id, it, ja, la, nl, ru, sa, so, es, sr, sv, to, uk, vi","multi","1|M|<|n|<|1|0|M","None","False","False","False","cc-by-4.0","None","None","None" +"tyqiangz/multilingual-sentiments","tyqiangz","2022-08-21 11:04:38","2023-05-23 15:01:51+00:00","a3080a58e563138e9c7a61765d8120b388dc572d","2928","53508","45","None","text-classification, sentiment-analysis, sentiment-classification","Multilingual Sentiments Dataset A collection of multilingual sentiments datasets grouped into 3 classes -- positive, neutral, negative. Most multilingual sentiment datasets are either 2-class positive","None","de, en, es, fr, ja, zh, id, ar, hi, it, ms, pt","multi","1|0|0|K|<|n|<|1|M|||1|M|<|n|<|1|0|M","None","False","False","False","apache-2.0","None","None","sentiment-analysis, sentiment-classification" +"mutiyama/alt","mutiyama","2022-03-02 23:29:22","2024-01-09 12:07:24+00:00","afbd92e198bbcf17f660e03076fd2938f5a4bbb2","896","53209","17","None","translation, token-classification, parsing","Dataset Card for Asian Language Treebank (ALT) Dataset Summary The ALT project aims to advance the state-of-the-art Asian natural language processing (NLP) techniques through the open collaboration fo","None","bn, en, fil, hi, id, ja, km, lo, ms, my, th, vi, zh","multi","1|0|0|K|<|n|<|1|M|||1|0|K|<|n|<|1|0|0|K","None","False","False","False","cc-by-4.0","None","None","parsing" +"HuggingFaceGECLM/REDDIT_comments","HuggingFaceGECLM","2023-03-15 14:14:58","2023-03-17 07:52:51+00:00","54779d3d1f1c1b12e5989f695e13d38b394a558f","22104","53201","11","task_categories:text-generation, task_ids:dialogue-modeling, task_ids:language-modeling, annotations_creators:no-annotation, language_creators:found, multilinguality:monolingual, language:en, size_categories:100M|1|T","None","False","False","False","odc-by","2405.16712","None","None" +"GEM/wiki_cat_sum","GEM","2022-03-02 23:29:22","2022-10-24 15:31:11+00:00","e732d1703eaad5b34a56370fd137b9d09921a94b","2934","43974","4","None","summarization","Summarise the most important facts of a given entity in the Film, Company, and Animal domains from a cluster of related documents.","@inproceedings{perez2019generating, title={Generating Summaries with Topic Templates and Structured Convolutional Decoders}, author={Perez-Beltrachini, Laura and Liu, Yang and Lapata, Mirella}, booktitle={Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics}, pages={5107--5116}, year={2019} }","en","mono","u|n|k|n|o|w|n","None","False","False","False","cc-by-sa-3.0","1906.04687","None","None" +"sentence-transformers/all-nli","sentence-transformers","2024-04-25 12:49:03","2024-05-15 11:22:30+00:00","d482672c8e74ce18da116f430137434ba2e52fab","2615","43814","31","None","feature-extraction, sentence-similarity","Dataset Card for AllNLI This dataset is a concatenation of the SNLI and MultiNLI datasets. Despite originally being intended for Natural Language Inference (NLI), this dataset can be used for training","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","None","None","None","None" +"openbmb/UltraFeedback","openbmb","2023-09-23 15:41:04","2023-12-29 14:11:19+00:00","40b436560ca83a8dba36114c22ab3c66e43f6d5e","2143","43485","347","None","text-generation","Introduction GitHub Repo UltraRM-13b UltraCM-13b UltraFeedback is a large-scale, fine-grained, diverse preference dataset, used for training powerful reward models and critic models. We collect about ","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","mit","None","None","None" +"mteb/scidocs","mteb","2024-03-02 20:24:54","2024-03-03 11:42:33+00:00","f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88","1382","42973","3","None","text-retrieval, document-retrieval","None","None","en","mono","None","None","False","False","False","None","None","None","document-retrieval" +"lesserfield/4chan-datasets","lesserfield","2023-04-16 14:39:15","2023-05-07 16:38:40+00:00","cfb1ebc8bec1f9326f69c0bb563602836796e5d8","940","42953","21","None","text-generation","Please see repo to turn the text file into json/csv format Deleted some boards, since they are already archived by https://archive.4plebs.org/","None","en","mono","None","None","False","False","False","unlicense","None","None","None" +"edinburgh-dawg/mmlu-redux","edinburgh-dawg","2024-06-04 17:10:52","2024-08-09 18:31:31+00:00","2ebba46760ad1df575a6b9df42c280b1d5fb3c10","1542","42592","28","None","question-answering","Dataset Card for MMLU-Redux MMLU-Redux is a subset of 3,000 manually re-annotated questions across 30 MMLU subjects. Dataset Details Dataset Description Each data point in MMLU-Redux contains seven co","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-4.0","2406.04127","None","None" +"sentence-transformers/parallel-sentences-talks","sentence-transformers","2024-04-30 10:29:15","2024-06-18 19:45:50+00:00","0c70bc6714efb1df12f8a16b9056e4653563d128","3613","42438","9","None","feature-extraction, sentence-similarity","Dataset Card for Parallel Sentences - Talks This dataset contains parallel sentences (i.e. English sentence + the same sentences in another language) for numerous other languages. Most of the sentence","None","en, multilingual, ar, bg, ca, cs, da, de, el, es, et, fa, fi, fr, gl, gu, he, hi, hr, hu, hy, id, it, ja, ka, ko, ku, lt, lv, mk, mn, mr, ms, my, nb, nl, pl, pt, ro, ru, sk, sl, sq, sr, sv, th, tr, uk, ur, vi, zh","multi","1|M|<|n|<|1|0|M","None","False","False","False","None","None","None","None" +"Lin-Chen/MMStar","Lin-Chen","2024-04-02 06:56:56","2024-04-07 08:15:45+00:00","bc98d668301da7b14f648724866e57302778ab27","5133","42295","26","None","multiple-choice, question-answering, visual-question-answering","MMStar (Are We on the Right Way for Evaluating Large Vision-Language Models?) ๐ŸŒ Homepage | ๐Ÿค— Dataset | ๐Ÿค— Paper | ๐Ÿ“– arXiv | GitHub Dataset Details As shown in the figure below, existing benchmarks lack","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","None","None","None","None" +"OpenGVLab/GUI-Odyssey","OpenGVLab","2024-06-13 07:21:10","2024-11-20 12:34:13+00:00","2298cb628895d1c6248b8ead10c71429a76ce943","20196","41942","10","None","None","Dataset Card for GUI Odyssey Repository: https://github.com/OpenGVLab/GUI-Odyssey Paper: https://arxiv.org/abs/2406.08451 Point of Contact: Wenqi Shao Introduction GUI Odyssey is a comprehensive datas","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-4.0","2406.08451","None","None" +"Cohere/wikipedia-22-12-en-embeddings","Cohere","2023-01-14 20:36:11","2023-03-22 16:51:57+00:00","85c2eca83d4b9dcecc043c23748cb8c1047f683f","1967","41790","64","None","text-retrieval, document-retrieval","Wikipedia (en) embedded with cohere.ai multilingual-22-12 encoder We encoded Wikipedia (en) using the cohere.ai multilingual-22-12 embedding model. To get an overview how this dataset was created and ","None","en","mono","None","None","False","False","False","apache-2.0","None","None","document-retrieval" +"RMT-team/babilong","RMT-team","2024-03-07 14:15:14","2024-06-17 09:49:52+00:00","ee0d588794c7ac098062ee0d247c733d62e94fe2","11959","41497","14","None","None","BABILong (100 samples) : a long-context needle-in-a-haystack benchmark for LLMs Preprint is on arXiv and code for LLM evaluation is available on GitHub. BABILong Leaderboard with top-performing long-c","None","en","mono","None","None","False","False","False","None","2406.10149","None","None" +"HuggingFaceFV/finevideo","HuggingFaceFV","2024-09-09 17:56:30","2024-12-16 08:04:42+00:00","012c690b405b22b99611e4b55bb065415d35a237","2692","41440","291","None","visual-question-answering, video-text-to-text","FineVideo FineVideo Description Dataset Explorer Revisions Dataset Distribution How to download and use FineVideo Using datasets Using huggingface_hub Load a subset of the dataset Dataset StructureDat","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","auto","False","cc","None","None","None" +"bentrevett/multi30k","bentrevett","2023-03-19 22:38:35","2023-03-24 14:50:27+00:00","4589883f3d09d4ef6361784e03f0ead219836469","1148","40495","6","None","translation","Multi30k This dataset contains the ""multi30k"" dataset, which is the ""task 1"" dataset from here. Each example consists of an ""en"" and a ""de"" feature. ""en"" is an English sentence, and ""de"" is the German","None","en, de","bi","1|0|K|<|n|<|1|0|0|K","None","False","False","False","None","None","None","None" +"BAAI/Infinity-Instruct","BAAI","2024-06-13 12:17:03","2025-01-16 08:47:04+00:00","40353d346e04c94ed0de467f8b6c95061d1e7b89","5298","39161","585","task_categories:text-generation, language:en, language:zh, license:cc-by-sa-4.0, size_categories:10M|1|T","None","False","False","False","apache-2.0","2312.07488","None","None" +"McGill-NLP/stereoset","McGill-NLP","2022-03-02 23:29:22","2024-01-23 08:34:39+00:00","bf6e7ce50491784d094fb7afe60a70ecccb89035","968","37890","16","None","text-classification","Dataset Card for StereoSet Dataset Summary StereoSet is a dataset that measures stereotype bias in language models. StereoSet consists of 17,000 sentences that measures model preferences across gender","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-sa-4.0","2004.09456","None","None" +"mteb/msmarco","mteb","2024-03-02 20:25:33","2024-03-03 11:13:39+00:00","c5a29a104738b98a9e76336939199e264163d4a0","1841","37889","1","None","text-retrieval, document-retrieval","None","None","en","mono","None","None","False","False","False","None","None","None","document-retrieval" +"Babelscape/SREDFM","Babelscape","2023-06-13 18:35:19","2023-06-20 07:33:28+00:00","2732d2834e12e36510aeb2a468163ea2642d55db","2928","37479","14","None","token-classification","Relation Extraction (RE) is a task that identifies relationships between entities in a text, enabling the acquisition of relational facts and bridging the gap between natural language and structured k","@InProceedings{REDFM2023, author = {Huguet Cabot, Pere-Lluis and Tedeschi, Simone and Ngonga Ngomo, Axel-Cyrille and Navigli, Roberto}, title = {RED\\textsuperscript{FM}: a Filtered and Multilingual Relation Extraction Dataset}, booktitle = {Proceedings of the 2023 Conference on Association for Computational Linguistics}, year = {2023}, publisher = {Association for Computational Linguistics}, location = {Toronto, Canada}, }","ar, ca, de, el, en, es, fr, hi, it, ja, ko, nl, pl, pt, ru, sv, vi, zh","multi","1|0|M|<|n|<|1|0|0|M","None","False","False","False","cc-by-sa-4.0","2306.09802","None","None" +"microsoft/orca-math-word-problems-200k","microsoft","2024-03-01 00:56:17","2024-03-04 18:01:08+00:00","29255d1770cc4eac66e5e7fa378cba542c026350","1145","37397","428","None","question-answering","Dataset Card This dataset contains ~200K grade school math word problems. All the answers in this dataset is generated using Azure GPT4-Turbo. Please refer to Orca-Math: Unlocking the potential of SLM","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","mit","None","None","None" +"qmeeus/vp-er-10l","qmeeus","2024-02-19 20:45:14","2024-03-28 14:43:22+00:00","2c70f24900c5f93815a3ab16134dbe2e39126771","8125","36814","0","language:cs, language:de, language:en, language:es, language:fr, language:hu, language:it, language:nl, language:pl, language:ro, size_categories:100K|1|T","None","False","False","False","cc-by-nc-sa-4.0","2308.12477v1","None","language-modeling" +"tinyBenchmarks/tinyHellaswag","tinyBenchmarks","2024-02-22 11:33:39","2024-05-25 10:44:12+00:00","5bb903986ca293b54d364c8c0423c768e20fb801","1637","28064","4","None","None","tinyHellaswag Welcome to tinyHellaswag! This dataset serves as a concise version of the hellaswag dataset, offering a subset of 100 data points selected from the original compilation. tinyHellaswag is","None","en","mono","n|<|1|K","None","False","False","False","None","None","None","None" +"proj-persona/PersonaHub","proj-persona","2024-06-28 16:35:21","2024-10-05 04:04:28+00:00","c91f99f3efd4d0977e338f3b77abd251653cd405","4205","27580","499","None","text-generation, text-classification, token-classification, fill-mask, table-question-answering, text2text-generation","Scaling Synthetic Data Creation with 1,000,000,000 Personas This repo releases data introduced in our paper Scaling Synthetic Data Creation with 1,000,000,000 Personas: We propose a novel persona-driv","None","en, zh","bi","1|0|0|K|<|n|<|1|M","None","False","False","False","cc-by-nc-sa-4.0","None","None","None" +"mteb/sib200","mteb","2024-05-07 14:07:00","2024-05-07 14:59:53+00:00","a74d7350ea12af010cfb1c21e34f1f81fd2e615b","1212","27530","1","task_categories:text-classification, task_ids:topic-classification, annotations_creators:found, language_creators:expert-generated, multilinguality:multilingual, source_datasets:original, language:ace, language:acm, language:acq, language:aeb, language:af, language:ajp, language:ak, language:als, language:am, language:apc, language:ar, language:ars, language:ary, language:arz, language:as, language:ast, language:awa, language:ayr, language:azb, language:azj, language:ba, language:bm, language:ban, language:be, language:bem, language:bn, language:bho, language:bjn, language:bo, language:bs, language:bug, language:bg, language:ca, language:ceb, language:cs, language:cjk, language:ckb, language:crh, language:cy, language:da, language:de, language:dik, language:dyu, language:dz, language:el, language:en, language:eo, language:et, language:eu, language:ee, language:fo, language:fj, language:fi, language:fon, language:fr, language:fur, language:fuv, language:gaz, language:gd, language:ga, language:gl, language:gn, language:gu, language:ht, language:ha, language:he, language:hi, language:hne, language:hr, language:hu, language:hy, language:ig, language:ilo, language:id, language:is, language:it, language:jv, language:ja, language:kab, language:kac, language:kam, language:kn, language:ks, language:ka, language:kk, language:kbp, language:kea, language:khk, language:km, language:ki, language:rw, language:ky, language:kmb, language:kmr, language:knc, language:kg, language:ko, language:lo, language:lij, language:li, language:ln, language:lt, language:lmo, language:ltg, language:lb, language:lua, language:lg, language:luo, language:lus, language:lvs, language:mag, language:mai, language:ml, language:mar, language:min, language:mk, language:mt, language:mni, language:mos, language:mi, language:my, language:nl, language:nn, language:nb, language:npi, language:nqo, language:nso, language:nus, language:ny, language:oc, language:ory, language:pag, language:pa, language:pap, language:pbt, language:pes, language:plt, language:pl, language:pt, language:prs, language:quy, language:ro, language:rn, language:ru, language:sg, language:sa, language:sat, language:scn, language:shn, language:si, language:sk, language:sl, language:sm, language:sn, language:sd, language:so, language:st, language:es, language:sc, language:sr, language:ss, language:su, language:sv, language:swh, language:szl, language:ta, language:taq, language:tt, language:te, language:tg, language:tl, language:th, language:ti, language:tpi, language:tn, language:ts, language:tk, language:tum, language:tr, language:tw, language:tzm, language:ug, language:uk, language:umb, language:ur, language:uzn, language:vec, language:vi, language:war, language:wo, language:xh, language:ydd, language:yo, language:yue, language:zh, language:zsm, language:zu, license:cc-by-sa-4.0, size_categories:100K 4k tokens). Copied from ""Long Document Classificatio","None","e, n","bi","1|||0|||K|||<|||n|||<|||1|||0|||0|||K","None","False","False","False","None","None","None","multi-class-classification, topic-classification" +"maharshipandya/spotify-tracks-dataset","maharshipandya","2023-06-14 11:42:44","2023-12-01 13:29:39+00:00","635b034f69257814eff850a5c2b3346fe458134f","981","23585","85","None","feature-extraction, tabular-classification, tabular-regression","Content This is a dataset of Spotify tracks over a range of 125 different genres. Each track has some audio features associated with it. The data is in CSV format which is tabular and can be loaded qu","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","bsd","None","None","None" +"GEM/viggo","GEM","2022-03-02 23:29:22","2022-10-24 15:31:07+00:00","c851cd5ff2ee92f0137fcf24014e37427a2d30b7","786","23502","33","None","table-to-text","ViGGO was designed for the task of data-to-text generation in chatbots (as opposed to task-oriented dialogue systems), with target responses being more conversational than information-seeking, yet con","@inproceedings{juraska-etal-2019-viggo, title = ""{V}i{GGO}: A Video Game Corpus for Data-To-Text Generation in Open-Domain Conversation"", author = ""Juraska, Juraj and Bowden, Kevin and Walker, Marilyn"", booktitle = ""Proceedings of the 12th International Conference on Natural Language Generation"", month = oct # ""{--}"" # nov, year = ""2019"", address = ""Tokyo, Japan"", publisher = ""Association for Computational Linguistics"", url = ""https://aclanthology.org/W19-8623"", doi = ""10.18653/v1/W19-8623"", pages = ""164--172"", }","en","mono","u|n|k|n|o|w|n","None","False","False","False","cc-by-sa-4.0","None","None","None" +"mlfoundations/MINT-1T-ArXiv","mlfoundations","2024-06-29 23:50:55","2024-09-19 21:32:59+00:00","7c5b00ffd5b563071010c3bf2082b4a8f836eb72","2347","23486","49","None","image-to-text, text-generation","๐Ÿƒ MINT-1T:Scaling Open-Source Multimodal Data by 10x: A Multimodal Dataset with One Trillion Tokens ๐Ÿƒ MINT-1T is an open-source Multimodal INTerleaved dataset with 1 trillion text tokens and 3.4 billi","None","en","mono","1|0|0|B|<|n|<|1|T","None","False","False","False","cc-by-4.0","2406.11271","None","None" +"intfloat/multilingual_cc_news","intfloat","2023-03-22 08:25:34","2023-04-23 08:19:06+00:00","1c24f619ae44d0a4c87fde72efa741dc25f7530e","2171","23387","17","None","None","\\ Multilingual CC-News dataset. This is the processed version from https://huggingface.co/datasets/CloverSearch/cc-news-mutlilingual.","None","en, zh, fr, de, af, ar","multi","1|0|0|M|<|n|<|1|B","None","False","False","False","None","None","None","None" +"JailbreakBench/JBB-Behaviors","JailbreakBench","2024-06-12 12:57:23","2024-09-26 11:05:44+00:00","886acc352a31533ffbcf4ef22c744658688086fc","2829","23367","36","language:en, license:mit, size_categories:n<1K, format:csv, modality:tabular, modality:text, library:datasets, library:pandas, library:mlcroissant, library:polars, arxiv:2404.01318, arxiv:2311.03348, arxiv:2307.15043, arxiv:2402.04249, doi:10.57967/hf/2540, region:us, jailbreaks, large language models, harmful behaviors, ml safety","None","An Open Robustness Benchmark for Jailbreaking Language Models NeurIPS 2024 Datasets and Benchmarks Track Paper | Leaderboard | Benchmark code What is JailbreakBench? Jailbreakbench is an open-source r","None","en","mono","None","None","False","False","False","mit","2404.01318","None","None" +"Salesforce/blip3-kale","Salesforce","2024-08-27 20:53:04","2024-11-14 23:39:47+00:00","adbc857b863005dbed15596d49410ebcb0392922","9257","22543","34","task_categories:image-to-text, language:en, license:apache-2.0, size_categories:100M|1|T","None","False","auto","False","None","None","None","None" +"mxeval/multi-humaneval","mxeval","2023-03-14 21:37:18","2023-03-20 19:20:48+00:00","63e25957849ef5676e5b686ed0a60106697a73c0","1021","20962","11","None","text-generation","A collection of execution-based multi-lingual benchmark for code generation.","@article{mbxp_athiwaratkun2022, title = {Multi-lingual Evaluation of Code Generation Models}, author = {Athiwaratkun, Ben and Gouda, Sanjay Krishna and Wang, Zijian and Li, Xiaopeng and Tian, Yuchen and Tan, Ming and Ahmad, Wasi Uddin and Wang, Shiqi and Sun, Qing and Shang, Mingyue and Gonugondla, Sujan Kumar and Ding, Hantian and Kumar, Varun and Fulton, Nathan and Farahani, Arash and Jain, Siddhartha and Giaquinto, Robert and Qian, Haifeng and Ramanathan, Murali Krishna and Nallapati, Ramesh and Ray, Baishakhi and Bhatia, Parminder and Sengupta, Sudipta and Roth, Dan and Xiang, Bing}, doi = {10.48550/ARXIV.2210.14868}, url = {https://arxiv.org/abs/2210.14868}, keywords = {Machine Learning (cs.LG), Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} }","en","mono","None","None","False","False","False","apache-2.0","2210.14868","None","None" +"0x22almostEvil/multilingual-wikihow-qa-16k","0x22almostEvil","2023-04-29 03:37:09","2023-05-13 16:59:15+00:00","3211e4c50a3c64090a548bf5fc10c3caaeb26eff","965","20649","9","None","question-answering","Dataset Card for multilingual WikiHow with ~16.8K entries. ~(2-2.2)K for each language. Warning [1] The WikiHow team contacted me and made it clear that they forbid the use of their data for machine l","None","en, ru, pt, it, es, fr, de, nl","multi","1|0|K|<|n|<|1|0|0|K","None","False","False","False","cc-by-nc-3.0","None","None","None" +"free-law/Caselaw_Access_Project","free-law","2024-02-28 15:06:54","2024-03-16 20:01:40+00:00","cee53e98cb2b401865a017ef291a59fbb3407179","8908","20531","63","None","text-generation","The Caselaw Access Project In collaboration with Ravel Law, Harvard Law Library digitized over 40 million U.S. court decisions consisting of 6.7 million cases from the last 360 years into a dataset th","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","cc0-1.0","None","None","None" +"mteb/cqadupstack-physics","mteb","2024-03-02 19:36:35","2024-03-02 19:56:34+00:00","79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4","954","20386","1","None","text-retrieval, document-retrieval","None","None","en","mono","None","None","False","False","False","None","None","None","document-retrieval" +"common-canvas/commoncatalog-cc-by-nd","common-canvas","2023-10-19 02:10:04","2024-05-16 19:42:40+00:00","3991ff88ebf48e0435ec8d044d2f4b159f4f716e","2174","20377","2","None","text-to-image","Dataset Card for CommonCatalog CC-BY-ND This dataset is a large collection of high-resolution Creative Common images (composed of different licenses, see paper Table 1 in the Appendix) collected in 20","None","en","mono","None","None","False","False","False","cc-by-nd-4.0","2310.16825","None","None" +"rayliuca/WikidataLabels","rayliuca","2024-01-01 00:23:08","2024-01-11 04:17:57+00:00","122b626f6127ca111659a4354590845b30736f9a","5752","20169","1","task_categories:translation, task_categories:text2text-generation, language:en, language:fr, language:de, language:ja, language:zh, language:hi, language:ar, language:bn, language:ru, language:es, license:cc0-1.0, size_categories:100M|1|T","None","False","False","False","apache-2.0","None","None","None" +"llamafactory/tiny-supervised-dataset","llamafactory","2024-06-07 19:25:33","2024-06-10 07:41:37+00:00","2ff06c75e01ae4195ed34fe77606e15902ea0b0d","4269","19543","1","None","text-generation, question-answering","None","None","en, zh","bi","n|<|1|K","None","False","False","False","apache-2.0","None","None","None" +"imageomics/TreeOfLife-10M","imageomics","2024-01-23 21:06:53","2024-12-20 16:05:17+00:00","343c36b9b362494065ac427ac1da989b834b22cf","15790","19524","25","task_categories:image-classification, task_categories:zero-shot-classification, language:en, language:la, size_categories:1M|1|T","None","False","False","False","apache-2.0","None","None","None" +"ai4privacy/pii-masking-200k","ai4privacy","2023-11-06 03:34:07","2024-04-21 17:27:20+00:00","a157bb3aa654864f870dd728fd6d06cc0454fabc","1751","19274","91","task_categories:text-classification, task_categories:token-classification, task_categories:table-question-answering, task_categories:question-answering, task_categories:zero-shot-classification, task_categories:summarization, task_categories:feature-extraction, task_categories:text-generation, task_categories:text2text-generation, task_categories:translation, task_categories:fill-mask, task_categories:tabular-classification, task_categories:tabular-to-text, task_categories:table-to-text, task_categories:text-retrieval, task_categories:other, multilinguality:multilingual, source_datasets:original, language:en, language:fr, language:de, language:it, size_categories:100K=6.5 image dataset using CogVLM2-4bit with the ""laion-pop""-prompt to generate captions which were ""likely"" used in Stable Di","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","cc-by-nc-sa-4.0","None","None","None" +"k-mktr/improved-flux-prompts-photoreal-portrait","k-mktr","2024-09-28 08:22:13","2024-10-03 10:55:26+00:00","36cf6aac4216523e41c831517bc93ca42624cd58","1210","10625","103","None","text-classification","Photo Portrait Prompt Dataset for FLUX Overview This dataset contains a curated collection of prompts specifically designed for generating photo portraits using FLUX.1, an advanced text-to-image model","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","mit","None","None","None" +"andstor/methods2test_small","andstor","2023-12-17 20:26:53","2024-11-03 09:40:11+00:00","54eb57b50de55a15774d52192bae9d5d004b1c82","7366","10589","0","task_categories:text-generation, language:en, license:mit, size_categories:100K ChronoMagic-Pro.zip unzip ChronoMagic-Pro.zip [NeurIPS D&B 2024 Spotlight] ChronoMagic-Bench: A Benchmark for Metamorphic Evaluation of Text-to-Time-lapse Video ","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","apache-2.0","None","None","None" +"dell-research-harvard/headlines-semantic-similarity","dell-research-harvard","2023-06-07 18:55:18","2024-06-07 20:52:17+00:00","bbfbc8798150e65da7145653d8e73d33d5ca0a5d","1039","8190","11","None","sentence-similarity","Dataset Card for HEADLINES Dataset Summary HEADLINES is a massive English-language semantic similarity dataset, containing 396,001,930 pairs of different headlines for the same newspaper article, take","None","en","mono","1|0|0|M|<|n|<|1|B","None","False","False","False","cc-by-2.0","2306.17810","None","None" +"allenai/wildjailbreak","allenai","2024-06-18 04:31:27","2024-08-08 05:39:06+00:00","5ddc12a7894f842b0619b8e1c7ee496b198af009","1174","8070","28","None","text-generation","WildJailbreak Dataset Card WildJailbreak is an open-source synthetic safety-training dataset with 262K vanilla (direct harmful requests) and adversarial (complex adversarial jailbreaks) prompt-respons","None","en","mono","1|0|0|M|<|n|<|1|B","None","False","auto","False","odc-by","None","None","None" +"Team-PIXEL/rendered-bookcorpus","Team-PIXEL","2022-05-11 14:41:02","2022-08-03 12:03:32+00:00","a17263cdc77c46cecb979e5b997bc23853065c29","1119","8014","4","annotations_creators:no-annotation, language_creators:found, multilinguality:monolingual, source_datasets:rendered|BookCorpusOpen, language:en, license:unknown, size_categories:1M|1|T","None","False","False","False","apache-2.0","None","None","None" +"audioshake/jam-alt","audioshake","2023-10-29 11:04:32","2024-12-30 15:34:37+00:00","0e839d1ee8a3b4baba3665b17be35312fdc4a172","1251","7557","11","None","automatic-speech-recognition","JamALT: A Readability-Aware Lyrics Transcription Benchmark JamALT is a revision of the JamendoLyrics dataset (80 songs in 4 languages), adapted for use as an automatic lyrics transcription (ALT) bench","None","en, fr, de, es","multi","None","None","False","False","False","None","2408.06370","None","None" +"kernelmachine/open-license-corpus","kernelmachine","2023-08-08 23:21:52","2023-08-09 03:14:36+00:00","384d5e19d19361803630ce4d382604267d3951d2","1757","7549","16","None","text-generation","PubText Welcome to the Open License Corpus (OLC), a 228B token corpus for training permissively-licensed language models. Disclaimer: OLC should not be considered a universally safe-to-use dataset. We","None","en","mono","1|0|0|B|<|n|<|1|T","None","False","False","False","apache-2.0","None","None","None" +"bigbio/biomrc","bigbio","2022-11-13 22:06:42","2022-12-22 15:43:44+00:00","09938b2d42d6e4ecd4d5282657d0bb5f8791950b","1274","7515","3","None","None","We introduce BIOMRC, a large-scale cloze-style biomedical MRC dataset. Care was taken to reduce noise, compared to the previous BIOREAD dataset of Pappas et al. (2018). Experiments show that simple he","@inproceedings{pappas-etal-2020-biomrc, title = ""{B}io{MRC}: A Dataset for Biomedical Machine Reading Comprehension"", author = ""Pappas, Dimitris and Stavropoulos, Petros and Androutsopoulos, Ion and McDonald, Ryan"", booktitle = ""Proceedings of the 19th SIGBioMed Workshop on Biomedical Language Processing"", month = jul, year = ""2020"", address = ""Online"", publisher = ""Association for Computational Linguistics"", url = ""https://www.aclweb.org/anthology/2020.bionlp-1.15"", pages = ""140--149"", }","en","mono","None","None","False","False","False","unknown","None","None","None" +"NousResearch/hermes-function-calling-v1","NousResearch","2024-08-14 01:22:36","2024-08-30 06:07:08+00:00","8f025148382537ba84cd325e1834b706e1461692","845","7459","234","None","text-generation, question-answering, feature-extraction","Hermes Function-Calling V1 This dataset is the compilation of structured output and function calling data used in the Hermes 2 Pro series of models. This repository contains a structured output datase","None","en","mono","None","None","False","False","False","apache-2.0","None","None","None" +"maxidl/FineNews-unfiltered","maxidl","2024-06-14 20:04:11","2024-06-16 19:48:41+00:00","58c6690fbaf643226bd5dae383a5de158ddf4c1a","987","7456","1","None","text-generation","FineNews WIP. Like FineWeb, but built from Common Crawl News instead of main web. For languages not listed as a split, check the data/ directory. For now, it contains the 2024-05 (May),-04 (April),-03","None","en, de, fr, pl, es, ru, it, ar, pt, tr, el, vi, ro, zh, uk, ko, hi, nl","multi","None","None","False","False","False","odc-by","None","None","None" +"TIGER-Lab/MMEB-train","TIGER-Lab","2024-10-08 04:05:01","2025-01-28 21:38:52+00:00","76dd0a440b6d4c02776830a804443fffbb2d0bfa","1063","7380","10","None","None","Massive Multimodal Embedding Benchmark The training data split used for training VLM2Vec models in the paper VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks (ICLR 2025)","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","apache-2.0","2410.05160","None","None" +"Magpie-Align/Magpie-Reasoning-V1-150K","Magpie-Align","2024-07-11 22:02:20","2025-01-27 19:59:05+00:00","a4bedadca568ba8fa50cae618ae62ca34dd1d196","1616","7361","53","None","None","Project Web: https://magpie-align.github.io/ Arxiv Technical Report: https://arxiv.org/abs/2406.08464 Codes: https://github.com/magpie-align/magpie Abstract Click Here High-quality instruction data is","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","llama3","2406.08464","None","None" +"philschmid/amazon-product-descriptions-vlm","philschmid","2024-09-28 14:26:19","2024-09-30 10:39:25+00:00","f08a021c69c51d6894cfe39206448e7785d6156b","977","7327","8","None","image-to-text","Amazon Multimodal Product dataset This is a modfied and slim verison of bprateek/amazon_product_description helpful to get started training multimodal LLMs. The description field was generated used Ge","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-nc-4.0","None","None","None" +"codesignal/sms-spam-collection","codesignal","2024-03-18 22:21:02","2024-03-18 23:14:07+00:00","93ffab6fa47849295146c709e341d169746582f3","1070","7250","1","None","None","SMS Spam Collection v.1 DESCRIPTION The SMS Spam Collection v.1 (hereafter the corpus) is a set of SMS tagged messages that have been collected for SMS Spam research. It contains one set of SMS messag","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-4.0","None","None","None" +"sci-benchmark/self-contradictory","sci-benchmark","2023-11-11 14:28:26","2024-08-05 06:09:58+00:00","5f4b7f23c089c034e75881126ce7c94cbc28ba7f","2444","7118","2","None","None","Introduction Official dataset of the ECCV24 paper, ""Dissecting Dissonance: Benchmarking Large Multimodal Models Against Self-Contradictory Instructions"". Website: https://selfcontradiction.github.io G","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","mit","None","None","None" +"apple/DataCompDR-12M","apple","2024-06-03 20:46:43","2024-07-22 22:42:50+00:00","bd23bbc361cc4b2ee0d2bd0431e107b2906b312f","1475","7042","27","None","text-to-image, image-to-text","Dataset Card for DataCompDR-12M This dataset contains synthetic captions, embeddings, and metadata for DataCompDR-12M. The metadata has been generated using pretrained image-text models on a 12M subse","None","en","mono","None","None","False","False","False","other","2311.17049","None","None" +"osunlp/AttributionBench","osunlp","2023-09-13 21:51:47","2024-02-26 22:00:35+00:00","62569e644f4186606f54f742178a4517431b42e1","1266","7029","3","None","text-classification","Dataset Card for AttributionBench Github repository: [Github] Paper: AttributionBench: How Hard is Automatic Attribution Evaluation? Point of Contact: Yifei Li Dataset Overview We constructed this dat","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","apache-2.0","2402.15089","None","None" +"AlexBlck/ANAKIN","AlexBlck","2023-08-17 12:33:16","2023-09-21 10:37:04+00:00","978fe26e78ecaaf6562aa249be43c2c690484d0a","1740","7010","1","None","video-classification, visual-question-answering","ANAKIN is a dataset of mANipulated videos and mAsK annotatIoNs.","@misc{black2023vader, title={VADER: Video Alignment Differencing and Retrieval}, author={Alexander Black and Simon Jenni and Tu Bui and Md. Mehrab Tanjim and Stefano Petrangeli and Ritwik Sinha and Viswanathan Swaminathan and John Collomosse}, year={2023}, eprint={2303.13193}, archivePrefix={arXiv}, primaryClass={cs.CV} }","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-4.0","2303.13193","None","None" +"Malikeh1375/medical-question-answering-datasets","Malikeh1375","2023-10-27 16:21:07","2023-11-02 03:13:38+00:00","4d69c9f6874bc5ec8be3817e9d86b7333a8a5070","1270","6951","33","None","question-answering","None","None","en","mono","None","None","False","False","False","None","None","None","None" +"parler-tts/mls-eng-speaker-descriptions","parler-tts","2024-06-13 12:37:25","2024-08-08 12:55:57+00:00","5179a15a06e43172aea0ac73d34de65536663745","816","6903","3","None","automatic-speech-recognition, text-to-speech, text-to-audio","Dataset Card for Annotations of English MLS This dataset consists in annotations of the English subset of the Multilingual LibriSpeech (MLS) dataset. MLS dataset is a large multilingual corpus suitabl","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","cc-by-4.0","None","None","None" +"taesiri/imagenet-hard-4K","taesiri","2023-05-21 17:33:17","2023-06-11 00:37:29+00:00","b70f7a9c60cc2cab806f1053815491c269484060","1537","6800","3","None","image-classification","Dataset Card for ""Imagenet-Hard-4K"" Project Page - Paper - Github ImageNet-Hard-4K is 4K version of the original ImageNet-Hard dataset, which is a new benchmark that comprises 10,980 images collected ","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","mit","2304.05538","None","None" +"Teklia/IAM-line","Teklia","2024-01-12 12:20:17","2024-03-14 16:19:29+00:00","fbdad97500ce54635c0d1ba306bf535cb40656cf","810","6791","7","None","image-to-text","IAM - line level Dataset Summary The IAM Handwriting Database contains forms of handwritten English text which can be used to train and test handwritten text recognizers and to perform writer identifi","None","en","mono","None","None","False","False","False","mit","None","None","None" +"DenyTranDFW/SEC_10K_FSNoNDS_Eat","DenyTranDFW","2024-08-07 08:12:12","2024-08-21 02:17:44+00:00","e4af1ff7fbaaf6fc75374aa7a5ce2fad8c6e13fb","1566","6776","0","language:en, license:gpl, size_categories:100K|1|T","None","False","False","False","apache-2.0","2409.12941","None","None" +"Cohere/miracl-en-corpus-22-12","Cohere","2023-02-02 23:21:21","2023-02-06 11:54:52+00:00","b9dbb115531962e828cc9c771a2f47ed5dbe55de","1272","6471","2","None","text-retrieval, document-retrieval","MIRACL (en) embedded with cohere.ai multilingual-22-12 encoder We encoded the MIRACL dataset using the cohere.ai multilingual-22-12 embedding model. The query embeddings can be found in Cohere/miracl-","None","en","mono","None","None","False","False","False","apache-2.0","None","None","document-retrieval" +"bespokelabs/Bespoke-Stratos-17k","bespokelabs","2025-01-21 09:38:20","2025-01-28 08:07:02+00:00","729410217b3ab45ff11d5f5a0160c318d805a3a2","6406","6415","123","language:en, license:apache-2.0, size_categories:10K|1|T","None","False","False","False","apache-2.0","None","None","None" +"harvard-lil/cold-cases","harvard-lil","2023-09-12 17:29:50","2024-03-26 15:50:21+00:00","5d8d0d8457ef63b6463af9737da21d3badd924ad","806","5609","22","None","None","Collaborative Open Legal Data (COLD) - Cases COLD Cases is a dataset of 8.3 million United States legal decisions with text and metadata, formatted as compressed parquet files. If you'd like to view a","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","cc0-1.0","None","None","None" +"yoshitomo-matsubara/srsd-feynman_medium","yoshitomo-matsubara","2022-06-08 06:22:10","2024-03-05 07:22:12+00:00","de7af9e2638494678ebe18d1174bbc6a02037eec","900","5562","1","None","tabular-regression","Dataset Card for SRSD-Feynman (Medium set) Dataset Summary Our SRSD (Feynman) datasets are designed to discuss the performance of Symbolic Regression for Scientific Discovery. We carefully reviewed th","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","cc-by-4.0","2206.10540","None","None" +"imageomics/fish-vista","imageomics","2024-06-10 19:09:08","2024-11-08 00:15:07+00:00","ced15685f2deceeabaed3e8a75e07e5737e45495","1286","5551","11","task_categories:image-classification, task_categories:image-segmentation, language:en, size_categories:100K' option will only load the CSV files. To download the entire dataset, including all processed images and seg","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","None","None","None","None" +"turing-motors/CoVLA-Dataset","turing-motors","2024-11-26 03:28:49","2025-01-24 00:58:35+00:00","3c88d564b017b652acd446eec8d3a17d78d40952","3725","5507","6","None","None","CoVLA-Dataset WACV 2025 Oral CoVLA-Dataset is a dataset comprising real-world driving videos spanning more than 80 hours. This dataset leverages a novel, scalable approach based on automated data proc","None","en","mono","None","None","False","auto","False","None","None","None","None" +"sentence-transformers/parallel-sentences-opensubtitles","sentence-transformers","2024-04-30 08:24:04","2024-06-18 19:45:43+00:00","d86a387587ab6f2fd9ec7453b2765cec68111c87","1412","5482","1","None","feature-extraction, sentence-similarity","Dataset Card for Parallel Sentences - OpenSubtitles This dataset contains parallel sentences (i.e. English sentence + the same sentences in another language) for numerous other languages. Most of the ","None","en, multilingual, ar, bg, ca, cs, da, de, el, es, et, fa, fi, fr, gl, he, hi, hr, hu, hy, id, it, ja, ka, ko, lt, lv, mk, ms, nl, pl, pt, ro, ru, sk, sl, sq, sr, sv, th, tr, uk, ur, vi, zh","multi","1|0|0|M|<|n|<|1|B","None","False","False","False","None","None","None","None" +"squeezebits/dynamic_sonnet_llama3","squeezebits","2024-08-14 10:27:58","2024-08-14 10:28:21+00:00","9edfa5537e3fd3032b9aeb0871db961186a31ae4","919","5316","1","None","question-answering, text-generation","Dynamic Sonnet - Llama3 Curated dataset for benchmarking LLM serving systems In real-world service scenarios, each request comes with varying input token lengths. Some requests generate only a few tok","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","None","None","None","None" +"LongSafari/open-genome","LongSafari","2024-07-06 15:54:59","2024-07-10 05:59:55+00:00","84369c058d192dcb607086d71679b877421e3250","3333","5283","12","task_categories:text-generation, language:en, license:apache-2.0, size_categories:10M|1|T","None","False","auto","False","odc-by","None","None","None" +"matteogabburo/mWikiQA","matteogabburo","2024-06-14 13:25:07","2024-07-27 09:26:35+00:00","3960a850ef4ba50778cea29ed6030701b67d4807","831","5109","0","None","question-answering","Dataset Description mWikiQA is a translated version of WikiQA. It contains 3,047 questions sampled from Bing query logs. The candidate answer sentences are extracted from Wikipedia and then manually l","None","en, fr, de, it, es, pt","multi","1|0|0|K|<|n|<|1|M","None","False","False","False","other","2406.10172","None","None" +"mesolitica/fineweb-filter-malaysian-context","mesolitica","2024-08-07 15:23:05","2024-08-13 08:38:23+00:00","c8df876812cc99955138c3f3691c2fc3b7933d6c","2034","5076","0","None","None","HuggingFaceFW/fineweb filter Malaysian context What is it? We filter the original ๐Ÿท FineWeb dataset that consists more than 15T tokens on simple Malaysian keywords. Total tokens for the filtered datas","None","en","mono","None","None","False","False","False","None","None","None","None" +"AnimaLab/bias-test-gpt-sentences","AnimaLab","2023-08-27 01:07:55","2024-03-16 02:49:01+00:00","c4626cbc91aafcd4de0a92dfd0526922cb2df8f0","1433","5068","1","None","None","Dataset Card for ""BiasTestGPT: Generated Test Sentences"" Dataset of sentences for bias testing in open-sourced Pretrained Language Models generated using ChatGPT and other generative Language Models. ","None","en","mono","1|0|K|<|n|<|1|0|0|K","None","False","False","False","apache-2.0","None","None","None" +"sentence-transformers/parallel-sentences-tatoeba","sentence-transformers","2024-04-30 11:19:18","2024-06-18 19:45:56+00:00","cec1343ab5a7a8befe99af4a2d0ca847b6c84743","825","5039","0","None","feature-extraction, sentence-similarity","Dataset Card for Parallel Sentences - Tatoeba This dataset contains parallel sentences (i.e. English sentence + the same sentences in another language) for numerous other languages. Most of the senten","None","en, multilingual, ar, bg, ca, cs, da, de, el, es, et, fa, fi, fr, gl, gu, he, hi, hu, hy, id, it, ja, ka, ko, ku, lt, lv, mk, mn, mr, ms, my, nb, nl, pl, pt, ro, ru, sk, sl, sq, sr, sv, th, tr, uk, ur, vi, zh","multi","1|M|<|n|<|1|0|M","None","False","False","False","None","None","None","None" +"BAAI/DenseFusion-1M","BAAI","2024-06-04 02:58:10","2024-10-17 08:00:41+00:00","b935d5ad34311e351bc2f3eb33667451e331e167","788","5002","29","None","question-answering, visual-question-answering","[Paper] https://arxiv.org/abs/2407.08303 [GitHub] https://github.com/baaivision/DenseFusion Introduction An image is worth a thousand words"". Comprehensive image descriptions are essential for multi-m","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","cc-by-4.0","2407.08303","None","None" +"espnet/floras","espnet","2024-10-24 01:40:36","2024-11-29 20:12:19+00:00","0cc744b3fad853c054031df2ef110c51f28a78e5","1606","4998","4","task_categories:automatic-speech-recognition, task_categories:translation, task_categories:summarization, language:en, language:es, language:fr, language:de, language:nl, language:it, language:pt, language:hu, language:fi, language:el, language:ca, language:eo, language:et, language:da, language:la, language:sv, language:cy, language:gl, language:ru, language:pl, language:uk, language:ro, language:cs, language:sl, language:sk, language:hr, language:bg, language:bs, language:ka, language:tr, language:fa, language:ar, language:uz, language:az, language:ku, language:ky, language:hi, language:ta, language:ur, language:bn, language:id, language:vi, language:th, language:mi, language:ms, language:ja, language:zh, license:cc-by-3.0, size_categories:1K|1|T","None","False","False","False","apache-2.0","None","None","None" +"him1411/polymath","him1411","2024-10-06 20:16:27","2024-10-24 07:55:57+00:00","15a27cc960d339f926e732ec3fdc7452c01e71dd","3081","3918","7","None","multiple-choice","Paper Information We present PolyMATH, a challenging benchmark aimed at evaluating the general cognitive reasoning abilities of MLLMs. PolyMATH comprises 5,000 manually collected high-quality images o","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","afl-3.0","2410.14702","None","None" +"TIGER-Lab/M-BEIR","TIGER-Lab","2023-12-01 00:08:47","2024-08-07 15:15:33+00:00","023f7cc5425e15e3c93f6146f7295c28eae767dc","899","3864","16","None","text-retrieval, text-to-image, image-to-text, visual-question-answering","UniIR: Training and Benchmarking Universal Multimodal Information Retrievers (ECCV 2024) ๐ŸŒ Homepage | ๐Ÿค— Model(UniIR Checkpoints) | ๐Ÿค— Paper | ๐Ÿ“– arXiv | GitHub How to download the M-BEIR Dataset ๐Ÿ””News ๐Ÿ”ฅ","None","en","mono","None","None","False","False","False","mit","None","None","None" +"eltorio/ROCOv2-radiology","eltorio","2024-11-11 18:34:08","2024-11-13 08:49:36+00:00","80ffeef4eb8d34d27cb5c2815305f1d8aee8a83c","1626","3852","42","None","None","ROCOv2: Radiology Object in COntext version 2 Introduction ROCOv2 is a multimodal dataset consisting of radiological images and associated medical concepts and captions extracted from the PMC Open Acc","None","en","mono","None","None","False","False","False","cc-by-nc-sa-4.0","None","None","None" +"Bin1117/AnyEdit","Bin1117","2024-12-12 09:15:37","2025-01-18 01:56:16+00:00","662e0178625f3fab3adae6868196eebcebbc2226","2693","3751","3","None","text-to-image, image-to-image","Celebrate! AnyEdit resolved the data alignment with re-uploading process (but view filter is still not working:(, though it has 25 edit types). You can view the validation split for a quick look. [Not","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","cc-by-4.0","2411.15738","None","None" +"launch/ampere","launch","2022-07-01 02:29:23","2022-11-09 01:57:52+00:00","e81e6a9ac798674b1a72239936bc4f71c4fa2c4e","1065","3721","0","None","text-classification","Dataset Card for AMPERE Dataset Description This dataset is released together with our NAACL 2019 Paper ""Argument Mining for Understanding Peer Reviews"". If you find our work useful, please cite: @inp","None","en","mono","None","None","False","False","False","cc-by-4.0","None","None","None" +"hpprc/jsick","hpprc","2023-04-08 16:02:06","2023-04-11 06:18:09+00:00","6f27df527556f0c2774f45297cfca7780477ad75","943","3692","7","None","sentence-similarity, text-classification, natural-language-inference, semantic-similarity-scoring","Japanese Sentences Involving Compositional Knowledge (JSICK) Dataset. JSICK is the Japanese NLI and STS dataset by manually translating the English dataset SICK (Marelli et al., 2014) into Japanese. W","@article{yanaka-mineshima-2022-compositional, title = ""Compositional Evaluation on {J}apanese Textual Entailment and Similarity"", author = ""Yanaka, Hitomi and Mineshima, Koji"", journal = ""Transactions of the Association for Computational Linguistics"", volume = ""10"", year = ""2022"", address = ""Cambridge, MA"", publisher = ""MIT Press"", url = ""https://aclanthology.org/2022.tacl-1.73"", doi = ""10.1162/tacl_a_00518"", pages = ""1266--1284"", }","ja, en","bi","1|0|K|<|n|<|1|0|0|K","None","False","False","False","cc-by-sa-4.0","None","None","natural-language-inference, semantic-similarity-scoring" +"Snowflake/msmarco-v2.1-snowflake-arctic-embed-l","Snowflake","2024-06-20 14:58:59","2024-08-07 16:57:50+00:00","7467d18cb02e4b85a4f0ccbe1f98c574d057437b","959","3559","0","None","question-answering","Snowflake Arctic Embed L Embeddings for MSMARCO V2.1 for TREC-RAG This dataset contains the embeddings for the MSMARCO-V2.1 dataset which is used as the corpora for TREC RAG All embeddings are created","None","en","mono","1|0|0|M|<|n|<|1|B","None","False","False","False","None","None","None","None" +"avaliev/chat_doctor","avaliev","2024-03-06 18:05:55","2024-03-06 20:55:49+00:00","19646f30de72c3890c6e0bc67579cbb538076822","818","3549","11","None","question-answering","This dataset was formed from the three data sources from the ChatDoctor work. 100k real conversations between patients and doctors from HealthCareMagic.com HealthCareMagic-100k. - ADDED 10k real conve","None","en","mono","1|0|0|K|<|n|<|1|M","None","False","False","False","apache-2.0","None","None","None" +"lmg-anon/vntl-leaderboard","lmg-anon","2024-06-08 20:33:07","2025-01-02 16:34:32+00:00","cf3d232d77458394857dbf8411de95fd3a894aef","876","3503","23","None","translation","VNTL Leaderboard The VNTL leaderboard ranks Large Language Models (LLMs) based on their performance in translating Japanese Visual Novels into English. Please be aware that the current results are pre","None","en, ja","bi","n|<|1|K","None","False","False","False","None","None","None","None" +"textdetox/multilingual_toxicity_dataset","textdetox","2024-02-01 15:44:46","2024-06-10 09:43:48+00:00","684bfd68651993aafc5ca653c35747e43f5b5c43","883","3460","20","None","text-classification","For the shared task CLEF TextDetox 2024, we provide a compilation of binary toxicity classification datasets for each language. Namely, for each language, we provide 5k subparts of the datasets -- 2.5","None","en, ru, uk, de, es, am, zh, ar, hi","multi","1|0|K|<|n|<|1|0|0|K","None","False","False","False","openrail++","None","None","None" +"krr-oxford/OntoLAMA","krr-oxford","2023-03-02 00:45:25","2024-11-20 23:18:29+00:00","2d9c6bfa5de20cd5c73f122f29ee331296832c0d","1316","3442","4","None","text-classification","OntoLAMA: LAnguage Model Analysis for Ontology Subsumption Inference Dataset Summary OntoLAMA is a set of language model (LM) probing datasets for ontology subsumption inference. The work follows the ","None","en","mono","1|M|<|n|<|1|0|M","None","False","False","False","apache-2.0","2302.06761","None","None" +"maum-ai/COMMAND","maum-ai","2024-12-08 17:02:58","2024-12-16 00:31:04+00:00","926be7d19b13a64eb84a7a5f7b5c253af72ad5f1","2357","3432","2","None","robotics","COMMAND Dataset ๐Ÿ˜ƒ Dataset Structure ๐Ÿ—‚๏ธ The COMMAND_dataset is organized as follows: COMMAND_dataset/ โ”œโ”€ README.md # Documentation for the dataset โ”œโ”€ map_asset/ # Directory containing map assets ๐Ÿ—บ๏ธ โ”‚ โ”œ","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-nc-4.0","None","None","None" +"arsaporta/symile-m3","arsaporta","2024-11-12 05:26:09","2024-11-26 00:51:57+00:00","9ec61c44424ffda11b1547e9d506056b7e98870a","2172","3402","5","None","zero-shot-classification, zero-shot-image-classification","Dataset Card for Symile-M3 Symile-M3 is a multilingual dataset of (audio, image, text) samples. The dataset is specifically designed to test a model's ability to capture higher-order information betwe","None","ar, el, en, hi, ja, ko, te, th, uk, zh","multi","1|0|M|<|n|<|1|0|0|M","None","False","False","False","cc-by-nc-sa-4.0","2411.01053","None","None" +"blanchon/EuroSAT_RGB","blanchon","2023-12-05 12:56:11","2023-12-05 13:02:42+00:00","655ef66ef2be07f89aec61407f24c772802eb87d","790","3310","3","None","image-classification","EuroSAT RGB EUROSAT RGB is the RGB version of the EUROSAT dataset based on Sentinel-2 satellite images covering 13 spectral bands and consisting of 10 classes with 27000 labeled and geo-referenced sam","None","e, n","bi","1|0|K|<|n|<|1|0|0|K","None","False","False","False","unknown","1709.00029","None","None" +"Nan-Do/code-search-net-python","Nan-Do","2023-05-14 00:42:57","2023-05-15 00:55:15+00:00","39db91866dd0f251f3b0c7f42c0f85634101df6e","859","3308","22","None","text-generation, text2text-generation, summarization","Dataset Card for ""code-search-net-python"" Dataset Description Homepage: None Repository: https://huggingface.co/datasets/Nan-Do/code-search-net-python Paper: None Leaderboard: None Point of Contact: @","None","en","mono","None","None","False","False","False","apache-2.0","None","None","None" +"ServiceNow/repliqa","ServiceNow","2024-06-11 20:23:34","2024-12-09 01:32:45+00:00","c1b92b58d6f3ce4c4d4dc5225ccf871fb2ac5515","1018","3299","8","None","question-answering, text-classification","RepLiQA - Repository of Likely Question-Answer for benchmarking Dataset Summary RepLiQA is an evaluation dataset that contains Context-Question-Answer triplets, where contexts are non-factual but natu","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-4.0","2406.11811","None","None" +"davanstrien/magpie-preference","davanstrien","2024-06-24 15:10:18","2025-01-28 11:32:55+00:00","63643ae537105e5a7aa84d084d3be0aa0ad66f7e","1095","3284","13","None","None","Dataset Card for Magpie Preference Dataset Dataset Description The Magpie Preference Dataset is a crowdsourced collection of human preferences on synthetic instruction-response pairs generated using t","None","en","mono","None","None","False","False","False","llama3","None","None","None" +"strickvl/isafpressreleases","strickvl","2024-03-23 14:17:21","2024-06-15 12:10:35+00:00","74875b0589d607cdb4356d513d4c9435286df9f4","1333","3246","6","None","feature-extraction, summarization, question-answering, text-classification, fill-mask, zero-shot-classification, named-entity-recognition, topic-classification, news-articles-summarization","ISAF Press Releases Dataset Description Homepage: [N/A] Repository: [N/A] Paper: A Knock on the Door: 22 Months of ISAF Press Releases Point of Contact: Alex Strick van Linschoten (@strickvl) Dataset ","None","en","mono","1|K|<|n|<|1|0|K","None","False","False","False","cc-by-sa-4.0","None","None","named-entity-recognition, topic-classification, news-articles-summarization" +"Rapidata/text-2-image-Rich-Human-Feedback","Rapidata","2025-01-06 18:21:13","2025-01-11 13:23:04+00:00","e77afd00e481d9d2ca41a5b5c4f89cb704de45c6","3215","3215","27","task_categories:text-to-image, task_categories:text-classification, task_categories:image-classification, task_categories:image-to-text, task_categories:image-segmentation, language:en, license:apache-2.0, size_categories:10K