MosaicBERT A collection of BERT-based models of different sequence lengths trained on the C4 dataset. Details: https://mosaicbert.github.io/ mosaicml/mosaic-bert-base Fill-Mask • Updated Mar 5, 2024 • 110 • 47 mosaicml/mosaic-bert-base-seqlen-256 Fill-Mask • Updated Mar 5, 2024 • 38 • 2 mosaicml/mosaic-bert-base-seqlen-512 Fill-Mask • Updated Mar 5, 2024 • 23 • 4 mosaicml/mosaic-bert-base-seqlen-1024 Fill-Mask • Updated Mar 5, 2024 • 30 • 15
MPT The MPT collections is a series of decoder-style transformer models trained from scratch by MosaicML. Details: https://www.mosaicml.com/mpt mosaicml/mpt-7b-8k Text Generation • Updated Mar 5, 2024 • 1.84k • 26 mosaicml/mpt-7b Text Generation • Updated Mar 5, 2024 • 18.6k • 1.17k mosaicml/mpt-7b-instruct Text Generation • Updated Mar 5, 2024 • 7.76k • 470 mosaicml/mpt-7b-8k-chat Text Generation • Updated Mar 5, 2024 • 1.86k • 40
MosaicBERT A collection of BERT-based models of different sequence lengths trained on the C4 dataset. Details: https://mosaicbert.github.io/ mosaicml/mosaic-bert-base Fill-Mask • Updated Mar 5, 2024 • 110 • 47 mosaicml/mosaic-bert-base-seqlen-256 Fill-Mask • Updated Mar 5, 2024 • 38 • 2 mosaicml/mosaic-bert-base-seqlen-512 Fill-Mask • Updated Mar 5, 2024 • 23 • 4 mosaicml/mosaic-bert-base-seqlen-1024 Fill-Mask • Updated Mar 5, 2024 • 30 • 15
MPT The MPT collections is a series of decoder-style transformer models trained from scratch by MosaicML. Details: https://www.mosaicml.com/mpt mosaicml/mpt-7b-8k Text Generation • Updated Mar 5, 2024 • 1.84k • 26 mosaicml/mpt-7b Text Generation • Updated Mar 5, 2024 • 18.6k • 1.17k mosaicml/mpt-7b-instruct Text Generation • Updated Mar 5, 2024 • 7.76k • 470 mosaicml/mpt-7b-8k-chat Text Generation • Updated Mar 5, 2024 • 1.86k • 40