MosaicBERT A collection of BERT-based models of different sequence lengths trained on the C4 dataset. Details: https://mosaicbert.github.io/ mosaicml/mosaic-bert-base Fill-Mask • Updated Mar 5, 2024 • 89 • 47 mosaicml/mosaic-bert-base-seqlen-256 Fill-Mask • Updated Mar 5, 2024 • 77 • 2 mosaicml/mosaic-bert-base-seqlen-512 Fill-Mask • Updated Mar 5, 2024 • 29 • 4 mosaicml/mosaic-bert-base-seqlen-1024 Fill-Mask • Updated Mar 5, 2024 • 124 • 15
MPT The MPT collections is a series of decoder-style transformer models trained from scratch by MosaicML. Details: https://www.mosaicml.com/mpt mosaicml/mpt-7b-8k Text Generation • Updated Mar 5, 2024 • 2.59k • 26 mosaicml/mpt-7b Text Generation • Updated Mar 5, 2024 • 23.1k • 1.17k mosaicml/mpt-7b-instruct Text Generation • Updated Mar 5, 2024 • 8.21k • 470 mosaicml/mpt-7b-8k-chat Text Generation • Updated Mar 5, 2024 • 797 • 40
MosaicBERT A collection of BERT-based models of different sequence lengths trained on the C4 dataset. Details: https://mosaicbert.github.io/ mosaicml/mosaic-bert-base Fill-Mask • Updated Mar 5, 2024 • 89 • 47 mosaicml/mosaic-bert-base-seqlen-256 Fill-Mask • Updated Mar 5, 2024 • 77 • 2 mosaicml/mosaic-bert-base-seqlen-512 Fill-Mask • Updated Mar 5, 2024 • 29 • 4 mosaicml/mosaic-bert-base-seqlen-1024 Fill-Mask • Updated Mar 5, 2024 • 124 • 15
MPT The MPT collections is a series of decoder-style transformer models trained from scratch by MosaicML. Details: https://www.mosaicml.com/mpt mosaicml/mpt-7b-8k Text Generation • Updated Mar 5, 2024 • 2.59k • 26 mosaicml/mpt-7b Text Generation • Updated Mar 5, 2024 • 23.1k • 1.17k mosaicml/mpt-7b-instruct Text Generation • Updated Mar 5, 2024 • 8.21k • 470 mosaicml/mpt-7b-8k-chat Text Generation • Updated Mar 5, 2024 • 797 • 40