MosaicBERT A collection of BERT-based models of different sequence lengths trained on the C4 dataset. Details: https://mosaicbert.github.io/ mosaicml/mosaic-bert-base Fill-Mask • Updated Mar 5, 2024 • 113 • 47 mosaicml/mosaic-bert-base-seqlen-256 Fill-Mask • Updated Mar 5, 2024 • 21 • 2 mosaicml/mosaic-bert-base-seqlen-512 Fill-Mask • Updated Mar 5, 2024 • 69 • 4 mosaicml/mosaic-bert-base-seqlen-1024 Fill-Mask • Updated Mar 5, 2024 • 42 • 15
MPT The MPT collections is a series of decoder-style transformer models trained from scratch by MosaicML. Details: https://www.mosaicml.com/mpt mosaicml/mpt-7b-8k Text Generation • Updated Mar 5, 2024 • 1.43k • 26 mosaicml/mpt-7b Text Generation • Updated Mar 5, 2024 • 13.5k • 1.18k mosaicml/mpt-7b-instruct Text Generation • Updated Mar 5, 2024 • 6.36k • 471 mosaicml/mpt-7b-8k-chat Text Generation • Updated Mar 5, 2024 • 1.18k • 40
MosaicBERT A collection of BERT-based models of different sequence lengths trained on the C4 dataset. Details: https://mosaicbert.github.io/ mosaicml/mosaic-bert-base Fill-Mask • Updated Mar 5, 2024 • 113 • 47 mosaicml/mosaic-bert-base-seqlen-256 Fill-Mask • Updated Mar 5, 2024 • 21 • 2 mosaicml/mosaic-bert-base-seqlen-512 Fill-Mask • Updated Mar 5, 2024 • 69 • 4 mosaicml/mosaic-bert-base-seqlen-1024 Fill-Mask • Updated Mar 5, 2024 • 42 • 15
MPT The MPT collections is a series of decoder-style transformer models trained from scratch by MosaicML. Details: https://www.mosaicml.com/mpt mosaicml/mpt-7b-8k Text Generation • Updated Mar 5, 2024 • 1.43k • 26 mosaicml/mpt-7b Text Generation • Updated Mar 5, 2024 • 13.5k • 1.18k mosaicml/mpt-7b-instruct Text Generation • Updated Mar 5, 2024 • 6.36k • 471 mosaicml/mpt-7b-8k-chat Text Generation • Updated Mar 5, 2024 • 1.18k • 40