Cerebras REAP Collection Sparse MoE models compressed using REAP (Router-weighted Expert Activation Pruning) method • 19 items • Updated 18 days ago • 77
VTP Collection Towards Scalable Pre-training of Visual Tokenizers for Generation • 4 items • Updated 21 days ago • 39
Teacher Logits Collection Logits captured from large models to act as the teacher for distillation • 3 items • Updated 22 days ago • 7
Ministral 3 Collection Mistral Ministral 3: new multimodal models in Base, Instruct, and Reasoning variants, available in 3B, 8B, and 14B sizes. • 36 items • Updated 13 days ago • 26
Ministral 3 Collection A collection of edge models, with Base, Instruct and Reasoning variants, in 3 different sizes: 3B, 8B and 14B. All with vision capabilities. • 9 items • Updated Dec 2, 2025 • 136
Trinity Collection Collection of Arcee AI models in the Trinity family • 8 items • Updated 25 days ago • 21
Olmo 3 Pre-training Collection All artifacts related to Olmo 3 pre-training • 10 items • Updated 14 days ago • 32
BERT Hash Nano Models Collection Set of BERT models with a modified embeddings layer • 4 items • Updated 14 days ago • 9
TOUCAN: Synthesizing 1.5M Tool-Agentic Data from Real-World MCP Environments Paper • 2510.01179 • Published Oct 1, 2025 • 25
💧 LFM2 Collection LFM2 is a new generation of hybrid models, designed for on-device deployment. • 26 items • Updated about 19 hours ago • 131
view article Article Welcome EmbeddingGemma, Google's new efficient embedding model +4 Sep 4, 2025 • 267
Tfree-HAT-7b-pretrained Collection Tokenizer free models based on Hierarchical Autoregressive Transformer (https://arxiv.org/abs/2501.10322) trained from scratch. • 2 items • Updated Aug 1, 2025 • 10