Following up on LLaDA 2.0 , the paper is now out on Daily Papers🔥 It has sparked a lot of discussion in the community for showing how discrete diffusion LLMs can scale to 100B and run faster than traditional AR models. LLaDA2.0: Scaling Up Diffusion Language Models to 100B (2512.15745)
✨ Built from real enterprise data (Enron + financial institutions), not synthetic tasks ✨ Tests end-to-end finance workflows ✨ Multimodal & cross-file reasoning ✨ Expert annotated (700+ hours) and genuinely challenging hard
✨ Any-to-Any & World-Model : one step forward to the real world - BAAI Emu 3.5 - Antgroup Ming-flash-omni - HunyuanWorld-Mirror: 3D
Aligning with the “world model” globally
✨ Audio & Speech + Video & Visual: released from entertainment labs to delivery platforms - SoulX-Podcast TTS - LongCat-Audio-Codec & LongCat-Video by Meituan delivery paltform - xiabs DreamOmni 2
✨ 48B total/ 3B active - MIT license ✨ Up to 1M context ✨ 84.3 on RULER (128k) with 3.98× speedup ✨ Hybrid KDA + MLA architecture for peak throughput & quality
🤖 Did you know your voice might be cloned without your consent from just *one sentence* of audio? That's not great. So with @frimelle , we brainstormed a new idea for developers who want to curb malicious use: ✨The Voice Consent Gate.✨ Details, code, here: https://huggingface.co/blog/voice-consent-gate
✨ Compresses long sequences visually to bypass token limits ✨ Reduces computational and memory costs ✨ Preserves meaning through multimodal encoding ✨ Built on GLM-4.1V-9B-Base
✨ Any prior in → 3D world out ✨ Mix camera, intrinsics, depth as priors ✨ Predict point clouds, normals, Gaussians & more in one pass ✨ Unified architecture for all 3D task
✨ Trained on Honey-Data-15M, a 15M-sample SFT corpus with dual-level CoT reasoning ✨ Backed by HoneyPipe, a transparent & reproducible open data curation suite
🌎 AI ethics and sustainability are two sides of the same coin.
In our new blog post with Dr. Sasha Luccioni, we argue that separating them (as is too often the case) means missing the bigger picture of how AI systems impact both people and the planet.
Ethical and sustainable AI development can’t be pursued in isolation. The same choices that affect who benefits or is harmed by AI systems also determine how much energy and resources they consume.
We explore how two key concepts, evaluation and transparency, can serve as bridges between these domains:
📊 Evaluation, by moving beyond accuracy or performance metrics to include environmental and social costs, as we’ve done with tools like the AI Energy Score.
🔍 Transparency, by enabling reproducibility, accountability, and environmental reporting through open tools like the Environmental Transparency Space.
AI systems mirror our priorities. If we separate ethics from sustainability, we risk building technologies that are efficient but unjust, or fair but unsustainable.
✨1T total / 50B active params per token ✨20T+ reasoning-dense tokens (Evo-CoT) ✨128K context via YaRN ✨FP8 training: 15%+ faster, same precision as BF16 ✨Hybrid Syntax-Function-Aesthetics reward for front-end & visual generation