Head Research Scientist at Cerebras
I'm a Head Research Scientist at Cerebras with nearly a decade of AI/ML experience. I designed MoE recipes from the ground up for Cerebras hardware and now lead training at unprecedented scale across hardware, software, and ML teams. From this hands-on work, I distill practical MoE recipes into The MoE 101 Guide.
My focus is building efficient, scalable AI systems end to end, spanning data, models, and infrastructure. This includes SlimPajama, a 627B-token dataset with over 1M downloads, and BTLM, a 3B model achieving 7B-class quality while using 3× less inference compute.
Previously, I built ML systems used by millions. At Yandex, I invented the YATI model that powers Yandex Search. At Google, I focused on improving Google Assistant's ASR and built models that were deployed in Google Captions and Google Gboard.
Full course is available at: https://www.cerebras.ai/moe-guide
I regularly advise top AI labs on MoE architecture and training dynamics. For consulting inquiries or early-stage collaborations, please reach out via email: dariamsoboleva@gmail.com.
Speaker, organizer, and moderator for both panels.
Also on Google Scholar