Training LLMs at Scale - Deepak Narayanan | Stanford MLSys #83
59:17
Serving 100s of LLMs on 1 GPU with LoRAX - Travis Addair | Stanford MLSys #84
58:06
Stanford Webinar - Large Language Models Get the Hype, but Compound Systems Are the Future of AI
24:04
Efficient Large-Scale Language Model Training on GPU Clusters Using Megatron-LM | Jared Casper
56:32
Monarch Mixer: Making Foundation Models More Efficient - Dan Fu | Stanford MLSys #86
1:04:33
Frontiers: Building Multimodal, Document-Grounded LLM Agents for Conversational AI in Education
1:19:06
Hardware-aware Algorithms for Sequence Modeling - Tri Dao | Stanford MLSys #87
24:19
A friendly introduction to distributed training (ML Tech Talks)
58:25