Deep Dive: Hugging Face models on AWS AI Accelerators
47:19
Deep Dive: Advanced distributed training with Hugging Face LLMs and AWS Trainium
1:05:15
Run performant and cost-effective GenAI Applications with AWS Graviton and Arcee AI
36:12
Deep Dive: Optimizing LLM inference
45:23
AWS Trainium and Inferentia // Kamran Khan and Matthew McClean // MLOps Podcast #238
44:06
LLM inference optimization: Architecture, KV cache and Flash attention
1:55:27
Worst Fails of the Year | Try Not to Laugh 💩
58:19
AWS re:Invent 2023 - A deep dive on AWS infrastructure powering the generative AI boom (CMP201)
39:05