Memory Optimization for On-Device LLMs
58:43
LLMs Quantization Crash Course for Beginners
28:44
3 Shocking Ways Custom Code Assistants Will Change Your Life!
30:25
Exploring the Latency/Throughput & Cost Space for LLM Inference // Timothée Lacroix // CTO Mistral
21:47
Fine-tuning Llama 3.2 on Your Data with a single GPU | Training LLM for Sentiment Analysis
37:01
Capitole Tech Talk - Software architectures to capitalize on LLMs
36:12
Deep Dive: Optimizing LLM inference
19:46
Quantization vs Pruning vs Distillation: Optimizing NNs for Inference
1:44:31