MiniLLM: Knowledge Distillation of Large Language Models
19:46
Quantization vs Pruning vs Distillation: Optimizing NNs for Inference
16:49
Better not Bigger: Distilling LLMs into Specialized Models
1:00:11
EfficientML.ai Lecture 9 - Knowledge Distillation (MIT 6.5940, Fall 2023)
36:25
Direct Preference Optimization (DPO): Your Language Model is Secretly a Reward Model Explained
6:47
Simply explained! KAN: Kolmogorov–Arnold Networks is interpretable! Mathematics and Physics
1:09:57
RetNet: A Successor to Transformer for Large Language Models Explained
57:03
Comment Deux Amis Changent Un CHÂTEAU Abandonné En HÔTEL 4☆ | @chateaudutheil
14:03