Deep dive: model merging (part 1)

32:15
Deep dive: model merging, part 2

17:04
Merge LLMs with No Code Mergekit GUI

22:20
Merge LLMs using Mergekit: Create your own Medical Mixture of Experts

2:11:12
How I use LLMs

40:54
Deep dive - Better Attention layers for Transformer models

57:45
Visualizing transformers and attention | Talk for TNG Big Tech Day '24

44:06
LLM inference optimization: Architecture, KV cache and Flash attention

11:48