PyTorch 2.5 Live Q&A
![](https://i.ytimg.com/vi/1OopuwTq6oE/mqdefault.jpg)
30:06
PyTorch 2.6 Release Live Q&A
![](https://i.ytimg.com/vi/ibgW_ali0Gc/mqdefault.jpg)
20:37
State of PyTorch - Ji Li & Damien Sereni, Meta
![](https://i.ytimg.com/vi/cvIeT4MlIx4/mqdefault.jpg)
29:18
PyTorch Expert Exchange Hacker Cup AI
![](https://i.ytimg.com/vi/HTcnp9NEHGY/mqdefault.jpg)
33:29
How does batching work on modern GPUs?
![](https://i.ytimg.com/vi/jk2FsJxZFo8/mqdefault.jpg)
44:06
LLM inference optimization: Architecture, KV cache and Flash attention
![](https://i.ytimg.com/vi/wd57g2IM3C4/mqdefault.jpg)
24:21
Running State-of-Art Gen AI Models on-Device with NPU Acceleration - Felix Baum, Qualcomm
![](https://i.ytimg.com/vi/Bh-jlh5vlF0/mqdefault.jpg)
32:03
DistServe: disaggregating prefill and decoding for goodput-optimized LLM inference
![](https://i.ytimg.com/vi/wjZofJX0v4M/mqdefault.jpg)
27:14