How Do Transformers Learn Variable Binding?

1:05:00
Automating scientific discovery and hypothesis generation with language model agents

1:11:10
How DeepSeek changes the LLM story

57:45
Visualizing transformers and attention | Talk for TNG Big Tech Day '24

1:01:46
More accurate behavioral predictions with hybrid Bayesian-Transformer models

18:08
Transformer Neural Networks Derived from Scratch

53:31
Charlie Snell, UC Berkeley. Title: Scaling LLM Test-Time Compute

1:09:41
Neural algorithms of human language

42:13