RWKV: Reinventing RNNs for the Transformer Era (Paper Explained)
4:43:55
Virologe Christian Drosten über die Lehren aus der Pandemie - Jung & Naiv: Folge 744
10:30
Why Neural Networks can learn (almost) anything
40:40
Mamba: Linear-Time Sequence Modeling with Selective State Spaces (Paper Explained)
57:45
Visualizing transformers and attention | Talk for TNG Big Tech Day '24
34:32
Mixtral of Experts (Paper Explained)
1:11:58
Hallucination-Free? Assessing the Reliability of Leading AI Legal Research Tools (Paper Explained)
32:27
Efficient Streaming Language Models with Attention Sinks (Paper Explained)
27:48