Lightning Talk: Accelerated Inference in PyTorch 2.X with Torch...- George Stefanakis & Dheeraj Peri
13:34
Lightning Talk: The Fastest Path to Production: PyTorch Inference in Python - Mark Saroufim, Meta
1:30:36
PyTorch 2.0 Live Q&A Series: A Deep Dive on TorchDynamo
18:52
TensorRT for Beginners: A Tutorial on Deep Learning Inference Optimization
15:09
How to use TensorRT C++ API for high performance GPU inference by Cyrus Behroozi
22:30
vLLM: Easy, Fast, and Cheap LLM Serving, Woosuk Kwon, UC Berkeley
15:23
PyTorch 2.0: Unlocking the Power of Deep Learning with the Torch Compile API - Christian Keller
15:16
Lightning Talk: AOTInductor: Ahead-of-Time Compilation for PT2 Exported Models - Bin Bao, Meta
11:48