Cover Image for vLLM compile deep dive (it's all torch.compile)
Cover Image for vLLM compile deep dive (it's all torch.compile)
15 Going
Namgyu Youn
invites you to join

vLLM compile deep dive (it's all torch.compile)

Hosted by Ayush Satyam, Daniel & Network School
Registration
Welcome! To join the event, please register below.
About Event

Join us for a focused deep dive into compiling vLLM models with torch.compile. We'll walk through why torch.compile matters for large-model performance and how vLLM leverages it, demonstrate concrete compilation strategies and flags, and profile real-world gains on inference workloads.

This session is best for people comfortable with vLLM or PyTorch and basic LLM concepts who want actionable techniques to speed up inference. This is mostly a theoretical session and we are open for discussion afterwards; questions and specific pain points are welcome. Meet at NS library on March 9.

Location
NS library
15 Going