Cover Image for High-Performance Inference for Open LLMs with Modal, Qwen and SGLang
Cover Image for High-Performance Inference for Open LLMs with Modal, Qwen and SGLang
Avatar for Modal
Presented by
Modal
AI infrastructure that developers love

High-Performance Inference for Open LLMs with Modal, Qwen and SGLang

Registration
Past Event
Please click on the button below to join the waitlist. You will be notified if additional spots become available.
About Event

When do open models and inference engines beat proprietary solutions?

Join Modal, Qwen, and SGLang for an evening optimizing performance and cost for LLM inference. Our speakers will cover:

  • The Cold Start Issue — how can AI infrastructure enable seamless AI experiences, right from the start?

  • ​Accelerating Open Models — how do inference engines work with model developers to achieve benchmarking goals?

  • Choosing the Best Model — how should developers choose the most effective model for their use case?

​With:


​​Agenda

We're excited to bring together founders, AI engineers, and ML systems researchers for an evening with:

  • ​​Demos & Lightning Talks

  • ​​​Community, Pizza, Drinks

Your hosts,

​Modal, Qwen & SGLang

Location
375 Alabama St
San Francisco, CA 94110, USA
Suite 490
Avatar for Modal
Presented by
Modal
AI infrastructure that developers love