

Stockholm vLLM Inference Meetup
vLLM Inference Meetup in Sweden
Hosted by Red Hat AI, AMD, and AI Sweden, this event takes place on 25 March 2026 in Stockholm, Sweden.
What to Expect
Deep technical sessions from vLLM maintainers, committers, and teams using vLLM at scale
Live demos focused on real workflows
Great networking with food and drinks
Who Should Attend
vLLM users and contributors
ML and infra engineers working on inference and serving
Platform teams running GenAI in production
Anyone curious about efficient inference across local, cloud, and Kubernetes
Agenda (Subject to More Awesomeness)
17:00 – 17:30 — Doors Open, Check-In
17:30 – 17:40 — Welcome and Opening Remarks
Ted Henriksson, Wizard, AI Labs, AI Sweden
Sasa Zelenovic, Sr. Technical Marketing Manager, Red Hat AI
17:40 – 18:00 — Intro to vLLM and Project Update
Michael Goin, vLLM Core Committer and Principal Engineer, Red Hat AI
18:00 – 18:30 — Accurate LLM Compression for Fast & Efficient Inference
Eldar Kurtić, Principal Research Scientist, Red Hat AI & ISTA
18:30 – 19:00 — vLLM Inference Optimization on AMD GPUs
Chang Liu, AI Product Application Engineer, AMD
19:00 – 19:15 — Inference endpoints in the SVEA project
Adam Ek, Research Engineer and AI Lead at SVEA, AI Sweden
19:15 – 19:45 — Discussion and Q&A
19:45 – 21:00 — Networking, Food and Drinks
Important information
Registration closes 24 hours before the event. We cannot admit unregistered attendees.
Please bring a photo ID to verify your registration on arrival.
See you in Stockholm
If you are building, deploying, or scaling inference, this is the room to be in.
See you soon!