Cover Image for Bigger is not always better: RAG with small LLM
Cover Image for Bigger is not always better: RAG with small LLM
Avatar for Ario
Presented by
Ario
Ario: Building the Plaid for every type of data

Bigger is not always better: RAG with small LLM

Register to See Address
Palo Alto, California
Registration
Past Event
Please click on the button below to join the waitlist. You will be notified if additional spots become available.
About Event

LLMs are pretty amazing at understanding natural language. When you combine them with RAG (Retrieval-Augmented Generation), it’s like a match made in heaven. You can use this combo for all sorts of things like answering questions on a knowledge base, interactive chatbots, smart coding assistants, and more.

But if you’re worried about privacy, using cloud-based LLMs like GPT or Gemini can be a problem.

What if you could get an astonishingly good RAG using a small LLM?

In this talk, Ariya Hidayat will walk through some code that uses Phi 2, a tiny but mighty LLM, to do RAG on your laptop. He will show you how to use some tricks like query transformation, classifier, and hypothetical embedding to make Phi 2 work like a charm for RAG, even without an internet connection.

Talk will begin around 6pm

Beer & pizza provided!

Location
Please register to see the exact location of this event.
Palo Alto, California
Avatar for Ario
Presented by
Ario
Ario: Building the Plaid for every type of data