Cover Image for Where Does the Slash in Attention Matrices Come From?
Cover Image for Where Does the Slash in Attention Matrices Come From?
Hosted By

Where Does the Slash in Attention Matrices Come From?

Hosted by NICE AI Talk
YouTube
Registration
Past Event
Welcome! To join the event, please register below.
About Event

Description:
Discover the origin of the “slash pattern” in LLM attention heatmaps. This talk reveals how Slash-Dominant Heads (SDH) arise from the interaction of RoPE positional encoding and model structure—and how they generalize to out-of-distribution inputs. Join us for insights from theory, mechanism, and empirical evidence.

Speaker: Fengzhuo Zhang, Postdoctoral Researcher at Yale CADMY, PhD from National University of Singapore. Research focus: statistical learning and optimization theory for LLMs.

Paper: Demystifying the Slash Pattern in Attention: The Role of RoPE

Host: Mingyu Jin, PhD Student at Rutgers University, specializing in language model interpretability and explainable reasoning. Homepage: https://mingyuj666.github.io/

Hosted By