How much attention does the LLM give to the System Prompt vs User Query vs Retrieved Documents? The U-Curve phenomenon.
System Prompt Attention visualizes where the model really "looks". The U-curve shows: Beginning and end of the context get the most attention, the middle is neglected. This has direct consequences for RAG and long documents.
System Prompts are the "personality" of an LLM. This visualization shows how strongly they shape attention distribution – and why Prompt Engineering is science, not art.
Anthropic's Claude and OpenAI's GPT use massive System Prompts (10K+ tokens). Understanding how these affect attention explains why sometimes instructions are "forgotten".