All 85+ interactive visualizations for exploring modern Large Language Models
From Tokens to Attention: The building blocks of modern LLMs
MoE, GQA, Flash Attention and Multimodal
Chain-of-Thought, o1/o3, DeepSeek R1, and Flexible Inference
KV-Cache, Positional Encodings, and Context Extension
ICL, System Prompts, and Few-Shot Learning
RLHF, Sampling, Quantization and Speculative Decoding
Benchmarks, Emergence and Attention Scaling
Calculators, References and Lookup Tools