Rick W / Monday, April 13, 2026 / Categories: Artificial Intelligence From Prompt to Prediction: Understanding Prefill, Decode, and the KV Cache in LLMs This article is divided into three parts; they are: • How Attention Works During Prefill • The Decode Phase of LLM Inference • KV Cache: How to Make Decode More Efficient Consider the prompt: Today’s weather is so . Previous Article GLM-5.1: Architecture, Benchmarks, Capabilities & How to Use It Next Article Top 10 Gemma 4 Projects That Will Blow Your Mind Print 2 Tags: LLM