Backpropagation — The Math Behind Learning
A complete derivation of backpropagation for MLPs — from chain rule intuition to delta propagation, with a worked numerical example showing exactly how errors flow backward through a network.
All the articles with the tag "optimization".
A complete derivation of backpropagation for MLPs — from chain rule intuition to delta propagation, with a worked numerical example showing exactly how errors flow backward through a network.
A narrative-first walkthrough of reinforcement learning, starting with everyday intuition and ending with the math behind Q-learning and DQN.
How to accelerate diffusion sampling and steer creativity. Learn the mechanics of DDIM, DPM-Solver, Classifier-Free Guidance (CFG), and the math of negative prompting.
A comprehensive deep-dive into production inference optimization, tracing the path of a request through LLM and diffusion model serving systems. Understanding the bottlenecks from gateway to GPU kernel execution.
How PagedAttention, Continuous Batching, Speculative Decoding, and Quantization unlock lightning-fast, reliable large language model serving.
A deep dive into XGBoost — how second-order Taylor approximations and sophisticated regularization make it the dominant algorithm for structured data, bridging mathematical rigor with system engineering excellence.