Backpropagation — The Math Behind Learning
A complete derivation of backpropagation for MLPs — from chain rule intuition to delta propagation, with a worked numerical example showing exactly how errors flow backward through a network.
All the articles with the tag "deep-learning".
A complete derivation of backpropagation for MLPs — from chain rule intuition to delta propagation, with a worked numerical example showing exactly how errors flow backward through a network.
A narrative-first walkthrough of reinforcement learning, starting with everyday intuition and ending with the math behind Q-learning and DQN.
A high level view on how modern vision-language models connect pixels and prose, from CLIP and BLIP to Flamingo, MiniGPT-4, Kosmos, and Gemini.
The evolution of image diffusion architectures. Learn how we moved from convolutional U-Nets to scalable Diffusion Transformers (DiT), and why treating images like language changed everything.
How to move from visual imitation to law-governed motion. Deep dive into injecting PDEs into neural networks, implicit physics extraction, and LLM-guided physical reasoning.
The fundamentals of video diffusion models. Learn how we extend 2D diffusion to time, the mechanics of temporal attention, and the architectural shifts required for motion consistency.