Reinforcement Learning — From Intuition to Algorithms
A narrative-first walkthrough of reinforcement learning, starting with everyday intuition and ending with the math behind Q-learning and DQN.
All the articles I've archived.
A narrative-first walkthrough of reinforcement learning, starting with everyday intuition and ending with the math behind Q-learning and DQN.
A structured articulation and pacing warm-up designed to help technologists speak with clarity and confidence in high-stakes meetings.
Why modern AI teams are handcrafting GPU kernels—from FlashAttention to TPU Pallas code—and how smarter tooling is making silicon-level tuning accessible.
How PagedAttention, Continuous Batching, Speculative Decoding, and Quantization unlock lightning-fast, reliable large language model serving.
A high level view on how modern vision-language models connect pixels and prose, from CLIP and BLIP to Flamingo, MiniGPT-4, Kosmos, and Gemini.
A reader-friendly guide to scaling AI models beyond the data pipeline—from training loops and distributed frameworks to checkpoints, mixed precision, and fault tolerance.
A collaborative 45-minute thinking algorithm tuned for Google-style coding interviews—classify the problem, co-design an optimal approach, code with confidence, and handle follow-ups with ease.
A deep dive into how datasets and dataloaders power modern AI—from the quiet pipeline that feeds models to the sophisticated tools that make training efficient. Understanding the hidden engine that keeps AI systems running.
A deep dive into XGBoost — how second-order Taylor approximations and sophisticated regularization make it the dominant algorithm for structured data, bridging mathematical rigor with system engineering excellence.
A clear introduction to diffusion and guided diffusion — how a simple physical process became a foundation for modern generative AI, from Stable Diffusion to robotics and protein design.
An intuitive introduction to the Transformer architecture — from the attention mechanism to self-attention and cross-attention, using language translation as a concrete example.
An intuitive introduction to Variational Autoencoders — how compressing data into probabilistic codes enables machines to generate realistic images, sounds, and structures.
Reflections on building production-grade behavior prediction systems at Zoox and Qualcomm — and why closed-loop reasoning is the bridge between perception and planning.
My research journey from wireless communication foundations to solving the camera calibration bottleneck that enables autonomous vehicle vision.
How we used deep learning to automatically calibrate traffic cameras by observing vehicle motion—work that won Best Paper Award at ACM BuildSys 2017.