AI Interview Prep

AI Interview Prep

Advanced Deep Learning Interview Questions #11 - The Bias-Weight Divergence Trap

When biases update but weights freeze, it exposes a forward-pass collapse that most engineers mistakenly attribute to gradient instability.

Hao Hoang's avatar
Hao Hoang
Apr 01, 2026
∙ Paid

You’re in a Senior ML Engineer interview at DeepMind. The interviewer sets a trap:

“During debugging, you notice your biases are updating rapidly, but your weight matrices are completely frozen, despite both sharing the exact same upstream gradient vector from the next layer. Looking at the isolated backprop equations for weight gradients versus bias gr…

User's avatar

Continue reading this post for free, courtesy of Hao Hoang.

Or purchase a paid subscription.
© 2026 Hao Hoang · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture