Published on July 9, 2018 by

The most popular optimization strategy in machine learning is called gradient descent. When gradient descent is applied to neural networks, its called back-propagation. In this video, i’ll use analogies, animations, equations, and code to give you an in-depth understanding of this technique. Once you feel comfortable with back-propagation, everything else becomes easier. It uses calculus to help us update our machine learning models. Enjoy!

Code for this video:

Please Subscribe! And like. And comment. That’s what keeps me going.

Want more education? Connect with me here:

This video is apart of my Machine Learning Journey course:

Join us in the Wizards Slack channel:

Sign up for the next course at The School of AI:

And please support me on Patreon:

Category Tag