This StatQuest picks up right here Part 1 left off, and this time we're going to go totally bonkers with The Chain Rule and optimize every single parameter in this simple Neural Network. BAM!!!
NOTE: This StatQuest assumes that you already know the main ideas behind Backpropagation: • Neural Networks Pt. 2:...
...and that also means you should be familiar with...
Neural Networks: • The Essential Main Ide...
The Chain Rule: • The Chain Rule
Gradient Descent: • Gradient Descent, Step...
LAST NOTE: When I was researching this 'Quest, I found this page by Sebastian Raschka to be helpful: sebastianraschka.com/faq/docs...
For a complete index of all the StatQuest videos, check out:
statquest.org/video-index/
If you'd like to support StatQuest, please consider...
Buying my book, The StatQuest Illustrated Guide to Machine Learning:
PDF - statquest.gumroad.com/l/wvtmc
Paperback - www.amazon.com/dp/B09ZCKR4H6
Kindle eBook - www.amazon.com/dp/B09ZG79HXC
Patreon: / statquest
...or...
KZitem Membership: / @statquest
...a cool StatQuest t-shirt or sweatshirt:
shop.spreadshirt.com/statques...
...buying one or two of my songs (or go large and get a whole album!)
joshuastarmer.bandcamp.com/
...or just donating to StatQuest!
www.paypal.me/statquest
Lastly, if you want to keep up with me as I research and create new StatQuests, follow me on twitter:
/ joshuastarmer
0:00 Awesome song and introduction
1:28 The derivative of the weight W1
5:58 The derivative of the bias b1
7:39 The derivatives of W2 and b2
9:21 Gradient Descent for all parameters
11:18 Fancy Gradient Descent Animation
#StatQuest #NeuralNetworks #Backpropagation
Негізгі бет Backpropagation Details Pt. 2: Going bonkers with The Chain Rule
Пікірлер: 481