Data Science Interview Question Challenge
Welcome to today's data science interview challenge! Here it goes:
1. When using gradient descent, why do we want to use a "batch" of examples, rather than one single example in the training dataset? What is a “mini-batch”?
2. Which derivatives need to be calculated first in order to update the weights through back propagation?
Keep reading with a 7-day free trial
Subscribe to The MLnotes Newsletter to keep reading this post and get 7 days of free access to the full post archives.