Why Gradient Descent for Optimization?


#1

Hello folks,

I have a question regarding the optimization technique used for updating the weights. People generally use gradient descent for the optimization whether its SGD or adaptive. Why can’t we use other techniques like Newton Raphson. Please help

Thanks in advance
Tarun Singh


#2

Hello,

Here you have a good explanation:

https://stackoverflow.com/questions/12066761/what-is-the-difference-between-gradient-descent-and-newtons-gradient-descent

Regards,
Carlos.