Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Gradient descent is just how neural networks (including auto-encoders) optimize parameters to minimize the loss function. They do this using derivatives to descend down the slope of the function. Autodiff is one way to compute the derivatives. Maybe we’re saying the same thing.


Yep, I was just asking Adam* to justify his loss function.

*pun intended :)




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: