Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Isn’t the current best practice to train highly over-parametrized models to zero training error? That’d be a global optima, no?

Unless we’re talking about the optima of test error.



If you find a zero in a non negative function, I would call that a global minima, yes.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: