Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

>an infinitely deep net (whatever that means) would not generalize on little data, and would get even harder to train the deeper it gets.

Not with proper priors/regularization.

>You've argued some ideas from intuition. There is little theoretical rigor around this, however.

There's this paper which goes more into theoretical depth on the idea: http://arxiv.org/abs/1412.0233



It goes into theoretical detail to show one fact, that local optima are close to the global optimum. It does not prove anything else.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: