美文网首页
Lecture 13 | (2/5) Recurrent Neu

Lecture 13 | (2/5) Recurrent Neu

作者: Ysgc | 来源:发表于2019-11-02 09:17 被阅读0次

https://www.youtube.com/watch?v=jaw5W0bCgUQ

dont want NN to blow up -> require BIBO

what if we introduce nonlinear activation???

sigmoid quickly saturate.
tanh and relu blow up or shink to 0
relu similar to standard linear system

these curves reflect what the NN remember as time goes by.

what about the input is vector rather than scalar???

tanh remember things for a while, but eventually the information saturates

another problem with RNN

this is a problem for any deep NN

maximum gradients of sigmoid, tanh and relu are all = 1

相关文章

网友评论

      本文标题:Lecture 13 | (2/5) Recurrent Neu

      本文链接:https://www.haomeiwen.com/subject/ijhcbctx.html