If we primarily use LSTMs over RNNs to solve the vanishing gradient problem, why can't we just use ReLUs/leaky ReLUs with RNNs instead?

后端 未结 0 1192
忘掉有多难
忘掉有多难 2021-01-18 13:46

We all knows that vanishing gradient problem occurs when we are using deep neural network with sigmoid and if we use relu , it solves this problem but it creates dead neuron

相关标签:
回答
  • 消灭零回复
提交回复
热议问题