Weight initialization for sigmoid function

前端 未结 0 739
隐瞒了意图╮
隐瞒了意图╮ 2020-11-21 19:44

I\'ve read the for a layer with hyperbolic tangent as its activation function, the weight initialization for that layer uses Xavier initialization. For ReLU, it uses He init

相关标签:
回答
  • 消灭零回复
提交回复
热议问题