I\'ve read the for a layer with hyperbolic tangent as its activation function, the weight initialization for that layer uses Xavier initialization. For ReLU, it uses He init