Why are the predicitions so bad even if the model's loss is so low?

后端 未结 0 1948
余生分开走
余生分开走 2020-12-08 04:19

I was training a Transformer model to convert English sentences to German. After training it for not even for 1 epoch, the loss went down to 0.009. This was

相关标签:
回答
  • 消灭零回复
提交回复
热议问题