How to reduce the inference time of Helsinki-NLP/opus-mt-es-en (translation model) from transformer

后端 未结 0 746
广开言路
广开言路 2021-01-03 00:31

Currently Helsinki-NLP/opus-mt-es-en model takes around 1.5sec for inference from transformer. How can that be reduced? Also when trying to convert it to onxx runtime gettin

相关标签:
回答
  • 消灭零回复
提交回复
热议问题