Fine-tune Bert for specific domain (unsupervised)

前端 未结 1 1858
没有蜡笔的小新
没有蜡笔的小新 2021-01-21 14:38

I want to fine-tune BERT on texts that are related to a specific domain (in my case related to engineering). The training should be unsupervised since I don\'t have any labels o

相关标签:
1条回答
  • 2021-01-21 15:09

    What you in fact want to is continue pre-training BERT on text from your specific domain. What you do in this case is to continue training the model as masked language model, but on your domain-specific data.

    You can use the run_mlm.py script from the Huggingface's Transformers.

    0 讨论(0)
提交回复
热议问题