FastText using pre-trained word vector for text classification

后端 未结 2 752
后悔当初
后悔当初 2021-02-20 03:05

I am working on a text classification problem, that is, given some text, I need to assign to it certain given labels.

I have tried using fast-text library by Facebook, w

2条回答
  •  面向向阳花
    2021-02-20 03:51

    FastText supervised training has -pretrainedVectors argument which can be used like this:

    $ ./fasttext supervised -input train.txt -output model -epoch 25 \
           -wordNgrams 2 -dim 300 -loss hs -thread 7 -minCount 1 \
           -lr 1.0 -verbose 2 -pretrainedVectors wiki.ru.vec
    

    Few things to consider:

    • Chosen dimension of embeddings must fit the one used in pretrained vectors. E.g. for Wiki word vectors is must be 300. It is set by -dim 300 argument.
    • As of mid-February 2018, Python API (v0.8.22) doesn't support training using pretrained vectors (the corresponding parameter is ignored). So you must use CLI (command line interface) version for training. However, a model trained by CLI with pretrained vectors can be loaded by Python API and used for predictions.
    • For large number of classes (in my case there were 340 of them) even CLI may break with an exception so you will need to use hierarchical softmax loss function (-loss hs)
    • Hierarchical softmax is worse in performance than normal softmax so it can give up all the gain you've got from pretrained embeddings.
    • The model trained with pretrained vectors can be several times larger than one trained without.
    • In my observation, the model trained with pretrained vectors gets overfitted faster than one trained without

提交回复
热议问题