As the title states: Is a countvectorizer
the same as tfidfvectorizer
with use_idf=false ? If not why not ?
So does this also mean that adding
No, they're not the same. TfidfVectorizer
normalizes its results, i.e. each vector in its output has norm 1:
>>> CountVectorizer().fit_transform(["foo bar baz", "foo bar quux"]).A
array([[1, 1, 1, 0],
[1, 0, 1, 1]])
>>> TfidfVectorizer(use_idf=False).fit_transform(["foo bar baz", "foo bar quux"]).A
array([[ 0.57735027, 0.57735027, 0.57735027, 0. ],
[ 0.57735027, 0. , 0.57735027, 0.57735027]])
This is done so that dot-products on the rows are cosine similarities. Also TfidfVectorizer
can use logarithmically discounted frequencies when given the option sublinear_tf=True
.
To make TfidfVectorizer
behave as CountVectorizer
, give it the constructor options use_idf=False, normalize=None
.