Does sample= 0
in Gensim word2vec mean that no downsampling is being used during my training? The documentation says just that
"useful range is (0, 1e-5)"
However putting the threshold to 0 would cause P(wi) to be equal to 1, meaning that no word would be discarded, am I understanding it right or not?
I'm working on a relatively small dataset of 7597 Facebook posts (18945 words) and my embeddings perform far better using sample= 0
rather than anything else within the recommended range. Is there any particular reason? Text size?