Yue
2018-11-07 05:24:09 UTC
Hello, I used the DBOW model without training word vectors, which
corresponds to the setting dm=0, dbow_words=0 in the model. After I trained
the model in this setting, what is happening when I call infer_vector on an
unseen document?
Was there some kind of back propagation going on that tries to maximize
some probability in the objective function (I used negative sampling)? If
so, why was infer_vector so quick given that it also goes through 30
iterations (I used 30 in training the model). I noticed that I got a vector
output immediately after I call infer_vector.
Also since the word vectors are left in their initial randomized state, are
we actually only mapping documents to their relative positions in space to
reflect their semantic relationship, while the word vectors don't reflect
any semantic relationship between the words? If so, what is making this
mode working well compared to the mixed mode where we also train the word
vectors simultaneously (dm=0, dbow_words=1)? I don't get how we could get
document vectors that reflect semantic relationship between documents
without having word vectors reflecting semantic relationship between words.
Thank you,
Yue
corresponds to the setting dm=0, dbow_words=0 in the model. After I trained
the model in this setting, what is happening when I call infer_vector on an
unseen document?
Was there some kind of back propagation going on that tries to maximize
some probability in the objective function (I used negative sampling)? If
so, why was infer_vector so quick given that it also goes through 30
iterations (I used 30 in training the model). I noticed that I got a vector
output immediately after I call infer_vector.
Also since the word vectors are left in their initial randomized state, are
we actually only mapping documents to their relative positions in space to
reflect their semantic relationship, while the word vectors don't reflect
any semantic relationship between the words? If so, what is making this
mode working well compared to the mixed mode where we also train the word
vectors simultaneously (dm=0, dbow_words=1)? I don't get how we could get
document vectors that reflect semantic relationship between documents
without having word vectors reflecting semantic relationship between words.
Thank you,
Yue
--
You received this message because you are subscribed to the Google Groups "Gensim" group.
To unsubscribe from this group and stop receiving emails from it, send an email to gensim+***@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
You received this message because you are subscribed to the Google Groups "Gensim" group.
To unsubscribe from this group and stop receiving emails from it, send an email to gensim+***@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.