You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
What steps will reproduce the problem?
1. Download attached text_simple train file
2. Compile word2vec.c as: gcc word2vec.c -o word2vec -lm -pthread
3. Run: ./word2vec -train text_simple -save-vocab vocab.txt
What is the expected output? What do you see instead?
Expect in saved vocab.txt file:
===============
</s> 0
and 12
the 11
four 10
in 8
used 5
war 5
one 5
nine 9
===============
What is really seen in the file
===============
</s> 0
and 12
the 11
four 10
in 8
used 5
war 5
one 5
===============
The last element "nine 5" wass missing.
What version of the product are you using? On what operating system?
MacOS, gcc version 4.2.1 (Based on Apple Inc. build 5658) (LLVM build
2336.11.00)
Please provide any additional information below.
This is NOT really a bug report because I am confused to understand the format
of train_file and how the vocab is constructed from it.
Based on the source code of word2vec.c, when reading from train_file, it will
1. insert </s> as the first element in vocab
2. scan each word (or </s> for newline) in train_file, add it to vocab, and
hash it in vocab_hash
So far the vocab_size = the number of words in vocab, INCLUDING </s> at the head
3. sort the words in vocab based on their counts, but keep </s> as the first of
vocab
Now the vocab_size because the number of words in vocab, EXCLUDING the leading
</s>. And if there is no newline character in train_file, </s> won't even be
hashed in vocab_hash
So there is a inconsistency here between vocab_size and the actual size of
vocab (including </s>). It could be a bug because later when the vocab is being
iterated, it is always done by iterating the elements from 0 to vocab_size-1,
like in SaveVocab(). This results in that the leading </s> will be saved, but
the last element in vocab will be ignored. At least that's what it looks with a
simple train file "text_simple" as attached here.
Original issue reported on code.google.com by [email protected] on 25 Aug 2013 at 2:38
The same confusion as in function CreateBinaryTree() in word2vec.c, where the
array representation of tree uses vocab_size*2+1 elements, which I understand
is essentially (len(vocab)-1)*2+1? That makes sense as only n-1 nodes are
needed for a full binary tree with n leaf nodes?
Many thanks if someone can clarify this a little bit.
Sorry the expected output of saved vocab.txt should be
===============
</s> 0
and 12
the 11
four 10
in 8
used 5
war 5
one 5
nine 5
===============
It is a typo in the last line
also looking for the reason. there are words that are missing from the trained
model, which are expected to be in the vocabulary since i put the min_count =
1.
(am working on CentOS)
Original issue reported on code.google.com by
[email protected]
on 25 Aug 2013 at 2:38Attachments:
The text was updated successfully, but these errors were encountered: