-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
All tokens are out of vocabulary in 2.0 #1204
Comments
Thanks. I didn't update the You can fix these by writing to the attributes:
|
If I understand the suggestion correctly, it won't work in this case, I can't write to the attributes because it is precisely their correct values that I need. (For now I am just using Gensim with fastText-pretrained models for this functionality.) |
@bittlingmayer Sorry I missed this reply. I meant that you might get the values from the v1 model, and import them. Once you've saved out the model, the correct values will be there. The stable models will have the correct vocab data. This is an issue with the model files rather than a bug in the codebase, so I'll relabel this. |
Merging this with #1457, as it'll be part of the same fix! |
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |
vector
is fine, but theis_oov
bit is off.The text was updated successfully, but these errors were encountered: