-
Notifications
You must be signed in to change notification settings - Fork 131
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to use fastText for out of sample words? #26
Comments
Facebook distribute 2 type of files:
next time please ask in mailing list mailing list |
@menshikh-iv is this clear from our documentation? I see people confused about these formats, how to load them and what can be done with them, all the time. A clear, authoritative docs section would help us with support too (just point with hyperlink). |
@piskvorky I agree this situation happens sometimes, it worth to make a tutorial. |
A tutorial would be ideal, but a simple paragraph in the docs would go a long way. Can you add it? |
This is not working for me with gensim 3.5, python 3.6, and a FB model:
I get: |
@scottlittle please read an exception again: we really don't support supervised fasttext models |
What is meant by supervised fasttext models and how to train for unsupervised? |
Exactly what supervised learning means. FB implementation have supervised-mode support (gensim - only unsupervised)
Just read an Gensim FastText documentation |
When downloading fastText with this method, we get a folder with a file in standard word2vec format, which can be loaded with
model = KeyedVectors.load_word2vec_format(path, binary=False)
But not with
from gensim.models import FastText
model = FastText.load_fasttext_format(path, binary=False)
This disables the ability to get vectors for out-of-vocabulary words.
How can this be done correcly?
The text was updated successfully, but these errors were encountered: