We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I have source and target files that are 15GB each and the system crashes after allocating all the 32GB RAM I have and the 2GB of swap disk.
I am now trying with 15 pairs of 1GB files...
Is there a way to tell XNMT how to train on very large files?
The text was updated successfully, but these errors were encountered:
I found an answer to my problem in the doc...
"sample_train_sents – If given, load a random subset of training sentences before each epoch. Useful when training data does not fit in memory."
I guess we need to read a large corpus by consecutive chunks as well, to make sure we cover the entire data set.
Sorry, something went wrong.
No branches or pull requests
I have source and target files that are 15GB each and the system crashes after allocating all the 32GB RAM I have and the 2GB of swap disk.
I am now trying with 15 pairs of 1GB files...
Is there a way to tell XNMT how to train on very large files?
The text was updated successfully, but these errors were encountered: