-
Notifications
You must be signed in to change notification settings - Fork 508
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Updated training? #37
Comments
I would rephrase the question as follows: when initializing from a checkpoint, what parameters/data can be changed, and what parameters will/must stay the same? |
When using In theory you could use a new dataset when training with |
I noticed, that "checkpoint_every" also does not change. |
Just so I'm clear then, the current purpose of |
Yes that's correct; the learning rate and learning rate decay could be On Thursday, March 17, 2016, Simon Elvery [email protected] wrote:
|
Just noticed, that the letters/digits may be reassigned (idx_to_token) after subsequent running preprocess.py (unless I miss/confuse something else). I do not have the previous version of .json file (will save it next time), but now I'm getting lots of "don1t"s and "it1s" instead of "don't" ad "it's" in the output. So it looks like "1" took the place of "'", "'" took the place of "!", etc. Can the tokens be lexicographically ordered by preprocess.py, so such issues can be avoided? |
I would appreciate it if the data generation were deterministic. When your RNN takes weeks to train and you decide you need to change something in the data, it'd be nice if you didn't have to start over. |
We are working on a new accessory script that can encode new data using a JSON schema from a previous dataset to address this. Going full deterministic can be really slow and cause lots of problems, so this is the next best thing. |
I'm very new to the whole world of neural networks so please forgive any silly questions.
Is there a way to update a network by training it on new text? The undocumented
-init_from
flag looks like it might do that, but I can't quite be sure.The text was updated successfully, but these errors were encountered: