Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Textual Inversion's dlr0 diverges to inf #222

Closed
and1011 opened this issue Feb 22, 2023 · 2 comments
Closed

Textual Inversion's dlr0 diverges to inf #222

and1011 opened this issue Feb 22, 2023 · 2 comments

Comments

@and1011
Copy link

and1011 commented Feb 22, 2023

I didn't test the algorithm before implementation of d-adaptation, but now, when I try to train a textual inversion, dlr0 starts low, which I suppose is correct anyway, but quickly starts increasing (exponentially) with each step until it reaches the float limit.
Also, after the value diverges to inf the embedding becomes unusable.
If I understood correctly, what could be wrong, but in theory shouldn't the dlr converge to some value?
Anyone else having this problem or who could help?

@bmaltais
Copy link
Owner

This is usually because the LR is too high... You could try to lower it a bit.

@and1011
Copy link
Author

and1011 commented Feb 23, 2023

I saw here that you fixed it in this commit 34ab844

Now that I've updated this bug has been fixed, I can now use AdamW normally.
Thank you.

@and1011 and1011 closed this as completed Feb 24, 2023
Cauldrath pushed a commit to Cauldrath/kohya_ss that referenced this issue Apr 5, 2023
fix training instability issue, add metadata
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants