Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

No module 'DAdaptAdan' #770

Closed
SileNTViP opened this issue May 11, 2023 · 7 comments
Closed

No module 'DAdaptAdan' #770

SileNTViP opened this issue May 11, 2023 · 7 comments
Labels
enhancement New feature or request

Comments

@SileNTViP
Copy link

Replace 'DAdaptAdan' with 'DAdaptAdam'

@bmaltais
Copy link
Owner

We're you trying to train using the new DadapAdan?

@SileNTViP
Copy link
Author

Yes. AttributeError: module 'dadaptation' has no attribute 'DAdaptAdan'. Did you mean: 'DAdaptAdam'?

@bmaltais
Copy link
Owner

Look like the version of the dadaptation module installed does not support the new optimizers... I will test upgrading to the newer one and bring it in the next release if there are no issues with it.

@bmaltais bmaltais added the enhancement New feature or request label May 11, 2023
@bmaltais
Copy link
Owner

Look like the new dadaptation module is causing issues: RuntimeError: Setting different lr values in different parameter groups is only supported for values of 0

I will wait for kohya to sort out the issues with dadaptation before upgrading. Until then those new dadaptation methods will not work.

@idlebg
Copy link

idlebg commented May 11, 2023

pip(3) install -U dadaptation

or force.. here it did not update for some reason..

testing it with my upcoming Fusion AI model
image
image
image

impressive difference on some situations.... figuring out people and faces with Adan

will try to summarize as still investigating as huge difference on difference when playing additionally with the Scheduler
wdecay and the 3 betas for the Adan
image
from left to right
normal da . then only Adan with different betas and scheduler tuning

@bmaltais
Copy link
Owner

The v2.0 version no longer alow to set LoRA Text Encoder LR and Uner LR to be different... this is why I felt I should not upgrade it... Have you been training LoRA using Dadaptation with different LR of TE and UNet?

@idlebg
Copy link

idlebg commented May 11, 2023

nt LR of TE and UNet?

currently trying all at lr1
to figure out how it handles compared to regular v2 if LR is 1 for all.
also "weight_decay=0.01" "betas=0.99,0.9,0.99" vs
"weight_decay=0.02" "betas=0.98,0.92,0.99" "eps=1e-8"
and testing how adan works with all Schedulers and offcet noice variants.
so far without much knowledge, it gets easily 0.092 with def. instalation

give me an hour to have more results how it actually handles small and big sets as rates in the long run.

I started experimenting as ver 1 did NOT support multi GPU sharing out of the box without deep seed and additional code change.... while adafactor worked out of the box with a few a100 ("relative_step=True" "scale_parameter=True" )

Now eager test v2 new Added support for PyTorch's Fully Sharded Data Parallel.

But prior that testing a normal 4900 to see what it does.

Thanks for the heads up about the LR.
will move and update on all findings later on
so far it does not handle LR with more epoch repeats....
(above is more)
image

need some time to summarize and test if the shareded GPU training works now

bmaltais pushed a commit that referenced this issue Aug 20, 2023
update doc and minor fix
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants