Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inquiry about fine-tuning speed on Dynamicrafter #7

Open
yujiangpu20 opened this issue Jul 25, 2024 · 1 comment
Open

Inquiry about fine-tuning speed on Dynamicrafter #7

yujiangpu20 opened this issue Jul 25, 2024 · 1 comment

Comments

@yujiangpu20
Copy link

Hi, Thank you for the excellent work!

I noticed that in Table 5 of the appendix of your paper, you provided the computation efficiency, mentioning that fine-tuning Dynamicrafter for 20k iterations on 8*A800 takes 8 hours (averaging about 1.4 seconds per iter). However, the comments in the train_512.yaml file indicate about 3 seconds per step. I was wondering what is the reason for this discrepancy.

Additionally, I tried to fine-tune on 8*A6000 (49G), and the monitor showed that one step takes about 7 seconds (batch_size=48). I would greatly appreciate if you could provide any suggestions for speeding this up.

@zhuhz22
Copy link
Collaborator

zhuhz22 commented Jul 27, 2024

Hi @yujiangpu20 ,
Thank you for your kind words! train_512.yaml was copied from the original DynamiCrafter repo, and the 3s comment was succeeded from it. Since our training configs were different from the original DynamiCrafter repo, this 3s comment isn't aplicable to our training process. We've deleted this comment from the file now.

As for speeding up, maybe the accelerate package ZeRO2 will help.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants