Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

problem about batch_size and gradient_accumulation_steps #218

Open
ShowLo opened this issue Nov 5, 2024 · 0 comments
Open

problem about batch_size and gradient_accumulation_steps #218

ShowLo opened this issue Nov 5, 2024 · 0 comments

Comments

@ShowLo
Copy link

ShowLo commented Nov 5, 2024

In the paper, it's show that MuseTalk training was performed on 2 NVIDIA H20 GPUs, and the Unet model was initially trained with L1 and perceptual losses for 200,000 steps. However, the paper doesn't specify the batch_size and gradient_accumulation_steps, which impact training speed. Could you provide the specific numbers used?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant