-
Notifications
You must be signed in to change notification settings - Fork 51
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0! #30
Comments
Traceback (most recent call last): |
I am running into the same issue when trying to finetune with LoRA on multiple GPUs. It works well if I apply LoRA only on |
I have a multi GPU setup with A100 40GB and I am getting the same problem. Here is the command I am using -
I have set up the WORLD_SIZE=8 as environment var. How do we solve this? This is preventing me from using this library to fine tune things. I tried to run using This throws CUDA OOM error... How can I run it using distributed settings? Please help |
Reduce the bs. However, is the multi gpu setting working? |
Will multi GPUs work has anyone been able to use this for 2 GPUs I ask because if 40B only requires 40GB of VRAM I would assume but could be wrong that 2x3090s or 2x4090s should work? |
how to solve it?
The text was updated successfully, but these errors were encountered: