Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FLUX finetuning speedup #2942

Open
eftSharptooth opened this issue Nov 2, 2024 · 0 comments
Open

FLUX finetuning speedup #2942

eftSharptooth opened this issue Nov 2, 2024 · 0 comments

Comments

@eftSharptooth
Copy link

Just looking for other users (primarily docker) to confirm these changes.

Environment: Docker on linux, Cuda 12.4 on base OS.
finetuning 1152 px images went from 10.6s/it to 6.4s/it by bumping versions to more closely match sd-scripts current versions. This also resulted in an almost 11GB drop in VRAM usage (I believe during the backwards pass). I changed no settings in-between the tests of s/it and VRAM.

changes:
torch to 2.4.0
torchvision to 0.19.0
xformers to whatever is more recent and compatible. In my case it was 0.0.27.post2 auto selected.
transformers to 4.36.2

command to run in docker container: pip install -U --extra-index-url https://download.pytorch.org/whl/cu121 --extra-index-url https://pypi.nvidia.com torch==2.4.0 torchvision==0.19.0 xformers transformers==4.36.2

Just wanted to see if it worked similarly for others without causing problems before submitting a PR to bump the version in linux-docker and Dockerfile.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant