Skip to content

Best practice with distributed training? #895

Answered by rwightman
songtianhui asked this question in Q&A
Discussion options

You must be logged in to vote

@songtianhui pretty much all models featured here that were trained with OpenCLIP are using --local-loss --gather-with-grad .. it's the only option that scales. Back when we first implemented it, we verified that w/ the gradient through gather, the local loss results were equivalence to doing the global loss.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by songtianhui
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants
Converted from issue

This discussion was converted from issue #894 on June 12, 2024 16:41.