Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to sync distribute model paramaters when training with continual learning fashion? #3421

Open
Iranb opened this issue Mar 5, 2025 · 0 comments

Comments

@Iranb
Copy link

Iranb commented Mar 5, 2025

When performing distributed continual learning tasks, it is common to expand model parameters as tasks increase. For example, I have defined an expand_classifier() method with random initialization to increase the parameters of the classifier.

How can I ensure that the newly added parameters are initialized the same on each GPU model?

If i do

if self.accelerator.is_main_process:
    self.model.module.prompt.expand_classifier()

How can i sync classifier across all distributed model?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant