Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Growing parameter capacity as training progress #1539

Merged
merged 1 commit into from
Sep 22, 2023

Conversation

emailweixu
Copy link
Contributor

This is done through Optimizer. Two arguments are added for optimizer:

capacity_ratio: scheduler for controlling the number of training elements of a parameter.
min_capacity: minimal number elements of each parameter being traing

To dynamically change capacity, we assign a random number for each element of the parameter. An element is turned on if its assigned random number is less than capacity_ratio. To save memory, we don't store the random numbers. Instead, we save the random number generator state.

This is done through Optimizer. Two arguments are added for optimizer:

capacity_ratio: scheduler for controlling the number of training elements of a parameter.
min_capacity: minimal number elements of each parameter being traing

To dynamically change capacity, we assign a random number for each element of
the parameter. An element is turned on if its assigned random number
is less than capacity_ratio. To save memory, we don't store the
random numbers. Instead, we save the random number generator state.
@emailweixu emailweixu merged commit 61f1565 into pytorch Sep 22, 2023
2 checks passed
@emailweixu emailweixu deleted the PR_dynamic_capacity branch September 22, 2023 01:00
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants