-
Notifications
You must be signed in to change notification settings - Fork 282
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Avoid calling _free_fp16_param_shard() too early #1164
base: ngoyal_changes_for_pp_fp8
Are you sure you want to change the base?
Avoid calling _free_fp16_param_shard() too early #1164
Commits on Jan 12, 2024
-
Configuration menu - View commit details
-
Copy full SHA for a4f02ef - Browse repository at this point
Copy the full SHA a4f02efView commit details
Commits on Feb 21, 2024
-
Configuration menu - View commit details
-
Copy full SHA for f2bb56f - Browse repository at this point
Copy the full SHA f2bb56fView commit details
Commits on Mar 25, 2024
-
Added requires_grad check for params_with_grad method (#1171)
Co-authored-by: Jie Wang <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 1307b1d - Browse repository at this point
Copy the full SHA 1307b1dView commit details
Commits on Apr 1, 2024
-
Changed to only run reshard hook if all gradients computed (#1166)
* Changed to only run reshard hook if all gradients computed * Fix decreasing it/s with multi-grad hook
Configuration menu - View commit details
-
Copy full SHA for 5faca97 - Browse repository at this point
Copy the full SHA 5faca97View commit details
Commits on Apr 5, 2024
-
Add cast input argument (#1175)
Co-authored-by: Jie Wang <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 7bcbc80 - Browse repository at this point
Copy the full SHA 7bcbc80View commit details
Commits on May 16, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 9cbb4a7 - Browse repository at this point
Copy the full SHA 9cbb4a7View commit details
Commits on Jun 10, 2024
-
[FSDPv1] Optimize memory usage for optimize_backward_concat=True (#1186)
* optimize memory * clean up * allocate GPU memory directly * add comment * rename fp32_grads to fp32_flat_grad * update doc
Configuration menu - View commit details
-
Copy full SHA for b73fffe - Browse repository at this point
Copy the full SHA b73fffeView commit details