Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Don't recompute scaling factor in activation checkpointing (#951)
Summary: Pull Request resolved: #951 Add the policy "layer_based_auto_wrap_policy_float8_training". It skips the recompute of float8 scaling factor (a scaler) to improve the latency. To enable it, change the config file like: P1690229394 Reviewed By: yoyoyocmu Differential Revision: D65360604 fbshipit-source-id: bd8c052fcf3c8af48775c08ef66a1e367397f09b
- Loading branch information