Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat(fp8): [Work In Progress] enable FP8 training #369

Open
wants to merge 6 commits into
base: develop
Choose a base branch
from

Conversation

zigzagcai
Copy link
Collaborator

@zigzagcai zigzagcai commented Nov 6, 2024

Motivation

Try to enable FP8 to speedup training on Hooper platform, via torchao.

Modification

internlm/core
internlm/quantization

BC-breaking (Optional)

None

Use cases (Optional)

None

Checklist

Before PR:

  • Pre-commit or other linting tools are used to fix the potential lint issues.
  • Bug fixes are fully covered by unit tests, the case that causes the bug should be added in the unit tests.
  • The modification is covered by complete unit tests. If not, please add more unit test to ensure the correctness.
  • The documentation has been modified accordingly, like docstring or example tutorials.

After PR:

  • If the modification has potential influence on downstream or other related projects, this PR should be tested with those projects.
  • CLA has been signed and all committers have signed the CLA in this PR.

@zigzagcai zigzagcai changed the title feat(fp8): enable integration with FP8 library feat(fp8): [Work In Progress] enable FP8 training Nov 6, 2024
@zigzagcai
Copy link
Collaborator Author

zigzagcai commented Nov 7, 2024

Current Status:

  • Workable with FP8 dynamic scaling + torch.compile, delivering 10%~15% throughput speedup. In the test case of 7B InternLM1 model (layers was reduced from 32 to 16, to make it runnable on 1 node with 2x GPUs) on 2x H100 GPUs
  • Workable with FP8 delayed scaling (input, weight) + torch.compile, also delivering 10%~15% throughput speedup
  • Meet compilation error with FP8 delayed scaling (input, weight, grad) + torch.compile, seems to be some error with torch compiler in the backward pass of all_reduce amax values, currently waiting for the response from torchao developers or trying on my own to solve this issue.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants