Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fsdp 2.0 integration #206

Draft
wants to merge 44 commits into
base: main
Choose a base branch
from
Draft

Fsdp 2.0 integration #206

wants to merge 44 commits into from

Conversation

le1nux
Copy link
Member

@le1nux le1nux commented Aug 4, 2024

What does this PR do?

This PR ..

General Changes

  • ..

Breaking Changes

  • ..

TODOs

  • Need to check if we want to use tied weights for the embedding and LM head and how this would be parallelized with FSDP and tensor parallelism

Checklist before submitting final PR

  • My PR is minimal and addresses one issue in isolation
  • I have merged the latest version of the target branch into this feature branch
  • I have reviewed my own code w.r.t. correct implementation, missing type hints, proper documentation, etc.
  • I have run a sample config for model training
  • I have checked that all tests run through (python tests/tests.py)

@le1nux le1nux added the enhancement New feature or request label Aug 23, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant