Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sync upstream/bach #5

Open
wants to merge 144 commits into
base: main
Choose a base branch
from
Open

Sync upstream/bach #5

wants to merge 144 commits into from

Conversation

bachvudinh
Copy link
Collaborator

@bachvudinh bachvudinh commented Nov 21, 2024

Issue: Sync Torchtune fork from upstream

Description:

This PR involves syncing the training code with the latest updates from the upstream repository to ensure our codebase remains up-to-date.
Tasks:

  • Pull the latest changes from the upstream repository.
  • Resolve any merge conflicts.
  • Test the synchronized code to ensure all functionality remains intact.
  • Update dependent files or configurations if necessary.

Expected Outcome:

The training code will be synchronized with the upstream repository, ensuring we are aligned with the latest features, improvements, and bug fixes.

Some Major new Feature

  • Qwen2.5 model family has landed (Qwen2.5 pytorch/torchtune#1863)

    • added builders for Qwen2.5, the cutting-edge models from the Qwen family of models! In their own words "Compared to Qwen2, Qwen2.5 has acquired significantly more knowledge (MMLU: 85+) and has greatly improved capabilities in coding (HumanEval 85+) and mathematics (MATH 80+)."
  • Change the Lr scheduler module path to torch.training.lr_scheduler

  • added model builders and configs for the 90B version of Llama3.2V

  • and fix some bug here for more details: https://github.com/pytorch/torchtune/releases.

Test plan

Running some training with Ichigo Llama3.1-s and MiniIchigo llama3.2-s. ALso do some testing relate to vram spike when do distributed training.

SalmanMohammadi and others added 30 commits October 2, 2024 12:57
Differential Revision: D64089998

Pull Request resolved: pytorch#1805
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.