Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add arctic model support by adding w2 to all_reduce #6856

Merged
merged 7 commits into from
Dec 18, 2024

Conversation

pi314ever
Copy link
Contributor

As title says.

Default behavior of arctic model produces shape issues with AutoTP due to the MLP layer performing w2 * act(w1*w3). However, method provided to fix Mixtral-7x8b in #5257 does not work since the MLP for Arctic is also used within a ModuleList for the MoE. This results in MLP weights hiding behind individual experts as layers #.w#, which is not caught by the fix in #5257. This adds the check directly within replace, where it can check for actual layer names for the w2 key in the model to patch with all_reduce.

@pi314ever pi314ever force-pushed the arctic-enabling-upstream branch from 2c2084b to 96eb813 Compare December 11, 2024 23:24
@pi314ever
Copy link
Contributor Author

@microsoft-github-policy-service agree company="Intel"

@loadams loadams requested review from jeffra and removed request for awan-10 December 12, 2024 00:38
@jeffra
Copy link
Collaborator

jeffra commented Dec 16, 2024

@RezaYazdaniAminabadi @sfc-gh-reyazda can you take a look?

@loadams loadams merged commit 0b25630 into microsoft:master Dec 18, 2024
10 of 11 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants