You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Verify the equivalence of model weights and outputs post-conversion. Something to look out for are discrepancies between the order of FFN, LayerNorm, and Dropout layers in Fast-LLM's GPT and OLMoE, i.e.
🧐 Problem Description
Fast-LLM doesn't yet support importing or exporting OLMoE models such as https://huggingface.co/allenai/OLMoE-1B-7B-0924.
💡 Proposed Solution
Add an OLMoE HF converter that offers both expert and import functionality:
Make it possible to export a Fast-LLM OLMoE-like model to HF's
OlmoeForCausalLM
format (see https://github.com/huggingface/transformers/blob/main/src/transformers/models/olmoe/modeling_olmoe.py).Load HF OLMoE models into Fast-LLM.
Verify the equivalence of model weights and outputs post-conversion. Something to look out for are discrepancies between the order of FFN, LayerNorm, and Dropout layers in Fast-LLM's GPT and OLMoE, i.e.
Fast-LLM/fast_llm/layers/transformer/transformer.py
Line 83 in 436d8d2
🔄 Alternatives Considered
It might be possible to export OLMoE-like models in HF Mixtral format.
📈 Potential Benefits
Allows for:
📝 Additional Context
The text was updated successfully, but these errors were encountered: