-
Notifications
You must be signed in to change notification settings - Fork 7.1k
[model] add gpt oss #8826
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[model] add gpt oss #8826
Conversation
what is the weight precision after finetune? is it still mxfp4 for the moe layer? |
When will full-parameter fine-tuning be supported? |
@ziheng0924 full finetuning is supported |
@yuimo the lora weights will be fp32 format |
VLLM推理什么时候支持呢? |
Hi, I just saw this PR to support gpt-oss. IIUC, with this recipe, only the Generally, targeting just the MoE layers may be fine, but the majority of parameters reside in the MoE layers (90% for 20b), so users may want to target those too. On the other hand, this will be much more expensive memory-wise, so there is a trade-off here. If you have questions about the new PEFT feature, LMK. |
@BenjaminBossan Sure, I agree with you. I'm excited to dive into exploring these new PEFT features. Thank you for pointing it out. |
vLLM doesn't support bf16 ckpt yet |
it will be bf16, if triton < 3.4.0 |
After finetuning the gpt-oss model, I have a BF16 weight. Is there any tool to create the MXFP4 weight from BF16 weight? |
+1, is there any method to convert a trained SFT model in bf16 to mxfp4, or to directly train using native mxfp4? |
微调gpt-oss-20b显存占用大概是什么样子呢,在A100上占用大概为50,单是为什么4*3090 24G会oom |
微调gpt-oss-20b看起来好像不支持 linger-kernel,input context length比较长,内存会爆掉,请问有解决办法来支持linger-kernel吗 |
请问gpt-oss-120b全参数微调最少要多少卡? |
Does this allow us to train with the reasoning content within the prompt? Or do we have to process the prompt to allow the gpt template to take in reasoning content as well? |
Apply LoRA Fine-Tuning on GPT-OSS model in 3 steps
1. Install LlamaFactory and transformers
2. Train GPT-OSS on a single GPU (> 44GB) (multi-GPU is also supported)
3. Merge the LoRA weight into the base model
llamafactory-cli export --model_name_or_path openai/gpt-oss-20b --adapter_name_or_path saves/gpt-20b/lora/sft --export_dir gpt_merged
(Optional) Chat with the fine-tuned model
Full fine-tuning recipes
See #8837
Use Web UI to fine-tune the model:
