Taming GPT-OSS 20B: LoRA's Wild Ride on OpenAI's MoE Beast
OpenAI drops a 20B MoE monster into open source, and suddenly fine-tuning isn't just for billion-dollar labs. One practitioner's gritty guide reveals LoRA hacks that make it feasible on everyday rigs.
⚡ Key Takeaways
- LoRA rank 32 with expert-targeted modules tames GPT-OSS 20B MoE efficiently.
- Freeze routers, use BF16 + gradient accum for single-GPU wins.
- Data curation trumps compute — hard negatives prevent MoE pitfalls.
Worth sharing?
Get the best AI stories of the week in your inbox — no noise, no spam.
Originally reported by Towards AI