ST-MoE-L 4.1B (fine-tuned)

1 papers with code • 0 benchmarks • 0 datasets

This task has no description! Would you like to contribute one?

Most implemented papers

ST-MoE: Designing Stable and Transferable Sparse Expert Models

tensorflow/mesh 17 Feb 2022

But advancing the state-of-the-art across a broad set of natural language tasks has been hindered by training instabilities and uncertain quality during fine-tuning.