Search Results for author: Mincong Huang

Found 2 papers, 0 papers with code

Dynamic Activation Pitfalls in LLaMA Models: An Empirical Study

no code implementations15 May 2024 Chi Ma, Mincong Huang, Chao Wang, Yujie Wang, Lei Yu

In this work, we systematically investigate the efficacy of dynamic activation mechanisms within the LLaMA family of language models.

Re-evaluating the Memory-balanced Pipeline Parallelism: BPipe

no code implementations4 Jan 2024 Mincong Huang, Chao Wang, Chi Ma, Yineng Zhang, Peng Zhang, Lei Yu

Pipeline parallelism is an essential technique in the training of large-scale Transformer models.

Cannot find the paper you are looking for? You can Submit a new open access paper.