Training a Large Video Model on a Single Machine in a Day

28 Sep 2023  ยท  Yue Zhao, Philipp Krรคhenbรผhl ยท

Videos are big, complex to pre-process, and slow to train on. State-of-the-art large-scale video models are trained on clusters of 32 or more GPUs for several days. As a consequence, academia largely ceded the training of large video models to industry. In this paper, we show how to still train a state-of-the-art video model on a single machine with eight consumer-grade GPUs in a day. We identify three bottlenecks, IO, CPU, and GPU computation, and optimize each. The result is a highly efficient video training pipeline. For comparable architectures, our pipeline achieves higher accuracies with $\frac{1}{8}$ of the computation compared to prior work. Code is available at https://github.com/zhaoyue-zephyrus/AVION.

PDF Abstract

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Multi-Instance Retrieval EPIC-KITCHENS-100 Avion (ViT-L) mAP(V2T) 57.9 # 1
mAP(T2V) 51.1 # 1
mAP (Avg) 54.5 # 1
nDCG (V2T) 70.4 # 1
nDCG (T2V) 67.6 # 1
nDCG (Avg) 69.0 # 1
Action Recognition EPIC-KITCHENS-100 Avion (ViT-L) Action@1 54.4 # 1
Verb@1 73.0 # 1
Noun@1 65.4 # 2

Methods


No methods listed for this paper. Add relevant methods here