ShengYun (Anthony) Peng

Inference Compute-Optimal Video Vision Language Models

ACL, 2025

Peiqi Wang
Xuewen Zhang
Hanchao Yu
Yibo Yang
Lifu Huang
Fujun Liu
Qifan Wang

Abstract

Video Vision Language Models (VLMs) have achieved remarkable performance across a wide range of video tasks. While significant efforts have been made to enhance model performance, limited attention has been given to identifying the optimal model architecture under a fixed inference compute budget. This study explores how to best allocate inference compute across three critical scaling factors in video VLMs: language model size, frame count, and the number of visual tokens per frame. Through extensive large-scale training sweeps and careful parametric modeling of task performance, we identify the inference compute-optimal frontier. Our experimental results illuminate the relationship between task performance and scaling factors, as well as the impact of finetuning data size on this efficiency frontier. These insights provide practical guidance for selecting the optimal scaling factors in video vision language models.

IsoPerformance Contours. Contours show average task performance as a function of a scaling factor and finetuning data size n, derived from the star sweep. We construct the star sweep by starting with a inference compute-intensive "center" x, varying one factor at a time while keeping the others fixed, and finetuning on different data sizes. Performance improves as scaling factors and n increase, albeit at a diminishing rate.

BibTeX

			
@article{wang2025inference,
  title={Inference Compute-Optimal Video Vision Language Models},
  author={Peiqi Wang, ShengYun Peng, Xuewen Zhang, Hanchao Yu, Yibo Yang, Lifu Huang, Fujun Liu, Qifan Wang},
  journal={ACL},
  year={2025}
}