ExplorerComputer VisionComputer Vision
Research PaperResearchia:202603.13006

EVATok: Adaptive Length Video Tokenization for Efficient Visual Autoregressive Generation

Tianwei Xiong

Abstract

Autoregressive (AR) video generative models rely on video tokenizers that compress pixels into discrete token sequences. The length of these token sequences is crucial for balancing reconstruction quality against downstream generation computational cost. Traditional video tokenizers apply a uniform token assignment across temporal blocks of different videos, often wasting tokens on simple, static, or repetitive segments while underserving dynamic or complex ones. To address this inefficiency, we...

Submitted: March 13, 2026Subjects: Computer Vision; Computer Vision

Description / Details

Autoregressive (AR) video generative models rely on video tokenizers that compress pixels into discrete token sequences. The length of these token sequences is crucial for balancing reconstruction quality against downstream generation computational cost. Traditional video tokenizers apply a uniform token assignment across temporal blocks of different videos, often wasting tokens on simple, static, or repetitive segments while underserving dynamic or complex ones. To address this inefficiency, we introduce EVATok\textbf{EVATok}, a framework to produce E\textbf{E}fficient V\textbf{V}ideo A\textbf{A}daptive Tok\textbf{Tok}enizers. Our framework estimates optimal token assignments for each video to achieve the best quality-cost trade-off, develops lightweight routers for fast prediction of these optimal assignments, and trains adaptive tokenizers that encode videos based on the assignments predicted by routers. We demonstrate that EVATok delivers substantial improvements in efficiency and overall quality for video reconstruction and downstream AR generation. Enhanced by our advanced training recipe that integrates video semantic encoders, EVATok achieves superior reconstruction and state-of-the-art class-to-video generation on UCF-101, with at least 24.4% savings in average token usage compared to the prior state-of-the-art LARP and our fixed-length baseline.


Source: arXiv:2603.12267v1 - http://arxiv.org/abs/2603.12267v1 PDF: https://arxiv.org/pdf/2603.12267v1 Original Link: http://arxiv.org/abs/2603.12267v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
Mar 13, 2026
Topic:
Computer Vision
Area:
Computer Vision
Comments:
0
Bookmark
EVATok: Adaptive Length Video Tokenization for Efficient Visual Autoregressive Generation | Researchia