ExplorerArtificial IntelligenceAI
Research PaperResearchia:202605.05001

SpecKV: Adaptive Speculative Decoding with Compression-Aware Gamma Selection

Shikhar Shukla

Abstract

Speculative decoding accelerates large language model (LLM) inference by using a small draft model to propose candidate tokens that a larger target model verifies. A critical hyperparameter in this process is the speculation length~$γ$, which determines how many tokens the draft model proposes per step. Nearly all existing systems use a fixed~$γ$ (typically~4), yet empirical evidence suggests that the optimal value varies across task types and, crucially, depends on the compression level applied...

Submitted: May 5, 2026Subjects: AI; Artificial Intelligence

Description / Details

Speculative decoding accelerates large language model (LLM) inference by using a small draft model to propose candidate tokens that a larger target model verifies. A critical hyperparameter in this process is the speculation length~γγ, which determines how many tokens the draft model proposes per step. Nearly all existing systems use a fixed~γγ (typically4), yet empirical evidence suggests that the optimal value varies across task types and, crucially, depends on the compression level applied to the target model. In this paper, we present \textbf{SpecKV}, a lightweight adaptive controller that selectsγγ per speculation step using signals extracted from the draft model itself. We profile speculative decoding across 4task categories, 4speculation lengths, and 3compression levels (FP16, INT8, NF4), collecting 5,112 step-level records with per-step acceptance rates, draft entropy, and draft confidence. We demonstrate that the optimalγγ shifts across compression regimes and that draft model confidence and entropy are strong predictors of acceptance rate (correlation~0.56\approx 0.56). SpecKV uses a small MLP trained on these signals to maximize expected tokens per speculation step, achieving a 56.0% improvement over the fixed-γγ=4 baseline with only 0.34,ms overhead per decision (<<0.5% of step time). The improvement is statistically significant (p<0.001p < 0.001, paired bootstrap test). We release all profiling data, trained models, and notebooks as open-source artifacts.


Source: arXiv:2605.02888v1 - http://arxiv.org/abs/2605.02888v1 PDF: https://arxiv.org/pdf/2605.02888v1 Original Link: http://arxiv.org/abs/2605.02888v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
May 5, 2026
Topic:
Artificial Intelligence
Area:
AI
Comments:
0
Bookmark
SpecKV: Adaptive Speculative Decoding with Compression-Aware Gamma Selection | Researchia