ExplorerData ScienceMachine Learning
Research PaperResearchia:202604.18065

Optimal algorithmic complexity of inference in quantum kernel methods

Elies Gil-fuster

Abstract

Quantum kernel methods are among the leading candidates for achieving quantum advantage in supervised learning. A key bottleneck is the cost of inference: evaluating a trained model on new data requires estimating a weighted sum $\sum_{i=1}^N α_i k(x,x_i)$ of $N$ kernel values to additive precision $\varepsilon$, where $α$ is the vector of trained coefficients. The standard approach estimates each term independently via sampling, yielding a query complexity of $O(N\lVertα\rVert_2^2/\varepsilon^2...

Submitted: April 18, 2026Subjects: Machine Learning; Data Science

Description / Details

Quantum kernel methods are among the leading candidates for achieving quantum advantage in supervised learning. A key bottleneck is the cost of inference: evaluating a trained model on new data requires estimating a weighted sum i=1Nαik(x,xi)\sum_{i=1}^N α_i k(x,x_i) of NN kernel values to additive precision ε\varepsilon, where αα is the vector of trained coefficients. The standard approach estimates each term independently via sampling, yielding a query complexity of O(Nα22/ε2)O(N\lVertα\rVert_2^2/\varepsilon^2). In this work, we identify two independent axes for improvement: (1) How individual kernel values are estimated (sampling versus quantum amplitude estimation), and (2) how the sum is approximated (term-by-term versus via a single observable), and systematically analyze all combinations thereof. The query-optimal combination, encoding the full inference sum as the expectation value of a single observable and applying quantum amplitude estimation, achieves a query complexity of O(α1/ε)O(\lVertα\rVert_1/\varepsilon), removing the dependence on NN from the query count and yielding a quadratic improvement in both α1\lVertα\rVert_1 and ε\varepsilon. We prove a matching lower bound of Ω(α1/ε)Ω(\lVertα\rVert_1/\varepsilon), establishing query-optimality of our approach up to logarithmic factors. Beyond query complexity, we also analyze how these improvements translate into gate costs and show that the query-optimal strategy is not always optimal in practice from the perspective of gate complexity. Our results provide both a query-optimal algorithm and a practically optimal choice of strategy depending on hardware capabilities, along with a complete landscape of intermediate methods to guide practitioners. All algorithms require only amplitude estimation as a subroutine and are thus natural candidates for early-fault-tolerant implementations.


Source: arXiv:2604.15214v1 - http://arxiv.org/abs/2604.15214v1 PDF: https://arxiv.org/pdf/2604.15214v1 Original Link: http://arxiv.org/abs/2604.15214v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
Apr 18, 2026
Topic:
Data Science
Area:
Machine Learning
Comments:
0
Bookmark
Optimal algorithmic complexity of inference in quantum kernel methods | Researchia