Duality for the Adversarial Total Variation
Abstract
Adversarial training of binary classifiers can be reformulated as regularized risk minimization involving a nonlocal total variation. Building on this perspective, we establish a characterization of the subdifferential of this total variation using duality techniques. To achieve this, we derive a dual representation of the nonlocal total variation and a related integration of parts formula, involving a nonlocal gradient and divergence. We provide such duality statements both in the space of cont...
Description / Details
Adversarial training of binary classifiers can be reformulated as regularized risk minimization involving a nonlocal total variation. Building on this perspective, we establish a characterization of the subdifferential of this total variation using duality techniques. To achieve this, we derive a dual representation of the nonlocal total variation and a related integration of parts formula, involving a nonlocal gradient and divergence. We provide such duality statements both in the space of continuous functions vanishing at infinity on proper metric spaces and for the space of essentially bounded functions on Euclidean domains. Furthermore, under some additional conditions we provide characterizations of the subdifferential in these settings.
Source: arXiv:2604.18540v1 - http://arxiv.org/abs/2604.18540v1 PDF: https://arxiv.org/pdf/2604.18540v1 Original Link: http://arxiv.org/abs/2604.18540v1
Please sign in to join the discussion.
No comments yet. Be the first to share your thoughts!
Apr 21, 2026
Mathematics
Mathematics
0