On the Global Optima of Kernelized Adversarial Representation Learning

Loading...
Thumbnail Image

Date

Journal Title

Journal ISSN

Volume Title

Publisher

IEEE

Access Rights

info:eu-repo/semantics/openAccess

Abstract

Adversarial representation learning is a promising paradigm for obtaining data representations that are invariant to certain sensitive attributes while retaining the information necessary for predicting target attributes. Existing approaches solve this problem through iterative adversarial minimax optimization and lack theoretical guarantees. In this paper, we first study the linear form of this problem i.e., the setting where all the players are linear functions. We show that the resulting optimization problem is both non-convex and non-differentiable. We obtain an exact closed-form expression for its global optima through spectral learning and provide performance guarantees in terms of analytical bounds on the achievable utility and invariance. We then extend this solution and analysis to non-linear functions through kernel representation. Numerical experiments on UCI, Extended Yale B and CIFAR-100 datasets indicate that, (a) practically, our solution is ideal for imparting provable invariance to any biased pre-trained data representation, and (b) the global optima of the kernel form can provide a comparable trade-off between utility and invariance in comparison to iterative minimax optimization of existing deep neural network based approaches, but with provable guarantees.

Description

IEEE/CVF International Conference on Computer Vision (ICCV) -- OCT 27-NOV 02, 2019 -- Seoul, SOUTH KOREA

Keywords

Journal or Series

2019 Ieee/Cvf International Conference on Computer Vision (Iccv 2019)

WoS Q Value

Scopus Q Value

Volume

Issue

Citation

Endorsement

Review

Supplemented By

Referenced By