We prove bounds on the population risk of the maximum margin algorithm for
two-class linear classification. For linearly separable training data, the
maximum margin algorithm has been shown in previous work to be equivalent to a
limit of training with logistic loss using gradient descent, as the training
error is driven to zero. We analyze this algorithm applied to random data
including misclassification noise. Our assumptions on the clean data include
the case in which the class-conditional distributions are standard normal
distributions. The misclassification noise may be chosen by an adversary,
subject to a limit on the fraction of corrupted labels. Our bounds show that,
with sufficient over-parameterization, the maximum margin algorithm trained on
noisy data can achieve nearly optimal population risk.
Description
[2004.12019] Finite-sample analysis of interpolating linear classifiers in the overparameterized regime
%0 Journal Article
%1 chatterji2020finitesample
%A Chatterji, Niladri S.
%A Long, Philip M.
%D 2020
%K adversarial deep-learning max-margin noise readings theory
%T Finite-sample analysis of interpolating linear classifiers in the
overparameterized regime
%U http://arxiv.org/abs/2004.12019
%X We prove bounds on the population risk of the maximum margin algorithm for
two-class linear classification. For linearly separable training data, the
maximum margin algorithm has been shown in previous work to be equivalent to a
limit of training with logistic loss using gradient descent, as the training
error is driven to zero. We analyze this algorithm applied to random data
including misclassification noise. Our assumptions on the clean data include
the case in which the class-conditional distributions are standard normal
distributions. The misclassification noise may be chosen by an adversary,
subject to a limit on the fraction of corrupted labels. Our bounds show that,
with sufficient over-parameterization, the maximum margin algorithm trained on
noisy data can achieve nearly optimal population risk.
@article{chatterji2020finitesample,
abstract = {We prove bounds on the population risk of the maximum margin algorithm for
two-class linear classification. For linearly separable training data, the
maximum margin algorithm has been shown in previous work to be equivalent to a
limit of training with logistic loss using gradient descent, as the training
error is driven to zero. We analyze this algorithm applied to random data
including misclassification noise. Our assumptions on the clean data include
the case in which the class-conditional distributions are standard normal
distributions. The misclassification noise may be chosen by an adversary,
subject to a limit on the fraction of corrupted labels. Our bounds show that,
with sufficient over-parameterization, the maximum margin algorithm trained on
noisy data can achieve nearly optimal population risk.},
added-at = {2020-05-03T04:14:52.000+0200},
author = {Chatterji, Niladri S. and Long, Philip M.},
biburl = {https://www.bibsonomy.org/bibtex/2148e96e693abd437ec16690abf91f9c4/kirk86},
description = {[2004.12019] Finite-sample analysis of interpolating linear classifiers in the overparameterized regime},
interhash = {17b2899c8da5568bd992d3b25292d677},
intrahash = {148e96e693abd437ec16690abf91f9c4},
keywords = {adversarial deep-learning max-margin noise readings theory},
note = {cite arxiv:2004.12019},
timestamp = {2020-05-03T04:14:52.000+0200},
title = {Finite-sample analysis of interpolating linear classifiers in the
overparameterized regime},
url = {http://arxiv.org/abs/2004.12019},
year = 2020
}