TY - JOUR
T1 - On the Precise Error Analysis of Support Vector Machines
AU - Kammoun, Abla
AU - Alouini, Mohamed-Slim
N1 - KAUST Repository Item: Exported on 2021-02-04
PY - 2021
Y1 - 2021
N2 - This paper investigates the asymptotic behavior of the soft-margin and hard-margin support vector machine (SVM) classifiers for simultaneously high-dimensional and numerous data (large n and large $p$ with $n/p\to\delta$) drawn from a Gaussian mixture distribution. Sharp predictions of the classification error rate of the hard-margin and soft-margin SVM are provided, as well as asymptotic limits of as such important parameters as the margin and the bias. As a further outcome, the analysis allows for the identification of the maximum number of training samples that the hard-margin SVM is able to separate. The precise nature of our results allows for an accurate performance comparison of the hard-margin and soft-margin SVM as well as a better understanding of the involved parameters (such as the number of measurements and the margin parameter) on the classification performance. Our analysis, confirmed by a set of numerical experiments, builds upon the convex Gaussian min-max Theorem, and extends its scope to new problems never studied before by this framework.
AB - This paper investigates the asymptotic behavior of the soft-margin and hard-margin support vector machine (SVM) classifiers for simultaneously high-dimensional and numerous data (large n and large $p$ with $n/p\to\delta$) drawn from a Gaussian mixture distribution. Sharp predictions of the classification error rate of the hard-margin and soft-margin SVM are provided, as well as asymptotic limits of as such important parameters as the margin and the bias. As a further outcome, the analysis allows for the identification of the maximum number of training samples that the hard-margin SVM is able to separate. The precise nature of our results allows for an accurate performance comparison of the hard-margin and soft-margin SVM as well as a better understanding of the involved parameters (such as the number of measurements and the margin parameter) on the classification performance. Our analysis, confirmed by a set of numerical experiments, builds upon the convex Gaussian min-max Theorem, and extends its scope to new problems never studied before by this framework.
UR - http://hdl.handle.net/10754/662462
UR - https://ieeexplore.ieee.org/document/9324992/
U2 - 10.1109/OJSP.2021.3051849
DO - 10.1109/OJSP.2021.3051849
M3 - Article
SN - 2644-1322
SP - 1
EP - 1
JO - IEEE Open Journal of Signal Processing
JF - IEEE Open Journal of Signal Processing
ER -