Linear SVM with general regularization
$$
\def\w{\mathbf{w}}
$$
Description
This package is an implementation of a linear svm solver with a wide
class of regularizations on the svm weight vector (l1, l2, mixed norm
l1-lq, adaptive lasso). We provide solvers for the
classical single task svm problem and for multi-task with joint
feature selection or similarity promoting term.
Note that this toolbox has been designed to be efficient for dense
data whereas most of the existing linear svm solvers have been
designed for sparse datasets.
This is the code that has been used for sensor selection in the paper:
R. Flamary, N. Jrad, R. Phlypo, M. Congedo, A. Rakotomamonjy, Mixed-Norm Regularization for Brain Decoding, Computational and Mathematical Methods in Medicine, Vol. 2014, N. 1, pp 1-13, 2014.
[Abstract]
[BibTeX]
[DOI]
[PDF]
[Slides]
[Code]
Abstract: This work investigates the use of mixed-norm regularization for sensor selection in event-related potential (ERP) based brain-computer interfaces (BCI). The classification problem is cast as a discriminative optimization framework where sensor selection is induced through the use of mixed-norms. This framework is extended to the multitask learning situation where several similar classification tasks related to different subjects are learned simultaneously. In this case, multitask learning helps in leveraging data scarcity issue yielding to more robust classifiers. For this purpose, we have introduced a regularizer that induces both sensor selection and classifier similarities. The different regularization approaches are compared on three ERP datasets showing the interest of mixed-norm regularization in terms of sensor selection. The multitask approaches are evaluated when a small number of learning examples are available yielding to significant performance improvements especially for subjects performing poorly.
BibTeX:
@article{flamary2014mixed,
author = {Flamary, R. and Jrad, N. and Phlypo, R. and Congedo, M. and Rakotomamonjy, A.},
title = {Mixed-Norm Regularization for Brain Decoding},
journal = {Computational and Mathematical Methods in Medicine},
volume = {2014},
number = {1},
pages = {1-13},
editor = {},
year = {2014}
}
Solver
We provide a general solver for squared hinge loss SVM of the form:
$$
\begin{equation*}
\min_{\w,b} \sum_{i=1}^{n} \max(0,1-y_i(\mathbf{x}_i^T\w+b))^2 + \Omega(\w)
\end{equation*}
$$
where $\Omega(\w)$ can be :
- l1 norm : $\Omega(\w)=\sum_i |w_i|$
- l2 norm (squared or not) : $\Omega(\w)=\sum_i |w_i|^2$
- l1-l2 mixed norm: $\Omega(\w)=\sum_g ||w_g||_2$ where $g$ denotes groups of features
- l1-lp mixed norm ($1