http://www.cnr.it/ontology/cnr/individuo/prodotto/ID52553
Data representations and generalization error in kernel based learning machines (Articolo in rivista)
- Type
- Label
- Data representations and generalization error in kernel based learning machines (Articolo in rivista) (literal)
- Anno
- 2006-01-01T00:00:00+01:00 (literal)
- Alternative label
- Http://www.cnr.it/ontology/cnr/pubblicazioni.owl#autori
- N. Ancona, R. Maglietta and E. Stella (literal)
- Pagina inizio
- Pagina fine
- Http://www.cnr.it/ontology/cnr/pubblicazioni.owl#numeroVolume
- Rivista
- Note
- ISI Web of Science (WOS) (literal)
- Http://www.cnr.it/ontology/cnr/pubblicazioni.owl#affiliazioni
- Istituto di Studi sui Sistemi Intelligenti per l'Automazione--C.N.R., Via Amendola 122/D-I--70126 Bari, Italy. (literal)
- Titolo
- Data representations and generalization error in kernel based learning machines (literal)
- Abstract
- This paper focuses on the problem of how data representation
influences the generalization error of kernel based learning
machines like Support Vector Machines (SVM) for classification.
Frame theory provides a well founded mathematical framework for
representing data in many different ways. We analyze the effects
of sparse and dense data representations on the generalization
error of such learning machines measured by using leave-one-out
error given a finite amount of training data. We show that, in the
case of sparse data representations, the generalization error of
an SVM trained by using polynomial or Gaussian kernel functions is
equal to the one of a linear SVM. This is equivalent to saying
that the capacity of separating points of functions belonging to
hypothesis spaces induced by polynomial or Gaussian kernel
functions reduces to the capacity of a separating hyperplane in
the input space. Moreover we show that, in general, sparse data
representations increase or leave unchanged the generalization
error of kernel based methods.
%as long as the representation is not too sparse, as in the case of
%very large dictionaries. Very sparse representations increase
%drastically the generalization error of kernel based methods.
Dense data representations, on the contrary, reduce the
generalization error in the case of very large frames. We use two
different schemes for representing data in overcomplete systems of
Haar and Gabor functions, and measure SVM generalization error on
benchmarked data sets. (literal)
- Prodotto di
- Autore CNR
Incoming links:
- Prodotto
- Autore CNR di
- Http://www.cnr.it/ontology/cnr/pubblicazioni.owl#rivistaDi