VC dimension of neural networks. Sontag, E. In Neural Networks and Machine Learning, pages 69-95. Springer, Berlin, 1998.
abstract   bibtex   
The Vapnik-Chervonenkis (VC) dimension is an integer which helps to characterize distribution-independent learning of binary concepts from positive and negative samples. This paper, based on lectures delivered at the Isaac Newton Institute in August of 1997, presents a brief introduction, establishes various elementary results, and discusses how to estimate the VC dimension in several examples of interest in neural network theory. (It does not address the learning and estimation-theoretic applications of VC dimension, and the applications to uniform convergence theorems for empirical probabilities, for which many suitable references are available.)
@INCOLLECTION{newtonVC,
   AUTHOR       = {E.D. Sontag},
   BOOKTITLE    = {Neural Networks and Machine Learning},
   PUBLISHER    = {Springer, Berlin},
   TITLE        = {VC dimension of neural networks},
   YEAR         = {1998},
   OPTADDRESS   = {},
   OPTCHAPTER   = {},
   OPTCROSSREF  = {},
   OPTEDITION   = {},
   EDITOR       = {C.M. Bishop},
   OPTMONTH     = {},
   OPTNOTE      = {},
   OPTNUMBER    = {},
   PAGES        = {69-95},
   OPTSERIES    = {},
   OPTTYPE      = {},
   OPTVOLUME    = {},
   KEYWORDS     = {neural networks, VC dimension, learning, 
      neural networks, shattering},
   PDF          = {../../FTPDIR/vc-expo.pdf},
   ABSTRACT     = { The Vapnik-Chervonenkis (VC) dimension is an integer 
      which helps to characterize distribution-independent learning of 
      binary concepts from positive and negative samples. This paper, based 
      on lectures delivered at the Isaac Newton Institute in August of 
      1997, presents a brief introduction, establishes various elementary 
      results, and discusses how to estimate the VC dimension in several 
      examples of interest in neural network theory. (It does not address 
      the learning and estimation-theoretic applications of VC dimension, 
      and the applications to uniform convergence theorems for empirical 
      probabilities, for which many suitable references are available.) }
}

Downloads: 0