|
![]() |
|||
|
||||
OverviewPattern recognition presents one of the most significant challenges for scientists and engineers, and many different approaches have been proposed. The aim of this book is to provide a self-contained account of probabilistic analysis of these approaches. The book includes a discussion of distance measures, nonparametric methods based on kernels or nearest neighbors, Vapnik-Chervonenkis theory, epsilon entropy, parametric classification, error estimation, free classifiers, and neural networks. Wherever possible, distribution-free properties and inequalities are derived. A substantial portion of the results or the analysis is new. Over 430 problems and exercises complement the material. Full Product DetailsAuthor: Luc Devroye , Laszlo Györfi , Gabor LugosiPublisher: Springer-Verlag New York Inc. Imprint: Springer-Verlag New York Inc. Edition: 1st ed. 1996. Corr. 2nd printing 1997 Volume: 31 Dimensions: Width: 15.50cm , Height: 3.40cm , Length: 23.50cm Weight: 2.420kg ISBN: 9780387946184ISBN 10: 0387946187 Pages: 638 Publication Date: 04 April 1996 Audience: College/higher education , Professional and scholarly , Postgraduate, Research & Scholarly , Professional & Vocational Format: Hardback Publisher's Status: Active Availability: In Print ![]() This item will be ordered in for you from one of our suppliers. Upon receipt, we will promptly dispatch it out to you. For in store availability, please contact us. Table of ContentsPreface * Introduction * The Bayes Error * Inequalities and alternate distance measures * Linear discrimination * Nearest neighbor rules * Consistency * Slow rates of convergence Error estimation * The regular histogram rule * Kernel rules Consistency of the k-nearest neighbor rule * Vapnik-Chervonenkis theory * Combinatorial aspects of Vapnik- Chervonenkis theory * Lower bounds for empirical classifier selection * The maximum likelihood principle * Parametric classification * Generalized linear discrimination * Complexity regularization * Condensed and edited nearest neighbor rules * Tree classifiers * Data- dependent partitioning * Splitting the data * The resubstitution estimate * Deleted estimates of the error probability * Automatic kernel rules * Automatic nearest neighbor rules * Hypercubes and discrete spaces * Epsilon entropy and totally bounded sets * Uniform laws of large numbers * Neural networks * Other error estimates * Feature extraction * Appendix * Notation * References * IndexReviewsAuthor InformationTab Content 6Author Website:Countries AvailableAll regions |