|
![]() |
|||
|
||||
OverviewSupport vector machines (SVMs) are one of the most successful algorithms on small and medium-sized data sets, but on large-scale data sets their training and predictions become computationally infeasible. The author considers a spatially defined data chunking method for large-scale learning problems, leading to so-called localized SVMs, and implements an in-depth mathematical analysis with theoretical guarantees, which in particular include classification rates. The statistical analysis relies on a new and simple partitioning based technique and takes well-known margin conditions into account that describe the behavior of the data-generating distribution. It turns out that the rates outperform known rates of several other learning algorithms under suitable sets of assumptions. From a practical point of view, the author shows that a common training and validation procedure achieves the theoretical rates adaptively, that is, without knowing the margin parameters in advance. Full Product DetailsAuthor: Ingrid Karin BlaschzykPublisher: Springer Fachmedien Wiesbaden Imprint: Springer Spektrum Edition: 1st ed. 2020 Weight: 0.454kg ISBN: 9783658295905ISBN 10: 3658295902 Pages: 126 Publication Date: 19 March 2020 Audience: Professional and scholarly , Professional & Vocational Format: Paperback Publisher's Status: Active Availability: Manufactured on demand ![]() We will order this item for you from a manufactured on demand supplier. Table of ContentsIntroduction to Statistical Learning Theory.- Histogram Rule: Oracle Inequality and Learning Rates.- Localized SVMs: Oracle Inequalities and Learning Rates.ReviewsAuthor InformationIngrid Karin Blaschzyk is a postdoctoral researcher in the Department of Mathematics at the University of Stuttgart, Germany. Tab Content 6Author Website:Countries AvailableAll regions |