Empirical Inference (eBook)

Festschrift in Honor of Vladimir N. Vapnik
eBook Download: PDF
2013 | 2013
XIX, 287 Seiten
Springer Berlin (Verlag)
978-3-642-41136-6 (ISBN)

Lese- und Medienproben

Empirical Inference -
Systemvoraussetzungen
53,49 inkl. MwSt
  • Download sofort lieferbar
  • Zahlungsarten anzeigen

This book honours the outstanding contributions of Vladimir Vapnik, a rare example of a scientist for whom the following statements hold true simultaneously: his work led to the inception of a new field of research, the theory of statistical learning and empirical inference; he has lived to see the field blossom; and he is still as active as ever. He started analyzing learning algorithms in the 1960s and he invented the first version of the generalized portrait algorithm. He later developed one of the most successful methods in machine learning, the support vector machine (SVM) - more than just an algorithm, this was a new approach to learning problems, pioneering the use of functional analysis and convex optimization in machine learning.

 

Part I of this book contains three chapters describing and witnessing some of Vladimir Vapnik's contributions to science. In the first chapter, Léon Bottou discusses the seminal paper published in 1968 by Vapnik and Chervonenkis that lay the foundations of statistical learning theory, and the second chapter is an English-language translation of that original paper. In the third chapter, Alexey Chervonenkis presents a first-hand account of the early history of SVMs and valuable insights into the first steps in the development of the SVM in the framework of the generalised portrait method.

 

The remaining chapters, by leading scientists in domains such as statistics, theoretical computer science, and mathematics, address substantial topics in the theory and practice of statistical learning theory, including SVMs and other kernel-based methods, boosting, PAC-Bayesian theory, online and transductive learning, loss functions, learnable function classes, notions of complexity for function classes, multitask learning, and hypothesis selection. These contributions include historical and context notes, short surveys, and comments on future research directions.

 

This book will be of interest to researchers, engineers, and graduate students engaged with all aspects of statistical learning.

Part I - History of Statistical Learning Theory.- Chap. 1 - In Hindsight: Doklady Akademii Nauk SSSR, 181(4), 1968.- Chap. 2 - On the Uniform Convergence of the Frequencies of Occurrence of Events to Their Probabilities.- Chap. 3 - Early History of Support Vector Machines.- Part II - Theory and Practice of Statistical Learning Theory.- Chap. 4 - Some Remarks on the Statistical Analysis of SVMs and Related Methods.- Chap. 5 - Explaining AdaBoost.- Chap. 6 - On the Relations and Differences Between Popper Dimension, Exclusion Dimension and VC-Dimension.- Chap. 7 - On Learnability, Complexity and Stability.- Chap. 8 - Loss Functions.- Chap. 9 - Statistical Learning Theory in Practice.- Chap. 10 - PAC-Bayesian Theory.- Chap. 11 - Kernel Ridge Regression.- Chap. 12 - Multi-task Learning for Computational Biology: Overview and Outlook.- Chap. 13 - Semi-supervised Learning in Causal and Anticausal Settings.- Chap. 14 - Strong Universal Consistent Estimate of the Minimum Mean-Squared Error.- Chap. 15 - The Median Hypothesis.- Chap. 16 - Efficient Transductive Online Learning via Randomized Rounding.- Chap. 17 - Pivotal Estimation in High-Dimensional Regression via Linear Programming.- Chap. 18 - Some Observations on Sparsity Inducing Regularization Methods for Machine Learning.- Chap. 19 - Sharp Oracle Inequalities in Low Rank Estimation.- Chap. 20 - On the Consistency of the Bootstrap Approach for Support Vector Machines and Related Kernel-Based Methods.- Chap. 21 - Kernels, Pre-images and Optimization.- Chap. 22 - Efficient Learning of Sparse Ranking Functions.- Chap. 23 - Direct Approximation of Divergences Between Probability Distributions.- Index.

Erscheint lt. Verlag 11.12.2013
Zusatzinfo XIX, 287 p. 33 illus., 26 illus. in color.
Verlagsort Berlin
Sprache englisch
Themenwelt Informatik Theorie / Studium Künstliche Intelligenz / Robotik
Mathematik / Informatik Mathematik Statistik
Technik
Schlagworte Bayesian theory • kernels • learning • machine learning • Optimization • statistical learning theory • Support vector machines (SVMs) • VC (Vapnik-Chervonenkis) dimension
ISBN-10 3-642-41136-3 / 3642411363
ISBN-13 978-3-642-41136-6 / 9783642411366
Haben Sie eine Frage zum Produkt?
Wie bewerten Sie den Artikel?
Bitte geben Sie Ihre Bewertung ein:
Bitte geben Sie Daten ein:
PDFPDF (Wasserzeichen)
Größe: 3,9 MB

DRM: Digitales Wasserzeichen
Dieses eBook enthält ein digitales Wasser­zeichen und ist damit für Sie persona­lisiert. Bei einer missbräuch­lichen Weiter­gabe des eBooks an Dritte ist eine Rück­ver­folgung an die Quelle möglich.

Dateiformat: PDF (Portable Document Format)
Mit einem festen Seiten­layout eignet sich die PDF besonders für Fach­bücher mit Spalten, Tabellen und Abbild­ungen. Eine PDF kann auf fast allen Geräten ange­zeigt werden, ist aber für kleine Displays (Smart­phone, eReader) nur einge­schränkt geeignet.

Systemvoraussetzungen:
PC/Mac: Mit einem PC oder Mac können Sie dieses eBook lesen. Sie benötigen dafür einen PDF-Viewer - z.B. den Adobe Reader oder Adobe Digital Editions.
eReader: Dieses eBook kann mit (fast) allen eBook-Readern gelesen werden. Mit dem amazon-Kindle ist es aber nicht kompatibel.
Smartphone/Tablet: Egal ob Apple oder Android, dieses eBook können Sie lesen. Sie benötigen dafür einen PDF-Viewer - z.B. die kostenlose Adobe Digital Editions-App.

Buying eBooks from abroad
For tax law reasons we can sell eBooks just within Germany and Switzerland. Regrettably we cannot fulfill eBook-orders from other countries.

Mehr entdecken
aus dem Bereich
der Praxis-Guide für Künstliche Intelligenz in Unternehmen - Chancen …

von Thomas R. Köhler; Julia Finkeissen

eBook Download (2024)
Campus Verlag
38,99