The aim of this book is to discuss the fundamental ideas which lie behind the statistical theory of learning and generalization. It considers learning as a general problem of function estimation based on empirical data. Omitting proofs and technical details, the author concentrates on discussing the main results of learning theory and their connections to fundamental problems in statistics. This second edition contains three new chapters devoted to further development of the learning theory and SVM techniques. Written in a readable and concise style, the book is intended for statisticians, mathematicians, physicists, and computer scientists.
Le informazioni nella sezione "Riassunto" possono far riferimento a edizioni diverse di questo titolo.
From the reviews of the second edition:
ZENTRALBLATT MATH
"...written in a concise style. It must be recommended to scientists of statistics, mathematics, physics, and computer science."
SHORT BOOK REVIEWS
"This interesting book helps a reader to understand the interconnections between various streams in the empirical modeling realm and may be recommended to any reader who feels lost in modern terminology, such as artificial intelligence, neural networks, machine learning etcetera."
"The book by Vapnik focuses on how to estimate a function of parameters from empirical data ... . The book is concisely written and is intended to be useful to statisticians, computer scientists, mathematicians, and physicists. ... This book is very well written at a very high level of abstract thinking and comprehension. The references are up-to-date." (Ramalingam Shanmugam, Journal of Statistical Computation and Simulation, Vol. 75 (2), February, 2005)
"The aim of the book is to introduce a wide range of readers to the fundamental ideas of statistical learning theory. ... Each chapter is supplemented by ‘Reasoning and Comments’ which describe the relations between classical research in mathematical statistics and research in learning theory. ... The book is well suited to promote the ideas of statistical learning theory and can be warmly recommended to all who are interested in computer learning problems." (S. Vogel, Metrika, June, 2002)
Informal Reasoning and Comments * Consistency of Learning Processes * Bounds on the Rate of Convergence of Learing Processes * Controlling the Generalization Ability of Learning Processes * Methods of Pattern Recognition * Methods of Function Estimation * Direct Methods in Statistical Learning Theory * The Vicinal Risk Minimization Principle and the SVMs
Le informazioni nella sezione "Su questo libro" possono far riferimento a edizioni diverse di questo titolo.
GRATIS per la spedizione in U.S.A.
Destinazione, tempi e costiEUR 3,53 per la spedizione in U.S.A.
Destinazione, tempi e costiDa: ThriftBooks-Atlanta, AUSTELL, GA, U.S.A.
Paperback. Condizione: Good. No Jacket. Pages can have notes/highlighting. Spine may show signs of wear. ~ ThriftBooks: Read More, Spend Less 1.04. Codice articolo G1441931600I3N00
Quantità: 1 disponibili
Da: Lucky's Textbooks, Dallas, TX, U.S.A.
Condizione: New. Codice articolo ABLIING23Mar2411530294748
Quantità: Più di 20 disponibili
Da: Ria Christie Collections, Uxbridge, Regno Unito
Condizione: New. In. Codice articolo ria9781441931603_new
Quantità: Più di 20 disponibili
Da: moluna, Greven, Germania
Condizione: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. The aim of this book is to discuss the fundamental ideas which lie behind the statistical theory of learning and generalization. Written in readable and concise style and devoted to key learning problems, the book is intended for statisticians, mathematicia. Codice articolo 4173616
Quantità: Più di 20 disponibili
Da: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germania
Taschenbuch. Condizione: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -The aim of this book is to discuss the fundamental ideas which lie behind the statistical theory of learning and generalization. It considers learning as a general problem of function estimation based on empirical data. Omitting proofs and technical details, the author concentrates on discussing the main results of learning theory and their connections to fundamental problems in statistics. These include: \* the setting of learning problems based on the model of minimizing the risk functional from empirical data \* a comprehensive analysis of the empirical risk minimization principle including necessary and sufficient conditions for its consistency \* non-asymptotic bounds for the risk achieved using the empirical risk minimization principle \* principles for controlling the generalization ability of learning machines using small sample sizes based on these bounds \* the Support Vector methods that control the generalization ability when estimating function using small sample size. The second edition of the book contains three new chapters devoted to further development of the learning theory and SVM techniques. These include: \* the theory of direct method of learning based on solving multidimensional integral equations for density, conditional probability, and conditional density estimation \* a new inductive principle of learning. Written in a readable and concise style, the book is intended for statisticians, mathematicians, physicists, and computer scientists. Vladimir N. Vapnik is Technology Leader AT&T Labs-Research and Professor of London University. He is one of the founders of 336 pp. Englisch. Codice articolo 9781441931603
Quantità: 2 disponibili
Da: California Books, Miami, FL, U.S.A.
Condizione: New. Codice articolo I-9781441931603
Quantità: Più di 20 disponibili
Da: AHA-BUCH GmbH, Einbeck, Germania
Taschenbuch. Condizione: Neu. Druck auf Anfrage Neuware - Printed after ordering - The aim of this book is to discuss the fundamental ideas which lie behind the statistical theory of learning and generalization. It considers learning as a general problem of function estimation based on empirical data. Omitting proofs and technical details, the author concentrates on discussing the main results of learning theory and their connections to fundamental problems in statistics. These include: \* the setting of learning problems based on the model of minimizing the risk functional from empirical data \* a comprehensive analysis of the empirical risk minimization principle including necessary and sufficient conditions for its consistency \* non-asymptotic bounds for the risk achieved using the empirical risk minimization principle \* principles for controlling the generalization ability of learning machines using small sample sizes based on these bounds \* the Support Vector methods that control the generalization ability when estimating function using small sample size. The second edition of the book contains three new chapters devoted to further development of the learning theory and SVM techniques. These include: \* the theory of direct method of learning based on solving multidimensional integral equations for density, conditional probability, and conditional density estimation \* a new inductive principle of learning. Written in a readable and concise style, the book is intended for statisticians, mathematicians, physicists, and computer scientists. Vladimir N. Vapnik is Technology Leader AT&T Labs-Research and Professor of London University. He is one of the founders of. Codice articolo 9781441931603
Quantità: 1 disponibili
Da: Mispah books, Redhill, SURRE, Regno Unito
Paperback. Condizione: Like New. Like New. book. Codice articolo ERICA77314419316006
Quantità: 1 disponibili