9780262111935-0262111934-An Introduction to Computational Learning Theory

An Introduction to Computational Learning Theory

ISBN-13: 9780262111935
ISBN-10: 0262111934
Author: Umesh Vazirani, Michael J. Kearns
Publication date: 1994
Publisher: The MIT Press
Format: Hardcover 221 pages
FREE US shipping

Book details

ISBN-13: 9780262111935
ISBN-10: 0262111934
Author: Umesh Vazirani, Michael J. Kearns
Publication date: 1994
Publisher: The MIT Press
Format: Hardcover 221 pages

Summary

An Introduction to Computational Learning Theory (ISBN-13: 9780262111935 and ISBN-10: 0262111934), written by authors Umesh Vazirani, Michael J. Kearns, was published by The MIT Press in 1994. With an overall rating of 3.9 stars, it's a notable title among other AI & Machine Learning (Schools & Teaching, Workbooks, Study Guides & Workbooks, Computer Science) books. You can easily purchase or rent An Introduction to Computational Learning Theory (Hardcover) from BooksRun, along with many other new and used AI & Machine Learning books and textbooks. And, if you're looking to sell your copy, our current buyback offer is $6.8.

Description

Emphasizing issues of computational efficiency, Michael Kearns and Umesh Vazirani introduce a number of central topics in computational learning theory for researchers and students in artificial intelligence, neural networks, theoretical computer science, and statistics.

Emphasizing issues of computational efficiency, Michael Kearns and Umesh Vazirani introduce a number of central topics in computational learning theory for researchers and students in artificial intelligence, neural networks, theoretical computer science, and statistics. Computational learning theory is a new and rapidly expanding area of research that examines formal models of induction with the goals of discovering the common methods underlying efficient learning algorithms and identifying the computational impediments to learning. Each topic in the book has been chosen to elucidate a general principle, which is explored in a precise formal setting. Intuition has been emphasized in the presentation to make the material accessible to the nontheoretician while still providing precise arguments for the specialist. This balance is the result of new proofs of established theorems, and new presentations of the standard proofs. The topics covered include the motivation, definitions, and fundamental results, both positive and negative, for the widely studied L. G. Valiant model of Probably Approximately Correct Learning; Occam's Razor, which formalizes a relationship between learning and data compression; the Vapnik-Chervonenkis dimension; the equivalence of weak and strong learning; efficient learning in the presence of noise by the method of statistical queries; relationships between learning and cryptography, and the resulting computational limitations on efficient learning; reducibility between learning problems; and algorithms for learning finite automata from active experimentation.

Rate this book Rate this book

We would LOVE it if you could help us and other readers by reviewing the book