Neural Network Learning: Theoretical Foundations

Neural Network Learning: Theoretical Foundations

ISBN-10:
052157353X
ISBN-13:
9780521573535
Pub. Date:
11/04/1999
Publisher:
Cambridge University Press
ISBN-10:
052157353X
ISBN-13:
9780521573535
Pub. Date:
11/04/1999
Publisher:
Cambridge University Press
Neural Network Learning: Theoretical Foundations

Neural Network Learning: Theoretical Foundations

$149.0
Current price is , Original price is $149.0. You
$149.00 
  • SHIP THIS ITEM
    Qualifies for Free Shipping
  • PICK UP IN STORE
    Check Availability at Nearby Stores

Overview

This important work describes recent theoretical advances in the study of artificial neural networks. It explores probabilistic models of supervised learning problems, and addresses the key statistical and computational questions. Chapters survey research on pattern classification with binary-output networks, including a discussion of the relevance of the Vapnik Chervonenkis dimension, and of estimates of the dimension for several neural network models. In addition, Anthony and Bartlett develop a model of classification by real-output networks, and demonstrate the usefulness of classification with a "large margin." The authors explain the role of scale-sensitive versions of the Vapnik Chervonenkis dimension in large margin classification, and in real prediction. Key chapters also discuss the computational complexity of neural network learning, describing a variety of hardness results, and outlining two efficient, constructive learning algorithms. The book is self-contained and accessible to researchers and graduate students in computer science, engineering, and mathematics.

Product Details

ISBN-13: 9780521573535
Publisher: Cambridge University Press
Publication date: 11/04/1999
Pages: 404
Product dimensions: 5.98(w) x 9.02(h) x 1.06(d)

Table of Contents

1. Introduction; Part I. Pattern Recognition with Binary-output Neural Networks: 2. The pattern recognition problem; 3. The growth function and VC-dimension; 4. General upper bounds on sample complexity; 5. General lower bounds; 6. The VC-dimension of linear threshold networks; 7. Bounding the VC-dimension using geometric techniques; 8. VC-dimension bounds for neural networks; Part II. Pattern Recognition with Real-output Neural Networks: 9. Classification with real values; 10. Covering numbers and uniform convergence; 11. The pseudo-dimension and fat-shattering dimension; 12. Bounding covering numbers with dimensions; 13. The sample complexity of classification learning; 14. The dimensions of neural networks; 15. Model selection; Part III. Learning Real-Valued Functions: 16. Learning classes of real functions; 17. Uniform convergence results for real function classes; 18. Bounding covering numbers; 19. The sample complexity of learning function classes; 20. Convex classes; 21. Other learning problems; Part IV. Algorithmics: 22. Efficient learning; 23. Learning as optimisation; 24. The Boolean perceptron; 25. Hardness results for feed-forward networks; 26. Constructive learning algorithms for two-layered networks.
From the B&N Reads Blog

Customer Reviews