Cantitate/Preț
Produs

Neural Network Learning: Theoretical Foundations

Autor Martin Anthony, Peter L. Bartlett
en Limba Engleză Paperback – 19 aug 2009
This book describes theoretical advances in the study of artificial neural networks. It explores probabilistic models of supervised learning problems, and addresses the key statistical and computational questions. Research on pattern classification with binary-output networks is surveyed, including a discussion of the relevance of the Vapnik–Chervonenkis dimension, and calculating estimates of the dimension for several neural network models. A model of classification by real-output networks is developed, and the usefulness of classification with a 'large margin' is demonstrated. The authors explain the role of scale-sensitive versions of the Vapnik–Chervonenkis dimension in large margin classification, and in real prediction. They also discuss the computational complexity of neural network learning, describing a variety of hardness results, and outlining two efficient constructive learning algorithms. The book is self-contained and is intended to be accessible to researchers and graduate students in computer science, engineering, and mathematics.
Citește tot Restrânge

Toate formatele și edițiile

Toate formatele și edițiile Preț Express
Paperback (1) 36223 lei  43-57 zile
  Cambridge University Press – 19 aug 2009 36223 lei  43-57 zile
Hardback (1) 81155 lei  43-57 zile
  Cambridge University Press – 3 noi 1999 81155 lei  43-57 zile

Preț: 36223 lei

Preț vechi: 45278 lei
-20% Nou

Puncte Express: 543

Preț estimativ în valută:
6932 7201$ 5758£

Carte tipărită la comandă

Livrare economică 03-17 februarie 25

Preluare comenzi: 021 569.72.76

Specificații

ISBN-13: 9780521118620
ISBN-10: 052111862X
Pagini: 404
Dimensiuni: 152 x 229 x 23 mm
Greutate: 0.59 kg
Ediția:New.
Editura: Cambridge University Press
Colecția Cambridge University Press
Locul publicării:Cambridge, United Kingdom

Cuprins

1. Introduction; Part I. Pattern Recognition with Binary-output Neural Networks: 2. The pattern recognition problem; 3. The growth function and VC-dimension; 4. General upper bounds on sample complexity; 5. General lower bounds; 6. The VC-dimension of linear threshold networks; 7. Bounding the VC-dimension using geometric techniques; 8. VC-dimension bounds for neural networks; Part II. Pattern Recognition with Real-output Neural Networks: 9. Classification with real values; 10. Covering numbers and uniform convergence; 11. The pseudo-dimension and fat-shattering dimension; 12. Bounding covering numbers with dimensions; 13. The sample complexity of classification learning; 14. The dimensions of neural networks; 15. Model selection; Part III. Learning Real-Valued Functions: 16. Learning classes of real functions; 17. Uniform convergence results for real function classes; 18. Bounding covering numbers; 19. The sample complexity of learning function classes; 20. Convex classes; 21. Other learning problems; Part IV. Algorithmics: 22. Efficient learning; 23. Learning as optimisation; 24. The Boolean perceptron; 25. Hardness results for feed-forward networks; 26. Constructive learning algorithms for two-layered networks.

Recenzii

'The book is a useful and readable mongraph. For beginners it is a nice introduction to the subject, for experts a valuable reference.' Zentralblatt MATH

Descriere

This book describes theoretical advances in the study of artificial neural networks.