Kernel Methods and Machine Learning
Offering a fundamental basis in kernel-based learning theory, this book covers both statistical and algebraic principles. It provides over 30 major theorems for kernel-based supervised and unsupervised learning models. The first of the theorems establishes a condition, arguably necessary and sufficient, for the kernelization of learning models. In addition, several other theorems are devoted to proving mathematical equivalence between seemingly unrelated models. With over 25 closed-form and iterative algorithms, the book provides a step-by-step guide to algorithmic procedures and analysing which factors to consider in tackling a given problem, enabling readers to improve specifically designed learning algorithms, build models for new applications and develop efficient techniques suitable for green machine learning technologies. Numerous real-world examples and over 200 problems, several of which are Matlab-based simulation exercises, make this an essential resource for graduate students and professionals in computer science, electrical and biomedical engineering. Solutions to problems are provided online for instructors.
- Electronic book text
- 31 Mar 2014
- CAMBRIDGE UNIVERSITY PRESS
- Cambridge University Press (Virtual Publishing)
- Cambridge, United Kingdom
- 136 b/w illus. 21 tables
Table of contents
Part I. Machine Learning and Kernel Vector Spaces: 1. Fundamentals of machine learning; 2. Kernel-induced vector spaces; Part II. Dimension-Reduction: Feature Selection and PCA/KPCA: 3. Feature selection; 4. PCA and Kernel-PCA; Part III. Unsupervised Learning Models for Cluster Analysis: 5. Unsupervised learning for cluster discovery; 6. Kernel methods for cluster discovery; Part IV. Kernel Ridge Regressors and Variants: 7. Kernel-based regression and regularization analysis; 8. Linear regression and discriminant analysis for supervised classification; 9. Kernel ridge regression for supervised classification; Part V. Support Vector Machines and Variants: 10. Support vector machines; 11. Support vector learning models for outlier detection; 12. Ridge-SVM learning models; Part VI. Kernel Methods for Green Machine Learning Technologies: 13. Efficient kernel methods for learning and classifcation; Part VII. Kernel Methods and Statistical Estimation Theory: 14. Statistical regression analysis and errors-in-variables models; 15: Kernel methods for estimation, prediction, and system identification; Part VIII. Appendices: Appendix A. Validation and test of learning models; Appendix B. kNN, PNN, and Bayes classifiers; References; Index.
About S. y. Kung
S. Y. Kung is a Professor in the Department of Electrical Engineering at Princeton University. His research areas include VLSI array/parallel processors, system modeling and identification, wireless communication, statistical signal processing, multimedia processing, sensor networks, bioinformatics, data mining and machine learning.