A Probabilistic Theory of Pattern RecognitionSpringer Science & Business Media, 20.2.1997 - 638 sivua Pattern recognition presents one of the most significant challenges for scientists and engineers, and many different approaches have been proposed. The aim of this book is to provide a self-contained account of probabilistic analysis of these approaches. The book includes a discussion of distance measures, nonparametric methods based on kernels or nearest neighbors, Vapnik-Chervonenkis theory, epsilon entropy, parametric classification, error estimation, free classifiers, and neural networks. Wherever possible, distribution-free properties and inequalities are derived. A substantial portion of the results or the analysis is new. Over 430 problems and exercises complement the material. |
Kirjan sisältä
Tulokset 1 - 5 kokonaismäärästä 85
Sivu ii
Valitettavasti tämän sivun sisältö on rajoitettu.
Valitettavasti tämän sivun sisältö on rajoitettu.
Sivu x
... Linear Discriminants 44 4.3 The Fisher Linear Discriminant 46 4.4 The Normal Distribution 47 4.5 Empirical Risk Minimization 49 4.6 Minimizing Other Criteria 54 Problems and Exercises 56 5 Nearest Neighbor Rules 61 5.1 Introduction 61 ...
... Linear Discriminants 44 4.3 The Fisher Linear Discriminant 46 4.4 The Normal Distribution 47 4.5 Empirical Risk Minimization 49 4.6 Minimizing Other Criteria 54 Problems and Exercises 56 5 Nearest Neighbor Rules 61 5.1 Introduction 61 ...
Sivu xi
... Linear and Generalized Linear Discrimination Rules 224 13.4 Convex Sets and Monotone Layers 226 Problems and Exercises 229 14 Lower Bounds for Empirical Classifier Selection 233 14.1 Minimax Lower Bounds 234 14.2 The Case Lc = 0 234 ...
... Linear and Generalized Linear Discrimination Rules 224 13.4 Convex Sets and Monotone Layers 226 Problems and Exercises 229 14 Lower Bounds for Empirical Classifier Selection 233 14.1 Minimax Lower Bounds 234 14.2 The Case Lc = 0 234 ...
Sivu xii
... Linear Discrimination 279 17.1 Fourier Series Classification 280 17.2 Generalized Linear Classification 285 Problems and Exercises 287 18 Complexity Regularization 289 18.1 Structural Risk Minimization 290 18.2 Poor Approximation ...
... Linear Discrimination 279 17.1 Fourier Series Classification 280 17.2 Generalized Linear Classification 285 Problems and Exercises 287 18 Complexity Regularization 289 18.1 Structural Risk Minimization 290 18.2 Poor Approximation ...
Sivu 8
Valitettavasti tämän sivun sisältö on rajoitettu.
Valitettavasti tämän sivun sisältö on rajoitettu.
Sisältö
Introduction | 1 |
The Bayes Error | 9 |
Inequalities and Alternate Distance Measures | 21 |
Linear Discrimination | 39 |
Nearest Neighbor Rules | 61 |
Consistency | 91 |
Slow Rates of Convergence | 111 |
Error Estimation | 121 |
Tree Classifiers | 315 |
DataDependent Partitioning | 363 |
Splitting the Data | 387 |
The Resubstitution Estimate | 397 |
Deleted Estimates of the Error Probability | 407 |
Automatic Kernel Rules | 423 |
Automatic Nearest Neighbor Rules | 451 |
Hypercubes and Discrete Spaces | 461 |
The Regular Histogram Rule | 133 |
Kernel Rules | 147 |
Consistency of the fcNearest Neighbor Rule | 169 |
VapnikChervonenkis Theory | 187 |
Combinatorial Aspects of VapnikChervonenkis Theory | 215 |
Lower Bounds for Empirical Classifier Selection | 233 |
The Maximum Likelihood Principle | 249 |
Parametric Classification | 263 |
Generalized Linear Discrimination | 279 |
Complexity Regularization | 289 |
Condensed and Edited Nearest Neighbor Rules | 303 |
Epsilon Entropy and Totally Bounded Sets | 479 |
Uniform Laws of Large Numbers | 489 |
Neural Networks | 507 |
Other Error Estimates | 549 |
Feature Extraction | 561 |
Appendix | 575 |
Notation | 591 |
619 | |
627 | |
Muita painoksia - Näytä kaikki
A Probabilistic Theory of Pattern Recognition Luc Devroye,Laszlo Györfi,Gabor Lugosi Rajoitettu esikatselu - 2013 |
A Probabilistic Theory of Pattern Recognition Luc Devroye,Laszlo Gyorfi,Gábor Lugosi Esikatselu ei käytettävissä - 2014 |
A Probabilistic Theory of Pattern Recognition Luc Devroye,László Györfi,Gabor Lugosi Esikatselu ei käytettävissä - 2013 |
Yleiset termit ja lausekkeet
apply Assume Bayes bound called cells Chapter classification rule classifier Clearly close coefficients complexity components computational Consider constant contains continuous convergence corresponding covering cuts decision defined definition denotes density depends discrimination distance distribution empirical error error probability estimate example exists expected fact Figure finite fixed function given hyperplane implies independent inequality integer introduce Lemma linear measure method minimizing nearest neighbor rule Note Observe obtained otherwise pairs parameters partition performance pick points positive possible probability of error Problem proof Prove random variables rectangles regions Remark respect result risk sample satisfies selected sequence shatter Show simple smoothing space split subsets suffices term Theorem tree uniform universally consistent upper bound values vc dimension vector weights zero