Explores the nearest neighbor classifier method, discussing its limitations in high-dimensional spaces and the importance of spatial correlation for effective predictions.
Delves into the trade-off between model flexibility and bias-variance in error decomposition, polynomial regression, KNN, and the curse of dimensionality.
Covers the k-Nearest-Neighbor classifier, hand-written digit recognition, multi-class k-NN, data reduction, applications, graph construction, limitations, and the curse of dimensionality.
Covers Principal Component Analysis for dimensionality reduction, exploring its applications, limitations, and importance of choosing the right components.
Covers PCA and LDA for dimensionality reduction, explaining variance maximization, eigenvector problems, and the benefits of Kernel PCA for nonlinear data.
Covers the Nearest Neighbor search algorithm and the Johnson-Lindenstrauss lemma for dimensionality reduction, exploring preprocessing techniques and locality-sensitive hashing.