Explores the concept of entropy expressed in bits and its relation to probability distributions, focusing on information gain and loss in various scenarios.
Explores Decision Trees, from induction to pruning, emphasizing interpretability and automatic feature selection strengths, while addressing challenges like overfitting.
Covers information measures like entropy, Kullback-Leibler divergence, and data processing inequality, along with probability kernels and mutual information.