Discusses minima in error functions, multiple minima, saddle points, weight space symmetry, and near-equivalent good solutions in deep neural networks.
Covers the significance of subtracting the mean reward in policy gradient methods for deep reinforcement learning, reducing noise in the stochastic gradient.