Discusses optimization techniques in machine learning, focusing on stochastic gradient descent and its applications in constrained and non-convex problems.
Explores the theory and applications of convex optimization, covering topics such as log-determinant function, affine transformations, and relative entropy.
Explores KKT conditions in convex optimization, covering dual problems, logarithmic constraints, least squares, matrix functions, and suboptimality of covering ellipsoids.