18.S096 Matrix Calculus for Machine Learning and Beyond
18.S096 Matrix Calculus for Machine Learning and Beyond (IAP 2023, MIT OCW). Instructors: Prof. Alan Edelman and Prof. Steven G. Johnson. This class covers a coherent approach to matrix calculus showing techniques that allow you to think of a matrix holistically (not just as an array of scalars), generalize and compute derivatives of important matrix factorizations and many other complicated-looking operations, and understand how differentiation formulas must be reimagined in large-scale computing. We will discuss reverse/adjoint/backpropagation differentiation, custom vector-Jacobian products, and how modern automatic differentiation is more computer science than calculus (it is neither symbolic formulas nor finite differences. (from ocw.mit.edu)
Lecture 07 - Part2: Second Derivatives, Bilinear Forms, and Hessian Matrices |
Instructors: Prof. Alan Edelman and Prof. Steven G. Johnson. First derivatives are linear operators, so second derivatives are bilinear forms, sometimes called "Hessians" (especially for scalar-valued functions of column vectors, where the Hessian is simply a symmetric matrix).
Go to the Course Home or watch other lectures: