Vector and Matrix Derivatives
Matrix CalculusVector DerivativesMatrix PropertiesDeterminants & TracesEigenvalues & EigenvectorsMatrix Decompositions (LU, QR, SVD)Orthogonality & ProjectionsOptimizationCommon Loss FunctionsGradient DescentConvexityLinear AlgebraSupervised LearningSupervised Learning (Regression/Classification)
Vector and Matrix Derivatives: Essential Tools for Optimization
In many areas of machine learning and optimization, especially when dealing with gradient-based methods, it is crucial to know the derivatives of vector and matrix functions. In this article, we summarize some common derivative formulas, specify the dimensions involved, and discuss how these results are applied in practice.
Dimensions and Notation
We use the following dimensions throughout:
Vector Derivatives
Matrix Derivatives
Norm Derivatives
Other Common Derivatives
where denotes the Frobenius norm of .
Useful Resources
For further reading on matrix calculus, a highly recommended resource is the Matrix Cookbook. Other valuable references include:
- Matrix Differential Calculus with Applications in Statistics and Econometrics by Magnus and Neudecker.
- Online resources and lecture notes on advanced linear algebra and optimization, such as MIT OpenCourseWare's Linear Algebra course by Gilbert Strang.
