Lecturer(s)
|
-
Machalová Jitka, doc. RNDr. Ph.D., MBA
-
Burkotová Jana, Mgr. Ph.D.
|
Course content
|
1. Systems of linear equations, condition number, eigenvalue and eigenvectors of matrices. 2. Matrix decompositions - LU, QR and SVD. 3. Orthogonalization and least squares method. 4. Introduction to large sparse systems of equations and their application. Direct and iterative methods. 5. Preconditioning, computer realization, termination criteria. 6. Matrix computations in statistics. 7. Special matrices in statistics, optimization and machine learning.
|
Learning activities and teaching methods
|
unspecified
|
Learning outcomes
|
Understand and be able to use the methods for matrix computations.
Knowledge Gain knowledge about basic and advanced methods in matrix calculus.
|
Prerequisites
|
Standard knowledge from matrix calculus.
|
Assessment methods and criteria
|
unspecified
Colloquium: elaboration of a selected problem, defense in the form of a presentation.
|
Recommended literature
|
-
A. Björck. (2015). Numerical Methods in Matrix Computations. Springer.
-
C. C. Aggarwal. (2020). Linear algebra and optimization for machine learning. Springer.
-
D. Bertaccini, F. Durastante. (2018). Iterative Methods and Preconditioning for Large and Sparse Linear Systems with Applications. Chapman and Hall/CRC.
-
D.A. Harville. (1997). Matrix Algebra From a Statistician's Perspective. Springer.
-
G. H. Golub, CH. F. Van Loan. (2013). Matrix Computations. Johns Hopkins University Press, Baltimore.
-
G. Strang. (2019). Linear algebra and learning from data. Wellesley - Cambridge Press, Wellesley, MA.
-
S. Puntanen, G.P.H. Styan, J. Isotalo. (2011). Matrix Tricks for Linear Statistical Models. Springer.
|