This paper discusses the solution of large-scale linear matrix equations using the Induced Dimension reduction method (IDR(s)). IDR(s) was originally presented to solve system of linear equations, and is based on the IDR(s) theorem. We generalize the IDR(s) theorem to solve linear problems in any finite-dimensional space. This generalization allows us to develop IDR(s) algorithms to approximate the solution of linear matrix equations. The IDR(s) method presented here has two main advantages; firstly, it does not require the computation of inverses of any matrix, and secondly, it allows incorporation of preconditioners. Additionally, we present a simple preconditioner to solve the Sylvester equation based on a fixed point iteration. Several numerical examples illustrate the performance of IDR(s) for solving linear matrix equations. We also present the software implementation.
Astudillo, R., & Van Gijzen, M. B. (2016). Induced Dimension Reduction method for solving linear matrix equations. In Procedia Computer Science (Vol. 80, pp. 222–232). Elsevier B.V. https://doi.org/10.1016/j.procs.2016.05.313