Consider a column-wise partitioned matrix:
If the above matrix is full column rank, the Moore–Penrose inverse matrices of it and its transpose are
This computation of the pseudoinverse requires (n + p)-square matrix inversion and does not take advantage of the block form.
To reduce computational costs to n- and p-square matrix inversions and to introduce parallelism, treating the blocks separately, one derives 1
where orthogonal projection matrices are defined by
The above formulas are not necessarily valid if [ A B ] {\displaystyle {\begin{bmatrix}\mathbf {A} &\mathbf {B} \end{bmatrix}}} does not have full rank – for example, if A ≠ 0 {\displaystyle \mathbf {A} \neq 0} , then
Given the same matrices as above, we consider the following least squares problems, which appear as multiple objective optimizations or constrained problems in signal processing. Eventually, we can implement a parallel algorithm for least squares based on the following results.
Suppose a solution x = [ x 1 x 2 ] {\displaystyle \mathbf {x} ={\begin{bmatrix}\mathbf {x} _{1}\\\mathbf {x} _{2}\\\end{bmatrix}}} solves an over-determined system:
Using the block matrix pseudoinverse, we have
Therefore, we have a decomposed solution:
Suppose a solution x {\displaystyle \mathbf {x} } solves an under-determined system:
The minimum-norm solution is given by
Instead of ( [ A B ] T [ A B ] ) − 1 {\displaystyle \mathbf {\left({\begin{bmatrix}\mathbf {A} &\mathbf {B} \end{bmatrix}}^{\textsf {T}}{\begin{bmatrix}\mathbf {A} &\mathbf {B} \end{bmatrix}}\right)} ^{-1}} , we need to calculate directly or indirectly[original research?]
In a dense and small system, we can use singular value decomposition, QR decomposition, or Cholesky decomposition to replace the matrix inversions with numerical routines. In a large system, we may employ iterative methods such as Krylov subspace methods.
Considering parallel algorithms, we can compute ( A T A ) − 1 {\displaystyle \left(\mathbf {A} ^{\textsf {T}}\mathbf {A} \right)^{-1}} and ( B T B ) − 1 {\displaystyle \left(\mathbf {B} ^{\textsf {T}}\mathbf {B} \right)^{-1}} in parallel. Then, we finish to compute ( A T P B ⊥ A ) − 1 {\displaystyle \left(\mathbf {A} ^{\textsf {T}}\mathbf {P} _{B}^{\perp }\mathbf {A} \right)^{-1}} and ( B T P A ⊥ B ) − 1 {\displaystyle \left(\mathbf {B} ^{\textsf {T}}\mathbf {P} _{A}^{\perp }\mathbf {B} \right)^{-1}} also in parallel.
J.K. Baksalary and O.M. Baksalary (2007). "Particular formulae for the Moore–Penrose inverse of a columnwise partitioned matrix". Linear Algebra Appl. 421: 16–23. doi:10.1016/j.laa.2006.03.031. /wiki/Jerzy_Baksalary ↩