Skip to content
BY 4.0 license Open Access Published by De Gruyter Open Access July 6, 2020

The core inverse and constrained matrix approximation problem

  • Hongxing Wang EMAIL logo and Xiaoyan Zhang
From the journal Open Mathematics

Abstract

In this article, we study the constrained matrix approximation problem in the Frobenius norm by using the core inverse:

| | M x b | | F = min subject to x ( M ) ,

where M n CM . We get the unique solution to the problem, provide two Cramer’s rules for the unique solution and establish two new expressions for the core inverse.

MSC 2010: 15A24; 15A29; 15A57

1 Introduction

Let M , ( M ) and N ( M ) stand for the conjugate transpose, range space and null space of M m × n , respectively. The symbol M ( i b ) denotes a matrix from M by replacing the ith column of M by b n . The symbol e i denotes the ith column of I n in which 1 i n . The Moore-Penrose inverse of M is the unique matrix X n × m satisfying the relations: M X M = M , X M X = X , ( M X ) = M X and ( X M ) = X M and is denoted by X = M [1,2,3].

Let M n × n be singular. The smallest positive integer k for which rk ( M k + 1 ) = rk ( M k ) is called the index of M and is denoted by Ind ( M ) . The index of a non-singular matrix is 0 and the index of a null matrix is 1. Furthermore,

(1.1) n CM = { M | Ind ( M ) 1 , M n × n } .

Let M n × n with Ind ( M ) = k . A matrix X is the Drazin inverse of M if M X M k = M k , X M X = X and M X = X M . We write X = M D for the Drazin inverse of M. In particular, when M n CM , the matrix X is the group inverse of M and is denoted by X = M # [1,2,3].

The core inverse of M n CM is defined as the unique matrix X n × n satisfying the equations: M X M = M , M X 2 = X and ( M X ) = M X and is denoted by X = M [4,5]. It is noteworthy that the core inverse is a “least squares” inverse [6,7]. Moreover, it is proved that M = M # M M [4].

Recently, the relevant conclusions of the core inverse are very rich. In [7,8,9,10], generalizations of core inverse are introduced, for example, the core-EP inverse and the weak group inverse. In [11,12,13,14,15], their algebraic properties and calculating methods are studied. In [16,17], the studying of them is extended to some new fields, for example, ring and operator. Moreover, those inverses are used to study partial orders in [4,5,10,18,19].

Consider the following equation:

(1.2) M x = b .

Let M n × n with Ind ( M ) = k and b ( M k ) . Campbell and Meyer [20] show that x = M D b is the unique solution of (1.2) with respect to x ( M k ) . Wei [21] gets the minimal P-norm solution of (1.2), where P is nonsingular, P 1 M P is the Jordan canonical form of M and x p = P 1 x 2 . Furthermore, let M m × n . Wei [22] considered the unique solution of

W M W x = b subject to x ( ( M W ) k 1 ) ,

where W n × m , k 1 = Ind ( M W ) , k 2 = Ind ( W M ) and b ( ( W M ) k 2 ) . More results of (1.2) under some certain conditions can be found in [3,21,23,24,25,26,27].

It is well known that b ( M ) if and only if (1.2) is solvable. Let b ( M ) and the index of M is 1, then x = M # b is the unique solution with x ( M ) [20]. It follows from M = M # M M that M # b = M b [28]. Furthermore, the unique solution x = M b is given by Cramer’s rule [28, Theorem 3.3].

When b ( M ) , (1.2) is unsolvable, yet, it has least-squares solutions. Motivated by the aforementioned works, it is natural to consider the least-squares solutions of (1.2) under the certain condition x ( M ) , i.e.,

(1.3) M x b F = min subject to x ( M ) ,

where M n CM , rk ( M ) = r < n and b n .

2 Preliminaries

Lemma 2.1

[1] Let M n × n be idempotent. Then, M = P ( M ) , N ( M ) with M N M = n . In contrast, if F G = n , then there exists an idempotent P F , G such that ( P F , G ) = F and N ( P F , G ) = G .

Furthermore, I n P F , G = P G , F .

Lemma 2.2

[3] Let M n × n . Then, Ind ( M ) = k if and only if

(2.1) ( M k ) N ( M k ) = n .

Lemma 2.3

[3] Let M X M = M and X M X = X . Then,

X M = P ( X ) , N ( M ) and M X = P ( M ) , N ( X ) .

Lemma 2.4

[3] Let F G = n . Then,

  1. P F , G M = M ( M ) F ;

  2. M P F , G = M N ( M ) G .

Lemma 2.5

[14] Let M n CM with rk ( M ) = r . Then, there exists a unitary matrix V such that

(2.2) M = V T S 0 0 V ,

where T r × r is nonsingular. Furthermore,

(2.3) M = V T 1 0 0 0 V .

3 Main results

3.1 Solution of (1.3)

Theorem 3.1

Let M n CM and b n . Then,

(3.1) x = M b

is the unique solution of ( 1.3 ).

Proof

From x ( M ) , it follows that there exists y n for which x = M y . Then, x is the solution of (1.3) if and only if y is the solution of

(3.2) M 2 y b F = min .

Let the decomposition of M be as in (2.2). Denote

(3.3) V y = y 1 y 2 , V b = b 1 b 2 and M b = V T 1 b 1 0 ,

where y 1 , b 1 and T 1 b 1 rk ( M ) . It follows that

M x b F 2 = T 2 y 1 + T S y 2 b 1 b 2 F 2 = T 2 y 1 + T S y 2 b 1 F 2 + b 2 F 2 .

Since T is invertible, we have min y 1 , y 2 T 2 y 1 + T S y 2 b 1 F 2 = 0 , that is, M 2 y b F = min = b 2 F , in which y 2 n rk ( M ) is arbitrary, and y 1 = T 2 b 1 T 1 S y 2 . It follows that

x = M y = V T S 0 0 V y = V T y 1 + S y 2 0 = V T 1 b 1 0 = M b ,

that is, (3.1) is the unique solution of (1.3).□

3.2 Determinantal formulas

When M n × n is nonsingular, it is well known that the solution of (1.2) is unique and x = M 1 b . Let x = ( x 1 , x 2 , , x n ) T . Then,

(3.4) x i = det ( M ( i b ) ) det ( M ) , i = 1 , 2 , , n ,

is called Cramer’s rule for solving (1.2). In [29], Ben-Israel gets a Cramer’s rule for obtaining the least-norm solution of the consistent linear system (1.2),

x i = det M ( i b ) U V ( i 0 ) 0 det M U V 0 , i = 1 , 2 , , n ,

where U and V are of full column rank, ( U ) = N ( M ) and ( V ) = N ( M ) . In [26], Wang gives a Cramer’s rule for the unique solution x ( M k ) of (1.2), where b ( M k ) and Ind ( M ) = k . In [30], Ji proposes two new condensed Cramer’s rules for the unique solution x ( M k ) of (1.2), where b ( M k ) and Ind ( M ) = k . More details of Cramer’s rules for finding restricted solutions of (1.2) can be found in [1,3,31,32,33,34,35,36]. In Theorems 3.4 and 3.6, we will give two Cramer’s rules for the unique solution of (1.3).

First of all, we give the following two lemmas to prepare for a Cramer’s rule for core inverse in Theorem 3.4.

Lemma 3.2

Let M n CM with rk ( M ) = r , and let L n × ( n r ) with rk ( L ) = n r and ( L ) = N ( M ) . Then,

(3.5) M M + ( I n M M ) L ( L L ) 1 L = I n .

Proof

Let M be as in (2.2), applying Lemma 2.2, we see that

(3.6) ( M ) N ( M ) = n .

Denote M 1 = I n M M and M 2 = L ( L L ) 1 L .

Applying Lemmas 2.1, 2.3 and M M = M # M , we have

(3.7) M M = P ( M ) , N ( M ) ,

(3.8) M 1 = I M M = P N ( M ) , ( M ) .

Since ( L ( L L ) 1 ) L ( L ( L L ) 1 ) = L ( L L ) 1 and L ( L ( L L ) 1 ) L = L , applying Lemma 2.3, we obtain

(3.9) M 2 = P ( L ) , N ( L ) = P N ( M ) , ( M ) .

Since ( L ) = N ( M ) , we obtain M 2 M 1 = M 2 and

(3.10) M 1 M 2 = P N ( M ) , ( M ) .

Therefore, applying Lemma 2.1, (3.7) and (3.10), we gain

M M + M 1 M 2 = P ( M ) , N ( M ) + P N ( M ) , ( M ) = I n ,

i.e., (3.5).□

In [28, Theorems 3.2 and 3.3], let M n C M , b n and b ( M ) , and let M b and M c be of the full column ranks with N ( M ) = ( M b ) and N ( M c ) = ( M ) . Then,

M M b M c 0

is invertible and the unique solution x = M b of (1.2) satisfying

x i = det M i b M b M c i 0 0 / det M M b M c 0 ,

where i = 1 , 2 , , n . In Lemma 3.3 and Theorem 3.4, we give the unique least-squares solution of (1.3) in a similar way.

Lemma 3.3

Let M and L be as in Lemma 3.2. Then,

(3.11) G = M L L 0

is invertible and

(3.12) G 1 = M ( I n M M ) L ( L L ) 1 ( L L ) 1 L 0 .

Proof

Since ( L ) = N ( M ) and M = M # M M , we have M L = M # M M L = 0 and ( L L ) 1 L M = 0 . Furthermore, applying (3.5), we have

M ( I n M M ) L ( L L ) 1 ( L L ) 1 L 0 M L L 0 = M M + ( I n M M ) L ( L L ) 1 L M L ( L L ) 1 L M ( L L ) 1 L L = I 2 n r ,

that is, G is invertible and G 1 is of the form (3.12).□

Based on Lemmas 3.2 and 3.3, we get a Cramer’s rule for the unique solution of (1.3).

Theorem 3.4

Let M and b be as in Lemma 3.2, and let L be as in Lemma 3.2. Then, (1.3) has the unique solution x = ( x 1 , x 2 , , x n ) T satisfying

(3.13) x i = det M ( i b ) L L ( i 0 ) 0 / det M L L 0 ,

where i = 1 , 2 , , n .

Proof

Since G is invertible, applying Lemma 3.3, we get the unique solution x ˆ = G 1 b ˆ of G x ˆ = b ˆ , in which x ˆ = [ x y ] and b ˆ = b 0 . It follows from (3.12) that

x y = M ( I n M M ) L ( L L ) 1 ( L L ) 1 L 0 b 0 = M b ( L L ) 1 L b .

Applying (3.4) we obtain (3.13).□

In the following theorem, we give a characterization of the core inverse and prepare for a Cramer’s rule for the core inverse in Theorem 3.6.

Theorem 3.5

Let M and L be as in Lemma 3.2. Then,

(3.14) M = ( M M M + L L ) 1 M M .

Proof

Since ( L ) = N ( M ) , M n CM and ( M ) = N ( M ) , we obtain

( L L ) ( L L ) = P N ( M ) , ( M ) , ( M M M ) ( M M M ) = P ( M ) , N ( M )

and

( M M M + L L ) ( ( M M M ) + ( L L ) ( M M M ) M M M ( L L ) ) = ( M M M ) ( M M M ) + ( L L ) ( L L ) = P ( M ) , N ( M ) + P N ( M ) , ( M ) = I n .

Therefore, M M M + L L is invertible.

Since ( L L ) M M = 0 and ( M M M ) M M = M , we have

( M M M + L L ) 1 M M = ( M M M ) M M + ( L L ) M M = M .

It follows that we get (3.14).□

Theorem 3.6

Let M and L be as in Lemma 3.2. Then, (1.3) has the unique solution x = ( x 1 , x 2 , , x n ) T satisfying

(3.15) x j = det ( M M M + L L ) ( j M M b ) det ( M M M + L L ) ,

where j = 1 , 2 , , n .

Proof

Applying Theorems 3.5 to 3.1, we have

x = ( M M M + L L ) 1 M M b ,

that is,

( M M M + L L ) x = M M b .

It follows from (3.4) that we get (3.15).□

In [30], Ji obtains the condensed determinantal expressions of M and M D . By using Theorem 3.5, we get a condensed determinantal expression of M .

Theorem 3.7

Let M and L be defined as in (3.11). Then, the core inverse M is given by:

(3.16) M i , j = det ( M M M + L L ) ( i ( M M ) e j ) det ( M M M + L L ) ,

where 1 i , j n .

Proof

Since M M M + L L is invertible, we consider

( M M M + L L ) x = ( M M ) e j

and get the solution

e i T x = det ( M M M + L L ) ( i ( M M ) e j ) det ( M M M + L L ) ,

in which i , j = 1 , , n .

It follows from (3.14) and M i , j = e i T M e j that we get (3.16).□

3.3 Examples

In the following examples, we show that our results are effective.

Example 3.1

Let M = 1 2 2 0 0 0 0 0 0 , L = 0 0 1 0 0 1 and b = 1 1 1 . It is easy to check that ( L ) = N ( M ) . By applying Lemma 3.3, we have M = 1 0 0 0 0 0 0 0 0 . Then, ( I n M M ) L ( L L ) 1 = 2 2 1 0 0 1 , ( L L ) 1 L = 0 1 0 0 0 1 , G = 1 2 2 0 0 0 0 0 1 0 0 0 0 0 1 0 1 0 0 0 0 0 1 0 0 , det ( G ) = 1 and G 1 = 1 0 0 2 2 0 0 0 1 0 0 0 0 0 1 0 1 0 0 0 0 0 1 0 0 .

By applying Theorem 3.1, we get the solution of (1.3) is x = M b = 1 0 0 .

For det 1 0 0 2 2 1 0 0 1 0 1 0 0 0 1 0 1 0 0 0 0 0 1 0 0 = 1 and det 1 1 2 0 0 0 1 0 1 0 0 1 0 0 1 0 0 0 0 0 0 0 1 0 0 = det 1 2 1 0 0 0 0 1 1 0 0 0 1 0 1 0 1 0 0 0 0 0 0 0 0 = 0 , by applying Theorem 3.4, we get x 1 = 1 1 , x 2 = 0 1 and x 3 = 0 1 . Therefore, the solution of (1.3) is x = 1 0 0 .

For det ( M M M + L L ) = det 9 18 18 0 1 0 0 0 1 = 9 , M M b = 9 0 0 ,

det ( M M M + L L ) ( 1 M M b ) = det 9 18 18 0 1 0 0 0 1 = 9 ,

det ( M M M + L L ) ( 2 M M b ) = 0 , det ( M M M + L L ) ( 3 M M b ) = 0 , and by applying Theorem 3.6, we get x 1 = 9 9 , x 2 = 0 9 and x 3 = 0 9 . Therefore, the solution of (1.3) is x = 1 0 0 .

Example 3.2

Let

M = 1 / 4 1 / 8 1 / 8 1 / 4 1 / 8 1 / 8 1 / 4 0 1 / 4 .

Then,

M # = 1 3 2 1 3 2 1 5 6 , M # = 4 / 3 4 / 3 0 8 / 3 8 / 3 4 4 / 3 4 / 3 4 , M = 2 2 2 2 2 2 2 2 6

and

L = 1 1 0

with rk ( L ) = n r and ( L ) = N ( M ) . It is easy to check that

M M M + L L = 137 / 128 125 / 128 3 / 64 119 / 128 131 / 128 3 / 64 5 / 64 3 / 128 7 / 128 .

By applying Theorem 3.5, we get

( M M M + L L ) 1 M M = 2 2 2 2 2 2 2 2 6 = M .

For

det ( M M M + L L ) = 3 / 4096 , det ( M M M + L L ) ( 1 ( M M ) e 1 ) = 3 / 2048 , det ( M M M + L L ) ( 1 ( M M ) e 2 ) = 3 / 2048 , det ( M M M + L L ) ( 1 ( M M ) e 3 ) = 3 / 2048 , det ( M M M + L L ) ( 2 ( M M ) e 1 ) = 3 / 2048 , det ( M M M + L L ) ( 2 ( M M ) e 2 ) = 3 / 2048 , det ( M M M + L L ) ( 2 ( M M ) e 3 ) = 3 / 2048 , det ( M M M + L L ) ( 3 ( M M ) e 1 ) = 3 / 2048 , det ( M M M + L L ) ( 3 ( M M ) e 2 ) = 3 / 2048 , det ( M M M + L L ) ( 3 ( M M ) e 3 ) = 9 / 2048 ,

by applying Theorem 3.7, we get

M 11 = 2 , M 12 = 2 , M 13 = 2 , M 21 = 2 , M 22 = 2 , M 23 = 2 , M 31 = 2 , M 32 = 2 , M 33 = 6 ,

that is,

M = 2 2 2 2 2 2 2 2 6 .

Acknowledgments

Hongxing Wang was supported partially by the Guangxi Natural Science Foundation (grant number 2018GXNSFAA138181), the Special Fund for Science and Technological Bases and Talents of Guangxi (grant number GUIKE AD19245148), the Xiangsihu Young Scholars Innovative Research Team of Guangxi University for Nationalities (grant number 2019RSCXSHQN03) and the Special Fund for Bagui Scholars of Guangxi (grant number 2016A17). Xiaoyan Zhang was supported partially by the National Natural Science Foundation of China (grant number 11361009) and High Level Innovation Teams and Distinguished Scholars in Guangxi Universities (grant number GUIJIAOREN201642HAO).

  1. Conflict of interest: The authors report no potential conflict of interest.

References

[1] A. Ben-Israel and T. N. E. Greville, Generalized Inverses: Theory and Applications, 2nd edn, Springer-Verlag, New York, 2003.Search in Google Scholar

[2] D. S. Cvetković Ilić and Y. Wei, Algebraic Properties of Generalized Inverses, Springer, Singapore, 2017.10.1007/978-981-10-6349-7Search in Google Scholar

[3] G. Wang, Y. Wei, and S. Qiao, Generalized Inverses: Theory and Computations, 2nd edn, Springer, Singapore, 2018.10.1007/978-981-13-0146-9Search in Google Scholar

[4] O. M. Baksalary and G. Trenkler, Core inverse of matrices, Linear Multilinear Algebra 58 (2010), no. 5–6, 681–697, 10.1080/03081080902778222.Search in Google Scholar

[5] H. Wang and X. Liu, Characterizations of the core inverse and the core partial ordering, Linear Multilinear Algebra 63 (2015), no. 9, 1829–1836, 10.1080/03081087.2014.975702.Search in Google Scholar

[6] R. E. Cline, Inverses of rank invariant powers of a matrix, SIAM J. Numer. Anal. 5 (1968), 182–197, 10.1137/0705015.Search in Google Scholar

[7] S. B. Malik and N. Thome, On a new generalized inverse for matrices of an arbitrary index, Appl. Math. Comput. 226 (2014), 575–580, 10.1016/j.amc.2013.10.060.Search in Google Scholar

[8] O. M. Baksalary and G. Trenkler, On a generalized core inverse, Appl. Math. Comput. 236 (2014), 450–457, 10.1016/j.amc.2014.03.048.Search in Google Scholar

[9] K. Manjunatha Prasad and K. S. Mohana, Core-EP inverse, Linear Multilinear Algebra 62 (2014), no. 6, 792–802, 10.1080/03081087.2013.791690.Search in Google Scholar

[10] H. Wang and J. Chen, Weak group inverse, Open Math. 16 (2018), 1218–1232, 10.1515/math-2018-0100.Search in Google Scholar

[11] Ivan Kyrchei, Determinantal representations of the quaternion core inverse and its generalizations, Adv. Appl. Clifford Algebr. 29 (2019), 104, 10.1007/s00006-019-1024-6.Search in Google Scholar

[12] H. Ma, Optimal perturbation bounds for the core inverse, Appl. Math. Comput. 336 (2018), 176–181, 10.1016/j.amc.2018.04.059.Search in Google Scholar

[13] K. Manjunatha Prasad and M. D. Raj, Bordering method to compute core-EP inverse, Spec. Matrices 6 (2018), 193–200, 10.1515/spma-2018-0016.Search in Google Scholar

[14] Hongxing Wang, Core-EP decomposition and its applications, Linear Algebra Appl. 508 (2016), 289–300, 10.1016/j.laa.2016.08.008.Search in Google Scholar

[15] H. Wang, J. Chen, and G. Yan, Generalized Cayley-Hamilton theorem for core-EP inverse matrix and DMP inverse matrix, J. Southeast Univer. (Engl. Ed.) 1 (2018), no. 4, 135–138, 10.3969/j.issn.1003-7985.2018.01.019.Search in Google Scholar

[16] Y. Gao and J. Chen, Pseudo core inverses in rings with involution, Comm. Algebra 46 (2018), no. 1, 38–50, 10.1080/00927872.2016.1260729.Search in Google Scholar

[17] D. S. Rakić, N. Č. Dinčić, and D. S. Djordjević, Core inverse and core partial order of Hilbert space operators, Appl. Math. Comput. 244 (2014), 283–302, 10.1016/j.amc.2014.06.112.Search in Google Scholar

[18] I. Kyrchei, Determinantal representations of the core inverse and its generalizations with applications, J. Math. 2019 (2019), 1–13, 10.1155/2019/1631979.Search in Google Scholar

[19] H. Wang and X. Liu, A partial order on the set of complex matrices with index one, Linear Multilinear Algebra 66 (2018), no. 1, 206–216, 10.1080/03081087.2017.1292995.Search in Google Scholar

[20] S. L. Campbell and C. D. Meyer, Generalized Inverses of Linear Transformations, Society for Industrial and Applied Mathematics (SIAM), Philadelphia, 2009.10.1137/1.9780898719048Search in Google Scholar

[21] Y. Wei, Index splitting for the Drazin inverse and the singular linear system, Appl. Math. Comput. 95 (1998), no. 2–3, 115–124, 10.1016/S0096-3003(97)10098-4.Search in Google Scholar

[22] Y. Wei, A characterization for the W-weighted Drazin inverse and a Cramer rule for the W-weighted Drazin inverse solution, Appl. Math. Comput. 125 (2002), no. 2–3, 303–310, 10.1016/S0096-3003(00)00132-6.Search in Google Scholar

[23] Y. L. Chen, Representations and Cramer rules for the solution of a restricted matrix equation, Linear and Multilinear Algebra 35 (1993), no. 3–4, 339–354, 10.1080/03081089308818266.Search in Google Scholar

[24] K. Morikuni and M. Rozluŏzník, On GMRES for singular EP and GP systems, SIAM J. Matrix Anal. Appl. 39 (2018), no. 2, 1033–1048, 10.1137/17M1128216.Search in Google Scholar

[25] F. Toutounian and R. Buzhabadi, New methods for computing the Drazin-inverse solution of singular linear systems, Appl. Math. Comput. 294 (2017), 343–352, 10.1016/j.amc.2016.09.013.Search in Google Scholar

[26] G. Wang, A Cramer rule for finding the solution of a class of singular equations, Linear Algebra Appl. 116 (1989), 27–34, 10.1016/0024-3795(89)90395-9.Search in Google Scholar

[27] Y. Wei and H. Wu, Convergence properties of Krylov subspace methods for singular linear systems with arbitrary index, J. Comput. Appl. Math. 114 (2000), no. 2, 305–318, 10.1016/S0377-0427(99)90237-6.Search in Google Scholar

[28] H. Ma and T. Li, Characterizations and representations of the core inverse and its applications, Linear Multilinear Algebra (2019), 10.1080/03081087.2019.1588847.Search in Google Scholar

[29] A. Ben-Israel, A Cramer rule for least-norm solutions of consistent linear equations, Linear Algebra Appl. 43 (1982), 223–226, 10.1016/0024-3795(82)90255-5.Search in Google Scholar

[30] J. Ji, Explicit expressions of the generalized inverses and condensed Cramer rules, Linear Algebra Appl. 404 (2005), 183–192, 10.1016/j.laa.2005.02.025.Search in Google Scholar

[31] I. Kyrchei, Analogs of Cramer’s rule for the minimum norm least squares solutions of some matrix equations, Appl. Math. Comput. 218 (2012), no. 11, 6375–6384, 10.1016/j.amc.2011.12.004.Search in Google Scholar

[32] I. Kyrchei, Explicit formulas for determinantal representations of the Drazin inverse solutions of some matrix and differential matrix equations, Appl. Math. Comput. 219 (2013), no. 14, 7632–7644, 10.1016/j.amc.2013.01.050.Search in Google Scholar

[33] I. Kyrchei, Cramer’s rule for generalized inverse solutions, in: I. Kyrchei (Ed.), Advances in Linear Algebra Research, pp. 79–132, Nova Science Publishers, New York, 2015.Search in Google Scholar

[34] I. Kyrchei, Weighted singular value decomposition and determinantal representations of the quaternion weighted Moore-Penrose inverse, Appl. Math. Comput. 309 (2017), 1–16, 10.1016/j.amc.2017.03.048.Search in Google Scholar

[35] J. Ji, A condensed Cramer’s rule for the minimum-norm least-squares solution of linear equations, Linear Algebra Appl. 437 (2012), no. 9, 2173–2178, 10.1016/j.laa.2012.06.012.Search in Google Scholar

[36] G. Wang and Z. Xu, Solving a kind of restricted matrix equations and Cramer rule, Appl. Math. Comput. 162 (2005), no. 1, 329–338, 10.1016/j.amc.2003.12.118.Search in Google Scholar

Received: 2019-08-30
Revised: 2020-03-07
Accepted: 2020-04-28
Published Online: 2020-07-06

© 2020 Hongxing Wang and Xiaoyan Zhang, published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 23.4.2024 from https://www.degruyter.com/document/doi/10.1515/math-2020-0178/html
Scroll to top button