Is the set ${X in mathcal{M}({m times n}) : rho(M-NX) < 1} $ connected?
Suppose $M in mathcal M(n times n; mathbb R)$ and $N in mathcal M(n times m; mathbb R)$ are fixed with $Nneq 0$. Let
begin{align*}
E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1},
end{align*}
where $rho(cdot)$ denotes the spectral radius of a square matrix. We assume $E$ is not empty.
This set is open. (see my other question concerning the closure). I would also like to know whether it is connected. In this case, equivalently, is the set path-connected?
linear-algebra general-topology connectedness path-connected
add a comment |
Suppose $M in mathcal M(n times n; mathbb R)$ and $N in mathcal M(n times m; mathbb R)$ are fixed with $Nneq 0$. Let
begin{align*}
E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1},
end{align*}
where $rho(cdot)$ denotes the spectral radius of a square matrix. We assume $E$ is not empty.
This set is open. (see my other question concerning the closure). I would also like to know whether it is connected. In this case, equivalently, is the set path-connected?
linear-algebra general-topology connectedness path-connected
2
I feel so that yes, but I can't prove it.
– peterh
Jun 1 at 22:59
Do you have any examples, or other form of empirical evidence, that would point in the direction of connected? Possibly in special cases, e.g., when $M$ is symmetric?
– Oskar Limka
Jun 10 at 16:54
@OskarLimka: Sorry, I only have some intuitive feeling. The set seems 'simple' but when concerning the topological properties, the only thing I know is openness.
– user1101010
Jun 11 at 16:55
add a comment |
Suppose $M in mathcal M(n times n; mathbb R)$ and $N in mathcal M(n times m; mathbb R)$ are fixed with $Nneq 0$. Let
begin{align*}
E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1},
end{align*}
where $rho(cdot)$ denotes the spectral radius of a square matrix. We assume $E$ is not empty.
This set is open. (see my other question concerning the closure). I would also like to know whether it is connected. In this case, equivalently, is the set path-connected?
linear-algebra general-topology connectedness path-connected
Suppose $M in mathcal M(n times n; mathbb R)$ and $N in mathcal M(n times m; mathbb R)$ are fixed with $Nneq 0$. Let
begin{align*}
E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1},
end{align*}
where $rho(cdot)$ denotes the spectral radius of a square matrix. We assume $E$ is not empty.
This set is open. (see my other question concerning the closure). I would also like to know whether it is connected. In this case, equivalently, is the set path-connected?
linear-algebra general-topology connectedness path-connected
linear-algebra general-topology connectedness path-connected
edited Jun 5 at 23:33
asked Jun 1 at 17:01
user1101010
7551630
7551630
2
I feel so that yes, but I can't prove it.
– peterh
Jun 1 at 22:59
Do you have any examples, or other form of empirical evidence, that would point in the direction of connected? Possibly in special cases, e.g., when $M$ is symmetric?
– Oskar Limka
Jun 10 at 16:54
@OskarLimka: Sorry, I only have some intuitive feeling. The set seems 'simple' but when concerning the topological properties, the only thing I know is openness.
– user1101010
Jun 11 at 16:55
add a comment |
2
I feel so that yes, but I can't prove it.
– peterh
Jun 1 at 22:59
Do you have any examples, or other form of empirical evidence, that would point in the direction of connected? Possibly in special cases, e.g., when $M$ is symmetric?
– Oskar Limka
Jun 10 at 16:54
@OskarLimka: Sorry, I only have some intuitive feeling. The set seems 'simple' but when concerning the topological properties, the only thing I know is openness.
– user1101010
Jun 11 at 16:55
2
2
I feel so that yes, but I can't prove it.
– peterh
Jun 1 at 22:59
I feel so that yes, but I can't prove it.
– peterh
Jun 1 at 22:59
Do you have any examples, or other form of empirical evidence, that would point in the direction of connected? Possibly in special cases, e.g., when $M$ is symmetric?
– Oskar Limka
Jun 10 at 16:54
Do you have any examples, or other form of empirical evidence, that would point in the direction of connected? Possibly in special cases, e.g., when $M$ is symmetric?
– Oskar Limka
Jun 10 at 16:54
@OskarLimka: Sorry, I only have some intuitive feeling. The set seems 'simple' but when concerning the topological properties, the only thing I know is openness.
– user1101010
Jun 11 at 16:55
@OskarLimka: Sorry, I only have some intuitive feeling. The set seems 'simple' but when concerning the topological properties, the only thing I know is openness.
– user1101010
Jun 11 at 16:55
add a comment |
1 Answer
1
active
oldest
votes
$newcommand{NN}{{mathbb{N}}}newcommand{CC}{{mathbb{C}}}newcommand{RR}{{mathbb{R}}}newcommand{QQ}{{mathbb Q}}newcommand{ra}{rightarrow}newcommand{ds}{displaystyle}newcommand{mat}[1]{left(begin{matrix}#1end{matrix}right)}newcommand{lam}{lambda}renewcommand{P}{{mathcal P}}newcommand{N}{{mathcal N}}newcommand{M}{{mathcal M}}renewcommand{L}{{mathcal L}}newcommand{EQ}[2]{begin{equation} {{#2}label{#1}} end{equation}}newcommand{eps}{varepsilon}newcommand{tabb}[1]{begin{matrix}#1end{matrix}}
DeclareMathOperator{rank}{rank}
DeclareMathOperator{bldiag}{bldiag}$
Theorem: $E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1}$ is connected.
The proof is surprisingly difficult (at least mine). After some preliminaries, I will first treat the case $rank(N)=1$ which is relatively
simple, then show that the case $rank(N)=2$ implies the general case, next treat the case $n=3$ as a preparation and finally
prove the Theorem for $rank(N)=2$ and all $n$.
The case $rank(N)=n$ (that is essentially $M=0$ and $N=I$, see below) is trivial,
because any $Xin E$ can be connected
to $0$ by the path $X_s=sX$ in $E$. It will not be mentioned again.
As explained in the linked answer to the question of the closure of $E$, we can
assume that $N=mat{I_m\0}$ and that the first $m$ rows of
$M$ vanish.
Here a further reduction is useful. Consider the matrix $tilde M(lam)=lambldiag(0,I_{n-m})-M$
which equals $lam I-M$ with the first $m$ diagonal elements replaced by zeros. Consider also the gcd $G(lam)$
of all subdeterminants of $tilde M(lam)$ of size $n-m$.
Claim 1: It is sufficient to treat the case that $G(lam)=1$.
This observation simplifies the presentation here; in the linked answer it was not needed.
Proof of Claim 1. Suppose that $G(lam)neq 1$ and suppose first that it has a real zero $mu$.
Since all subdeterminants of size $n-m$ of $tilde M(mu)$ vanish, its last $n-m$ rows are linearly dependent. Thus there
exists a nonzero row vector $v=(0,dots,0,v_{m+1},dots,v_n)$ such that $v,tilde M(mu)=0$ and hence $v,M=mu,v$. We can assume that $v_nneq0$: There exists some $k$ such that $v_kneq0$. Then we exchange rows $k$ and $n$ and columns $k$ and $n$, respectively, and the components $v_k$ and $v_n$.
Consider now the matrix $T$ with last row $v$ and first $n-1$ rows $e_j$, $j=1,dots,n-1$, where $e_j$ denotes the $j$th unit vector.
Let $M'=T,M,T^{-1}$. A small calculation (using $M',T=T,M$) shows that the first $m$ rows of $M'$ vanish and the last row of $M'$
is $mu e_n$. Since $TN=N$, the matrices $M-NX$ and $M'-NX'$ with $X'=XT^{-1}$ are similar and hence have the same spectral radius.
By construction, $M'-NX'$ is block triangular for all $X'$ and always has the eigenvalue $mu$. The latter must have $|mu|<1$ because
there exists $X'_0$ such that $rho(M'-NX'_0)<1$. Therefore the last row and the last column of $M'-NX'$ are irrelevant for the decision
whether $rho(M'-NX')<1$ or not. We have shown that $n$ can be reduced if $G(lam)$ has a real zero.
If $G(lam)$ has a nonreal zero $mu=alpha+beta i$ then there is a left eigenvector $v=r+i s$, where the first
$m$ components of $r$ and $s$ vanish. Now $(r+i s)M=(alpha +beta i)(r + is)$ means that
$rM=alpha r-beta s$ and $sM=beta r +alpha s$. We proceed as above with $T$ having the last two rows $r,s$. The result
is a block triangular matrix $M'$ with lower right block $mat{alpha&-beta\beta&alpha}$ and again $n$ can be reduced.
As we have shown that $n$ can be reduced if $G(lam)neq 1$, this can be repeated until $G(lam)=1$ and Claim 1 is proved.
Case $mathbf {m=1}$: First, we recall considerations of the linked answer. We use the Laplace (or cofactor) expansion
for the first row of the determinant giving the characteristic polynomial $p_{M-NX}(lam)$
$${det(lam I - M+ NX)=(x_1+lam)C_1(lam)+sum_{k=2}^n x_kC_k(lam)}$$
where $X=(x_1,dots,x_n)$ and $C_j(lam)$, $j=1,dots,n$ are the $1,j$ cofactors of $lam I - M$, that is
$C_j(lam)=(-1)^{j+1}M_{1j}(lam)$, where $M_{1j}(lam)$ is the determinant of the $n-1$ by $n-1$
matrix obtained from $lam I-M$ by deleting its first row and its $j$-th column.
Observe that $C_1(lam)in lam^{n-1}+P_{n-1}$ and $C_j(lam)inP_{n-1}$, where $P_k$ denotes the
vector space of all polynomials of degree less than $k$.
Before continuing the proof, we need an auxiliary statement.
Lemma 1 Let $D(lam)$ denote the gcd of $C_j(lam)$, $j=1,dots,n$ and let $d$ denote its
degree. Then the mapping
$${L:RR^nto D(lam)P_{n-d}, L(X)=sum_{k=1}^n x_kC_k(lam)mbox{ if } X=(x_1,dots,x_n)}$$
is surjective (or onto).
Remark: The characteristic polynomial of $M-NX$ is the given by $p_{M-NX}(lam)=lam C_1(lam)+L(X)$.
For the proof, see the linked answer.
Now we continue with the proof in the case $m=1$. By Claim 1, we can assume that $G(lam)=1$ and hence
by Lemma 1, $L$ is an isomorphism. It induces a bijection between $X$ and the characteristic polynomial
$p_{M-NX}(lam)$ as stated in the above remark.
Consider now any element $X_1$ of $E$ and factor $q_1(lam):=p_{M-NX_1}(lam)=prod_{k=1}^{n}(lam-mu_k)$
with certain (not necessarily real) $mu_k$, $|mu_k|<1$.
Now define for $sin[0,1]$ the polynomials $q_s(lam)=prod_{k=1}^{n}(lam-smu_k)$.
By construction, their coefficients are real and their zeros are in the open unit disk.
Therefore the matrices $X_s$ corresponding to them by $q_s(lam)=lam C_1(lam)+L(X_s)$
are elements of $E$ and connect $X_1$ to the uniquely determined $X_0$ with
$p_{M-NX_0}(lam)=lam^n$. This shows that all elements of $E$ are connected to $X_0$ within $E$
and proves the Theorem for $m=1$.
Actually, this allows to prove that $E$ is simply connected in the case $m=1$.
The case $mathbf{m=2}$ implies the Theorem for all $mathbf m$. Again, we suppose that $G(lam)=1$
(see Claim 1). Consider two elements
$X_a,X_b$ of $E$, i.e. we have $rho(M-NX_a)<1$ and $rho(M-NX_b)<1$. By going over to
appropriate very close matrices, we can assume that the elements of $X_a,X_b$ are algebraically
independent transcendents over $QQ(M)$. Then we construct for $k=2,dots,m-1$ matrices
$X_k$ such that $X_a$ can be connected to $X_b$ in $E$ via these auxiliary matrices,
modifying only two rows at one time.
For $kin{2,dots,m-1}$, we first use the $m$ by $n$ matrix $tilde X_k$ with the
first $k$ rows from $X_b$ and the last $m-k$ rows from $X_a$. The formula is
$tilde X_k=bldiag(I_k,0)X_b+bldiag(0,I_{n-k})X_a$. There is no reason why
$tilde X_k$ should be in $E$, but we claim
Claim 2: The gcd of all $1,j$ cofactors $C_j^{(k)}(lam)$ of
$lam I - (M-Ntilde X_k)$ is 1.
The proof of this Claim is very similar to the considerations below Corollary 2 in the
linked answer. Consider the gcd $D(lam)$ of $C_j^{(k)}(lam)$, $j=1,dots,n$.
A priori, $D(lam)$ is a polynomial in $lam$ with coefficients in $QQ(M,hat X_k)$ because the
$C_j^{(k)}(lam)$ are; here $hat X_k$ denotes the matrix obtained from $tilde X_k$ by erasing the
first row. The $C_j^{(k)}(lam)$ are actually in $K[hat X_k][lam]$, where $K=QQ(M)$.
Since $K$ is a field, $K[hat X_k]$ is a unique factorisation domain because the elements of
$hat X_k$ are algebraically independent transcendents.
We multiply $D(lam)$ by the lcm
of the denominators of the coefficients (seen as elements of $K(hat X_k)$), and then divide by
its content (that is the gcd of the coefficients seen as elements of $K[hat X_k]$).
We call the result $D(lam)$ again.
Then $D(lam)in K[hat X_k][lam]$ has content 1 and divides every $C_j^{(k)}(lam)$ in
$K(hat X_k)[lam]$ by definition. By Gauss's Lemma,
it follows that $D(lam)$ divides every $C_j^{(k)}(lam)$ in $K[hat X_k][lam]$.
Now, for $j=1,dots,n$,
$C_j^{(k)}(lam)$ does not contain any element of the $j$-th column of $hat X_k$.
Therefore $D(lam)$ must actually be an element of $K[lam]$ and it divides
in $K[hat X_k]$ the $n$ cofactors $C_j^{(k)}(lam)$.
Therefore $D(lam)$ must also divide the $1,j$ cofactors $bar C_j(lam)$ of
$lam I-(M-Nbar X)$, where $bar X$ is any $m$ by $n$ matrix.
This follows by substitution of the elements of $hat{bar X}$ for the corresponding
elements of $hat X_k$.
By choosing the rows of $hat{bar X}$ as appropriate unit vectors, we can achieve that
one of the cofactors is any of the $n-m$ by $n-m$ subdeterminants of $M$ we want.
Therefore $D(lam)$ divides any of the latter subdeterminants and we have assumed
that their gcd $G(lam)$ is 1.
This proves Claim 2.
At this point, Lemma 1 applies and yields that the mapping $L^{(k)}:RR^nto P_n$,
$V_1=(x_1,dots,x_n)mapsto L^{(k)}(V_1)=sum_{j=1}^n x_j C_j^{(k)}$ is bijective.
We choose a vector $V_1^{(k)}$ such that $lam C_1^{(k)}+L^{(k)}(V_1^{(k)})=lam^n$.
This means that the matrix $X_k$ obtained from $hat X_k$ by adding the first row
$V_1^{(k)}$ gives a nilpotent $M-NX_k$; in any case $X_k$ is an element of $E$.
We complete $X_2,dots,X_{m-1}$ by $X_1=X_a$ and $X_m=X_b$.
For any $kin{2,dots,m}$, the matrices $X_{k-1}$ and $X_{k}$ have only $two$ different rows:
the first and the $k$th. Both are in $E$ as we have seen. Here the case $m=2$ can be applied
after exchanging the second and the $k$th rows and columns, respectively, of the two matrices.
Hence they are connected by a path in $E$. Combining these $m-1$ paths, we obtain a path
in $E$ connecting $X_a$ and $X_b$. This proves the Theorem for all $mgeq3$ if it is true for $m=2$.
Case $mathbf{n=3, m=2}$. We denote the third row of $M$
by $mat{a&b&c}$, the second row of $X$ by $mat{d&e&f}$.
By Claim 1, we can assume that $aneq0$ or $bneq0$. In order to apply Lemma 1, it is interesting
to study the gcd of all $2times2$ submatrices of $mat{d&lam+e&f\-a&-b&lam-c}$,
in particular to establish a condition when this gcd is 1.
As above, it is not equal to 1 iff there is a number $mu$ such that
$B(mu)=mat{d&mu+e&f\-a&-b&mu-c}$ has rank 1.
If $a=0$, then $bneq0$ and $B(mu)$ has rank 2 for all $mu$ iff $dneq0$.
If $aneq0$, we first add a multiple of the second row of $B(mu)$ to the first
to obtain $mat{0&mu+e-frac{db}a&f+frac da mu-frac{dc}a\-a&-b&mu-c}$ and then
add a multiple of the second column to the third one to obtain
$tilde B(mu)=mat{0&mu+e-frac{db}a&f-frac{de}a+frac{d^2b}{a^2}-
frac{dc}a\-a&-b&mu-c+frac{db}a}$,
a matrix of the same rank as $B(mu)$. It has rank 2 for all $mu$ iff
the third element of the first row is nonzero, that is iff
$Delta(X)=a^2f-{de}a+{d^2b}-{dc}aneq0$.
Actually the condition in the case $a=0$ can be included in the latter one since $bneq0$.
Observe that $Delta(X)$ is closely related to the determinant of the matrix associated to
the operator $L$ of Lemma 1 with respect to canonical bases of $RR^n$ and $P_n$.
This is not surprising because $L$ is invertible iff this determinant is nonzero and, by Lemma 1,
iff the gcd of the $1,j$ cofactors of $lam I-M+NX$ equals 1.
We call the set $Delta(X)=0$ the singular surface.
Consider now two matrices $X_1,X_2in E$. If we can find a path connecting
their second rows $mat{d_1&e_1&f_1}$ and $mat{d_2&e_2&f_2}$ on which $Delta$ does not vanish
then we can extend it to a path connecting $X_1$ and $X_2$ in $E$. Indeed,
we can first find path connecting the characteristic polynomials
of $M-NX_1$ and $M-NX_2$ (for example via $lam^3$) and then the corresponding
path connecting the first rows of $X_1,X_2$ by Lemma 1.
More precisely, consider a path $Z_s=mat{d_s&e_s&f_s}$, $sin[1,2]$ in $RR^3$ such that
$a^2f_s-{d_se_s}a+{d_s^2b}-{d_sc}aneq0$ for all $s$.
We denote the determinants of Lemma 1 by $C_j^s$, the operator associated to $Z_s$ by $L_s$.
Observe that $L_s$ and $L_s^{-1}$ depend continuously upon $s$.
If the characteristic polynomials of
$X_i$ are $p_i(lam)=prod_{k=1}^3(lam-mu_k^i)$, we connect them by $p_s(lam)$
defined by
$$p_s(lam)=left{begin{array}lprod_{k=1}^3left(lam-(3-2s)mu_k^1right)mbox{ if }1leq sleq 1.5\[0.2cm]
prod_{k=1}^3left(lam-(2s-3)mu_k^2right)
mbox{ if }1.5leq sleq 2end{array}right.$$
Observe that for every $s$, the zeros of $p_s$ are in the open unit disk.
Then put $Y_s=L_s^{-1}left(p_s(lam)-lam C_1^sright)$ and let $X_s$ be the matrices with
rows $Y_s$ and $Z_s$. By construction and the remark below Lemma 1, for every $s$,
the characteristic polynomial of $M-NX_s$ is then $p_s(lam)$, therefore $X_sin E$.
It is easy to see that such a path connecting the second rows of $X_1, X_2$ exists
if $Delta(X_j)=a^2f_j-{d_je_j}a+{d_j^2b}-{d_jc}a$, $j=1,2$, are nonzero and have the same sign.
If $aneq0$, then we can first increase the modulus of $f$ sufficiently, then modify $d_1,e_1$
into $d_2,e_2$ and finally reduce $f$ to $f_2$. The case $a=0$ is even simpler and omitted.
Since $E$ is open, we can assume that both $Delta(X_j)neq0.$ It remains to treat
the case that both have opposite sign. If $aneq0$ and $Delta(X)=0$ then the above calculation
shows that the gcd $D(lam)$ of the $2times 2$ submatrices of $B(lam)$
for general $X$ with second row $mat{d&e&f}$ is $D(lam)=lam+e-frac{db}a$.
As $D(lam)$ is a factor of the characteristic polynomial of $M-NX$, a path connecting
$X_1,X_2$ in $E$ must cross the singular surface at a point $X$ where $|e-frac{db}a|<1$.
This makes it clear that the singular surface causes problems in finding a path in $E$
between two given elements, not only for $n=3$, but also for general $n$.
Still in the case $aneq0$, $Delta(X_1)>0>Delta(X_2)$, we choose as crossing point
a point $X^{c}$ with second row $mat{0&0&0}$ (hence $D(lam)=lam$)
and such that $rho(M-NX^c)<1$. This is possible
using Lemma 1. Then we consider
two matrices $X_j^c$, $j=1,2$, which differ from $X^c$ only in the second row:
it is $mat{0&0&+eps}$ for $X_1^c$ and $mat{0&0&-eps}$ for $X_2^c$. If $eps>0$ is sufficiently
small then both matrices are in $E$. As seen above, we can connect $X_1$ and $X_1^c$ in $E$ as
they are both in the set $Delta>0$ and we can connect $X_2$ and $X_2^c$ in $E$
as $Delta<0$ for both. Obviously, $X_1^c$ and $X_2^c$ are connected in $E$ via the point
$X^c$ on the singular surface. This completes the proof in the case $aneq0$.
In the case $a=0$, the proof is quite similar and we only mention the differences.
The singular surface is simply $d=0$ and on it, we have $D(lam)=detmat{lam+e&f\-b&lam-c}$.
Since $bneq0$, we can use $X^c$ with second row $mat{0&c&c^2/b}$ (so that $D^c(lam)=lam^2$)
and with vanishing first row: the characteristic polynomial of $M-NX^c$ is $lam^3$.
Then we use $X_j^c$ with second row $mat{pmeps&c&c^2/b}$ and small $eps$.
This completes the proof in the case $n=3$, $m=2$.
General $mathbf n$ and $mathbf{m=2}$. We suppose again that $G(lam)=1$ (see Claim 1).
As above, we can associate to any row vector $ZinRR^n$ the gcd $D^Z(lam)$ of the subdeterminants
$C_j^Z(lam)$, $j=1,dots,n$ of the $n-1$ by $n$ matrix $bar M(Z,lam)=
mat{lam e_2+Z\tilde M(lam)}$
(see before Claim 1 for $tilde M(lam)$). It is 1 iff the operator $L^Z$ associated to the
$C_j^Z(lam)$ by Lemma 1 is invertible and this is the case iff the determinant $Delta(Z)$
of the matrix associated to $L^Z$ for the canonical bases of $RR^n$ and $P_n$
has nonzero determinant. The singular surface is again $Delta(Z)=0$.
Consider two matrices $X_i$, $i=1,2$, in $E$ and let $Y_i,Z_i$ denote the two rows of each of them.
As for $n=3$, we can assume that $Delta(Z_i)neq0$.
As in the case $n=3,m=2$ we can prove
Lemma 2: If there is a path $Z_s$, $sin[1,2]$ in $RR^n$ connecting $Z_1$ and $Z_2$ such that
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk, then it can be completed
by a path $Y_s$ such that $X_s=mat{Y_s\Z_s}$ connects $X_1$ and $X_2$ in $E$.
Using this Lemma, the Theorem is proved once we show
Claim 3: There always exists a path connecting $Z_1,Z_2$ such that,
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk.
Proof: We prove the Claim by induction over $n$. For $n=3$, it had already been shown
above. Suppose now that it has been shown for $n-1$.
We consider two cases: a) the first column of $ M$ vanishes,
b) it contains a nonzero element.
case a) It is almost identical to the case $n=3,m=2,a=0$.
Denote by $check M$ the $(n-2)times(n-1)$ matrix obtained from $M$ by deleting
the first column. Then the gcd of the $n-2$ by $n-2$ subdeterminants
of $lam(0,I_{n-2})-check M$ is the same as the one for $tilde M(lam)$ and hence
equals $G(lam)=1$. Therefore the condition $Delta(Z)=0$
is equivalent to the first component of $Z$ being 0.This follows by cofactor expansion with respect to the first column. If $Delta(Z)=0$
then $D(lam)$ is the characteristic polynomial of $mat{ -check Z\check M }$,
where $check ZinRR^{n-1}$ is obtained from $Z$ by erasing the first component.
Since the gcd of the $n-2$ by $n-2$ subdeterminants of $lam(0,I_{n-2})-check M$
is $1$, we can find a vector $check Z_cinRR^{n-1}$ such that
the matrix $mat{ -check Z_c\check M }$ has a characteristic polynomial $q(lam)$
with zeros in the open unit disk. We just have to use Lemma 1 in dimension $n-1$.
Now recall $X_i$, $Y_i$, $Z_i$, $i=1,2$. If the first components of $Z_1$ and $Z_2$ have
the same sign, we can simply connect them by any path on which the first component does not
vanish. The critical case is when the first component $Z_{11}$ of $Z_1$ is positive, say,
and the corresponding $Z_{21}$ is negative. Then we can
connect $Z_1$ to the vector $(+eps,check Z)$ and $Z_2$ to the vector $(-eps,check Z)$
by a path that does not cross the singular surface.
Finally, we can simply connect $(pmeps,check Z)$ by the segment. It crosses the singular surface
at the point $(0,check Z)$ and by construction, $D^{(0,check Z)}(lam)=q(lam)$
has only zeros in the open unit disk. This completes the proof in case a).
case b) The first column of $M$ contains some nonzero element. This is, of course, the generic case
and it has taken me a long time to find a reduction - I had tried to find directly for any $n$
a path crossing the singular surface only at points $Z^c$ with $D^{Z^c}(lam)$ having
zeros $mu$, $|mu|<1$, but this turned out to be very complicated.
The present proof using induction is comparatively simple.
First, we prepare the matrix $M$ for the inductive step, more precisely, we transform the expression
$M-NX$ into similar matrices such that $N$ remains unchanged. The eigenvalues of $M-NX$ do
not change either in this way.
Suppose that the first element of the $k$th row of $M$ is nonzero. Then we exchange the third
and the $k$th row of $M$ and the third and the $k$th columns of $M$ and $X$. Since the row
operation does not have any effect on $N$, this is a similarity transformation for
$M-NX$ leaving $N$ unchanged. Next we multiply the third row of $M$ by a convenient number
and divide the third columns of $M$ and $X$ by the same number. So we can assume that
the first element of the third row of $M$ is 1.
If $alpha$ denotes the first element of the fourth row of $M$,
we now add $-alpha$ times the third row to it. This row operation is equivalent to multiplication
of $M$ from the left by some elementary matrix $T$. Observe that again, $N$ is unchanged by this
operation. Then we multiply from the right by the inverse of $T$. This amounts to adding
$alpha$ times the fourth columns of $M$ and $X$ to their third columns. Thus we can assume that
the fourth row of $M$ starts with a 0.
In the same way, we treat the remaining rows of $M$. So we can assume that the first
column of $M$ is the third unit vector.
It is also useful to add a certain multiple of the first column to the $k$th for $M$ and $X$ and subtract the same multiple
of the $k$tk row from the first row for $M$. This temporarily leads to a matrix $M$ the first row of which is not
zero, but this can be repaired later. We can do this without interfering with $N$ if $kgeq3$. As it is a
similarity transformation, the eigenvalues of $M-NX$ do not change. In the case $k=2$, the above transformation can still be
done, but we have to clarify how to deal with $N$. So denote by $T_n$ the $ntimes n$ matrix correponding to adding
a certain multiple of the first column to the second one. Then we modify $M-NX$ into
$T_n^{-1}M T_n- (T_n^{-1}N T_2) (T_2^{-1}XT_n)$. Since $T_n^{-1}N T_2=N$, this leads again to an expression of the same form
if we substract a multiple of the second row from the first not only for $M$, but also for $X$.
After having done all these modification, we simple incorporate the new first line into $X$.
So we can choose the elements of the third row of $M$ besides the leading 1 as we like - only $X$
has to be modified accordingly. Observe that all these operations can be undone.
In summary, we can assume from now on that:
1. The first column of $M$ is the third unit vector,
2. The elements of $M$ in the third row besides 1 are algebraically independent
transcendents over $Q(check M)$.
Here $check M$ denotes the matrix obtained from $M$ by removing the three first rows and the first column.
Let us write
$$M=mat{0&tabb{dots&0}\
0&tabb{dots&0}\
1& R\
0_{n-3}&check M},$$
where $R$ denotes a row vector of $n-1$ elements and $0_{n-3}$ indicates a column vector of $n-3$ zeros.
Consider first the $(n-1)times(n-1)$ matrix $L=mat{0\R\check M}$ and the $(n-1)times1$ matrix $K=mat{1\0_{n-2}}$.
With row vectors $UinRR^{n-1}$ we can consider the matrices $L-KU$. Here we are actually in the case $m=1$.
The gcd $F(lam)$ of the $(n-2)times(n-2)$ submatrices of $tilde L(lam)=lambldiag(0,I_{n-2})-L$
(analogous to $tilde M(lam)$) satisfies $F(lam)=1$.
This comes from the fact that the second row $R$ of $L$ consists of
algebraically independent transcendents over $Q(check M)$ and it had been shown in the proof
of Claim 2.
Therefore by Lemma 1, there exists a vector $U_c$ such that the characteristic polynomial of $L-KU_c$ is any prescribed polynomial
with roots in the open unit disk.
Recall again $X_i$, $Y_i$, $Z_i$, $i=1,2$, and denote $mu_i$ the first elements of $Z_i$. Then we want to connect
$Z_1$ and $Z_2$ by a path as required in Claim 3 by first connecting $Z_1$ and $(mu_1,U_c)$, then $(mu_1,U_c)$ and
$(mu_2,U_c)$ and finally $(mu_2,U_c)$ and $Z_2$. By slightly modifying $mu_1,,mu_2$ and $U_c$, if
necessary, we can assume that none of the four points is on the singular surface.
Connection of $(mu_1,U_c)$ and $(mu_2,U_c)$. Let us assume that $mu_2>mu_1$.
Then the straight line $sto(s,U_c)$, $sin[mu_1,mu_2]$, connects the two points.
As the $(mu_j,U_c)$ are not on the singular surface, this path can cross the
singular surface only at finitely many points because the equation for the
corresponding $s$ is polynomial.
If $s$ is some crossing point then the corresponding
gcd $D^{(s,U_c)}(lam)$ is a divisor of the subdeterminant formed by the columns $2$ to $n$ of
$bar M((s,U_c),lam)$
which happens to be the characteristic polynomial of $L-KU_c$ diskussed above. The vector
$U_c$ had been chosen such that all its zeros are in the open unit disk. Hence this is also
the case for the zeros of $D^{(s,U_c)}(lam)$.
Connection of $Z_1$ and $(mu_1,U_c)$. We want to achieve this by a path $(mu_1,U(s))$ where
$U:[0,1]toRR^{n-1}$ has to be determined such that
$U(1)=U_c$ and $(mu_1,U(0))=Z_1$. For convenience,
we write $U(s)=(U(s)_2,dots,U(s)_n)$, $R=(R_2,dots,R_n)$,
$check M=(m_{ij})_{igeq4,jgeq2}$ and abbreviate $D^{(mu_1,U(s))}(lam)$ by $D_s(lam).$
We want that the gcd $D_s(lam)$ of the $(n-1)times(n-1)$ subdeterminants of
$$bar M(s,lam)=left(
begin{array}{cccccc}
mu_1&lam+U(s)_2&U(s)_3&dots&dots&U(s)_n\-1&-R_2&lam-R_3&dots&dots&-R_n\
0&-m_{42}&-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&-m_{n3}&dots&dots&lam-m_{nn}end{array}right)$$
to have only zeros in the open unit disk if the path crosses the singular surface.
Here we use that the gcd of these subdeterminants does not change if we add a multiple of one
row or column to another row or column, respectively.
Precisely first, we add $mu_1$ times row 2 to the first row in order to have a
zero in the upper left corner; we obtain as first row
$$mat{0&tabb{lam+U(s)_2-mu_1R_2&U(s)_3+mu_1lam-mu_1R_3&dots&U(s)_n-mu_1R_n}}.$$
Then we add $-mu_1$ times column 2 to column 3 to remove the multiple of $lam$ from the third
component in row 1. We obtain as new first row
$mat{0&tabb{lam+tilde U(s)_2&tilde U(s)_3&dots&tilde U(s)_n}},$
where $tilde U(s)_j=U(s)_j-mu_1R_j$ for $jneq3$ and
$tilde U(s)_3=U(s)_3-mu_1U(s)_2-mu_1R_3+mu_1^2R_2.$ Observe that $tilde U(0)$ and
$tilde U(1)$ are determined by $Z_1$ and $U_c$ and that we have to find a path connecting
them with the property wanted in Claim 3.
Altogether, for the matrix
$$tilde{bar M}(s,lam)=left(
begin{array}{cccccc}
0&lam+tilde U(s)_2&tilde U(s)_3&tilde U(s)_4&dots&tilde U(s)_n\
-1&-R_2&lam-R_3+mu_1R_2&-R_4&dots&-R_n\
0&-m_{42}&mu_1m_{42}-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&mu_1m_{n2}-m_{n3}&-m_{n4}&dots&lam-m_{nn}end{array}right),$$
the gcd of its $(n-1)times (n-1)$ submatrices is the same as for $bar M(s,lam)$,
namely $D_s(lam)$.
This gcd is the same as the gcd of the $(n-2)times (n-2)$
subdeterminants for the matrix $hat M(s,lam)$
obtained from $tilde{bar M}(s,lam)$ by erasing the first column and the second row.
Indeed, any $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
containing the first column agrees with
the $(n-2)times (n-2)$ subdeterminant of $hat M(s,lam)$ obtained by erasing its
first column and second row. The $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
not containing the first column is a linear combination of the smaller determinants by
Laplace (or cofactor) expansion. It is convenient now to exchange the first two columns
of $hat M(s,lam)$ to have
$$hat M(s,lam)=left(
begin{array}{ccccc}
tilde U(s)_3&lam+tilde U(s)_2&tilde U(s)_4&dots&tilde U(s)_n\
mu_1m_{42}-m_{43}&-m_{42}&lam-m_{44}&dots&-m_{4n}\
vdots&&&ddots&vdots\
mu_1m_{n2}-m_{n3}&-m_{n2}&-m_{n4}&dots&lam-m_{nn}end{array}right).$$
Observe that, by construction, the gcd of its $(n-2)times(n-2)$ subdeterminants is
$D_s(lam)$.
$hat M(s,lam)$ is the analogue of $bar M(s,lam)$ for the $(n-3)times (n-1)$ matrix
$$hat M'=left(begin{array}{ccccc}
-mu_1m_{42}+m_{43}&m_{42}& m_{44}&dots&m_{4n}\
vdots&&&&vdots\
-mu_1m_{n2}+m_{n3}&m_{n2}&m_{n4}&dots&m_{nn}end{array}right).$$
Here the induction hypothesis applies and yields that there is a path
$sto(tilde U(s)_3,tilde U(s)_2,U(s)_4,dots,tilde U(s)_n)$ which crosses the singular surface
$D_s(lam)neq1$ only at points $s$ where the zeros of $D_s(lam)$ are in the open unit disk.
Tracing back the transformations we have made from $bar M(s,lam)$ to $hat M(s,lam)$, we have found
a path $sto(mu_1, U(s)_2, U(s)_3,dots, U(s)_n)$ connecting $Z_1$ and $(mu_1,U_c)$ which
crosses the singular surface only at points where the zeros of $D_s(lam)$ are in the open unit disk.
Observe that in the above reduction of the dimension, it was crucial that the first component
of the points on the path remained constant.
As the connection from $(mu_2,U_c)$ to $Z_2$ is completely analogous to the one
from $Z_1$ to $(mu_1,U_c)$, the proof of Claim 3, and therefore the proof of the Theorem,
is complete.
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2804569%2fis-the-set-x-in-mathcalmm-times-n-rhom-nx-1-connected%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$newcommand{NN}{{mathbb{N}}}newcommand{CC}{{mathbb{C}}}newcommand{RR}{{mathbb{R}}}newcommand{QQ}{{mathbb Q}}newcommand{ra}{rightarrow}newcommand{ds}{displaystyle}newcommand{mat}[1]{left(begin{matrix}#1end{matrix}right)}newcommand{lam}{lambda}renewcommand{P}{{mathcal P}}newcommand{N}{{mathcal N}}newcommand{M}{{mathcal M}}renewcommand{L}{{mathcal L}}newcommand{EQ}[2]{begin{equation} {{#2}label{#1}} end{equation}}newcommand{eps}{varepsilon}newcommand{tabb}[1]{begin{matrix}#1end{matrix}}
DeclareMathOperator{rank}{rank}
DeclareMathOperator{bldiag}{bldiag}$
Theorem: $E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1}$ is connected.
The proof is surprisingly difficult (at least mine). After some preliminaries, I will first treat the case $rank(N)=1$ which is relatively
simple, then show that the case $rank(N)=2$ implies the general case, next treat the case $n=3$ as a preparation and finally
prove the Theorem for $rank(N)=2$ and all $n$.
The case $rank(N)=n$ (that is essentially $M=0$ and $N=I$, see below) is trivial,
because any $Xin E$ can be connected
to $0$ by the path $X_s=sX$ in $E$. It will not be mentioned again.
As explained in the linked answer to the question of the closure of $E$, we can
assume that $N=mat{I_m\0}$ and that the first $m$ rows of
$M$ vanish.
Here a further reduction is useful. Consider the matrix $tilde M(lam)=lambldiag(0,I_{n-m})-M$
which equals $lam I-M$ with the first $m$ diagonal elements replaced by zeros. Consider also the gcd $G(lam)$
of all subdeterminants of $tilde M(lam)$ of size $n-m$.
Claim 1: It is sufficient to treat the case that $G(lam)=1$.
This observation simplifies the presentation here; in the linked answer it was not needed.
Proof of Claim 1. Suppose that $G(lam)neq 1$ and suppose first that it has a real zero $mu$.
Since all subdeterminants of size $n-m$ of $tilde M(mu)$ vanish, its last $n-m$ rows are linearly dependent. Thus there
exists a nonzero row vector $v=(0,dots,0,v_{m+1},dots,v_n)$ such that $v,tilde M(mu)=0$ and hence $v,M=mu,v$. We can assume that $v_nneq0$: There exists some $k$ such that $v_kneq0$. Then we exchange rows $k$ and $n$ and columns $k$ and $n$, respectively, and the components $v_k$ and $v_n$.
Consider now the matrix $T$ with last row $v$ and first $n-1$ rows $e_j$, $j=1,dots,n-1$, where $e_j$ denotes the $j$th unit vector.
Let $M'=T,M,T^{-1}$. A small calculation (using $M',T=T,M$) shows that the first $m$ rows of $M'$ vanish and the last row of $M'$
is $mu e_n$. Since $TN=N$, the matrices $M-NX$ and $M'-NX'$ with $X'=XT^{-1}$ are similar and hence have the same spectral radius.
By construction, $M'-NX'$ is block triangular for all $X'$ and always has the eigenvalue $mu$. The latter must have $|mu|<1$ because
there exists $X'_0$ such that $rho(M'-NX'_0)<1$. Therefore the last row and the last column of $M'-NX'$ are irrelevant for the decision
whether $rho(M'-NX')<1$ or not. We have shown that $n$ can be reduced if $G(lam)$ has a real zero.
If $G(lam)$ has a nonreal zero $mu=alpha+beta i$ then there is a left eigenvector $v=r+i s$, where the first
$m$ components of $r$ and $s$ vanish. Now $(r+i s)M=(alpha +beta i)(r + is)$ means that
$rM=alpha r-beta s$ and $sM=beta r +alpha s$. We proceed as above with $T$ having the last two rows $r,s$. The result
is a block triangular matrix $M'$ with lower right block $mat{alpha&-beta\beta&alpha}$ and again $n$ can be reduced.
As we have shown that $n$ can be reduced if $G(lam)neq 1$, this can be repeated until $G(lam)=1$ and Claim 1 is proved.
Case $mathbf {m=1}$: First, we recall considerations of the linked answer. We use the Laplace (or cofactor) expansion
for the first row of the determinant giving the characteristic polynomial $p_{M-NX}(lam)$
$${det(lam I - M+ NX)=(x_1+lam)C_1(lam)+sum_{k=2}^n x_kC_k(lam)}$$
where $X=(x_1,dots,x_n)$ and $C_j(lam)$, $j=1,dots,n$ are the $1,j$ cofactors of $lam I - M$, that is
$C_j(lam)=(-1)^{j+1}M_{1j}(lam)$, where $M_{1j}(lam)$ is the determinant of the $n-1$ by $n-1$
matrix obtained from $lam I-M$ by deleting its first row and its $j$-th column.
Observe that $C_1(lam)in lam^{n-1}+P_{n-1}$ and $C_j(lam)inP_{n-1}$, where $P_k$ denotes the
vector space of all polynomials of degree less than $k$.
Before continuing the proof, we need an auxiliary statement.
Lemma 1 Let $D(lam)$ denote the gcd of $C_j(lam)$, $j=1,dots,n$ and let $d$ denote its
degree. Then the mapping
$${L:RR^nto D(lam)P_{n-d}, L(X)=sum_{k=1}^n x_kC_k(lam)mbox{ if } X=(x_1,dots,x_n)}$$
is surjective (or onto).
Remark: The characteristic polynomial of $M-NX$ is the given by $p_{M-NX}(lam)=lam C_1(lam)+L(X)$.
For the proof, see the linked answer.
Now we continue with the proof in the case $m=1$. By Claim 1, we can assume that $G(lam)=1$ and hence
by Lemma 1, $L$ is an isomorphism. It induces a bijection between $X$ and the characteristic polynomial
$p_{M-NX}(lam)$ as stated in the above remark.
Consider now any element $X_1$ of $E$ and factor $q_1(lam):=p_{M-NX_1}(lam)=prod_{k=1}^{n}(lam-mu_k)$
with certain (not necessarily real) $mu_k$, $|mu_k|<1$.
Now define for $sin[0,1]$ the polynomials $q_s(lam)=prod_{k=1}^{n}(lam-smu_k)$.
By construction, their coefficients are real and their zeros are in the open unit disk.
Therefore the matrices $X_s$ corresponding to them by $q_s(lam)=lam C_1(lam)+L(X_s)$
are elements of $E$ and connect $X_1$ to the uniquely determined $X_0$ with
$p_{M-NX_0}(lam)=lam^n$. This shows that all elements of $E$ are connected to $X_0$ within $E$
and proves the Theorem for $m=1$.
Actually, this allows to prove that $E$ is simply connected in the case $m=1$.
The case $mathbf{m=2}$ implies the Theorem for all $mathbf m$. Again, we suppose that $G(lam)=1$
(see Claim 1). Consider two elements
$X_a,X_b$ of $E$, i.e. we have $rho(M-NX_a)<1$ and $rho(M-NX_b)<1$. By going over to
appropriate very close matrices, we can assume that the elements of $X_a,X_b$ are algebraically
independent transcendents over $QQ(M)$. Then we construct for $k=2,dots,m-1$ matrices
$X_k$ such that $X_a$ can be connected to $X_b$ in $E$ via these auxiliary matrices,
modifying only two rows at one time.
For $kin{2,dots,m-1}$, we first use the $m$ by $n$ matrix $tilde X_k$ with the
first $k$ rows from $X_b$ and the last $m-k$ rows from $X_a$. The formula is
$tilde X_k=bldiag(I_k,0)X_b+bldiag(0,I_{n-k})X_a$. There is no reason why
$tilde X_k$ should be in $E$, but we claim
Claim 2: The gcd of all $1,j$ cofactors $C_j^{(k)}(lam)$ of
$lam I - (M-Ntilde X_k)$ is 1.
The proof of this Claim is very similar to the considerations below Corollary 2 in the
linked answer. Consider the gcd $D(lam)$ of $C_j^{(k)}(lam)$, $j=1,dots,n$.
A priori, $D(lam)$ is a polynomial in $lam$ with coefficients in $QQ(M,hat X_k)$ because the
$C_j^{(k)}(lam)$ are; here $hat X_k$ denotes the matrix obtained from $tilde X_k$ by erasing the
first row. The $C_j^{(k)}(lam)$ are actually in $K[hat X_k][lam]$, where $K=QQ(M)$.
Since $K$ is a field, $K[hat X_k]$ is a unique factorisation domain because the elements of
$hat X_k$ are algebraically independent transcendents.
We multiply $D(lam)$ by the lcm
of the denominators of the coefficients (seen as elements of $K(hat X_k)$), and then divide by
its content (that is the gcd of the coefficients seen as elements of $K[hat X_k]$).
We call the result $D(lam)$ again.
Then $D(lam)in K[hat X_k][lam]$ has content 1 and divides every $C_j^{(k)}(lam)$ in
$K(hat X_k)[lam]$ by definition. By Gauss's Lemma,
it follows that $D(lam)$ divides every $C_j^{(k)}(lam)$ in $K[hat X_k][lam]$.
Now, for $j=1,dots,n$,
$C_j^{(k)}(lam)$ does not contain any element of the $j$-th column of $hat X_k$.
Therefore $D(lam)$ must actually be an element of $K[lam]$ and it divides
in $K[hat X_k]$ the $n$ cofactors $C_j^{(k)}(lam)$.
Therefore $D(lam)$ must also divide the $1,j$ cofactors $bar C_j(lam)$ of
$lam I-(M-Nbar X)$, where $bar X$ is any $m$ by $n$ matrix.
This follows by substitution of the elements of $hat{bar X}$ for the corresponding
elements of $hat X_k$.
By choosing the rows of $hat{bar X}$ as appropriate unit vectors, we can achieve that
one of the cofactors is any of the $n-m$ by $n-m$ subdeterminants of $M$ we want.
Therefore $D(lam)$ divides any of the latter subdeterminants and we have assumed
that their gcd $G(lam)$ is 1.
This proves Claim 2.
At this point, Lemma 1 applies and yields that the mapping $L^{(k)}:RR^nto P_n$,
$V_1=(x_1,dots,x_n)mapsto L^{(k)}(V_1)=sum_{j=1}^n x_j C_j^{(k)}$ is bijective.
We choose a vector $V_1^{(k)}$ such that $lam C_1^{(k)}+L^{(k)}(V_1^{(k)})=lam^n$.
This means that the matrix $X_k$ obtained from $hat X_k$ by adding the first row
$V_1^{(k)}$ gives a nilpotent $M-NX_k$; in any case $X_k$ is an element of $E$.
We complete $X_2,dots,X_{m-1}$ by $X_1=X_a$ and $X_m=X_b$.
For any $kin{2,dots,m}$, the matrices $X_{k-1}$ and $X_{k}$ have only $two$ different rows:
the first and the $k$th. Both are in $E$ as we have seen. Here the case $m=2$ can be applied
after exchanging the second and the $k$th rows and columns, respectively, of the two matrices.
Hence they are connected by a path in $E$. Combining these $m-1$ paths, we obtain a path
in $E$ connecting $X_a$ and $X_b$. This proves the Theorem for all $mgeq3$ if it is true for $m=2$.
Case $mathbf{n=3, m=2}$. We denote the third row of $M$
by $mat{a&b&c}$, the second row of $X$ by $mat{d&e&f}$.
By Claim 1, we can assume that $aneq0$ or $bneq0$. In order to apply Lemma 1, it is interesting
to study the gcd of all $2times2$ submatrices of $mat{d&lam+e&f\-a&-b&lam-c}$,
in particular to establish a condition when this gcd is 1.
As above, it is not equal to 1 iff there is a number $mu$ such that
$B(mu)=mat{d&mu+e&f\-a&-b&mu-c}$ has rank 1.
If $a=0$, then $bneq0$ and $B(mu)$ has rank 2 for all $mu$ iff $dneq0$.
If $aneq0$, we first add a multiple of the second row of $B(mu)$ to the first
to obtain $mat{0&mu+e-frac{db}a&f+frac da mu-frac{dc}a\-a&-b&mu-c}$ and then
add a multiple of the second column to the third one to obtain
$tilde B(mu)=mat{0&mu+e-frac{db}a&f-frac{de}a+frac{d^2b}{a^2}-
frac{dc}a\-a&-b&mu-c+frac{db}a}$,
a matrix of the same rank as $B(mu)$. It has rank 2 for all $mu$ iff
the third element of the first row is nonzero, that is iff
$Delta(X)=a^2f-{de}a+{d^2b}-{dc}aneq0$.
Actually the condition in the case $a=0$ can be included in the latter one since $bneq0$.
Observe that $Delta(X)$ is closely related to the determinant of the matrix associated to
the operator $L$ of Lemma 1 with respect to canonical bases of $RR^n$ and $P_n$.
This is not surprising because $L$ is invertible iff this determinant is nonzero and, by Lemma 1,
iff the gcd of the $1,j$ cofactors of $lam I-M+NX$ equals 1.
We call the set $Delta(X)=0$ the singular surface.
Consider now two matrices $X_1,X_2in E$. If we can find a path connecting
their second rows $mat{d_1&e_1&f_1}$ and $mat{d_2&e_2&f_2}$ on which $Delta$ does not vanish
then we can extend it to a path connecting $X_1$ and $X_2$ in $E$. Indeed,
we can first find path connecting the characteristic polynomials
of $M-NX_1$ and $M-NX_2$ (for example via $lam^3$) and then the corresponding
path connecting the first rows of $X_1,X_2$ by Lemma 1.
More precisely, consider a path $Z_s=mat{d_s&e_s&f_s}$, $sin[1,2]$ in $RR^3$ such that
$a^2f_s-{d_se_s}a+{d_s^2b}-{d_sc}aneq0$ for all $s$.
We denote the determinants of Lemma 1 by $C_j^s$, the operator associated to $Z_s$ by $L_s$.
Observe that $L_s$ and $L_s^{-1}$ depend continuously upon $s$.
If the characteristic polynomials of
$X_i$ are $p_i(lam)=prod_{k=1}^3(lam-mu_k^i)$, we connect them by $p_s(lam)$
defined by
$$p_s(lam)=left{begin{array}lprod_{k=1}^3left(lam-(3-2s)mu_k^1right)mbox{ if }1leq sleq 1.5\[0.2cm]
prod_{k=1}^3left(lam-(2s-3)mu_k^2right)
mbox{ if }1.5leq sleq 2end{array}right.$$
Observe that for every $s$, the zeros of $p_s$ are in the open unit disk.
Then put $Y_s=L_s^{-1}left(p_s(lam)-lam C_1^sright)$ and let $X_s$ be the matrices with
rows $Y_s$ and $Z_s$. By construction and the remark below Lemma 1, for every $s$,
the characteristic polynomial of $M-NX_s$ is then $p_s(lam)$, therefore $X_sin E$.
It is easy to see that such a path connecting the second rows of $X_1, X_2$ exists
if $Delta(X_j)=a^2f_j-{d_je_j}a+{d_j^2b}-{d_jc}a$, $j=1,2$, are nonzero and have the same sign.
If $aneq0$, then we can first increase the modulus of $f$ sufficiently, then modify $d_1,e_1$
into $d_2,e_2$ and finally reduce $f$ to $f_2$. The case $a=0$ is even simpler and omitted.
Since $E$ is open, we can assume that both $Delta(X_j)neq0.$ It remains to treat
the case that both have opposite sign. If $aneq0$ and $Delta(X)=0$ then the above calculation
shows that the gcd $D(lam)$ of the $2times 2$ submatrices of $B(lam)$
for general $X$ with second row $mat{d&e&f}$ is $D(lam)=lam+e-frac{db}a$.
As $D(lam)$ is a factor of the characteristic polynomial of $M-NX$, a path connecting
$X_1,X_2$ in $E$ must cross the singular surface at a point $X$ where $|e-frac{db}a|<1$.
This makes it clear that the singular surface causes problems in finding a path in $E$
between two given elements, not only for $n=3$, but also for general $n$.
Still in the case $aneq0$, $Delta(X_1)>0>Delta(X_2)$, we choose as crossing point
a point $X^{c}$ with second row $mat{0&0&0}$ (hence $D(lam)=lam$)
and such that $rho(M-NX^c)<1$. This is possible
using Lemma 1. Then we consider
two matrices $X_j^c$, $j=1,2$, which differ from $X^c$ only in the second row:
it is $mat{0&0&+eps}$ for $X_1^c$ and $mat{0&0&-eps}$ for $X_2^c$. If $eps>0$ is sufficiently
small then both matrices are in $E$. As seen above, we can connect $X_1$ and $X_1^c$ in $E$ as
they are both in the set $Delta>0$ and we can connect $X_2$ and $X_2^c$ in $E$
as $Delta<0$ for both. Obviously, $X_1^c$ and $X_2^c$ are connected in $E$ via the point
$X^c$ on the singular surface. This completes the proof in the case $aneq0$.
In the case $a=0$, the proof is quite similar and we only mention the differences.
The singular surface is simply $d=0$ and on it, we have $D(lam)=detmat{lam+e&f\-b&lam-c}$.
Since $bneq0$, we can use $X^c$ with second row $mat{0&c&c^2/b}$ (so that $D^c(lam)=lam^2$)
and with vanishing first row: the characteristic polynomial of $M-NX^c$ is $lam^3$.
Then we use $X_j^c$ with second row $mat{pmeps&c&c^2/b}$ and small $eps$.
This completes the proof in the case $n=3$, $m=2$.
General $mathbf n$ and $mathbf{m=2}$. We suppose again that $G(lam)=1$ (see Claim 1).
As above, we can associate to any row vector $ZinRR^n$ the gcd $D^Z(lam)$ of the subdeterminants
$C_j^Z(lam)$, $j=1,dots,n$ of the $n-1$ by $n$ matrix $bar M(Z,lam)=
mat{lam e_2+Z\tilde M(lam)}$
(see before Claim 1 for $tilde M(lam)$). It is 1 iff the operator $L^Z$ associated to the
$C_j^Z(lam)$ by Lemma 1 is invertible and this is the case iff the determinant $Delta(Z)$
of the matrix associated to $L^Z$ for the canonical bases of $RR^n$ and $P_n$
has nonzero determinant. The singular surface is again $Delta(Z)=0$.
Consider two matrices $X_i$, $i=1,2$, in $E$ and let $Y_i,Z_i$ denote the two rows of each of them.
As for $n=3$, we can assume that $Delta(Z_i)neq0$.
As in the case $n=3,m=2$ we can prove
Lemma 2: If there is a path $Z_s$, $sin[1,2]$ in $RR^n$ connecting $Z_1$ and $Z_2$ such that
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk, then it can be completed
by a path $Y_s$ such that $X_s=mat{Y_s\Z_s}$ connects $X_1$ and $X_2$ in $E$.
Using this Lemma, the Theorem is proved once we show
Claim 3: There always exists a path connecting $Z_1,Z_2$ such that,
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk.
Proof: We prove the Claim by induction over $n$. For $n=3$, it had already been shown
above. Suppose now that it has been shown for $n-1$.
We consider two cases: a) the first column of $ M$ vanishes,
b) it contains a nonzero element.
case a) It is almost identical to the case $n=3,m=2,a=0$.
Denote by $check M$ the $(n-2)times(n-1)$ matrix obtained from $M$ by deleting
the first column. Then the gcd of the $n-2$ by $n-2$ subdeterminants
of $lam(0,I_{n-2})-check M$ is the same as the one for $tilde M(lam)$ and hence
equals $G(lam)=1$. Therefore the condition $Delta(Z)=0$
is equivalent to the first component of $Z$ being 0.This follows by cofactor expansion with respect to the first column. If $Delta(Z)=0$
then $D(lam)$ is the characteristic polynomial of $mat{ -check Z\check M }$,
where $check ZinRR^{n-1}$ is obtained from $Z$ by erasing the first component.
Since the gcd of the $n-2$ by $n-2$ subdeterminants of $lam(0,I_{n-2})-check M$
is $1$, we can find a vector $check Z_cinRR^{n-1}$ such that
the matrix $mat{ -check Z_c\check M }$ has a characteristic polynomial $q(lam)$
with zeros in the open unit disk. We just have to use Lemma 1 in dimension $n-1$.
Now recall $X_i$, $Y_i$, $Z_i$, $i=1,2$. If the first components of $Z_1$ and $Z_2$ have
the same sign, we can simply connect them by any path on which the first component does not
vanish. The critical case is when the first component $Z_{11}$ of $Z_1$ is positive, say,
and the corresponding $Z_{21}$ is negative. Then we can
connect $Z_1$ to the vector $(+eps,check Z)$ and $Z_2$ to the vector $(-eps,check Z)$
by a path that does not cross the singular surface.
Finally, we can simply connect $(pmeps,check Z)$ by the segment. It crosses the singular surface
at the point $(0,check Z)$ and by construction, $D^{(0,check Z)}(lam)=q(lam)$
has only zeros in the open unit disk. This completes the proof in case a).
case b) The first column of $M$ contains some nonzero element. This is, of course, the generic case
and it has taken me a long time to find a reduction - I had tried to find directly for any $n$
a path crossing the singular surface only at points $Z^c$ with $D^{Z^c}(lam)$ having
zeros $mu$, $|mu|<1$, but this turned out to be very complicated.
The present proof using induction is comparatively simple.
First, we prepare the matrix $M$ for the inductive step, more precisely, we transform the expression
$M-NX$ into similar matrices such that $N$ remains unchanged. The eigenvalues of $M-NX$ do
not change either in this way.
Suppose that the first element of the $k$th row of $M$ is nonzero. Then we exchange the third
and the $k$th row of $M$ and the third and the $k$th columns of $M$ and $X$. Since the row
operation does not have any effect on $N$, this is a similarity transformation for
$M-NX$ leaving $N$ unchanged. Next we multiply the third row of $M$ by a convenient number
and divide the third columns of $M$ and $X$ by the same number. So we can assume that
the first element of the third row of $M$ is 1.
If $alpha$ denotes the first element of the fourth row of $M$,
we now add $-alpha$ times the third row to it. This row operation is equivalent to multiplication
of $M$ from the left by some elementary matrix $T$. Observe that again, $N$ is unchanged by this
operation. Then we multiply from the right by the inverse of $T$. This amounts to adding
$alpha$ times the fourth columns of $M$ and $X$ to their third columns. Thus we can assume that
the fourth row of $M$ starts with a 0.
In the same way, we treat the remaining rows of $M$. So we can assume that the first
column of $M$ is the third unit vector.
It is also useful to add a certain multiple of the first column to the $k$th for $M$ and $X$ and subtract the same multiple
of the $k$tk row from the first row for $M$. This temporarily leads to a matrix $M$ the first row of which is not
zero, but this can be repaired later. We can do this without interfering with $N$ if $kgeq3$. As it is a
similarity transformation, the eigenvalues of $M-NX$ do not change. In the case $k=2$, the above transformation can still be
done, but we have to clarify how to deal with $N$. So denote by $T_n$ the $ntimes n$ matrix correponding to adding
a certain multiple of the first column to the second one. Then we modify $M-NX$ into
$T_n^{-1}M T_n- (T_n^{-1}N T_2) (T_2^{-1}XT_n)$. Since $T_n^{-1}N T_2=N$, this leads again to an expression of the same form
if we substract a multiple of the second row from the first not only for $M$, but also for $X$.
After having done all these modification, we simple incorporate the new first line into $X$.
So we can choose the elements of the third row of $M$ besides the leading 1 as we like - only $X$
has to be modified accordingly. Observe that all these operations can be undone.
In summary, we can assume from now on that:
1. The first column of $M$ is the third unit vector,
2. The elements of $M$ in the third row besides 1 are algebraically independent
transcendents over $Q(check M)$.
Here $check M$ denotes the matrix obtained from $M$ by removing the three first rows and the first column.
Let us write
$$M=mat{0&tabb{dots&0}\
0&tabb{dots&0}\
1& R\
0_{n-3}&check M},$$
where $R$ denotes a row vector of $n-1$ elements and $0_{n-3}$ indicates a column vector of $n-3$ zeros.
Consider first the $(n-1)times(n-1)$ matrix $L=mat{0\R\check M}$ and the $(n-1)times1$ matrix $K=mat{1\0_{n-2}}$.
With row vectors $UinRR^{n-1}$ we can consider the matrices $L-KU$. Here we are actually in the case $m=1$.
The gcd $F(lam)$ of the $(n-2)times(n-2)$ submatrices of $tilde L(lam)=lambldiag(0,I_{n-2})-L$
(analogous to $tilde M(lam)$) satisfies $F(lam)=1$.
This comes from the fact that the second row $R$ of $L$ consists of
algebraically independent transcendents over $Q(check M)$ and it had been shown in the proof
of Claim 2.
Therefore by Lemma 1, there exists a vector $U_c$ such that the characteristic polynomial of $L-KU_c$ is any prescribed polynomial
with roots in the open unit disk.
Recall again $X_i$, $Y_i$, $Z_i$, $i=1,2$, and denote $mu_i$ the first elements of $Z_i$. Then we want to connect
$Z_1$ and $Z_2$ by a path as required in Claim 3 by first connecting $Z_1$ and $(mu_1,U_c)$, then $(mu_1,U_c)$ and
$(mu_2,U_c)$ and finally $(mu_2,U_c)$ and $Z_2$. By slightly modifying $mu_1,,mu_2$ and $U_c$, if
necessary, we can assume that none of the four points is on the singular surface.
Connection of $(mu_1,U_c)$ and $(mu_2,U_c)$. Let us assume that $mu_2>mu_1$.
Then the straight line $sto(s,U_c)$, $sin[mu_1,mu_2]$, connects the two points.
As the $(mu_j,U_c)$ are not on the singular surface, this path can cross the
singular surface only at finitely many points because the equation for the
corresponding $s$ is polynomial.
If $s$ is some crossing point then the corresponding
gcd $D^{(s,U_c)}(lam)$ is a divisor of the subdeterminant formed by the columns $2$ to $n$ of
$bar M((s,U_c),lam)$
which happens to be the characteristic polynomial of $L-KU_c$ diskussed above. The vector
$U_c$ had been chosen such that all its zeros are in the open unit disk. Hence this is also
the case for the zeros of $D^{(s,U_c)}(lam)$.
Connection of $Z_1$ and $(mu_1,U_c)$. We want to achieve this by a path $(mu_1,U(s))$ where
$U:[0,1]toRR^{n-1}$ has to be determined such that
$U(1)=U_c$ and $(mu_1,U(0))=Z_1$. For convenience,
we write $U(s)=(U(s)_2,dots,U(s)_n)$, $R=(R_2,dots,R_n)$,
$check M=(m_{ij})_{igeq4,jgeq2}$ and abbreviate $D^{(mu_1,U(s))}(lam)$ by $D_s(lam).$
We want that the gcd $D_s(lam)$ of the $(n-1)times(n-1)$ subdeterminants of
$$bar M(s,lam)=left(
begin{array}{cccccc}
mu_1&lam+U(s)_2&U(s)_3&dots&dots&U(s)_n\-1&-R_2&lam-R_3&dots&dots&-R_n\
0&-m_{42}&-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&-m_{n3}&dots&dots&lam-m_{nn}end{array}right)$$
to have only zeros in the open unit disk if the path crosses the singular surface.
Here we use that the gcd of these subdeterminants does not change if we add a multiple of one
row or column to another row or column, respectively.
Precisely first, we add $mu_1$ times row 2 to the first row in order to have a
zero in the upper left corner; we obtain as first row
$$mat{0&tabb{lam+U(s)_2-mu_1R_2&U(s)_3+mu_1lam-mu_1R_3&dots&U(s)_n-mu_1R_n}}.$$
Then we add $-mu_1$ times column 2 to column 3 to remove the multiple of $lam$ from the third
component in row 1. We obtain as new first row
$mat{0&tabb{lam+tilde U(s)_2&tilde U(s)_3&dots&tilde U(s)_n}},$
where $tilde U(s)_j=U(s)_j-mu_1R_j$ for $jneq3$ and
$tilde U(s)_3=U(s)_3-mu_1U(s)_2-mu_1R_3+mu_1^2R_2.$ Observe that $tilde U(0)$ and
$tilde U(1)$ are determined by $Z_1$ and $U_c$ and that we have to find a path connecting
them with the property wanted in Claim 3.
Altogether, for the matrix
$$tilde{bar M}(s,lam)=left(
begin{array}{cccccc}
0&lam+tilde U(s)_2&tilde U(s)_3&tilde U(s)_4&dots&tilde U(s)_n\
-1&-R_2&lam-R_3+mu_1R_2&-R_4&dots&-R_n\
0&-m_{42}&mu_1m_{42}-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&mu_1m_{n2}-m_{n3}&-m_{n4}&dots&lam-m_{nn}end{array}right),$$
the gcd of its $(n-1)times (n-1)$ submatrices is the same as for $bar M(s,lam)$,
namely $D_s(lam)$.
This gcd is the same as the gcd of the $(n-2)times (n-2)$
subdeterminants for the matrix $hat M(s,lam)$
obtained from $tilde{bar M}(s,lam)$ by erasing the first column and the second row.
Indeed, any $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
containing the first column agrees with
the $(n-2)times (n-2)$ subdeterminant of $hat M(s,lam)$ obtained by erasing its
first column and second row. The $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
not containing the first column is a linear combination of the smaller determinants by
Laplace (or cofactor) expansion. It is convenient now to exchange the first two columns
of $hat M(s,lam)$ to have
$$hat M(s,lam)=left(
begin{array}{ccccc}
tilde U(s)_3&lam+tilde U(s)_2&tilde U(s)_4&dots&tilde U(s)_n\
mu_1m_{42}-m_{43}&-m_{42}&lam-m_{44}&dots&-m_{4n}\
vdots&&&ddots&vdots\
mu_1m_{n2}-m_{n3}&-m_{n2}&-m_{n4}&dots&lam-m_{nn}end{array}right).$$
Observe that, by construction, the gcd of its $(n-2)times(n-2)$ subdeterminants is
$D_s(lam)$.
$hat M(s,lam)$ is the analogue of $bar M(s,lam)$ for the $(n-3)times (n-1)$ matrix
$$hat M'=left(begin{array}{ccccc}
-mu_1m_{42}+m_{43}&m_{42}& m_{44}&dots&m_{4n}\
vdots&&&&vdots\
-mu_1m_{n2}+m_{n3}&m_{n2}&m_{n4}&dots&m_{nn}end{array}right).$$
Here the induction hypothesis applies and yields that there is a path
$sto(tilde U(s)_3,tilde U(s)_2,U(s)_4,dots,tilde U(s)_n)$ which crosses the singular surface
$D_s(lam)neq1$ only at points $s$ where the zeros of $D_s(lam)$ are in the open unit disk.
Tracing back the transformations we have made from $bar M(s,lam)$ to $hat M(s,lam)$, we have found
a path $sto(mu_1, U(s)_2, U(s)_3,dots, U(s)_n)$ connecting $Z_1$ and $(mu_1,U_c)$ which
crosses the singular surface only at points where the zeros of $D_s(lam)$ are in the open unit disk.
Observe that in the above reduction of the dimension, it was crucial that the first component
of the points on the path remained constant.
As the connection from $(mu_2,U_c)$ to $Z_2$ is completely analogous to the one
from $Z_1$ to $(mu_1,U_c)$, the proof of Claim 3, and therefore the proof of the Theorem,
is complete.
add a comment |
$newcommand{NN}{{mathbb{N}}}newcommand{CC}{{mathbb{C}}}newcommand{RR}{{mathbb{R}}}newcommand{QQ}{{mathbb Q}}newcommand{ra}{rightarrow}newcommand{ds}{displaystyle}newcommand{mat}[1]{left(begin{matrix}#1end{matrix}right)}newcommand{lam}{lambda}renewcommand{P}{{mathcal P}}newcommand{N}{{mathcal N}}newcommand{M}{{mathcal M}}renewcommand{L}{{mathcal L}}newcommand{EQ}[2]{begin{equation} {{#2}label{#1}} end{equation}}newcommand{eps}{varepsilon}newcommand{tabb}[1]{begin{matrix}#1end{matrix}}
DeclareMathOperator{rank}{rank}
DeclareMathOperator{bldiag}{bldiag}$
Theorem: $E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1}$ is connected.
The proof is surprisingly difficult (at least mine). After some preliminaries, I will first treat the case $rank(N)=1$ which is relatively
simple, then show that the case $rank(N)=2$ implies the general case, next treat the case $n=3$ as a preparation and finally
prove the Theorem for $rank(N)=2$ and all $n$.
The case $rank(N)=n$ (that is essentially $M=0$ and $N=I$, see below) is trivial,
because any $Xin E$ can be connected
to $0$ by the path $X_s=sX$ in $E$. It will not be mentioned again.
As explained in the linked answer to the question of the closure of $E$, we can
assume that $N=mat{I_m\0}$ and that the first $m$ rows of
$M$ vanish.
Here a further reduction is useful. Consider the matrix $tilde M(lam)=lambldiag(0,I_{n-m})-M$
which equals $lam I-M$ with the first $m$ diagonal elements replaced by zeros. Consider also the gcd $G(lam)$
of all subdeterminants of $tilde M(lam)$ of size $n-m$.
Claim 1: It is sufficient to treat the case that $G(lam)=1$.
This observation simplifies the presentation here; in the linked answer it was not needed.
Proof of Claim 1. Suppose that $G(lam)neq 1$ and suppose first that it has a real zero $mu$.
Since all subdeterminants of size $n-m$ of $tilde M(mu)$ vanish, its last $n-m$ rows are linearly dependent. Thus there
exists a nonzero row vector $v=(0,dots,0,v_{m+1},dots,v_n)$ such that $v,tilde M(mu)=0$ and hence $v,M=mu,v$. We can assume that $v_nneq0$: There exists some $k$ such that $v_kneq0$. Then we exchange rows $k$ and $n$ and columns $k$ and $n$, respectively, and the components $v_k$ and $v_n$.
Consider now the matrix $T$ with last row $v$ and first $n-1$ rows $e_j$, $j=1,dots,n-1$, where $e_j$ denotes the $j$th unit vector.
Let $M'=T,M,T^{-1}$. A small calculation (using $M',T=T,M$) shows that the first $m$ rows of $M'$ vanish and the last row of $M'$
is $mu e_n$. Since $TN=N$, the matrices $M-NX$ and $M'-NX'$ with $X'=XT^{-1}$ are similar and hence have the same spectral radius.
By construction, $M'-NX'$ is block triangular for all $X'$ and always has the eigenvalue $mu$. The latter must have $|mu|<1$ because
there exists $X'_0$ such that $rho(M'-NX'_0)<1$. Therefore the last row and the last column of $M'-NX'$ are irrelevant for the decision
whether $rho(M'-NX')<1$ or not. We have shown that $n$ can be reduced if $G(lam)$ has a real zero.
If $G(lam)$ has a nonreal zero $mu=alpha+beta i$ then there is a left eigenvector $v=r+i s$, where the first
$m$ components of $r$ and $s$ vanish. Now $(r+i s)M=(alpha +beta i)(r + is)$ means that
$rM=alpha r-beta s$ and $sM=beta r +alpha s$. We proceed as above with $T$ having the last two rows $r,s$. The result
is a block triangular matrix $M'$ with lower right block $mat{alpha&-beta\beta&alpha}$ and again $n$ can be reduced.
As we have shown that $n$ can be reduced if $G(lam)neq 1$, this can be repeated until $G(lam)=1$ and Claim 1 is proved.
Case $mathbf {m=1}$: First, we recall considerations of the linked answer. We use the Laplace (or cofactor) expansion
for the first row of the determinant giving the characteristic polynomial $p_{M-NX}(lam)$
$${det(lam I - M+ NX)=(x_1+lam)C_1(lam)+sum_{k=2}^n x_kC_k(lam)}$$
where $X=(x_1,dots,x_n)$ and $C_j(lam)$, $j=1,dots,n$ are the $1,j$ cofactors of $lam I - M$, that is
$C_j(lam)=(-1)^{j+1}M_{1j}(lam)$, where $M_{1j}(lam)$ is the determinant of the $n-1$ by $n-1$
matrix obtained from $lam I-M$ by deleting its first row and its $j$-th column.
Observe that $C_1(lam)in lam^{n-1}+P_{n-1}$ and $C_j(lam)inP_{n-1}$, where $P_k$ denotes the
vector space of all polynomials of degree less than $k$.
Before continuing the proof, we need an auxiliary statement.
Lemma 1 Let $D(lam)$ denote the gcd of $C_j(lam)$, $j=1,dots,n$ and let $d$ denote its
degree. Then the mapping
$${L:RR^nto D(lam)P_{n-d}, L(X)=sum_{k=1}^n x_kC_k(lam)mbox{ if } X=(x_1,dots,x_n)}$$
is surjective (or onto).
Remark: The characteristic polynomial of $M-NX$ is the given by $p_{M-NX}(lam)=lam C_1(lam)+L(X)$.
For the proof, see the linked answer.
Now we continue with the proof in the case $m=1$. By Claim 1, we can assume that $G(lam)=1$ and hence
by Lemma 1, $L$ is an isomorphism. It induces a bijection between $X$ and the characteristic polynomial
$p_{M-NX}(lam)$ as stated in the above remark.
Consider now any element $X_1$ of $E$ and factor $q_1(lam):=p_{M-NX_1}(lam)=prod_{k=1}^{n}(lam-mu_k)$
with certain (not necessarily real) $mu_k$, $|mu_k|<1$.
Now define for $sin[0,1]$ the polynomials $q_s(lam)=prod_{k=1}^{n}(lam-smu_k)$.
By construction, their coefficients are real and their zeros are in the open unit disk.
Therefore the matrices $X_s$ corresponding to them by $q_s(lam)=lam C_1(lam)+L(X_s)$
are elements of $E$ and connect $X_1$ to the uniquely determined $X_0$ with
$p_{M-NX_0}(lam)=lam^n$. This shows that all elements of $E$ are connected to $X_0$ within $E$
and proves the Theorem for $m=1$.
Actually, this allows to prove that $E$ is simply connected in the case $m=1$.
The case $mathbf{m=2}$ implies the Theorem for all $mathbf m$. Again, we suppose that $G(lam)=1$
(see Claim 1). Consider two elements
$X_a,X_b$ of $E$, i.e. we have $rho(M-NX_a)<1$ and $rho(M-NX_b)<1$. By going over to
appropriate very close matrices, we can assume that the elements of $X_a,X_b$ are algebraically
independent transcendents over $QQ(M)$. Then we construct for $k=2,dots,m-1$ matrices
$X_k$ such that $X_a$ can be connected to $X_b$ in $E$ via these auxiliary matrices,
modifying only two rows at one time.
For $kin{2,dots,m-1}$, we first use the $m$ by $n$ matrix $tilde X_k$ with the
first $k$ rows from $X_b$ and the last $m-k$ rows from $X_a$. The formula is
$tilde X_k=bldiag(I_k,0)X_b+bldiag(0,I_{n-k})X_a$. There is no reason why
$tilde X_k$ should be in $E$, but we claim
Claim 2: The gcd of all $1,j$ cofactors $C_j^{(k)}(lam)$ of
$lam I - (M-Ntilde X_k)$ is 1.
The proof of this Claim is very similar to the considerations below Corollary 2 in the
linked answer. Consider the gcd $D(lam)$ of $C_j^{(k)}(lam)$, $j=1,dots,n$.
A priori, $D(lam)$ is a polynomial in $lam$ with coefficients in $QQ(M,hat X_k)$ because the
$C_j^{(k)}(lam)$ are; here $hat X_k$ denotes the matrix obtained from $tilde X_k$ by erasing the
first row. The $C_j^{(k)}(lam)$ are actually in $K[hat X_k][lam]$, where $K=QQ(M)$.
Since $K$ is a field, $K[hat X_k]$ is a unique factorisation domain because the elements of
$hat X_k$ are algebraically independent transcendents.
We multiply $D(lam)$ by the lcm
of the denominators of the coefficients (seen as elements of $K(hat X_k)$), and then divide by
its content (that is the gcd of the coefficients seen as elements of $K[hat X_k]$).
We call the result $D(lam)$ again.
Then $D(lam)in K[hat X_k][lam]$ has content 1 and divides every $C_j^{(k)}(lam)$ in
$K(hat X_k)[lam]$ by definition. By Gauss's Lemma,
it follows that $D(lam)$ divides every $C_j^{(k)}(lam)$ in $K[hat X_k][lam]$.
Now, for $j=1,dots,n$,
$C_j^{(k)}(lam)$ does not contain any element of the $j$-th column of $hat X_k$.
Therefore $D(lam)$ must actually be an element of $K[lam]$ and it divides
in $K[hat X_k]$ the $n$ cofactors $C_j^{(k)}(lam)$.
Therefore $D(lam)$ must also divide the $1,j$ cofactors $bar C_j(lam)$ of
$lam I-(M-Nbar X)$, where $bar X$ is any $m$ by $n$ matrix.
This follows by substitution of the elements of $hat{bar X}$ for the corresponding
elements of $hat X_k$.
By choosing the rows of $hat{bar X}$ as appropriate unit vectors, we can achieve that
one of the cofactors is any of the $n-m$ by $n-m$ subdeterminants of $M$ we want.
Therefore $D(lam)$ divides any of the latter subdeterminants and we have assumed
that their gcd $G(lam)$ is 1.
This proves Claim 2.
At this point, Lemma 1 applies and yields that the mapping $L^{(k)}:RR^nto P_n$,
$V_1=(x_1,dots,x_n)mapsto L^{(k)}(V_1)=sum_{j=1}^n x_j C_j^{(k)}$ is bijective.
We choose a vector $V_1^{(k)}$ such that $lam C_1^{(k)}+L^{(k)}(V_1^{(k)})=lam^n$.
This means that the matrix $X_k$ obtained from $hat X_k$ by adding the first row
$V_1^{(k)}$ gives a nilpotent $M-NX_k$; in any case $X_k$ is an element of $E$.
We complete $X_2,dots,X_{m-1}$ by $X_1=X_a$ and $X_m=X_b$.
For any $kin{2,dots,m}$, the matrices $X_{k-1}$ and $X_{k}$ have only $two$ different rows:
the first and the $k$th. Both are in $E$ as we have seen. Here the case $m=2$ can be applied
after exchanging the second and the $k$th rows and columns, respectively, of the two matrices.
Hence they are connected by a path in $E$. Combining these $m-1$ paths, we obtain a path
in $E$ connecting $X_a$ and $X_b$. This proves the Theorem for all $mgeq3$ if it is true for $m=2$.
Case $mathbf{n=3, m=2}$. We denote the third row of $M$
by $mat{a&b&c}$, the second row of $X$ by $mat{d&e&f}$.
By Claim 1, we can assume that $aneq0$ or $bneq0$. In order to apply Lemma 1, it is interesting
to study the gcd of all $2times2$ submatrices of $mat{d&lam+e&f\-a&-b&lam-c}$,
in particular to establish a condition when this gcd is 1.
As above, it is not equal to 1 iff there is a number $mu$ such that
$B(mu)=mat{d&mu+e&f\-a&-b&mu-c}$ has rank 1.
If $a=0$, then $bneq0$ and $B(mu)$ has rank 2 for all $mu$ iff $dneq0$.
If $aneq0$, we first add a multiple of the second row of $B(mu)$ to the first
to obtain $mat{0&mu+e-frac{db}a&f+frac da mu-frac{dc}a\-a&-b&mu-c}$ and then
add a multiple of the second column to the third one to obtain
$tilde B(mu)=mat{0&mu+e-frac{db}a&f-frac{de}a+frac{d^2b}{a^2}-
frac{dc}a\-a&-b&mu-c+frac{db}a}$,
a matrix of the same rank as $B(mu)$. It has rank 2 for all $mu$ iff
the third element of the first row is nonzero, that is iff
$Delta(X)=a^2f-{de}a+{d^2b}-{dc}aneq0$.
Actually the condition in the case $a=0$ can be included in the latter one since $bneq0$.
Observe that $Delta(X)$ is closely related to the determinant of the matrix associated to
the operator $L$ of Lemma 1 with respect to canonical bases of $RR^n$ and $P_n$.
This is not surprising because $L$ is invertible iff this determinant is nonzero and, by Lemma 1,
iff the gcd of the $1,j$ cofactors of $lam I-M+NX$ equals 1.
We call the set $Delta(X)=0$ the singular surface.
Consider now two matrices $X_1,X_2in E$. If we can find a path connecting
their second rows $mat{d_1&e_1&f_1}$ and $mat{d_2&e_2&f_2}$ on which $Delta$ does not vanish
then we can extend it to a path connecting $X_1$ and $X_2$ in $E$. Indeed,
we can first find path connecting the characteristic polynomials
of $M-NX_1$ and $M-NX_2$ (for example via $lam^3$) and then the corresponding
path connecting the first rows of $X_1,X_2$ by Lemma 1.
More precisely, consider a path $Z_s=mat{d_s&e_s&f_s}$, $sin[1,2]$ in $RR^3$ such that
$a^2f_s-{d_se_s}a+{d_s^2b}-{d_sc}aneq0$ for all $s$.
We denote the determinants of Lemma 1 by $C_j^s$, the operator associated to $Z_s$ by $L_s$.
Observe that $L_s$ and $L_s^{-1}$ depend continuously upon $s$.
If the characteristic polynomials of
$X_i$ are $p_i(lam)=prod_{k=1}^3(lam-mu_k^i)$, we connect them by $p_s(lam)$
defined by
$$p_s(lam)=left{begin{array}lprod_{k=1}^3left(lam-(3-2s)mu_k^1right)mbox{ if }1leq sleq 1.5\[0.2cm]
prod_{k=1}^3left(lam-(2s-3)mu_k^2right)
mbox{ if }1.5leq sleq 2end{array}right.$$
Observe that for every $s$, the zeros of $p_s$ are in the open unit disk.
Then put $Y_s=L_s^{-1}left(p_s(lam)-lam C_1^sright)$ and let $X_s$ be the matrices with
rows $Y_s$ and $Z_s$. By construction and the remark below Lemma 1, for every $s$,
the characteristic polynomial of $M-NX_s$ is then $p_s(lam)$, therefore $X_sin E$.
It is easy to see that such a path connecting the second rows of $X_1, X_2$ exists
if $Delta(X_j)=a^2f_j-{d_je_j}a+{d_j^2b}-{d_jc}a$, $j=1,2$, are nonzero and have the same sign.
If $aneq0$, then we can first increase the modulus of $f$ sufficiently, then modify $d_1,e_1$
into $d_2,e_2$ and finally reduce $f$ to $f_2$. The case $a=0$ is even simpler and omitted.
Since $E$ is open, we can assume that both $Delta(X_j)neq0.$ It remains to treat
the case that both have opposite sign. If $aneq0$ and $Delta(X)=0$ then the above calculation
shows that the gcd $D(lam)$ of the $2times 2$ submatrices of $B(lam)$
for general $X$ with second row $mat{d&e&f}$ is $D(lam)=lam+e-frac{db}a$.
As $D(lam)$ is a factor of the characteristic polynomial of $M-NX$, a path connecting
$X_1,X_2$ in $E$ must cross the singular surface at a point $X$ where $|e-frac{db}a|<1$.
This makes it clear that the singular surface causes problems in finding a path in $E$
between two given elements, not only for $n=3$, but also for general $n$.
Still in the case $aneq0$, $Delta(X_1)>0>Delta(X_2)$, we choose as crossing point
a point $X^{c}$ with second row $mat{0&0&0}$ (hence $D(lam)=lam$)
and such that $rho(M-NX^c)<1$. This is possible
using Lemma 1. Then we consider
two matrices $X_j^c$, $j=1,2$, which differ from $X^c$ only in the second row:
it is $mat{0&0&+eps}$ for $X_1^c$ and $mat{0&0&-eps}$ for $X_2^c$. If $eps>0$ is sufficiently
small then both matrices are in $E$. As seen above, we can connect $X_1$ and $X_1^c$ in $E$ as
they are both in the set $Delta>0$ and we can connect $X_2$ and $X_2^c$ in $E$
as $Delta<0$ for both. Obviously, $X_1^c$ and $X_2^c$ are connected in $E$ via the point
$X^c$ on the singular surface. This completes the proof in the case $aneq0$.
In the case $a=0$, the proof is quite similar and we only mention the differences.
The singular surface is simply $d=0$ and on it, we have $D(lam)=detmat{lam+e&f\-b&lam-c}$.
Since $bneq0$, we can use $X^c$ with second row $mat{0&c&c^2/b}$ (so that $D^c(lam)=lam^2$)
and with vanishing first row: the characteristic polynomial of $M-NX^c$ is $lam^3$.
Then we use $X_j^c$ with second row $mat{pmeps&c&c^2/b}$ and small $eps$.
This completes the proof in the case $n=3$, $m=2$.
General $mathbf n$ and $mathbf{m=2}$. We suppose again that $G(lam)=1$ (see Claim 1).
As above, we can associate to any row vector $ZinRR^n$ the gcd $D^Z(lam)$ of the subdeterminants
$C_j^Z(lam)$, $j=1,dots,n$ of the $n-1$ by $n$ matrix $bar M(Z,lam)=
mat{lam e_2+Z\tilde M(lam)}$
(see before Claim 1 for $tilde M(lam)$). It is 1 iff the operator $L^Z$ associated to the
$C_j^Z(lam)$ by Lemma 1 is invertible and this is the case iff the determinant $Delta(Z)$
of the matrix associated to $L^Z$ for the canonical bases of $RR^n$ and $P_n$
has nonzero determinant. The singular surface is again $Delta(Z)=0$.
Consider two matrices $X_i$, $i=1,2$, in $E$ and let $Y_i,Z_i$ denote the two rows of each of them.
As for $n=3$, we can assume that $Delta(Z_i)neq0$.
As in the case $n=3,m=2$ we can prove
Lemma 2: If there is a path $Z_s$, $sin[1,2]$ in $RR^n$ connecting $Z_1$ and $Z_2$ such that
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk, then it can be completed
by a path $Y_s$ such that $X_s=mat{Y_s\Z_s}$ connects $X_1$ and $X_2$ in $E$.
Using this Lemma, the Theorem is proved once we show
Claim 3: There always exists a path connecting $Z_1,Z_2$ such that,
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk.
Proof: We prove the Claim by induction over $n$. For $n=3$, it had already been shown
above. Suppose now that it has been shown for $n-1$.
We consider two cases: a) the first column of $ M$ vanishes,
b) it contains a nonzero element.
case a) It is almost identical to the case $n=3,m=2,a=0$.
Denote by $check M$ the $(n-2)times(n-1)$ matrix obtained from $M$ by deleting
the first column. Then the gcd of the $n-2$ by $n-2$ subdeterminants
of $lam(0,I_{n-2})-check M$ is the same as the one for $tilde M(lam)$ and hence
equals $G(lam)=1$. Therefore the condition $Delta(Z)=0$
is equivalent to the first component of $Z$ being 0.This follows by cofactor expansion with respect to the first column. If $Delta(Z)=0$
then $D(lam)$ is the characteristic polynomial of $mat{ -check Z\check M }$,
where $check ZinRR^{n-1}$ is obtained from $Z$ by erasing the first component.
Since the gcd of the $n-2$ by $n-2$ subdeterminants of $lam(0,I_{n-2})-check M$
is $1$, we can find a vector $check Z_cinRR^{n-1}$ such that
the matrix $mat{ -check Z_c\check M }$ has a characteristic polynomial $q(lam)$
with zeros in the open unit disk. We just have to use Lemma 1 in dimension $n-1$.
Now recall $X_i$, $Y_i$, $Z_i$, $i=1,2$. If the first components of $Z_1$ and $Z_2$ have
the same sign, we can simply connect them by any path on which the first component does not
vanish. The critical case is when the first component $Z_{11}$ of $Z_1$ is positive, say,
and the corresponding $Z_{21}$ is negative. Then we can
connect $Z_1$ to the vector $(+eps,check Z)$ and $Z_2$ to the vector $(-eps,check Z)$
by a path that does not cross the singular surface.
Finally, we can simply connect $(pmeps,check Z)$ by the segment. It crosses the singular surface
at the point $(0,check Z)$ and by construction, $D^{(0,check Z)}(lam)=q(lam)$
has only zeros in the open unit disk. This completes the proof in case a).
case b) The first column of $M$ contains some nonzero element. This is, of course, the generic case
and it has taken me a long time to find a reduction - I had tried to find directly for any $n$
a path crossing the singular surface only at points $Z^c$ with $D^{Z^c}(lam)$ having
zeros $mu$, $|mu|<1$, but this turned out to be very complicated.
The present proof using induction is comparatively simple.
First, we prepare the matrix $M$ for the inductive step, more precisely, we transform the expression
$M-NX$ into similar matrices such that $N$ remains unchanged. The eigenvalues of $M-NX$ do
not change either in this way.
Suppose that the first element of the $k$th row of $M$ is nonzero. Then we exchange the third
and the $k$th row of $M$ and the third and the $k$th columns of $M$ and $X$. Since the row
operation does not have any effect on $N$, this is a similarity transformation for
$M-NX$ leaving $N$ unchanged. Next we multiply the third row of $M$ by a convenient number
and divide the third columns of $M$ and $X$ by the same number. So we can assume that
the first element of the third row of $M$ is 1.
If $alpha$ denotes the first element of the fourth row of $M$,
we now add $-alpha$ times the third row to it. This row operation is equivalent to multiplication
of $M$ from the left by some elementary matrix $T$. Observe that again, $N$ is unchanged by this
operation. Then we multiply from the right by the inverse of $T$. This amounts to adding
$alpha$ times the fourth columns of $M$ and $X$ to their third columns. Thus we can assume that
the fourth row of $M$ starts with a 0.
In the same way, we treat the remaining rows of $M$. So we can assume that the first
column of $M$ is the third unit vector.
It is also useful to add a certain multiple of the first column to the $k$th for $M$ and $X$ and subtract the same multiple
of the $k$tk row from the first row for $M$. This temporarily leads to a matrix $M$ the first row of which is not
zero, but this can be repaired later. We can do this without interfering with $N$ if $kgeq3$. As it is a
similarity transformation, the eigenvalues of $M-NX$ do not change. In the case $k=2$, the above transformation can still be
done, but we have to clarify how to deal with $N$. So denote by $T_n$ the $ntimes n$ matrix correponding to adding
a certain multiple of the first column to the second one. Then we modify $M-NX$ into
$T_n^{-1}M T_n- (T_n^{-1}N T_2) (T_2^{-1}XT_n)$. Since $T_n^{-1}N T_2=N$, this leads again to an expression of the same form
if we substract a multiple of the second row from the first not only for $M$, but also for $X$.
After having done all these modification, we simple incorporate the new first line into $X$.
So we can choose the elements of the third row of $M$ besides the leading 1 as we like - only $X$
has to be modified accordingly. Observe that all these operations can be undone.
In summary, we can assume from now on that:
1. The first column of $M$ is the third unit vector,
2. The elements of $M$ in the third row besides 1 are algebraically independent
transcendents over $Q(check M)$.
Here $check M$ denotes the matrix obtained from $M$ by removing the three first rows and the first column.
Let us write
$$M=mat{0&tabb{dots&0}\
0&tabb{dots&0}\
1& R\
0_{n-3}&check M},$$
where $R$ denotes a row vector of $n-1$ elements and $0_{n-3}$ indicates a column vector of $n-3$ zeros.
Consider first the $(n-1)times(n-1)$ matrix $L=mat{0\R\check M}$ and the $(n-1)times1$ matrix $K=mat{1\0_{n-2}}$.
With row vectors $UinRR^{n-1}$ we can consider the matrices $L-KU$. Here we are actually in the case $m=1$.
The gcd $F(lam)$ of the $(n-2)times(n-2)$ submatrices of $tilde L(lam)=lambldiag(0,I_{n-2})-L$
(analogous to $tilde M(lam)$) satisfies $F(lam)=1$.
This comes from the fact that the second row $R$ of $L$ consists of
algebraically independent transcendents over $Q(check M)$ and it had been shown in the proof
of Claim 2.
Therefore by Lemma 1, there exists a vector $U_c$ such that the characteristic polynomial of $L-KU_c$ is any prescribed polynomial
with roots in the open unit disk.
Recall again $X_i$, $Y_i$, $Z_i$, $i=1,2$, and denote $mu_i$ the first elements of $Z_i$. Then we want to connect
$Z_1$ and $Z_2$ by a path as required in Claim 3 by first connecting $Z_1$ and $(mu_1,U_c)$, then $(mu_1,U_c)$ and
$(mu_2,U_c)$ and finally $(mu_2,U_c)$ and $Z_2$. By slightly modifying $mu_1,,mu_2$ and $U_c$, if
necessary, we can assume that none of the four points is on the singular surface.
Connection of $(mu_1,U_c)$ and $(mu_2,U_c)$. Let us assume that $mu_2>mu_1$.
Then the straight line $sto(s,U_c)$, $sin[mu_1,mu_2]$, connects the two points.
As the $(mu_j,U_c)$ are not on the singular surface, this path can cross the
singular surface only at finitely many points because the equation for the
corresponding $s$ is polynomial.
If $s$ is some crossing point then the corresponding
gcd $D^{(s,U_c)}(lam)$ is a divisor of the subdeterminant formed by the columns $2$ to $n$ of
$bar M((s,U_c),lam)$
which happens to be the characteristic polynomial of $L-KU_c$ diskussed above. The vector
$U_c$ had been chosen such that all its zeros are in the open unit disk. Hence this is also
the case for the zeros of $D^{(s,U_c)}(lam)$.
Connection of $Z_1$ and $(mu_1,U_c)$. We want to achieve this by a path $(mu_1,U(s))$ where
$U:[0,1]toRR^{n-1}$ has to be determined such that
$U(1)=U_c$ and $(mu_1,U(0))=Z_1$. For convenience,
we write $U(s)=(U(s)_2,dots,U(s)_n)$, $R=(R_2,dots,R_n)$,
$check M=(m_{ij})_{igeq4,jgeq2}$ and abbreviate $D^{(mu_1,U(s))}(lam)$ by $D_s(lam).$
We want that the gcd $D_s(lam)$ of the $(n-1)times(n-1)$ subdeterminants of
$$bar M(s,lam)=left(
begin{array}{cccccc}
mu_1&lam+U(s)_2&U(s)_3&dots&dots&U(s)_n\-1&-R_2&lam-R_3&dots&dots&-R_n\
0&-m_{42}&-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&-m_{n3}&dots&dots&lam-m_{nn}end{array}right)$$
to have only zeros in the open unit disk if the path crosses the singular surface.
Here we use that the gcd of these subdeterminants does not change if we add a multiple of one
row or column to another row or column, respectively.
Precisely first, we add $mu_1$ times row 2 to the first row in order to have a
zero in the upper left corner; we obtain as first row
$$mat{0&tabb{lam+U(s)_2-mu_1R_2&U(s)_3+mu_1lam-mu_1R_3&dots&U(s)_n-mu_1R_n}}.$$
Then we add $-mu_1$ times column 2 to column 3 to remove the multiple of $lam$ from the third
component in row 1. We obtain as new first row
$mat{0&tabb{lam+tilde U(s)_2&tilde U(s)_3&dots&tilde U(s)_n}},$
where $tilde U(s)_j=U(s)_j-mu_1R_j$ for $jneq3$ and
$tilde U(s)_3=U(s)_3-mu_1U(s)_2-mu_1R_3+mu_1^2R_2.$ Observe that $tilde U(0)$ and
$tilde U(1)$ are determined by $Z_1$ and $U_c$ and that we have to find a path connecting
them with the property wanted in Claim 3.
Altogether, for the matrix
$$tilde{bar M}(s,lam)=left(
begin{array}{cccccc}
0&lam+tilde U(s)_2&tilde U(s)_3&tilde U(s)_4&dots&tilde U(s)_n\
-1&-R_2&lam-R_3+mu_1R_2&-R_4&dots&-R_n\
0&-m_{42}&mu_1m_{42}-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&mu_1m_{n2}-m_{n3}&-m_{n4}&dots&lam-m_{nn}end{array}right),$$
the gcd of its $(n-1)times (n-1)$ submatrices is the same as for $bar M(s,lam)$,
namely $D_s(lam)$.
This gcd is the same as the gcd of the $(n-2)times (n-2)$
subdeterminants for the matrix $hat M(s,lam)$
obtained from $tilde{bar M}(s,lam)$ by erasing the first column and the second row.
Indeed, any $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
containing the first column agrees with
the $(n-2)times (n-2)$ subdeterminant of $hat M(s,lam)$ obtained by erasing its
first column and second row. The $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
not containing the first column is a linear combination of the smaller determinants by
Laplace (or cofactor) expansion. It is convenient now to exchange the first two columns
of $hat M(s,lam)$ to have
$$hat M(s,lam)=left(
begin{array}{ccccc}
tilde U(s)_3&lam+tilde U(s)_2&tilde U(s)_4&dots&tilde U(s)_n\
mu_1m_{42}-m_{43}&-m_{42}&lam-m_{44}&dots&-m_{4n}\
vdots&&&ddots&vdots\
mu_1m_{n2}-m_{n3}&-m_{n2}&-m_{n4}&dots&lam-m_{nn}end{array}right).$$
Observe that, by construction, the gcd of its $(n-2)times(n-2)$ subdeterminants is
$D_s(lam)$.
$hat M(s,lam)$ is the analogue of $bar M(s,lam)$ for the $(n-3)times (n-1)$ matrix
$$hat M'=left(begin{array}{ccccc}
-mu_1m_{42}+m_{43}&m_{42}& m_{44}&dots&m_{4n}\
vdots&&&&vdots\
-mu_1m_{n2}+m_{n3}&m_{n2}&m_{n4}&dots&m_{nn}end{array}right).$$
Here the induction hypothesis applies and yields that there is a path
$sto(tilde U(s)_3,tilde U(s)_2,U(s)_4,dots,tilde U(s)_n)$ which crosses the singular surface
$D_s(lam)neq1$ only at points $s$ where the zeros of $D_s(lam)$ are in the open unit disk.
Tracing back the transformations we have made from $bar M(s,lam)$ to $hat M(s,lam)$, we have found
a path $sto(mu_1, U(s)_2, U(s)_3,dots, U(s)_n)$ connecting $Z_1$ and $(mu_1,U_c)$ which
crosses the singular surface only at points where the zeros of $D_s(lam)$ are in the open unit disk.
Observe that in the above reduction of the dimension, it was crucial that the first component
of the points on the path remained constant.
As the connection from $(mu_2,U_c)$ to $Z_2$ is completely analogous to the one
from $Z_1$ to $(mu_1,U_c)$, the proof of Claim 3, and therefore the proof of the Theorem,
is complete.
add a comment |
$newcommand{NN}{{mathbb{N}}}newcommand{CC}{{mathbb{C}}}newcommand{RR}{{mathbb{R}}}newcommand{QQ}{{mathbb Q}}newcommand{ra}{rightarrow}newcommand{ds}{displaystyle}newcommand{mat}[1]{left(begin{matrix}#1end{matrix}right)}newcommand{lam}{lambda}renewcommand{P}{{mathcal P}}newcommand{N}{{mathcal N}}newcommand{M}{{mathcal M}}renewcommand{L}{{mathcal L}}newcommand{EQ}[2]{begin{equation} {{#2}label{#1}} end{equation}}newcommand{eps}{varepsilon}newcommand{tabb}[1]{begin{matrix}#1end{matrix}}
DeclareMathOperator{rank}{rank}
DeclareMathOperator{bldiag}{bldiag}$
Theorem: $E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1}$ is connected.
The proof is surprisingly difficult (at least mine). After some preliminaries, I will first treat the case $rank(N)=1$ which is relatively
simple, then show that the case $rank(N)=2$ implies the general case, next treat the case $n=3$ as a preparation and finally
prove the Theorem for $rank(N)=2$ and all $n$.
The case $rank(N)=n$ (that is essentially $M=0$ and $N=I$, see below) is trivial,
because any $Xin E$ can be connected
to $0$ by the path $X_s=sX$ in $E$. It will not be mentioned again.
As explained in the linked answer to the question of the closure of $E$, we can
assume that $N=mat{I_m\0}$ and that the first $m$ rows of
$M$ vanish.
Here a further reduction is useful. Consider the matrix $tilde M(lam)=lambldiag(0,I_{n-m})-M$
which equals $lam I-M$ with the first $m$ diagonal elements replaced by zeros. Consider also the gcd $G(lam)$
of all subdeterminants of $tilde M(lam)$ of size $n-m$.
Claim 1: It is sufficient to treat the case that $G(lam)=1$.
This observation simplifies the presentation here; in the linked answer it was not needed.
Proof of Claim 1. Suppose that $G(lam)neq 1$ and suppose first that it has a real zero $mu$.
Since all subdeterminants of size $n-m$ of $tilde M(mu)$ vanish, its last $n-m$ rows are linearly dependent. Thus there
exists a nonzero row vector $v=(0,dots,0,v_{m+1},dots,v_n)$ such that $v,tilde M(mu)=0$ and hence $v,M=mu,v$. We can assume that $v_nneq0$: There exists some $k$ such that $v_kneq0$. Then we exchange rows $k$ and $n$ and columns $k$ and $n$, respectively, and the components $v_k$ and $v_n$.
Consider now the matrix $T$ with last row $v$ and first $n-1$ rows $e_j$, $j=1,dots,n-1$, where $e_j$ denotes the $j$th unit vector.
Let $M'=T,M,T^{-1}$. A small calculation (using $M',T=T,M$) shows that the first $m$ rows of $M'$ vanish and the last row of $M'$
is $mu e_n$. Since $TN=N$, the matrices $M-NX$ and $M'-NX'$ with $X'=XT^{-1}$ are similar and hence have the same spectral radius.
By construction, $M'-NX'$ is block triangular for all $X'$ and always has the eigenvalue $mu$. The latter must have $|mu|<1$ because
there exists $X'_0$ such that $rho(M'-NX'_0)<1$. Therefore the last row and the last column of $M'-NX'$ are irrelevant for the decision
whether $rho(M'-NX')<1$ or not. We have shown that $n$ can be reduced if $G(lam)$ has a real zero.
If $G(lam)$ has a nonreal zero $mu=alpha+beta i$ then there is a left eigenvector $v=r+i s$, where the first
$m$ components of $r$ and $s$ vanish. Now $(r+i s)M=(alpha +beta i)(r + is)$ means that
$rM=alpha r-beta s$ and $sM=beta r +alpha s$. We proceed as above with $T$ having the last two rows $r,s$. The result
is a block triangular matrix $M'$ with lower right block $mat{alpha&-beta\beta&alpha}$ and again $n$ can be reduced.
As we have shown that $n$ can be reduced if $G(lam)neq 1$, this can be repeated until $G(lam)=1$ and Claim 1 is proved.
Case $mathbf {m=1}$: First, we recall considerations of the linked answer. We use the Laplace (or cofactor) expansion
for the first row of the determinant giving the characteristic polynomial $p_{M-NX}(lam)$
$${det(lam I - M+ NX)=(x_1+lam)C_1(lam)+sum_{k=2}^n x_kC_k(lam)}$$
where $X=(x_1,dots,x_n)$ and $C_j(lam)$, $j=1,dots,n$ are the $1,j$ cofactors of $lam I - M$, that is
$C_j(lam)=(-1)^{j+1}M_{1j}(lam)$, where $M_{1j}(lam)$ is the determinant of the $n-1$ by $n-1$
matrix obtained from $lam I-M$ by deleting its first row and its $j$-th column.
Observe that $C_1(lam)in lam^{n-1}+P_{n-1}$ and $C_j(lam)inP_{n-1}$, where $P_k$ denotes the
vector space of all polynomials of degree less than $k$.
Before continuing the proof, we need an auxiliary statement.
Lemma 1 Let $D(lam)$ denote the gcd of $C_j(lam)$, $j=1,dots,n$ and let $d$ denote its
degree. Then the mapping
$${L:RR^nto D(lam)P_{n-d}, L(X)=sum_{k=1}^n x_kC_k(lam)mbox{ if } X=(x_1,dots,x_n)}$$
is surjective (or onto).
Remark: The characteristic polynomial of $M-NX$ is the given by $p_{M-NX}(lam)=lam C_1(lam)+L(X)$.
For the proof, see the linked answer.
Now we continue with the proof in the case $m=1$. By Claim 1, we can assume that $G(lam)=1$ and hence
by Lemma 1, $L$ is an isomorphism. It induces a bijection between $X$ and the characteristic polynomial
$p_{M-NX}(lam)$ as stated in the above remark.
Consider now any element $X_1$ of $E$ and factor $q_1(lam):=p_{M-NX_1}(lam)=prod_{k=1}^{n}(lam-mu_k)$
with certain (not necessarily real) $mu_k$, $|mu_k|<1$.
Now define for $sin[0,1]$ the polynomials $q_s(lam)=prod_{k=1}^{n}(lam-smu_k)$.
By construction, their coefficients are real and their zeros are in the open unit disk.
Therefore the matrices $X_s$ corresponding to them by $q_s(lam)=lam C_1(lam)+L(X_s)$
are elements of $E$ and connect $X_1$ to the uniquely determined $X_0$ with
$p_{M-NX_0}(lam)=lam^n$. This shows that all elements of $E$ are connected to $X_0$ within $E$
and proves the Theorem for $m=1$.
Actually, this allows to prove that $E$ is simply connected in the case $m=1$.
The case $mathbf{m=2}$ implies the Theorem for all $mathbf m$. Again, we suppose that $G(lam)=1$
(see Claim 1). Consider two elements
$X_a,X_b$ of $E$, i.e. we have $rho(M-NX_a)<1$ and $rho(M-NX_b)<1$. By going over to
appropriate very close matrices, we can assume that the elements of $X_a,X_b$ are algebraically
independent transcendents over $QQ(M)$. Then we construct for $k=2,dots,m-1$ matrices
$X_k$ such that $X_a$ can be connected to $X_b$ in $E$ via these auxiliary matrices,
modifying only two rows at one time.
For $kin{2,dots,m-1}$, we first use the $m$ by $n$ matrix $tilde X_k$ with the
first $k$ rows from $X_b$ and the last $m-k$ rows from $X_a$. The formula is
$tilde X_k=bldiag(I_k,0)X_b+bldiag(0,I_{n-k})X_a$. There is no reason why
$tilde X_k$ should be in $E$, but we claim
Claim 2: The gcd of all $1,j$ cofactors $C_j^{(k)}(lam)$ of
$lam I - (M-Ntilde X_k)$ is 1.
The proof of this Claim is very similar to the considerations below Corollary 2 in the
linked answer. Consider the gcd $D(lam)$ of $C_j^{(k)}(lam)$, $j=1,dots,n$.
A priori, $D(lam)$ is a polynomial in $lam$ with coefficients in $QQ(M,hat X_k)$ because the
$C_j^{(k)}(lam)$ are; here $hat X_k$ denotes the matrix obtained from $tilde X_k$ by erasing the
first row. The $C_j^{(k)}(lam)$ are actually in $K[hat X_k][lam]$, where $K=QQ(M)$.
Since $K$ is a field, $K[hat X_k]$ is a unique factorisation domain because the elements of
$hat X_k$ are algebraically independent transcendents.
We multiply $D(lam)$ by the lcm
of the denominators of the coefficients (seen as elements of $K(hat X_k)$), and then divide by
its content (that is the gcd of the coefficients seen as elements of $K[hat X_k]$).
We call the result $D(lam)$ again.
Then $D(lam)in K[hat X_k][lam]$ has content 1 and divides every $C_j^{(k)}(lam)$ in
$K(hat X_k)[lam]$ by definition. By Gauss's Lemma,
it follows that $D(lam)$ divides every $C_j^{(k)}(lam)$ in $K[hat X_k][lam]$.
Now, for $j=1,dots,n$,
$C_j^{(k)}(lam)$ does not contain any element of the $j$-th column of $hat X_k$.
Therefore $D(lam)$ must actually be an element of $K[lam]$ and it divides
in $K[hat X_k]$ the $n$ cofactors $C_j^{(k)}(lam)$.
Therefore $D(lam)$ must also divide the $1,j$ cofactors $bar C_j(lam)$ of
$lam I-(M-Nbar X)$, where $bar X$ is any $m$ by $n$ matrix.
This follows by substitution of the elements of $hat{bar X}$ for the corresponding
elements of $hat X_k$.
By choosing the rows of $hat{bar X}$ as appropriate unit vectors, we can achieve that
one of the cofactors is any of the $n-m$ by $n-m$ subdeterminants of $M$ we want.
Therefore $D(lam)$ divides any of the latter subdeterminants and we have assumed
that their gcd $G(lam)$ is 1.
This proves Claim 2.
At this point, Lemma 1 applies and yields that the mapping $L^{(k)}:RR^nto P_n$,
$V_1=(x_1,dots,x_n)mapsto L^{(k)}(V_1)=sum_{j=1}^n x_j C_j^{(k)}$ is bijective.
We choose a vector $V_1^{(k)}$ such that $lam C_1^{(k)}+L^{(k)}(V_1^{(k)})=lam^n$.
This means that the matrix $X_k$ obtained from $hat X_k$ by adding the first row
$V_1^{(k)}$ gives a nilpotent $M-NX_k$; in any case $X_k$ is an element of $E$.
We complete $X_2,dots,X_{m-1}$ by $X_1=X_a$ and $X_m=X_b$.
For any $kin{2,dots,m}$, the matrices $X_{k-1}$ and $X_{k}$ have only $two$ different rows:
the first and the $k$th. Both are in $E$ as we have seen. Here the case $m=2$ can be applied
after exchanging the second and the $k$th rows and columns, respectively, of the two matrices.
Hence they are connected by a path in $E$. Combining these $m-1$ paths, we obtain a path
in $E$ connecting $X_a$ and $X_b$. This proves the Theorem for all $mgeq3$ if it is true for $m=2$.
Case $mathbf{n=3, m=2}$. We denote the third row of $M$
by $mat{a&b&c}$, the second row of $X$ by $mat{d&e&f}$.
By Claim 1, we can assume that $aneq0$ or $bneq0$. In order to apply Lemma 1, it is interesting
to study the gcd of all $2times2$ submatrices of $mat{d&lam+e&f\-a&-b&lam-c}$,
in particular to establish a condition when this gcd is 1.
As above, it is not equal to 1 iff there is a number $mu$ such that
$B(mu)=mat{d&mu+e&f\-a&-b&mu-c}$ has rank 1.
If $a=0$, then $bneq0$ and $B(mu)$ has rank 2 for all $mu$ iff $dneq0$.
If $aneq0$, we first add a multiple of the second row of $B(mu)$ to the first
to obtain $mat{0&mu+e-frac{db}a&f+frac da mu-frac{dc}a\-a&-b&mu-c}$ and then
add a multiple of the second column to the third one to obtain
$tilde B(mu)=mat{0&mu+e-frac{db}a&f-frac{de}a+frac{d^2b}{a^2}-
frac{dc}a\-a&-b&mu-c+frac{db}a}$,
a matrix of the same rank as $B(mu)$. It has rank 2 for all $mu$ iff
the third element of the first row is nonzero, that is iff
$Delta(X)=a^2f-{de}a+{d^2b}-{dc}aneq0$.
Actually the condition in the case $a=0$ can be included in the latter one since $bneq0$.
Observe that $Delta(X)$ is closely related to the determinant of the matrix associated to
the operator $L$ of Lemma 1 with respect to canonical bases of $RR^n$ and $P_n$.
This is not surprising because $L$ is invertible iff this determinant is nonzero and, by Lemma 1,
iff the gcd of the $1,j$ cofactors of $lam I-M+NX$ equals 1.
We call the set $Delta(X)=0$ the singular surface.
Consider now two matrices $X_1,X_2in E$. If we can find a path connecting
their second rows $mat{d_1&e_1&f_1}$ and $mat{d_2&e_2&f_2}$ on which $Delta$ does not vanish
then we can extend it to a path connecting $X_1$ and $X_2$ in $E$. Indeed,
we can first find path connecting the characteristic polynomials
of $M-NX_1$ and $M-NX_2$ (for example via $lam^3$) and then the corresponding
path connecting the first rows of $X_1,X_2$ by Lemma 1.
More precisely, consider a path $Z_s=mat{d_s&e_s&f_s}$, $sin[1,2]$ in $RR^3$ such that
$a^2f_s-{d_se_s}a+{d_s^2b}-{d_sc}aneq0$ for all $s$.
We denote the determinants of Lemma 1 by $C_j^s$, the operator associated to $Z_s$ by $L_s$.
Observe that $L_s$ and $L_s^{-1}$ depend continuously upon $s$.
If the characteristic polynomials of
$X_i$ are $p_i(lam)=prod_{k=1}^3(lam-mu_k^i)$, we connect them by $p_s(lam)$
defined by
$$p_s(lam)=left{begin{array}lprod_{k=1}^3left(lam-(3-2s)mu_k^1right)mbox{ if }1leq sleq 1.5\[0.2cm]
prod_{k=1}^3left(lam-(2s-3)mu_k^2right)
mbox{ if }1.5leq sleq 2end{array}right.$$
Observe that for every $s$, the zeros of $p_s$ are in the open unit disk.
Then put $Y_s=L_s^{-1}left(p_s(lam)-lam C_1^sright)$ and let $X_s$ be the matrices with
rows $Y_s$ and $Z_s$. By construction and the remark below Lemma 1, for every $s$,
the characteristic polynomial of $M-NX_s$ is then $p_s(lam)$, therefore $X_sin E$.
It is easy to see that such a path connecting the second rows of $X_1, X_2$ exists
if $Delta(X_j)=a^2f_j-{d_je_j}a+{d_j^2b}-{d_jc}a$, $j=1,2$, are nonzero and have the same sign.
If $aneq0$, then we can first increase the modulus of $f$ sufficiently, then modify $d_1,e_1$
into $d_2,e_2$ and finally reduce $f$ to $f_2$. The case $a=0$ is even simpler and omitted.
Since $E$ is open, we can assume that both $Delta(X_j)neq0.$ It remains to treat
the case that both have opposite sign. If $aneq0$ and $Delta(X)=0$ then the above calculation
shows that the gcd $D(lam)$ of the $2times 2$ submatrices of $B(lam)$
for general $X$ with second row $mat{d&e&f}$ is $D(lam)=lam+e-frac{db}a$.
As $D(lam)$ is a factor of the characteristic polynomial of $M-NX$, a path connecting
$X_1,X_2$ in $E$ must cross the singular surface at a point $X$ where $|e-frac{db}a|<1$.
This makes it clear that the singular surface causes problems in finding a path in $E$
between two given elements, not only for $n=3$, but also for general $n$.
Still in the case $aneq0$, $Delta(X_1)>0>Delta(X_2)$, we choose as crossing point
a point $X^{c}$ with second row $mat{0&0&0}$ (hence $D(lam)=lam$)
and such that $rho(M-NX^c)<1$. This is possible
using Lemma 1. Then we consider
two matrices $X_j^c$, $j=1,2$, which differ from $X^c$ only in the second row:
it is $mat{0&0&+eps}$ for $X_1^c$ and $mat{0&0&-eps}$ for $X_2^c$. If $eps>0$ is sufficiently
small then both matrices are in $E$. As seen above, we can connect $X_1$ and $X_1^c$ in $E$ as
they are both in the set $Delta>0$ and we can connect $X_2$ and $X_2^c$ in $E$
as $Delta<0$ for both. Obviously, $X_1^c$ and $X_2^c$ are connected in $E$ via the point
$X^c$ on the singular surface. This completes the proof in the case $aneq0$.
In the case $a=0$, the proof is quite similar and we only mention the differences.
The singular surface is simply $d=0$ and on it, we have $D(lam)=detmat{lam+e&f\-b&lam-c}$.
Since $bneq0$, we can use $X^c$ with second row $mat{0&c&c^2/b}$ (so that $D^c(lam)=lam^2$)
and with vanishing first row: the characteristic polynomial of $M-NX^c$ is $lam^3$.
Then we use $X_j^c$ with second row $mat{pmeps&c&c^2/b}$ and small $eps$.
This completes the proof in the case $n=3$, $m=2$.
General $mathbf n$ and $mathbf{m=2}$. We suppose again that $G(lam)=1$ (see Claim 1).
As above, we can associate to any row vector $ZinRR^n$ the gcd $D^Z(lam)$ of the subdeterminants
$C_j^Z(lam)$, $j=1,dots,n$ of the $n-1$ by $n$ matrix $bar M(Z,lam)=
mat{lam e_2+Z\tilde M(lam)}$
(see before Claim 1 for $tilde M(lam)$). It is 1 iff the operator $L^Z$ associated to the
$C_j^Z(lam)$ by Lemma 1 is invertible and this is the case iff the determinant $Delta(Z)$
of the matrix associated to $L^Z$ for the canonical bases of $RR^n$ and $P_n$
has nonzero determinant. The singular surface is again $Delta(Z)=0$.
Consider two matrices $X_i$, $i=1,2$, in $E$ and let $Y_i,Z_i$ denote the two rows of each of them.
As for $n=3$, we can assume that $Delta(Z_i)neq0$.
As in the case $n=3,m=2$ we can prove
Lemma 2: If there is a path $Z_s$, $sin[1,2]$ in $RR^n$ connecting $Z_1$ and $Z_2$ such that
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk, then it can be completed
by a path $Y_s$ such that $X_s=mat{Y_s\Z_s}$ connects $X_1$ and $X_2$ in $E$.
Using this Lemma, the Theorem is proved once we show
Claim 3: There always exists a path connecting $Z_1,Z_2$ such that,
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk.
Proof: We prove the Claim by induction over $n$. For $n=3$, it had already been shown
above. Suppose now that it has been shown for $n-1$.
We consider two cases: a) the first column of $ M$ vanishes,
b) it contains a nonzero element.
case a) It is almost identical to the case $n=3,m=2,a=0$.
Denote by $check M$ the $(n-2)times(n-1)$ matrix obtained from $M$ by deleting
the first column. Then the gcd of the $n-2$ by $n-2$ subdeterminants
of $lam(0,I_{n-2})-check M$ is the same as the one for $tilde M(lam)$ and hence
equals $G(lam)=1$. Therefore the condition $Delta(Z)=0$
is equivalent to the first component of $Z$ being 0.This follows by cofactor expansion with respect to the first column. If $Delta(Z)=0$
then $D(lam)$ is the characteristic polynomial of $mat{ -check Z\check M }$,
where $check ZinRR^{n-1}$ is obtained from $Z$ by erasing the first component.
Since the gcd of the $n-2$ by $n-2$ subdeterminants of $lam(0,I_{n-2})-check M$
is $1$, we can find a vector $check Z_cinRR^{n-1}$ such that
the matrix $mat{ -check Z_c\check M }$ has a characteristic polynomial $q(lam)$
with zeros in the open unit disk. We just have to use Lemma 1 in dimension $n-1$.
Now recall $X_i$, $Y_i$, $Z_i$, $i=1,2$. If the first components of $Z_1$ and $Z_2$ have
the same sign, we can simply connect them by any path on which the first component does not
vanish. The critical case is when the first component $Z_{11}$ of $Z_1$ is positive, say,
and the corresponding $Z_{21}$ is negative. Then we can
connect $Z_1$ to the vector $(+eps,check Z)$ and $Z_2$ to the vector $(-eps,check Z)$
by a path that does not cross the singular surface.
Finally, we can simply connect $(pmeps,check Z)$ by the segment. It crosses the singular surface
at the point $(0,check Z)$ and by construction, $D^{(0,check Z)}(lam)=q(lam)$
has only zeros in the open unit disk. This completes the proof in case a).
case b) The first column of $M$ contains some nonzero element. This is, of course, the generic case
and it has taken me a long time to find a reduction - I had tried to find directly for any $n$
a path crossing the singular surface only at points $Z^c$ with $D^{Z^c}(lam)$ having
zeros $mu$, $|mu|<1$, but this turned out to be very complicated.
The present proof using induction is comparatively simple.
First, we prepare the matrix $M$ for the inductive step, more precisely, we transform the expression
$M-NX$ into similar matrices such that $N$ remains unchanged. The eigenvalues of $M-NX$ do
not change either in this way.
Suppose that the first element of the $k$th row of $M$ is nonzero. Then we exchange the third
and the $k$th row of $M$ and the third and the $k$th columns of $M$ and $X$. Since the row
operation does not have any effect on $N$, this is a similarity transformation for
$M-NX$ leaving $N$ unchanged. Next we multiply the third row of $M$ by a convenient number
and divide the third columns of $M$ and $X$ by the same number. So we can assume that
the first element of the third row of $M$ is 1.
If $alpha$ denotes the first element of the fourth row of $M$,
we now add $-alpha$ times the third row to it. This row operation is equivalent to multiplication
of $M$ from the left by some elementary matrix $T$. Observe that again, $N$ is unchanged by this
operation. Then we multiply from the right by the inverse of $T$. This amounts to adding
$alpha$ times the fourth columns of $M$ and $X$ to their third columns. Thus we can assume that
the fourth row of $M$ starts with a 0.
In the same way, we treat the remaining rows of $M$. So we can assume that the first
column of $M$ is the third unit vector.
It is also useful to add a certain multiple of the first column to the $k$th for $M$ and $X$ and subtract the same multiple
of the $k$tk row from the first row for $M$. This temporarily leads to a matrix $M$ the first row of which is not
zero, but this can be repaired later. We can do this without interfering with $N$ if $kgeq3$. As it is a
similarity transformation, the eigenvalues of $M-NX$ do not change. In the case $k=2$, the above transformation can still be
done, but we have to clarify how to deal with $N$. So denote by $T_n$ the $ntimes n$ matrix correponding to adding
a certain multiple of the first column to the second one. Then we modify $M-NX$ into
$T_n^{-1}M T_n- (T_n^{-1}N T_2) (T_2^{-1}XT_n)$. Since $T_n^{-1}N T_2=N$, this leads again to an expression of the same form
if we substract a multiple of the second row from the first not only for $M$, but also for $X$.
After having done all these modification, we simple incorporate the new first line into $X$.
So we can choose the elements of the third row of $M$ besides the leading 1 as we like - only $X$
has to be modified accordingly. Observe that all these operations can be undone.
In summary, we can assume from now on that:
1. The first column of $M$ is the third unit vector,
2. The elements of $M$ in the third row besides 1 are algebraically independent
transcendents over $Q(check M)$.
Here $check M$ denotes the matrix obtained from $M$ by removing the three first rows and the first column.
Let us write
$$M=mat{0&tabb{dots&0}\
0&tabb{dots&0}\
1& R\
0_{n-3}&check M},$$
where $R$ denotes a row vector of $n-1$ elements and $0_{n-3}$ indicates a column vector of $n-3$ zeros.
Consider first the $(n-1)times(n-1)$ matrix $L=mat{0\R\check M}$ and the $(n-1)times1$ matrix $K=mat{1\0_{n-2}}$.
With row vectors $UinRR^{n-1}$ we can consider the matrices $L-KU$. Here we are actually in the case $m=1$.
The gcd $F(lam)$ of the $(n-2)times(n-2)$ submatrices of $tilde L(lam)=lambldiag(0,I_{n-2})-L$
(analogous to $tilde M(lam)$) satisfies $F(lam)=1$.
This comes from the fact that the second row $R$ of $L$ consists of
algebraically independent transcendents over $Q(check M)$ and it had been shown in the proof
of Claim 2.
Therefore by Lemma 1, there exists a vector $U_c$ such that the characteristic polynomial of $L-KU_c$ is any prescribed polynomial
with roots in the open unit disk.
Recall again $X_i$, $Y_i$, $Z_i$, $i=1,2$, and denote $mu_i$ the first elements of $Z_i$. Then we want to connect
$Z_1$ and $Z_2$ by a path as required in Claim 3 by first connecting $Z_1$ and $(mu_1,U_c)$, then $(mu_1,U_c)$ and
$(mu_2,U_c)$ and finally $(mu_2,U_c)$ and $Z_2$. By slightly modifying $mu_1,,mu_2$ and $U_c$, if
necessary, we can assume that none of the four points is on the singular surface.
Connection of $(mu_1,U_c)$ and $(mu_2,U_c)$. Let us assume that $mu_2>mu_1$.
Then the straight line $sto(s,U_c)$, $sin[mu_1,mu_2]$, connects the two points.
As the $(mu_j,U_c)$ are not on the singular surface, this path can cross the
singular surface only at finitely many points because the equation for the
corresponding $s$ is polynomial.
If $s$ is some crossing point then the corresponding
gcd $D^{(s,U_c)}(lam)$ is a divisor of the subdeterminant formed by the columns $2$ to $n$ of
$bar M((s,U_c),lam)$
which happens to be the characteristic polynomial of $L-KU_c$ diskussed above. The vector
$U_c$ had been chosen such that all its zeros are in the open unit disk. Hence this is also
the case for the zeros of $D^{(s,U_c)}(lam)$.
Connection of $Z_1$ and $(mu_1,U_c)$. We want to achieve this by a path $(mu_1,U(s))$ where
$U:[0,1]toRR^{n-1}$ has to be determined such that
$U(1)=U_c$ and $(mu_1,U(0))=Z_1$. For convenience,
we write $U(s)=(U(s)_2,dots,U(s)_n)$, $R=(R_2,dots,R_n)$,
$check M=(m_{ij})_{igeq4,jgeq2}$ and abbreviate $D^{(mu_1,U(s))}(lam)$ by $D_s(lam).$
We want that the gcd $D_s(lam)$ of the $(n-1)times(n-1)$ subdeterminants of
$$bar M(s,lam)=left(
begin{array}{cccccc}
mu_1&lam+U(s)_2&U(s)_3&dots&dots&U(s)_n\-1&-R_2&lam-R_3&dots&dots&-R_n\
0&-m_{42}&-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&-m_{n3}&dots&dots&lam-m_{nn}end{array}right)$$
to have only zeros in the open unit disk if the path crosses the singular surface.
Here we use that the gcd of these subdeterminants does not change if we add a multiple of one
row or column to another row or column, respectively.
Precisely first, we add $mu_1$ times row 2 to the first row in order to have a
zero in the upper left corner; we obtain as first row
$$mat{0&tabb{lam+U(s)_2-mu_1R_2&U(s)_3+mu_1lam-mu_1R_3&dots&U(s)_n-mu_1R_n}}.$$
Then we add $-mu_1$ times column 2 to column 3 to remove the multiple of $lam$ from the third
component in row 1. We obtain as new first row
$mat{0&tabb{lam+tilde U(s)_2&tilde U(s)_3&dots&tilde U(s)_n}},$
where $tilde U(s)_j=U(s)_j-mu_1R_j$ for $jneq3$ and
$tilde U(s)_3=U(s)_3-mu_1U(s)_2-mu_1R_3+mu_1^2R_2.$ Observe that $tilde U(0)$ and
$tilde U(1)$ are determined by $Z_1$ and $U_c$ and that we have to find a path connecting
them with the property wanted in Claim 3.
Altogether, for the matrix
$$tilde{bar M}(s,lam)=left(
begin{array}{cccccc}
0&lam+tilde U(s)_2&tilde U(s)_3&tilde U(s)_4&dots&tilde U(s)_n\
-1&-R_2&lam-R_3+mu_1R_2&-R_4&dots&-R_n\
0&-m_{42}&mu_1m_{42}-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&mu_1m_{n2}-m_{n3}&-m_{n4}&dots&lam-m_{nn}end{array}right),$$
the gcd of its $(n-1)times (n-1)$ submatrices is the same as for $bar M(s,lam)$,
namely $D_s(lam)$.
This gcd is the same as the gcd of the $(n-2)times (n-2)$
subdeterminants for the matrix $hat M(s,lam)$
obtained from $tilde{bar M}(s,lam)$ by erasing the first column and the second row.
Indeed, any $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
containing the first column agrees with
the $(n-2)times (n-2)$ subdeterminant of $hat M(s,lam)$ obtained by erasing its
first column and second row. The $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
not containing the first column is a linear combination of the smaller determinants by
Laplace (or cofactor) expansion. It is convenient now to exchange the first two columns
of $hat M(s,lam)$ to have
$$hat M(s,lam)=left(
begin{array}{ccccc}
tilde U(s)_3&lam+tilde U(s)_2&tilde U(s)_4&dots&tilde U(s)_n\
mu_1m_{42}-m_{43}&-m_{42}&lam-m_{44}&dots&-m_{4n}\
vdots&&&ddots&vdots\
mu_1m_{n2}-m_{n3}&-m_{n2}&-m_{n4}&dots&lam-m_{nn}end{array}right).$$
Observe that, by construction, the gcd of its $(n-2)times(n-2)$ subdeterminants is
$D_s(lam)$.
$hat M(s,lam)$ is the analogue of $bar M(s,lam)$ for the $(n-3)times (n-1)$ matrix
$$hat M'=left(begin{array}{ccccc}
-mu_1m_{42}+m_{43}&m_{42}& m_{44}&dots&m_{4n}\
vdots&&&&vdots\
-mu_1m_{n2}+m_{n3}&m_{n2}&m_{n4}&dots&m_{nn}end{array}right).$$
Here the induction hypothesis applies and yields that there is a path
$sto(tilde U(s)_3,tilde U(s)_2,U(s)_4,dots,tilde U(s)_n)$ which crosses the singular surface
$D_s(lam)neq1$ only at points $s$ where the zeros of $D_s(lam)$ are in the open unit disk.
Tracing back the transformations we have made from $bar M(s,lam)$ to $hat M(s,lam)$, we have found
a path $sto(mu_1, U(s)_2, U(s)_3,dots, U(s)_n)$ connecting $Z_1$ and $(mu_1,U_c)$ which
crosses the singular surface only at points where the zeros of $D_s(lam)$ are in the open unit disk.
Observe that in the above reduction of the dimension, it was crucial that the first component
of the points on the path remained constant.
As the connection from $(mu_2,U_c)$ to $Z_2$ is completely analogous to the one
from $Z_1$ to $(mu_1,U_c)$, the proof of Claim 3, and therefore the proof of the Theorem,
is complete.
$newcommand{NN}{{mathbb{N}}}newcommand{CC}{{mathbb{C}}}newcommand{RR}{{mathbb{R}}}newcommand{QQ}{{mathbb Q}}newcommand{ra}{rightarrow}newcommand{ds}{displaystyle}newcommand{mat}[1]{left(begin{matrix}#1end{matrix}right)}newcommand{lam}{lambda}renewcommand{P}{{mathcal P}}newcommand{N}{{mathcal N}}newcommand{M}{{mathcal M}}renewcommand{L}{{mathcal L}}newcommand{EQ}[2]{begin{equation} {{#2}label{#1}} end{equation}}newcommand{eps}{varepsilon}newcommand{tabb}[1]{begin{matrix}#1end{matrix}}
DeclareMathOperator{rank}{rank}
DeclareMathOperator{bldiag}{bldiag}$
Theorem: $E = {X in mathcal{M}(m times n; mathbb R) : rho(M-NX) < 1}$ is connected.
The proof is surprisingly difficult (at least mine). After some preliminaries, I will first treat the case $rank(N)=1$ which is relatively
simple, then show that the case $rank(N)=2$ implies the general case, next treat the case $n=3$ as a preparation and finally
prove the Theorem for $rank(N)=2$ and all $n$.
The case $rank(N)=n$ (that is essentially $M=0$ and $N=I$, see below) is trivial,
because any $Xin E$ can be connected
to $0$ by the path $X_s=sX$ in $E$. It will not be mentioned again.
As explained in the linked answer to the question of the closure of $E$, we can
assume that $N=mat{I_m\0}$ and that the first $m$ rows of
$M$ vanish.
Here a further reduction is useful. Consider the matrix $tilde M(lam)=lambldiag(0,I_{n-m})-M$
which equals $lam I-M$ with the first $m$ diagonal elements replaced by zeros. Consider also the gcd $G(lam)$
of all subdeterminants of $tilde M(lam)$ of size $n-m$.
Claim 1: It is sufficient to treat the case that $G(lam)=1$.
This observation simplifies the presentation here; in the linked answer it was not needed.
Proof of Claim 1. Suppose that $G(lam)neq 1$ and suppose first that it has a real zero $mu$.
Since all subdeterminants of size $n-m$ of $tilde M(mu)$ vanish, its last $n-m$ rows are linearly dependent. Thus there
exists a nonzero row vector $v=(0,dots,0,v_{m+1},dots,v_n)$ such that $v,tilde M(mu)=0$ and hence $v,M=mu,v$. We can assume that $v_nneq0$: There exists some $k$ such that $v_kneq0$. Then we exchange rows $k$ and $n$ and columns $k$ and $n$, respectively, and the components $v_k$ and $v_n$.
Consider now the matrix $T$ with last row $v$ and first $n-1$ rows $e_j$, $j=1,dots,n-1$, where $e_j$ denotes the $j$th unit vector.
Let $M'=T,M,T^{-1}$. A small calculation (using $M',T=T,M$) shows that the first $m$ rows of $M'$ vanish and the last row of $M'$
is $mu e_n$. Since $TN=N$, the matrices $M-NX$ and $M'-NX'$ with $X'=XT^{-1}$ are similar and hence have the same spectral radius.
By construction, $M'-NX'$ is block triangular for all $X'$ and always has the eigenvalue $mu$. The latter must have $|mu|<1$ because
there exists $X'_0$ such that $rho(M'-NX'_0)<1$. Therefore the last row and the last column of $M'-NX'$ are irrelevant for the decision
whether $rho(M'-NX')<1$ or not. We have shown that $n$ can be reduced if $G(lam)$ has a real zero.
If $G(lam)$ has a nonreal zero $mu=alpha+beta i$ then there is a left eigenvector $v=r+i s$, where the first
$m$ components of $r$ and $s$ vanish. Now $(r+i s)M=(alpha +beta i)(r + is)$ means that
$rM=alpha r-beta s$ and $sM=beta r +alpha s$. We proceed as above with $T$ having the last two rows $r,s$. The result
is a block triangular matrix $M'$ with lower right block $mat{alpha&-beta\beta&alpha}$ and again $n$ can be reduced.
As we have shown that $n$ can be reduced if $G(lam)neq 1$, this can be repeated until $G(lam)=1$ and Claim 1 is proved.
Case $mathbf {m=1}$: First, we recall considerations of the linked answer. We use the Laplace (or cofactor) expansion
for the first row of the determinant giving the characteristic polynomial $p_{M-NX}(lam)$
$${det(lam I - M+ NX)=(x_1+lam)C_1(lam)+sum_{k=2}^n x_kC_k(lam)}$$
where $X=(x_1,dots,x_n)$ and $C_j(lam)$, $j=1,dots,n$ are the $1,j$ cofactors of $lam I - M$, that is
$C_j(lam)=(-1)^{j+1}M_{1j}(lam)$, where $M_{1j}(lam)$ is the determinant of the $n-1$ by $n-1$
matrix obtained from $lam I-M$ by deleting its first row and its $j$-th column.
Observe that $C_1(lam)in lam^{n-1}+P_{n-1}$ and $C_j(lam)inP_{n-1}$, where $P_k$ denotes the
vector space of all polynomials of degree less than $k$.
Before continuing the proof, we need an auxiliary statement.
Lemma 1 Let $D(lam)$ denote the gcd of $C_j(lam)$, $j=1,dots,n$ and let $d$ denote its
degree. Then the mapping
$${L:RR^nto D(lam)P_{n-d}, L(X)=sum_{k=1}^n x_kC_k(lam)mbox{ if } X=(x_1,dots,x_n)}$$
is surjective (or onto).
Remark: The characteristic polynomial of $M-NX$ is the given by $p_{M-NX}(lam)=lam C_1(lam)+L(X)$.
For the proof, see the linked answer.
Now we continue with the proof in the case $m=1$. By Claim 1, we can assume that $G(lam)=1$ and hence
by Lemma 1, $L$ is an isomorphism. It induces a bijection between $X$ and the characteristic polynomial
$p_{M-NX}(lam)$ as stated in the above remark.
Consider now any element $X_1$ of $E$ and factor $q_1(lam):=p_{M-NX_1}(lam)=prod_{k=1}^{n}(lam-mu_k)$
with certain (not necessarily real) $mu_k$, $|mu_k|<1$.
Now define for $sin[0,1]$ the polynomials $q_s(lam)=prod_{k=1}^{n}(lam-smu_k)$.
By construction, their coefficients are real and their zeros are in the open unit disk.
Therefore the matrices $X_s$ corresponding to them by $q_s(lam)=lam C_1(lam)+L(X_s)$
are elements of $E$ and connect $X_1$ to the uniquely determined $X_0$ with
$p_{M-NX_0}(lam)=lam^n$. This shows that all elements of $E$ are connected to $X_0$ within $E$
and proves the Theorem for $m=1$.
Actually, this allows to prove that $E$ is simply connected in the case $m=1$.
The case $mathbf{m=2}$ implies the Theorem for all $mathbf m$. Again, we suppose that $G(lam)=1$
(see Claim 1). Consider two elements
$X_a,X_b$ of $E$, i.e. we have $rho(M-NX_a)<1$ and $rho(M-NX_b)<1$. By going over to
appropriate very close matrices, we can assume that the elements of $X_a,X_b$ are algebraically
independent transcendents over $QQ(M)$. Then we construct for $k=2,dots,m-1$ matrices
$X_k$ such that $X_a$ can be connected to $X_b$ in $E$ via these auxiliary matrices,
modifying only two rows at one time.
For $kin{2,dots,m-1}$, we first use the $m$ by $n$ matrix $tilde X_k$ with the
first $k$ rows from $X_b$ and the last $m-k$ rows from $X_a$. The formula is
$tilde X_k=bldiag(I_k,0)X_b+bldiag(0,I_{n-k})X_a$. There is no reason why
$tilde X_k$ should be in $E$, but we claim
Claim 2: The gcd of all $1,j$ cofactors $C_j^{(k)}(lam)$ of
$lam I - (M-Ntilde X_k)$ is 1.
The proof of this Claim is very similar to the considerations below Corollary 2 in the
linked answer. Consider the gcd $D(lam)$ of $C_j^{(k)}(lam)$, $j=1,dots,n$.
A priori, $D(lam)$ is a polynomial in $lam$ with coefficients in $QQ(M,hat X_k)$ because the
$C_j^{(k)}(lam)$ are; here $hat X_k$ denotes the matrix obtained from $tilde X_k$ by erasing the
first row. The $C_j^{(k)}(lam)$ are actually in $K[hat X_k][lam]$, where $K=QQ(M)$.
Since $K$ is a field, $K[hat X_k]$ is a unique factorisation domain because the elements of
$hat X_k$ are algebraically independent transcendents.
We multiply $D(lam)$ by the lcm
of the denominators of the coefficients (seen as elements of $K(hat X_k)$), and then divide by
its content (that is the gcd of the coefficients seen as elements of $K[hat X_k]$).
We call the result $D(lam)$ again.
Then $D(lam)in K[hat X_k][lam]$ has content 1 and divides every $C_j^{(k)}(lam)$ in
$K(hat X_k)[lam]$ by definition. By Gauss's Lemma,
it follows that $D(lam)$ divides every $C_j^{(k)}(lam)$ in $K[hat X_k][lam]$.
Now, for $j=1,dots,n$,
$C_j^{(k)}(lam)$ does not contain any element of the $j$-th column of $hat X_k$.
Therefore $D(lam)$ must actually be an element of $K[lam]$ and it divides
in $K[hat X_k]$ the $n$ cofactors $C_j^{(k)}(lam)$.
Therefore $D(lam)$ must also divide the $1,j$ cofactors $bar C_j(lam)$ of
$lam I-(M-Nbar X)$, where $bar X$ is any $m$ by $n$ matrix.
This follows by substitution of the elements of $hat{bar X}$ for the corresponding
elements of $hat X_k$.
By choosing the rows of $hat{bar X}$ as appropriate unit vectors, we can achieve that
one of the cofactors is any of the $n-m$ by $n-m$ subdeterminants of $M$ we want.
Therefore $D(lam)$ divides any of the latter subdeterminants and we have assumed
that their gcd $G(lam)$ is 1.
This proves Claim 2.
At this point, Lemma 1 applies and yields that the mapping $L^{(k)}:RR^nto P_n$,
$V_1=(x_1,dots,x_n)mapsto L^{(k)}(V_1)=sum_{j=1}^n x_j C_j^{(k)}$ is bijective.
We choose a vector $V_1^{(k)}$ such that $lam C_1^{(k)}+L^{(k)}(V_1^{(k)})=lam^n$.
This means that the matrix $X_k$ obtained from $hat X_k$ by adding the first row
$V_1^{(k)}$ gives a nilpotent $M-NX_k$; in any case $X_k$ is an element of $E$.
We complete $X_2,dots,X_{m-1}$ by $X_1=X_a$ and $X_m=X_b$.
For any $kin{2,dots,m}$, the matrices $X_{k-1}$ and $X_{k}$ have only $two$ different rows:
the first and the $k$th. Both are in $E$ as we have seen. Here the case $m=2$ can be applied
after exchanging the second and the $k$th rows and columns, respectively, of the two matrices.
Hence they are connected by a path in $E$. Combining these $m-1$ paths, we obtain a path
in $E$ connecting $X_a$ and $X_b$. This proves the Theorem for all $mgeq3$ if it is true for $m=2$.
Case $mathbf{n=3, m=2}$. We denote the third row of $M$
by $mat{a&b&c}$, the second row of $X$ by $mat{d&e&f}$.
By Claim 1, we can assume that $aneq0$ or $bneq0$. In order to apply Lemma 1, it is interesting
to study the gcd of all $2times2$ submatrices of $mat{d&lam+e&f\-a&-b&lam-c}$,
in particular to establish a condition when this gcd is 1.
As above, it is not equal to 1 iff there is a number $mu$ such that
$B(mu)=mat{d&mu+e&f\-a&-b&mu-c}$ has rank 1.
If $a=0$, then $bneq0$ and $B(mu)$ has rank 2 for all $mu$ iff $dneq0$.
If $aneq0$, we first add a multiple of the second row of $B(mu)$ to the first
to obtain $mat{0&mu+e-frac{db}a&f+frac da mu-frac{dc}a\-a&-b&mu-c}$ and then
add a multiple of the second column to the third one to obtain
$tilde B(mu)=mat{0&mu+e-frac{db}a&f-frac{de}a+frac{d^2b}{a^2}-
frac{dc}a\-a&-b&mu-c+frac{db}a}$,
a matrix of the same rank as $B(mu)$. It has rank 2 for all $mu$ iff
the third element of the first row is nonzero, that is iff
$Delta(X)=a^2f-{de}a+{d^2b}-{dc}aneq0$.
Actually the condition in the case $a=0$ can be included in the latter one since $bneq0$.
Observe that $Delta(X)$ is closely related to the determinant of the matrix associated to
the operator $L$ of Lemma 1 with respect to canonical bases of $RR^n$ and $P_n$.
This is not surprising because $L$ is invertible iff this determinant is nonzero and, by Lemma 1,
iff the gcd of the $1,j$ cofactors of $lam I-M+NX$ equals 1.
We call the set $Delta(X)=0$ the singular surface.
Consider now two matrices $X_1,X_2in E$. If we can find a path connecting
their second rows $mat{d_1&e_1&f_1}$ and $mat{d_2&e_2&f_2}$ on which $Delta$ does not vanish
then we can extend it to a path connecting $X_1$ and $X_2$ in $E$. Indeed,
we can first find path connecting the characteristic polynomials
of $M-NX_1$ and $M-NX_2$ (for example via $lam^3$) and then the corresponding
path connecting the first rows of $X_1,X_2$ by Lemma 1.
More precisely, consider a path $Z_s=mat{d_s&e_s&f_s}$, $sin[1,2]$ in $RR^3$ such that
$a^2f_s-{d_se_s}a+{d_s^2b}-{d_sc}aneq0$ for all $s$.
We denote the determinants of Lemma 1 by $C_j^s$, the operator associated to $Z_s$ by $L_s$.
Observe that $L_s$ and $L_s^{-1}$ depend continuously upon $s$.
If the characteristic polynomials of
$X_i$ are $p_i(lam)=prod_{k=1}^3(lam-mu_k^i)$, we connect them by $p_s(lam)$
defined by
$$p_s(lam)=left{begin{array}lprod_{k=1}^3left(lam-(3-2s)mu_k^1right)mbox{ if }1leq sleq 1.5\[0.2cm]
prod_{k=1}^3left(lam-(2s-3)mu_k^2right)
mbox{ if }1.5leq sleq 2end{array}right.$$
Observe that for every $s$, the zeros of $p_s$ are in the open unit disk.
Then put $Y_s=L_s^{-1}left(p_s(lam)-lam C_1^sright)$ and let $X_s$ be the matrices with
rows $Y_s$ and $Z_s$. By construction and the remark below Lemma 1, for every $s$,
the characteristic polynomial of $M-NX_s$ is then $p_s(lam)$, therefore $X_sin E$.
It is easy to see that such a path connecting the second rows of $X_1, X_2$ exists
if $Delta(X_j)=a^2f_j-{d_je_j}a+{d_j^2b}-{d_jc}a$, $j=1,2$, are nonzero and have the same sign.
If $aneq0$, then we can first increase the modulus of $f$ sufficiently, then modify $d_1,e_1$
into $d_2,e_2$ and finally reduce $f$ to $f_2$. The case $a=0$ is even simpler and omitted.
Since $E$ is open, we can assume that both $Delta(X_j)neq0.$ It remains to treat
the case that both have opposite sign. If $aneq0$ and $Delta(X)=0$ then the above calculation
shows that the gcd $D(lam)$ of the $2times 2$ submatrices of $B(lam)$
for general $X$ with second row $mat{d&e&f}$ is $D(lam)=lam+e-frac{db}a$.
As $D(lam)$ is a factor of the characteristic polynomial of $M-NX$, a path connecting
$X_1,X_2$ in $E$ must cross the singular surface at a point $X$ where $|e-frac{db}a|<1$.
This makes it clear that the singular surface causes problems in finding a path in $E$
between two given elements, not only for $n=3$, but also for general $n$.
Still in the case $aneq0$, $Delta(X_1)>0>Delta(X_2)$, we choose as crossing point
a point $X^{c}$ with second row $mat{0&0&0}$ (hence $D(lam)=lam$)
and such that $rho(M-NX^c)<1$. This is possible
using Lemma 1. Then we consider
two matrices $X_j^c$, $j=1,2$, which differ from $X^c$ only in the second row:
it is $mat{0&0&+eps}$ for $X_1^c$ and $mat{0&0&-eps}$ for $X_2^c$. If $eps>0$ is sufficiently
small then both matrices are in $E$. As seen above, we can connect $X_1$ and $X_1^c$ in $E$ as
they are both in the set $Delta>0$ and we can connect $X_2$ and $X_2^c$ in $E$
as $Delta<0$ for both. Obviously, $X_1^c$ and $X_2^c$ are connected in $E$ via the point
$X^c$ on the singular surface. This completes the proof in the case $aneq0$.
In the case $a=0$, the proof is quite similar and we only mention the differences.
The singular surface is simply $d=0$ and on it, we have $D(lam)=detmat{lam+e&f\-b&lam-c}$.
Since $bneq0$, we can use $X^c$ with second row $mat{0&c&c^2/b}$ (so that $D^c(lam)=lam^2$)
and with vanishing first row: the characteristic polynomial of $M-NX^c$ is $lam^3$.
Then we use $X_j^c$ with second row $mat{pmeps&c&c^2/b}$ and small $eps$.
This completes the proof in the case $n=3$, $m=2$.
General $mathbf n$ and $mathbf{m=2}$. We suppose again that $G(lam)=1$ (see Claim 1).
As above, we can associate to any row vector $ZinRR^n$ the gcd $D^Z(lam)$ of the subdeterminants
$C_j^Z(lam)$, $j=1,dots,n$ of the $n-1$ by $n$ matrix $bar M(Z,lam)=
mat{lam e_2+Z\tilde M(lam)}$
(see before Claim 1 for $tilde M(lam)$). It is 1 iff the operator $L^Z$ associated to the
$C_j^Z(lam)$ by Lemma 1 is invertible and this is the case iff the determinant $Delta(Z)$
of the matrix associated to $L^Z$ for the canonical bases of $RR^n$ and $P_n$
has nonzero determinant. The singular surface is again $Delta(Z)=0$.
Consider two matrices $X_i$, $i=1,2$, in $E$ and let $Y_i,Z_i$ denote the two rows of each of them.
As for $n=3$, we can assume that $Delta(Z_i)neq0$.
As in the case $n=3,m=2$ we can prove
Lemma 2: If there is a path $Z_s$, $sin[1,2]$ in $RR^n$ connecting $Z_1$ and $Z_2$ such that
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk, then it can be completed
by a path $Y_s$ such that $X_s=mat{Y_s\Z_s}$ connects $X_1$ and $X_2$ in $E$.
Using this Lemma, the Theorem is proved once we show
Claim 3: There always exists a path connecting $Z_1,Z_2$ such that,
in each of its intersection points $Z^c$ with the singular surface, the gcd
$D^{Z^c}(lam)$ has only zeros in the open unit disk.
Proof: We prove the Claim by induction over $n$. For $n=3$, it had already been shown
above. Suppose now that it has been shown for $n-1$.
We consider two cases: a) the first column of $ M$ vanishes,
b) it contains a nonzero element.
case a) It is almost identical to the case $n=3,m=2,a=0$.
Denote by $check M$ the $(n-2)times(n-1)$ matrix obtained from $M$ by deleting
the first column. Then the gcd of the $n-2$ by $n-2$ subdeterminants
of $lam(0,I_{n-2})-check M$ is the same as the one for $tilde M(lam)$ and hence
equals $G(lam)=1$. Therefore the condition $Delta(Z)=0$
is equivalent to the first component of $Z$ being 0.This follows by cofactor expansion with respect to the first column. If $Delta(Z)=0$
then $D(lam)$ is the characteristic polynomial of $mat{ -check Z\check M }$,
where $check ZinRR^{n-1}$ is obtained from $Z$ by erasing the first component.
Since the gcd of the $n-2$ by $n-2$ subdeterminants of $lam(0,I_{n-2})-check M$
is $1$, we can find a vector $check Z_cinRR^{n-1}$ such that
the matrix $mat{ -check Z_c\check M }$ has a characteristic polynomial $q(lam)$
with zeros in the open unit disk. We just have to use Lemma 1 in dimension $n-1$.
Now recall $X_i$, $Y_i$, $Z_i$, $i=1,2$. If the first components of $Z_1$ and $Z_2$ have
the same sign, we can simply connect them by any path on which the first component does not
vanish. The critical case is when the first component $Z_{11}$ of $Z_1$ is positive, say,
and the corresponding $Z_{21}$ is negative. Then we can
connect $Z_1$ to the vector $(+eps,check Z)$ and $Z_2$ to the vector $(-eps,check Z)$
by a path that does not cross the singular surface.
Finally, we can simply connect $(pmeps,check Z)$ by the segment. It crosses the singular surface
at the point $(0,check Z)$ and by construction, $D^{(0,check Z)}(lam)=q(lam)$
has only zeros in the open unit disk. This completes the proof in case a).
case b) The first column of $M$ contains some nonzero element. This is, of course, the generic case
and it has taken me a long time to find a reduction - I had tried to find directly for any $n$
a path crossing the singular surface only at points $Z^c$ with $D^{Z^c}(lam)$ having
zeros $mu$, $|mu|<1$, but this turned out to be very complicated.
The present proof using induction is comparatively simple.
First, we prepare the matrix $M$ for the inductive step, more precisely, we transform the expression
$M-NX$ into similar matrices such that $N$ remains unchanged. The eigenvalues of $M-NX$ do
not change either in this way.
Suppose that the first element of the $k$th row of $M$ is nonzero. Then we exchange the third
and the $k$th row of $M$ and the third and the $k$th columns of $M$ and $X$. Since the row
operation does not have any effect on $N$, this is a similarity transformation for
$M-NX$ leaving $N$ unchanged. Next we multiply the third row of $M$ by a convenient number
and divide the third columns of $M$ and $X$ by the same number. So we can assume that
the first element of the third row of $M$ is 1.
If $alpha$ denotes the first element of the fourth row of $M$,
we now add $-alpha$ times the third row to it. This row operation is equivalent to multiplication
of $M$ from the left by some elementary matrix $T$. Observe that again, $N$ is unchanged by this
operation. Then we multiply from the right by the inverse of $T$. This amounts to adding
$alpha$ times the fourth columns of $M$ and $X$ to their third columns. Thus we can assume that
the fourth row of $M$ starts with a 0.
In the same way, we treat the remaining rows of $M$. So we can assume that the first
column of $M$ is the third unit vector.
It is also useful to add a certain multiple of the first column to the $k$th for $M$ and $X$ and subtract the same multiple
of the $k$tk row from the first row for $M$. This temporarily leads to a matrix $M$ the first row of which is not
zero, but this can be repaired later. We can do this without interfering with $N$ if $kgeq3$. As it is a
similarity transformation, the eigenvalues of $M-NX$ do not change. In the case $k=2$, the above transformation can still be
done, but we have to clarify how to deal with $N$. So denote by $T_n$ the $ntimes n$ matrix correponding to adding
a certain multiple of the first column to the second one. Then we modify $M-NX$ into
$T_n^{-1}M T_n- (T_n^{-1}N T_2) (T_2^{-1}XT_n)$. Since $T_n^{-1}N T_2=N$, this leads again to an expression of the same form
if we substract a multiple of the second row from the first not only for $M$, but also for $X$.
After having done all these modification, we simple incorporate the new first line into $X$.
So we can choose the elements of the third row of $M$ besides the leading 1 as we like - only $X$
has to be modified accordingly. Observe that all these operations can be undone.
In summary, we can assume from now on that:
1. The first column of $M$ is the third unit vector,
2. The elements of $M$ in the third row besides 1 are algebraically independent
transcendents over $Q(check M)$.
Here $check M$ denotes the matrix obtained from $M$ by removing the three first rows and the first column.
Let us write
$$M=mat{0&tabb{dots&0}\
0&tabb{dots&0}\
1& R\
0_{n-3}&check M},$$
where $R$ denotes a row vector of $n-1$ elements and $0_{n-3}$ indicates a column vector of $n-3$ zeros.
Consider first the $(n-1)times(n-1)$ matrix $L=mat{0\R\check M}$ and the $(n-1)times1$ matrix $K=mat{1\0_{n-2}}$.
With row vectors $UinRR^{n-1}$ we can consider the matrices $L-KU$. Here we are actually in the case $m=1$.
The gcd $F(lam)$ of the $(n-2)times(n-2)$ submatrices of $tilde L(lam)=lambldiag(0,I_{n-2})-L$
(analogous to $tilde M(lam)$) satisfies $F(lam)=1$.
This comes from the fact that the second row $R$ of $L$ consists of
algebraically independent transcendents over $Q(check M)$ and it had been shown in the proof
of Claim 2.
Therefore by Lemma 1, there exists a vector $U_c$ such that the characteristic polynomial of $L-KU_c$ is any prescribed polynomial
with roots in the open unit disk.
Recall again $X_i$, $Y_i$, $Z_i$, $i=1,2$, and denote $mu_i$ the first elements of $Z_i$. Then we want to connect
$Z_1$ and $Z_2$ by a path as required in Claim 3 by first connecting $Z_1$ and $(mu_1,U_c)$, then $(mu_1,U_c)$ and
$(mu_2,U_c)$ and finally $(mu_2,U_c)$ and $Z_2$. By slightly modifying $mu_1,,mu_2$ and $U_c$, if
necessary, we can assume that none of the four points is on the singular surface.
Connection of $(mu_1,U_c)$ and $(mu_2,U_c)$. Let us assume that $mu_2>mu_1$.
Then the straight line $sto(s,U_c)$, $sin[mu_1,mu_2]$, connects the two points.
As the $(mu_j,U_c)$ are not on the singular surface, this path can cross the
singular surface only at finitely many points because the equation for the
corresponding $s$ is polynomial.
If $s$ is some crossing point then the corresponding
gcd $D^{(s,U_c)}(lam)$ is a divisor of the subdeterminant formed by the columns $2$ to $n$ of
$bar M((s,U_c),lam)$
which happens to be the characteristic polynomial of $L-KU_c$ diskussed above. The vector
$U_c$ had been chosen such that all its zeros are in the open unit disk. Hence this is also
the case for the zeros of $D^{(s,U_c)}(lam)$.
Connection of $Z_1$ and $(mu_1,U_c)$. We want to achieve this by a path $(mu_1,U(s))$ where
$U:[0,1]toRR^{n-1}$ has to be determined such that
$U(1)=U_c$ and $(mu_1,U(0))=Z_1$. For convenience,
we write $U(s)=(U(s)_2,dots,U(s)_n)$, $R=(R_2,dots,R_n)$,
$check M=(m_{ij})_{igeq4,jgeq2}$ and abbreviate $D^{(mu_1,U(s))}(lam)$ by $D_s(lam).$
We want that the gcd $D_s(lam)$ of the $(n-1)times(n-1)$ subdeterminants of
$$bar M(s,lam)=left(
begin{array}{cccccc}
mu_1&lam+U(s)_2&U(s)_3&dots&dots&U(s)_n\-1&-R_2&lam-R_3&dots&dots&-R_n\
0&-m_{42}&-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&-m_{n3}&dots&dots&lam-m_{nn}end{array}right)$$
to have only zeros in the open unit disk if the path crosses the singular surface.
Here we use that the gcd of these subdeterminants does not change if we add a multiple of one
row or column to another row or column, respectively.
Precisely first, we add $mu_1$ times row 2 to the first row in order to have a
zero in the upper left corner; we obtain as first row
$$mat{0&tabb{lam+U(s)_2-mu_1R_2&U(s)_3+mu_1lam-mu_1R_3&dots&U(s)_n-mu_1R_n}}.$$
Then we add $-mu_1$ times column 2 to column 3 to remove the multiple of $lam$ from the third
component in row 1. We obtain as new first row
$mat{0&tabb{lam+tilde U(s)_2&tilde U(s)_3&dots&tilde U(s)_n}},$
where $tilde U(s)_j=U(s)_j-mu_1R_j$ for $jneq3$ and
$tilde U(s)_3=U(s)_3-mu_1U(s)_2-mu_1R_3+mu_1^2R_2.$ Observe that $tilde U(0)$ and
$tilde U(1)$ are determined by $Z_1$ and $U_c$ and that we have to find a path connecting
them with the property wanted in Claim 3.
Altogether, for the matrix
$$tilde{bar M}(s,lam)=left(
begin{array}{cccccc}
0&lam+tilde U(s)_2&tilde U(s)_3&tilde U(s)_4&dots&tilde U(s)_n\
-1&-R_2&lam-R_3+mu_1R_2&-R_4&dots&-R_n\
0&-m_{42}&mu_1m_{42}-m_{43}&lam-m_{44}&dots&-m_{4n}\
vdots&vdots&&&ddots&vdots\
0&-m_{n2}&mu_1m_{n2}-m_{n3}&-m_{n4}&dots&lam-m_{nn}end{array}right),$$
the gcd of its $(n-1)times (n-1)$ submatrices is the same as for $bar M(s,lam)$,
namely $D_s(lam)$.
This gcd is the same as the gcd of the $(n-2)times (n-2)$
subdeterminants for the matrix $hat M(s,lam)$
obtained from $tilde{bar M}(s,lam)$ by erasing the first column and the second row.
Indeed, any $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
containing the first column agrees with
the $(n-2)times (n-2)$ subdeterminant of $hat M(s,lam)$ obtained by erasing its
first column and second row. The $(n-1)times(n-1)$ subdeterminant of $tilde{bar M}(s,lam)$
not containing the first column is a linear combination of the smaller determinants by
Laplace (or cofactor) expansion. It is convenient now to exchange the first two columns
of $hat M(s,lam)$ to have
$$hat M(s,lam)=left(
begin{array}{ccccc}
tilde U(s)_3&lam+tilde U(s)_2&tilde U(s)_4&dots&tilde U(s)_n\
mu_1m_{42}-m_{43}&-m_{42}&lam-m_{44}&dots&-m_{4n}\
vdots&&&ddots&vdots\
mu_1m_{n2}-m_{n3}&-m_{n2}&-m_{n4}&dots&lam-m_{nn}end{array}right).$$
Observe that, by construction, the gcd of its $(n-2)times(n-2)$ subdeterminants is
$D_s(lam)$.
$hat M(s,lam)$ is the analogue of $bar M(s,lam)$ for the $(n-3)times (n-1)$ matrix
$$hat M'=left(begin{array}{ccccc}
-mu_1m_{42}+m_{43}&m_{42}& m_{44}&dots&m_{4n}\
vdots&&&&vdots\
-mu_1m_{n2}+m_{n3}&m_{n2}&m_{n4}&dots&m_{nn}end{array}right).$$
Here the induction hypothesis applies and yields that there is a path
$sto(tilde U(s)_3,tilde U(s)_2,U(s)_4,dots,tilde U(s)_n)$ which crosses the singular surface
$D_s(lam)neq1$ only at points $s$ where the zeros of $D_s(lam)$ are in the open unit disk.
Tracing back the transformations we have made from $bar M(s,lam)$ to $hat M(s,lam)$, we have found
a path $sto(mu_1, U(s)_2, U(s)_3,dots, U(s)_n)$ connecting $Z_1$ and $(mu_1,U_c)$ which
crosses the singular surface only at points where the zeros of $D_s(lam)$ are in the open unit disk.
Observe that in the above reduction of the dimension, it was crucial that the first component
of the points on the path remained constant.
As the connection from $(mu_2,U_c)$ to $Z_2$ is completely analogous to the one
from $Z_1$ to $(mu_1,U_c)$, the proof of Claim 3, and therefore the proof of the Theorem,
is complete.
edited Dec 12 at 10:15
answered Dec 6 at 23:21
Helmut
664118
664118
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2804569%2fis-the-set-x-in-mathcalmm-times-n-rhom-nx-1-connected%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
2
I feel so that yes, but I can't prove it.
– peterh
Jun 1 at 22:59
Do you have any examples, or other form of empirical evidence, that would point in the direction of connected? Possibly in special cases, e.g., when $M$ is symmetric?
– Oskar Limka
Jun 10 at 16:54
@OskarLimka: Sorry, I only have some intuitive feeling. The set seems 'simple' but when concerning the topological properties, the only thing I know is openness.
– user1101010
Jun 11 at 16:55