10
$\begingroup$

Hello I am studying for the qualifying exam in algebra and I am having trouble solving this seemingly easy problem. If $A$ is a matrix whose minimal polynomial and characteristic polynomial agree, and $B$ commutes with $A$ then $B$ is a polynomial in $A$.

I have shown that the dimension of the subspace of polynomials in $A$ must be equal to the dimension of the underlying vector space. Clearly this subspace is contained in the subspace of matrices that commute with $A$. So if I can show the latter must have dimension less than or equal to the dimension of $V$, I'll be done. But I don't see how to show that.

Or is there an easier way?

$\endgroup$
3
  • $\begingroup$ Have you asked the question you intended? Where does $B$ come from? $\endgroup$ Commented Aug 13, 2011 at 22:18
  • $\begingroup$ The dimension in question (i.e. that of $K[A]$, where $K$ is the underlying field) is always equal to the degree of the minimal polynomial. In your case it's equal to (not less than) the dimension of the underlying vector space. (I agree with Geoff. I'm ignoring $B$ in this comment.) $\endgroup$ Commented Aug 13, 2011 at 22:18
  • $\begingroup$ @geoff Whoops. Yes B must commute with A. pierre. That's what I meant to write. Fixed both errors now. $\endgroup$ Commented Aug 13, 2011 at 22:20

3 Answers 3

20
$\begingroup$

Hint: 1) Show that there is a vector $e$ such that $(e,Ae,A^2e,\dots,A^{n-1}e)$ is a basis (this is another standard exercice).

Write $Be = \sum_{i=0}^{n-1} a_i A^i e$.

2) Show that $B$ and $\sum_{i=0}^{n-1} a_i A^i$ agree on the basis.

$\endgroup$
5
$\begingroup$

This becomes quite easy viewing a vector space equipped with a linear operator$~T$ as a module over the ring $K[X]$ (with $X$ acting as $T$), especially using the structure theory of finitely generated modules over a P.I.D. (here $K[X]$). If one chooses $A$ to give the linear operator$~T$ defining the $K[X]$-module structure, then the commutation of $B$ with $A$, and hence with any polynomial in$~A$, precisely means that $B$ defines an endomorphism of the so defined $K[X]$-module.

The argument now consist of two parts: (1) equality of characteristic and minimal polynomials implies that the $K[X]$-module is cyclic: for some fixed generating vector $v_0$, every $v\in V$ can be written $v=P\cdot v_0$ for some $P\in K[X]$; (2) an endomorphism of a cyclic $K[X]$-module is given by multiplication by some fixed $Q\in K[X]$. (In the end this will mean that $B=Q[A]$.)

The proof of (2) is straightforward: calling the endomorphism $\beta$ and writing $\beta(v_0)=Q\cdot v_0$ for some polynomial $Q$ (which is possible by the fact the $v_0$ is a generating vector for the cyclic module), one has for any $v\in V$, again writing $v=P\cdot v_0$ for some $P\in K[X]$, that $$\beta(v) =\beta(P\cdot v_0) =P\cdot\beta(v_0) =P\cdot Q\cdot v_0=Q\cdot(P\cdot v_0) =Q\cdot v, $$ which shows that $\beta$ coincides with multiplication by$~Q$. (In the application $\beta$ is multiplication by the matrix$~B$, which coincides with multiplication by $Q[A]$, so this must be the same matrix.) (This part corresponds to the answer by user10676, with $Q=\sum_{i=0}^{n-1}a_iX^i$ of that answer.)

The proof of (1) is easy with the mentioned structure theorem, which says that $V$ decomposes as a direct sum of cyclic modules, with the minimal polynomial of each summand dividing that of the next. Then the last minimal polynomial annihilates all summands and is therefore the global minimal polynomial; the characteristic polynomial is the product of those minimal polynomials. So minimal and characteristic polynomials coincide (if and) only if the decomposition contains no more then one (non-trivial) module.

If one wants to reason without using the structure theorem for f.g. modules over a P.I.D., (1) is still quite plausible. For every vector $v$ the minimal degree monic polynomial $M$ such that $M\cdot v=0$ divides the minimal polynomial$~\mu$, and for every strict divisor $D$ of$~\mu$ the subspace annihilated by $D$ is a proper subspace; this suggests that for most vectors $v$ one has $Q=\mu$, and given that $\deg(\mu)=\dim(V)$, this means that $v$ generates $V$ as a cyclic module. Indeed there are only finitely many divisors$~D$, so if $K$ is an infinite field this amounts to a valid argument. An complete argument valid for all fields is a bit more work, but can be found in this answer.

$\endgroup$
0
$\begingroup$

Let $\mathbb K\supseteq \mathbb F$ be a a splitting field for $A$'s characteristic polynomial (or the algebraic closure of $\mathbb F$ if preferred) and notice the maps
$T':M_n\big(\mathbb F\big)\rightarrow M_n\big(\mathbb F\big)\text{, } T:M_n\big(\mathbb K\big)\rightarrow M_n\big(\mathbb K\big)$, both given by $X\mapsto AX -XA$ necessarily have the same rank (hence the same kernel dimensions) since field extensions do not change rank.

Since $A$ has a degree $n$ minimal polynomial we know every eigenspace of $A$ has dimension $1$ and $\big\{I, A, \cdots, A^{n-1}\big\}$ contains $n$ linearly independent vectors in $\ker T$. So $\dim \ker T\geq n$ and it remains to show $n\geq\dim \ker T$. Working over $\mathbb K$, let $A$ have minimal polynomial $m\big(x\big)= \prod_{k=1}^n \big(x-\lambda_k\big)$ (for not necessarily distinct $\lambda_k$).

The map $A^{(\lambda_k)}: \ker T \longrightarrow \ker T$ given by $X\mapsto \big(A-\lambda_k I\big)X$ satisfies $\dim \ker A^{(\lambda_k)}\leq 1$. We know this since $\text{nullspace } \big(A-\lambda_k I\big)=\big\{\alpha \cdot \mathbf v_k\big\}$ so each column of a non-zero $X \in \ker A^{(\lambda_k)}$ must be $ \propto \mathbf v_k\implies X $ is rank one, of the form $X= \mathbf v_k \mathbf z^T$; $\text{left nullspace } \big(A-\lambda_k I\big) =\big\{\alpha \cdot \mathbf y_k^T\big\}$ and $X\big(A-\lambda_k I\big) = \big(A-\lambda_k I\big)X=\mathbf 0\implies \mathbf z^T \propto \mathbf y_k^T $. Thus $\ker A^{(\lambda_k)}\subseteq \big\{\beta\cdot \mathbf v_k \mathbf y_k^T\big\}$ which is $1$-dimensional.

For $k\in \big\{1,\dots,n\big\}$ define $S_k := A^{(\lambda_k)}\big(S_{k-1}\big)$ and $S_0:= \ker T$
Notice $S_n = A^{(\lambda_n)}\circ\dots A^{(\lambda_2)} \circ A^{(\lambda_1)}\big(S_0\big) = m\big(A\big)\circ \big(S_0\big)=\mathbf 0\circ \big(S_0\big) = \big\{\mathbf 0\big\}$

We trivially have $\dim S_0 \geq \dim \ker T -0$ by construction (Base Case) and for $1\leq k\leq n$
$\dim S_k \geq \dim S_{k-1}-1 \geq \Big(\dim\ker T-\big(k-1\big)\Big)-1=\dim\ker T-k$
where the first inequality is rank-nullity applied to the restriction $A_{\vert S_{k-1}}^{(\lambda_k)}:S_{k-1}\rightarrow \ker T$ and the second inequality is induction hypothesis.
$\implies 0 =\dim S_n \geq \dim \ker T-n\implies n\geq \dim\ker T$

remark:
if we flip the ordering of the indices, i.e. $V_{i}:= S_{n-i}$ for $0\leq i \leq n$ then the $V_i$ form a complete flag.

$\endgroup$

You must log in to answer this question.

Start asking to get answers

Find the answer to your question by asking.

Ask question

Explore related questions

See similar questions with these tags.