The associative law of multiplication enables us to write the product of three (or more) factors without any parentheses; in particular we may consider the product of any finite number, say, \(m\) , of factors all equal to \(A\) . This product depends only on \(A\) and on \(m\) (and not, as we just remarked, on any bracketing of the factors); we shall denote it by \(A^{m}\) . The justification for this notation is that, although in general transformation multiplication is not commutative, for the powers of one transformation we do have the usual laws of exponents, \(A^{n} A^{m}=A^{n+m}\) and \((A^{n})^{m}=A^{n m}\) . We observe that \(A^{1}=A\) ; it is customary also to write, by definition, \(A^{0}=1\) . With these definitions the calculus of powers of a single transformation is almost exactly the same as in ordinary arithmetic. We may, in particular, define polynomials in a linear transformation. Thus if \(p\) is any polynomial with scalar coefficients in a variable \(t\) , say \(p(t)=\alpha_{0}+\alpha_{1} t+\cdots+\alpha_{n} t^{n}\) , we may form the linear transformation

\[p(A)=\alpha_{0} 1+\alpha_{1} A+\cdots+\alpha_{n} A^{n}.\] 

The rules for the algebraic manipulation of such polynomials are easy. Thus \(p(t) q(t)=r(t)\) implies \(p(A) q(A)=r(A)\) (so that, in particular, any \(p(A)\) and \(q(A)\) are commutative); if \(p(t)=\alpha\) (identically), we shall usually write \(p(A)=\alpha\) (instead of \(p(A)=\alpha \cdot 1\) ); this is in harmony with the use of the symbols \(0\) and \(1\) for linear transformations.

If \(p\) is a polynomial in two variables and if \(A\) and \(B\) are linear transformations, it is not usually possible to give any sensible interpretation to \(p(A, B)\) . The trouble, of course, is that \(A\) and \(B\) may not commute, and even a simple monomial, such as \(s^{2} t\) , will cause confusion. If \(p(s, t)=s^{2} t\) , what should we mean by \(p(A, B)\) ? Should it be \(A^{2} B\) , or \(A B A\) , or \(B A^{2}\) ? It is important to recognize that there is a difficulty here; fortunately for us it is not necessary to try to get around it. We shall work with polynomials in several variables only in connection with commutative transformations, and then everything is simple. We observe that if \(A B=B A\) , then \(A^{n} B^{m}=B^{m} A^{n}\) , and therefore \(p(A, B)\) has an unambiguous meaning for every polynomial \(p\) . The formal properties of the correspondence between (commutative) transformations and polynomials are just as valid for several variables as for one; we omit the details.

For an example of the possible behavior of the powers of a transformation we look at the differentiation transformation \(D\) on \(\mathcal{P}\) (or, just as well, on \(\mathcal{P}_n\) , for some \(n\) ). It is easy to see that for every positive integer \(k\) , and for every polynomial \(x\) in \(\mathcal{P}\) , we have \((D^{k} x)(t)=\frac{d^{k} x}{d t^{k}}\) . We observe that whatever else \(D\) does, it lowers the degree of the polynomial on which it acts by exactly one unit (assuming, of course, that the degree is \(\geq 1\) ). Let \(x\) be a polynomial of degree \(n-1\) , say; what is \(D^{n} x\) ? Or put it another way: what is the product of the two (commutative) transformations \(D^{k}\) and \(D^{n-k}\) (where \(k\) is any integer between \(0\) and \(n\) ), considered on the space \(\mathcal{P}_n\) ? We mention this example to bring out the disconcerting fact implied by the answer to the last question; the product of two transformations may vanish even though neither one of them is zero. A non-zero transformation whose product with some non-zero transformation is zero is called a divisor of zero.

EXERCISES

Exercise 1. Calculate the linear transformations \(D^{n} S^{n}\) and \(S^{n} D^{n}\) , \(n=1,2,3, \ldots\) ; in other words, compute the effect of each such transformation on an arbitrary element of \(\mathcal{P}\) . (Here \(D\) and \(S\) denote the differentiation and integration transformations defined in Section: Linear transformations .)

Exercise 2. If \(A\) and \(B\) are linear transformations such that \(A B-B A\) commutes with \(A\) , then \[A^{k} B-B A^{k}=k A^{k-1}(A B-B A)\] for every positive integer \(k\) .

Exercise 3. Suppose that \(A x(t)=x(t+1)\) for every \(x\) in \(\mathcal{P}_n\) ; prove that if \(D\) is the differentiation operator, then \[1+\frac{D}{1!}+\frac{D^{2}}{2!}+\cdots+\frac{D^{n-1}}{(n-1)!}=A.\] 

Exercise 4. 

  1. If \(A\) is a linear transformation on an \(n\) -dimensional vector space, then there exists a non-zero polynomial \(p\) of degree \(\leq n^{2}\) such that \(p(A)=0\) .
  2. If \(A x=y_{0}(x) x_{0}\) (see Section: Linear transformations , (2)), find a non-zero polynomial \(p\) such that \(p(A)=0\) . What is the smallest possible degree \(p\) can have?

Exercise 5. The product of linear transformations between different vector spaces is defined only if they "match" in the following sense. Suppose that \(\mathcal{U}\) , \(\mathcal{V}\) , and \(\mathcal{W}\) are vector spaces over the same field, and suppose that \(A\) and \(B\) are linear transformations from \(\mathcal{U}\) to \(\mathcal{V}\) and from \(\mathcal{V}\) to \(\mathcal{W}\) , respectively. The product \(C=B A\) (the order is important) is defined to be the linear transformation from \(\mathcal{U}\) to \(\mathcal{W}\) given by \(C x=B(A x)\) . Interpret and prove as many as possible among the equations Section: Products , (1)–(5) for this concept of multiplication.

Exercise 6. Let \(A\) be a linear transformation on an \(n\) -dimensional vector space \(\mathcal{V}\) .

  1. Prove that the set of all those linear transformations \(B\) on \(\mathcal{V}\) for which \(A B=0\) is a subspace of the space of all linear transformations on \(\mathcal{V}\) .
  2. Show that by a suitable choice for \(A\) the dimension of the subspace described in (a) can be made to equal \(0\) , or \(n\) , or \(n^{2}\) . What values can this dimension attain?
  3. Can every subspace of the space of all linear transformations be obtained in the manner described in (a) (by the choice of a suitable \(A\) )?

Exercise 7. Let \(A\) be a linear transformation on a vector space \(\mathcal{V}\) , and consider the correspondence that assigns to each linear transformation \(X\) on \(\mathcal{V}\) the linear transformation \(A X\) . Prove that this correspondence is a linear transformation (on the space of all linear transformations). Can every linear transformation on that space be obtained in this manner (by the choice of a suitable \(A\) )?