Problems

Filters
Clear Filters

16 problems found

2019 Paper 3 Q3
D: 1500.0 B: 1500.0

The matrix A is given by $$\mathbf{A} = \begin{pmatrix} a & b \\ c & d \end{pmatrix}.$$

  1. You are given that the transformation represented by A has a line \(L_1\) of invariant points (so that each point on \(L_1\) is transformed to itself). Let \((x, y)\) be a point on \(L_1\). Show that \(((a - 1)(d - 1) - bc)xy = 0\). Show further that \((a - 1)(d - 1) = bc\). What can be said about A if \(L_1\) does not pass through the origin?
  2. By considering the cases \(b \neq 0\) and \(b = 0\) separately, show that if \((a - 1)(d - 1) = bc\) then the transformation represented by A has a line of invariant points. You should identify the line in the different cases that arise.
  3. You are given instead that the transformation represented by A has an invariant line \(L_2\) (so that each point on \(L_2\) is transformed to a point on \(L_2\)) and that \(L_2\) does not pass through the origin. If \(L_2\) has the form \(y = mx + k\), show that \((a - 1)(d - 1) = bc\).


Solution:

  1. Suppose \((x,y)\) is on the line of invariant points, then \begin{align*} &&\begin{pmatrix} x \\ y \end{pmatrix} &= \begin{pmatrix} a & b \\ c & d \end{pmatrix} \begin{pmatrix} x \\ y \end{pmatrix} \\ &&&= \begin{pmatrix} ax + by \\ cx + dy \end{pmatrix} \\ \Rightarrow && \begin{cases} (a-1)x + by = 0 \\ (cx + (d-1)y = 0 \end{cases} \tag{*} \end{align*} Therefore either \(x = 0, y = 0\) or \((a-1)(d-1)-bc = 0\) \(\Rightarrow ((a-1)(d-1)-bc)xy = 0\). We also know this is true for all values \(x,y\) on the line of invariant points. If there is one where both \(x \neq 0, y \neq 0\) we are done, otherwise the line of invariant points must be one of the axes. ie but then one of \(\begin{pmatrix} a \\ c \end{pmatrix} = \begin{pmatrix} 1 \\ 0 \end{pmatrix}\) or \(\begin{pmatrix} b \\ d \end{pmatrix} = \begin{pmatrix} 0 \\ 1 \end{pmatrix}\) is true and we'd also be done. If the line doesn't go through the origin then there are points on every line, not equal to the origin which are fixed. But then every point on those lines is fixed (since \(\mathbf{A}\) is a linear operator) and so every point is fixed. ie \(\mathbf{A} = \mathbf{I}\).
  2. Suppose \((a-1)(d-1) -bc = 0\) and \(b \neq 0\) then I claim that \(y = \frac{1-a}{b}x\) is a line of invariant points. It's clear that the first equation will be satisfied in \((*)\) so it suffices to check the second, but the first condition is equivalent to the equations being linearly dependent, ie both equations are satisfied. If \(b = 0\) then \((a-1)(d-1) = 0\), so our matrix must look like \(\begin{pmatrix} 1 & 0 \\ c & d\end{pmatrix}\) (if \(d \neq 1\))or \(\begin{pmatrix} * & 0 \\ * & 1\end{pmatrix}\). In the first case, the line \(y = \frac{c}{1-d}x\) and in the second \(x = 0\) is an invariant line.
  3. Suppose the invariant line is \(y = mx+k\) then we must have that \begin{align*} \begin{pmatrix} a & b \\ c & d \end{pmatrix} \begin{pmatrix} x \\ mx + k \end{pmatrix} &= \begin{pmatrix} (a + mb)x + bk \\ (c+dm)x + dk \end{pmatrix} \end{align*} and \((c+dm)x + dk = m((a + mb)x + bk) +k \Rightarrow k(d-mb-1) = x(-c+(a-d)m+m^2b)\) Since this equation must be true for all values of \(x\), and \(k \neq 0\) we can say that \(mb = d-1\) and \(-c+(a-d)m+m^2b = 0\), ie \(-c + (a-d)m + m(d-1) = 0 \Rightarrow (a-1)m-c = 0\) if \(m \neq 0\) then \((a-1)\frac{(d-1)}{b} - c = 0\) ie our desired relation is true. If \(m = 0\) then we must have that \(y = k\) is an invariant line, ie \(d-1=0\) and \(c=0\) which also satisfies our relation.

2000 Paper 3 Q5
D: 1700.0 B: 1500.0

Given two non-zero vectors $\mathbf{a}=\begin{pmatrix}a_{1}\\ a_{2} \end{pmatrix}\( and \)\mathbf{b}=\begin{pmatrix}b_{1}\\ b_{2} \end{pmatrix}\( define \)\Delta\!\! \left( \bf a, \bf b \right)\( by \)\Delta\!\! \left( \bf a, \bf b \right) = a_1 b_2 - a_2 b_1$. Let \(A\), \(B\) and \(C\) be points with position vectors \(\bf a\), \(\bf b\) and \(\bf c\), respectively, no two of which are parallel. Let \(P\), \(Q\) and \(R\) be points with position vectors \(\bf p\), \(\bf q\) and \(\bf r\), respectively, none of which are parallel.

  1. Show that there exists a \(2 \times 2\) matrix \(\bf M\) such that \(P\) and \(Q\) are the images of \(A\) and \(B\) under the transformation represented by \(\bf M\).
  2. Show that \( \Delta\!\! \left( \bf a, \bf b \right) \bf c + \Delta\!\! \left( \bf c, \bf a \right) \bf b + \Delta\!\! \left( \bf b, \bf c \right) \bf a = 0. \) Hence, or otherwise, prove that a necessary and sufficient condition for the points \(P\), \(Q\), and \(R\) to be the images of points \(A\), \(B\) and \(C\) under the transformation represented by some \(2 \times 2\) matrix \(\bf M\) is that \[ \Delta\!\! \left( \bf a, \bf b \right) : \Delta\!\! \left( \bf b, \bf c \right) : \Delta\!\! \left( \bf c, \bf a \right) = \Delta\!\! \left( \bf p, \bf q \right) : \Delta\!\! \left( \bf q, \bf r \right) : \Delta\!\! \left( \bf r, \bf p \right). \]


Solution:

  1. First notice that there is a matrix taking \((1,0)\) and \((0,1)\) to \(P\) and \(Q\). Notice there is also a matrix taking \((1,0)\) and \((0,1)\) to \(A\) and \(B\). Since \(A\) and \(B\) are not parallel, this map is invertible. Then we must be able to compose this inverse with the second map to obtain a matrix \(\mathbf{M}\) satisfying our conditions.
  2. \(\,\) \begin{align*} && LHS &= \Delta\!\! \left( \bf a, \bf b \right) \bf c + \Delta\!\! \left( \bf c, \bf a \right) \bf b + \Delta\!\! \left( \bf b, \bf c \right) \bf a \\ &&&= (a_1b_2-a_2b_1) \binom{c_1}{c_2} + (c_1a_2-c_2a_1)\binom{b_1}{b_2} + (b_1c_2-b_2c_1)\binom{a_1}{a_2} \\ &&&= \binom{a_1b_2c_1-a_2b_1c_1+c_1a_2b_1-c_2a_1b_1+b_1c_2a_1-b_2c_1a_1}{a_1b_2c_2-a_2b_1c_2+c_1a_2b_2-c_2a_1b_2+b_1c_1a_2-b_2c_1a_2} \\ &&&= \binom{0}{0} \\ &&&= \mathbf{0} \end{align*} First note that the matrix taking \(P\), \(Q\) to \(A\), \(B\) is unique. (\(\Rightarrow\)) Suppose \(\mathbf{Ma} = \mathbf{p}\) and \(\mathbf{Mb} = \mathbf{q}\) and \(\mathbf{Mc} = \mathbf{r}\). Then notice that \begin{align*} && \mathbf{0} &= \mathbf{M0} \\ &&&= \mathbf{M}\left ( \Delta\!\! \left( \bf a, \bf b \right) \bf c + \Delta\!\! \left( \bf c, \bf a \right) \bf b + \Delta\!\! \left( \bf b, \bf c \right) \bf a\right) \\ &&&= \Delta\!\! \left( \bf a, \bf b \right)\mathbf{M} \bf c + \Delta\!\! \left( \bf c, \bf a \right) \mathbf{M}\bf b + \Delta\!\! \left( \bf b, \bf c \right) \mathbf{M}\bf a\\ &&&= \Delta\!\! \left( \bf a, \bf b \right)\bf r + \Delta\!\! \left( \bf c, \bf a \right)\bf q + \Delta\!\! \left( \bf b, \bf c \right) \bf p\\ \end{align*} However, since \(\mathbf{p}, \mathbf{q}, \mathbf{r}\) are not parallel, then these coefficients must be a scalar multiples of \(\Delta(\mathbf{p}, \mathbf{q}), \cdots\) as required. \((\Leftarrow)\) Suppose we have this relationship, and \(\mathbf{Ma} = \mathbf{p}\) and \(\mathbf{Mb} = \mathbf{q}\), then \begin{align*} && \mathbf{0} &= \mathbf{M0} \\ &&&= \mathbf{M}\left ( \Delta\!\! \left( \bf a, \bf b \right) \bf c + \Delta\!\! \left( \bf c, \bf a \right) \bf b + \Delta\!\! \left( \bf b, \bf c \right) \bf a\right) \\ &&&= \Delta\!\! \left( \bf a, \bf b \right)\mathbf{M} \bf c + \Delta\!\! \left( \bf c, \bf a \right) \mathbf{M}\bf b + \Delta\!\! \left( \bf b, \bf c \right) \mathbf{M}\bf a\\ &&&= \Delta\!\! \left( \bf a, \bf b \right)\mathbf{Mc} + \Delta\!\! \left( \bf c, \bf a \right)\bf q + \Delta\!\! \left( \bf b, \bf c \right) \bf p\\ \end{align*} Since these are scalar multiples of \(\Delta(\mathbf{p}, \mathbf{q}), \cdots\) and we write this as \begin{align*} && \mathbf{0} &= \Delta(\mathbf{p}, \mathbf{q})\mathbf{Mc} + \Delta(\mathbf{r}, \mathbf{p})\mathbf{q} + \Delta (\mathbf{q}, \mathbf{r})\mathbf{p} \end{align*} But since \(\mathbf{p}, \mathbf{q}, \mathbf{r}\) are not parallel, this means that \(\mathbf{Mc}\) is uniquely defined to be \(\mathbf{r}\) as required.

1998 Paper 3 Q5
D: 1700.0 B: 1516.0

The exponential of a square matrix \({\bf A}\) is defined to be $$ \exp ({\bf A}) = \sum_{r=0}^\infty {1\over r!} {\bf A}^r \,, $$ where \({\bf A}^0={\bf I}\) and \(\bf I\) is the identity matrix. Let $$ {\bf M}=\left(\begin{array}{cc} 0 & -1 \\ 1 & \phantom{-} 0 \end{array} \right) \,. $$ Show that \({\bf M}^2=-{\bf I}\) and hence express \(\exp({\theta {\bf M}})\) as a single \(2\times 2\) matrix, where \(\theta\) is a real number. Explain the geometrical significance of \(\exp({\theta {\bf M}})\). Let $$ {\bf N}=\left(\begin{array}{rr} 0 & 1 \\ 0 & 0 \end{array}\right) \,. $$ Express similarly \(\exp({s{\bf N}})\), where \(s\) is a real number, and explain the geometrical significance of \(\exp({s{\bf N}})\). For which values of \(\theta\) does $$ \exp({s{\bf N}})\; \exp({\theta {\bf M}})\, = \, \exp({\theta {\bf M}})\;\exp({s{\bf N}}) $$ for all \(s\)? Interpret this fact geometrically.


Solution: \begin{align*} \mathbf{M}^2 &= \begin{pmatrix} 0 & - 1 \\ 1 & 0 \end{pmatrix}^2 \\ &= \begin{pmatrix} 0 \cdot 0 + (-1) \cdot 1 & 0 \cdot (-1) + (-1) \cdot 0 \\ 1 \cdot 0 + 0 \cdot 1 & 1 \cdot (-1) + 0 \cdot 0 \end{pmatrix} \\ &= \begin{pmatrix} -1 & 0 \\ 0 & -1\end{pmatrix} \\ &= - \mathbf{I} \end{align*} \begin{align*} \exp(\theta \mathbf{M}) &= \sum_{r=0}^\infty \frac1{r!} (\theta \mathbf{M})^r \\ &= \sum_{r=0}^\infty \frac{1}{r!} \theta^r \mathbf{M}^r \\ &= \cos \theta \mathbf{I} + \sin \theta \mathbf{M} \\ &= \begin{pmatrix} \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \end{pmatrix} \end{align*} This is a rotation of \(\theta\) degrees about the origin. \begin{align*} && \mathbf{N}^2 &= \begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix}^2 \\ && &= \begin{pmatrix} 0 & 0 \\ 0 & 0 \end{pmatrix} \\ \Rightarrow && \exp(s\mathbf{N}) &= \sum_{r=0}^\infty \frac{1}{r!} (s\mathbf{N})^r \\ &&&= \mathbf{I} + s \mathbf{N} \\ &&&= \begin{pmatrix} 1 &s \\ 0 & 1 \end{pmatrix} \end{align*} This is a shear, leaving the \(y\)-axis invariant, sending \((1,1)\) to \((1+s, 1)\). Suppose those matrices commute, for all \(s\), ie \begin{align*} && \begin{pmatrix} 1 &s \\ 0 & 1 \end{pmatrix}\begin{pmatrix} \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \end{pmatrix} &= \begin{pmatrix} \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \end{pmatrix}\begin{pmatrix} 1 &s \\ 0 & 1 \end{pmatrix} \\ \Rightarrow && \begin{pmatrix} \cos \theta - s \sin \theta & -\sin \theta + s \cos \theta \\ \sin \theta & \cos \theta \end{pmatrix} &= \begin{pmatrix} \cos \theta & s \cos \theta - \sin \theta \\ \sin \theta & s \sin \theta + \cos \theta \end{pmatrix} \\ \Rightarrow && \sin \theta &= 0 \\ \Rightarrow && \theta &=n \pi, n \in \mathbb{Z} \end{align*} Clearly it doesn't matter when we do nothing. If we are rotating by \(\pi\) then it also doesn't matter which order we do it in as the stretch happens in both directions equally.

1996 Paper 3 Q6
D: 1674.0 B: 1529.9

  1. Let \(S\) be the set of matrices of the form \[ \begin{pmatrix}a & a\\ a & a \end{pmatrix}, \] where \(a\) is any real non-zero number. Show that \(S\) is closed under matrix multiplication and, further, that \(S\) is a group under matrix multiplication.
  2. Let \(G\) be a set of \(n\times n\) matrices which is a group under matrix multiplication, with identity element \(\mathbf{E}.\) By considering equations of the form \(\mathbf{BC=D}\) for suitable elements \(\mathbf{B},\) \(\mathbf{C}\) and \(\mathbf{D}\) of \(G\), show that if a given element \(\mathbf{A}\) of \(G\) is a singular matrix (i.e. \(\det\mathbf{A}=0\)), then all elements of \(G\) are singular. Give, with justification, an example of such a group of singular matrices in the case \(n=3.\)


Solution:

  1. Let $\mathbf{A} = \begin{pmatrix}1 & 1\\ 1 & 1 \end{pmatrix}\(, then we need to show that \)(a\mathbf{A})(b\mathbf{A})\( is of the form \)cA\( where \)a, b, c \neq 0$. Since $\mathbf{A}^2 = \begin{pmatrix}2 & 2\\ 2 & 2 \end{pmatrix} = 2\mathbf{A}\( this is certainly the case, since \)(a\mathbf{A})(b\mathbf{A}) = 2ab\mathbf{A}$. To check that we have a group be need to check:
    • Closure (done)
    • Associativity (inherited from matrix multiplication)
    • Identity (\(\frac12 \mathbf{A}\))
    • Inverses the inverse of \(a\mathbf{A}\) is \(\frac{1}{4a}\mathbf{A}\)
  2. Suppose \(\mathbf{A}\) is singular (ie \(\det\mathbf{A}=0\)), then \(\mathbf{AA^{-1}B=B}\) (where inverse is the group inverse rather than the matrix inverse) for any matrix \(\mathbf{B}\). Taking determinants we have: \(\det(\mathbf{AA^{-1}B}) = \det(B) \Rightarrow \det(A) \det(A^{-1}B) = \det(B) \Rightarrow 0 = \det(B)\), ie all matrices are singular. Consider the set of non-zero multiples of \(\begin{pmatrix} 1 & 1 & 1 \\ 1 & 1 & 1 \\ 1 & 1 & 1 \end{pmatrix}\), then the same logic as part (i) will suffice

1995 Paper 3 Q7
D: 1654.7 B: 1516.0

Consider the following sets with the usual definition of multiplication appropriate to each. In each case you may assume that the multiplication is associative. In each case state, giving adequate reasons, whether or not the set is a group.

  1. the complex numbers of unit modulus;
  2. the integers modulo 4;
  3. the matrices \[ \mathrm{M}(\theta)=\begin{pmatrix}\cos\theta & -\sin\theta\\ \sin\theta & \cos\theta \end{pmatrix}, \] where \(0\leqslant\theta<2\pi\);
  4. the integers \(1,3,5,7\) modulo 8;
  5. the \(2\times2\) matrices all of whose entries are integers;
  6. the integers \(1,2,3,4\) modulo 5.
In the case of each pair of groups above state, with reasons, whether or not they are isomorphic.


Solution:

  1. \(\{ z \in \mathbb{C} : |z| = 1\}\) is a group.
    1. (Closure) \(|z_1z_2| = |z_1||z_2| = 1\). Set is closed under multiplication
    2. (Associativity) Multiplication of complex numbers is associative
    3. (Identity) \(|1| = 1\)
    4. (Inverses) \(| \frac{1}{z} | = \frac{1}{|z|} = \frac{1}{1} = 1\), the set contains inverses
  2. the integers \(\pmod{4}\) are not a group under multiplication, \(2\) has no inverse, since \(0 \times k \equiv 0 \pmod{4}\)
  3. The set of rotation matrices is a group:
    1. (Closure) \begin{align*} \begin{pmatrix}\cos\theta_1 & -\sin\theta_1\\ \sin\theta_1 & \cos\theta_1 \end{pmatrix} \begin{pmatrix}\cos\theta_2 & -\sin\theta_2\\ \sin\theta_2 & \cos\theta_2 \end{pmatrix} &= {\scriptscriptstyle\begin{pmatrix}\cos\theta_1 \cos \theta_2 - \sin \theta_1\sin \theta_2 & -\sin\theta_1\ \cos \theta_1 - \sin \theta_2\cos\theta_1\\ \sin\theta_1\ \cos \theta_1 + \sin \theta_2\cos\theta_1 & \cos\theta_1 \cos \theta_2 - \sin \theta_1\sin \theta_2 \end{pmatrix}} \\ &= \begin{pmatrix}\cos(\theta_1+\theta_2) & -\sin(\theta_1+\theta_2)\\ \sin(\theta_1+\theta_2) & \cos(\theta_1+\theta_2) \end{pmatrix} \end{align*} Since \(\cos, \sin\) are periodic with period \(2\pi\), we can find \(\theta_3 = \theta_1 + \theta_2 + 2k\pi\) such that \(0 \leq \theta_3 < 2 \pi\), so our set is closed
    2. (Associativity) Matrix multiplication is associative
    3. (Identity) Consider \(\theta = 0\)
    4. (Inverses) Consider \(2\pi - \theta\)
  4. \(\{1, 3, 5, 7\} \pmod{8}\) is a group:
    1357
    11357
    33175
    55713
    77531
    1. (Closure) See Cayley table
    2. (Associativity) Integer multiplication is associative
    3. (Identity) \(1\)
    4. (Inverses) \(x \mapsto x\) (See Cayley table)
  5. \(2\times2\) matrices are not a group, consider $0 = \begin{pmatrix} 0 & 0 \\ 0 & 0 \end{pmatrix}\(, then \)\mathbf{0}\mathbf{M} = \mathbf{0}$ for all other matrices.
  6. 1234
    11234
    22413
    33142
    44321
    1. (Closure) See Cayley table
    2. (Associativity) Integer multiplication is associative
    3. (Identity) \(1\)
    4. (Inverses) \(1 \mapsto 1, 2 \mapsto 3, 3 \mapsto 2, 4 \mapsto 4\) (See Cayley table)
(i)(iii)(iv)(vi)
(i)\(\checkmark\)\(\checkmark\) consider \(z \mapsto \begin{pmatrix} \cos \arg (z)- \sin \arg(z)
\sin \arg(z)\cos \arg(z) \end{pmatrix}\)not finitenot finite
(iii)\(\checkmark\)not finitenot finite
(iv)\(\checkmark\)no element order \(4\)
(vi)\(\checkmark\)

1994 Paper 3 Q8
D: 1700.0 B: 1516.0

Let \(a,b,c,d,p,q,r\) and \(s\) be real numbers. By considering the determinant of the matrix product \[ \begin{pmatrix}z_{1} & z_{2}\\ -z_{2}^{*} & z_{1}^{*} \end{pmatrix}\begin{pmatrix}z_{3} & z_{4}\\ -z_{4}^{*} & z_{3}^{*} \end{pmatrix}, \] where \(z_{1},z_{2},z_{3}\) and \(z_{4}\) are suitably chosen complex numbers, find expressions \(L_{1},L_{2},L_{3}\) and \(L_{4},\) each of which is linear in \(a,b,c\) and \(d\) and also linear in \(p,q,r\) and \(s,\) such that \[ (a^{2}+b^{2}+c^{2}+d^{2})(p^{2}+q^{2}+r^{2}+s^{2})=L_{1}^{2}+L_{2}^{2}+L_{3}^{2}+L_{4}^{2}. \]


Solution: Supppose \(z_1 = a+ib, z_2 = c+id, z_3 = p+iq, z_4 = r+is\) then: \begin{align*} && \det \left (\begin{pmatrix}z_{1} & z_{2}\\ -z_{2}^{*} & z_{1}^{*} \end{pmatrix}\begin{pmatrix}z_{3} & z_{4}\\ -z_{4}^{*} & z_{3}^{*} \end{pmatrix} \right) &= \det \begin{pmatrix}z_{1} & z_{2}\\ -z_{2}^{*} & z_{1}^{*} \end{pmatrix}\det\begin{pmatrix}z_{3} & z_{4}\\ -z_{4}^{*} & z_{3}^{*} \end{pmatrix} \\ && \det \begin{pmatrix}z_{1}z_3-z_2z_4^* & z_1z_4+z_2z_3^*\\ -z_2^*z_3-z_1^*z_4*& -z_2^*z_4+z_{1}^*z_3^* \end{pmatrix}&= (z_1z_1^*+z_2z_2^*)(z_3z_3^*+z_4z_4^*) \\ && |z_{1}z_3-z_2z_4^*|^2+|z_1z_4+z_2z_3^*|^2&= (a^2+b^2+c^2+d^2)(p^2+q^2+r^2+s^2) \\ && L_1^2 + L_2^2+L_3^2+L_4^2 &= \ldots \end{align*}

1993 Paper 3 Q3
D: 1700.0 B: 1516.0

The matrices \(\mathbf{A},\mathbf{B}\) and \(\mathbf{M}\) are given by \[ \mathbf{A}=\begin{pmatrix}a & 0 & 0\\ b & c & 0\\ d & e & f \end{pmatrix},\quad\mathbf{B}=\begin{pmatrix}1 & p & q\\ 0 & 1 & r\\ 0 & 0 & 1 \end{pmatrix},\quad\mathbf{M}=\begin{pmatrix}1 & 3 & 2\\ 4 & 13 & 5\\ 3 & 8 & 7 \end{pmatrix}, \] where \(a,b,\ldots,r\) are real numbers. Given that \(\mathbf{M=AB},\) show that \(a=1,b=4,c=1,d=3,e=1,f=-2,p=3,q=2\) and \(r=-3\) gives the unique solution for \(\mathbf{A}\) and \(\mathbf{B}.\) Evaluate \(\mathbf{A}^{-1}\) and \(\mathbf{B}^{-1},\) Hence, or otherwise, solve the simultaneous equations \begin{alignat*}{1} x+3y+2z & =7\\ 4x+13y+5z & =18\\ 3x+8y+7z & =25. \end{alignat*}


Solution: \begin{align*} && \begin{pmatrix}1 & 3 & 2\\ 4 & 13 & 5\\ 3 & 8 & 7 \end{pmatrix} &= \begin{pmatrix}a & 0 & 0\\ b & c & 0\\ d & e & f \end{pmatrix}\begin{pmatrix}1 & p & q\\ 0 & 1 & r\\ 0 & 0 & 1 \end{pmatrix} \\ &&&= \begin{pmatrix} a & ap & aq \\ b & pb + c & qb + cr\\ d & pd + e & qd + er +f \end{pmatrix} \\ \Rightarrow && a,b,d,p,q&=1,4,3,3,2\\ &&&= \begin{pmatrix} 1 & 3 & 2 \\ 4 & 12 + c & 8+ cr\\ 3 & 9 + e & 6 + er +f \end{pmatrix} \\ \Rightarrow && c, e&=1,-1\\ &&&= \begin{pmatrix} 1 & 3 & 2 \\ 4 & 13 & 8+ r\\ 3 & 8 & 6 -r +f \end{pmatrix} \\ \Rightarrow && r, f &= -3, -2 \end{align*} \begin{align*} \mathbf{A}^{-1} &= \begin{pmatrix} 1 & 0 & 0 \\ 4 & 1 & 0\\ 3 & -1 & -2 \end{pmatrix}^{-1} \\ &=\frac{1}{-2} \begin{pmatrix} -2 & 0 & 0 \\ 8 & -2 & 0\\ -7 & 1 & 1 \end{pmatrix} \\ \\ \mathbf{B}^{-1} &= \begin{pmatrix} 1 & 3 & 2 \\ 0 & 1 & -3\\ 0 & 0 & 1 \end{pmatrix}^{-1} \\ &= \begin{pmatrix} 1 & -3 & -11 \\ 0 & 1 & 3\\ 0 & 0 & 1 \end{pmatrix} \\ \end{align*} We want to solve \(\mathbf{M}\mathbf{v} = \begin{pmatrix} 7 \\ 18 \\ 25 \end{pmatrix}\), ie \begin{align*} \mathbf{v} &= \mathbf{M}^{-1}\begin{pmatrix} 7 \\ 18 \\ 25 \end{pmatrix} \\ &= \mathbf{B}^{-1} \mathbf{A}^{-1}\begin{pmatrix} 7 \\ 18 \\ 25 \end{pmatrix} \\ &= \frac{1}{-2}\mathbf{B}^{-1} \begin{pmatrix} -2 & 0 & 0 \\ 8 & -2 & 0\\ -7 & 1 & 1 \end{pmatrix} \begin{pmatrix} 7 \\ 18 \\ 25 \end{pmatrix} \\ &= \frac{1}{-2}\mathbf{B}^{-1} \begin{pmatrix} -14 \\ 20 \\ -6 \end{pmatrix} \\ &= \begin{pmatrix} 1 & -3 & -11 \\ 0 & 1 & 3\\ 0 & 0 & 1 \end{pmatrix}\begin{pmatrix} 7 \\ -10 \\ 3 \end{pmatrix} \\ &= \begin{pmatrix} 4\\ -1 \\ 3 \end{pmatrix} \end{align*} This algorithm is called the "LU-decomposition"

1993 Paper 3 Q5
D: 1730.5 B: 1466.6

The set \(S\) consists of ordered pairs of complex numbers \((z_1,z_2)\) and a binary operation \(\circ\) on \(S\) is defined by $$ (z_1,z_2)\circ(w_1,w_2)= (z_1w_1-z_2w^*_2, \; z_1w_2+z_2w^*_1). $$ Show that the operation \(\circ\) is associative and determine whether it is commutative. Evaluate \((z,0)\circ(w,0)\), \((z,0)\circ(0,w)\), \((0,z)\circ(w,0)\) and \((0,z)\circ(0,w)\). The set \(S_1\) is the subset of \(S\) consisting of \(A\), \(B\), \(\ldots\,\), \(H\), where \(A=(1,0)\), \(B=(0,1)\), \(C=(i,0)\), \(D=(0,i)\), \(E=(-1,0)\), \(F=(0,-1)\), \(G=(-i,0)\) and \(H=(0,-i)\). Show that \(S_1\) is closed under \(\circ\) and that it has an identity element. Determine the inverse and order of each element of \(S_1\). Show that \(S_1\) is a group under \(\circ\). \hfil\break [You are not required to compute the multiplication table in full.] Show that \(\{A,B,E,F\}\) is a subgroup of \(S_1\) and determine whether it is isomorphic to the group generated by the \(2\times2\) matrix $\begin{pmatrix}0 & 1\\ -1 & 0 \end{pmatrix}$ under matrix multiplication.

1993 Paper 3 Q10
D: 1700.0 B: 1484.0

The transformation \(T\) of the point \(P\) in the \(x\),\(y\) plane to the point \(P'\) is constructed as follows: \hfil\break Lines are drawn through \(P\) parallel to the lines \(y=mx\) and \(y=-mx\) to cut the line \(y=kx\) at \(Q\) and \(R\) respectively, \(m\) and \(k\) being given constants. \(P'\) is the fourth vertex of the parallelogram \(PQP'R\). Show that if \(P\) is \((x_1,y_1)\) then \(Q\) is $$ \left( {mx_1-y_1 \over m-k}, {k(mx_1-y_1)\over m-k}\right). $$ Obtain the coordinates of \(P'\) in terms of \(x_1\), \(y_1\), \(m\) and \(k\), and express \(T\) as a matrix transformation. Show that areas are transformed under \(T\) into areas of the same magnitude.

1992 Paper 2 Q5
D: 1577.1 B: 1470.1

Explain what is meant by the order of an element \(g\) of a group \(G\). The set \(S\) consists of all \(2\times2\) matrices whose determinant is \(1\). Find the inverse of the element \(\mathbf{A}\) of \(S\), where \[ \mathbf{A}=\begin{pmatrix}w & x\\ y & z \end{pmatrix}. \] Show that \(S\) is a group under matrix multiplication (you may assume that matrix multiplication is associative). For which elements \(\mathbf{A}\) is \(\mathbf{A}^{-1}=\mathbf{A}\)? Which element or elements have order 2? Show that the element \(\mathbf{A}\) of \(S\) has order 3 if, and only if, \(w+z+1=0.\) Write down one such element.


Solution: The order of an element \(g\) is the smallest positive number \(k\) such that \(g^k = e\). $\mathbf{A}^{-1} = \begin{pmatrix}z & -x\\ -y & w \end{pmatrix}$. Claim, \(S\) is a group. \begin{enumerate} \item (Closure) The product of two \(2\times2\) matrices is always a \(2\times 2\) matrix so we only need to check the determinant. Suppose \(\det(\mathbf{A}) = \det (\mathbf{B}) = 1\), then \(\det(AB) = \det(A)\det(B) = 1\), so our operation is closed \item (Associativity) Inherited from matrix multiplication \item (Identity) $\mathbf{I} =\begin{pmatrix}1 & 0\\ 1 & 1 \end{pmatrix}\( has determinant \)1$. \item (Inverses) The inverse is always fine since the matrix of cofactors always contains integers and the determinant is one, so we never end up with anything which isn't an integer. \end{itemize} If \(\mathbf{A}^-1 = \mathbf{A}\) then assuming $\mathbf{A} = \begin{pmatrix}a & b\\ c & d \end{pmatrix}\( then \)\mathbf{A}^{-1} = \begin{pmatrix}d & - b\\ -c & a \end{pmatrix}\( so we must have \)a=d, -b=b, -c=c\(, so \)b = c = 0\( and \)a = d\(. For the determinant to be \)1\( we must have \)ad = a^2 = 1\(, ie \)a = \pm 1\(. Therefore we must have \)\mathbf{A} = \begin{pmatrix}1 & 0\\ 0 & 1 \end{pmatrix}\( or \)\mathbf{A} = \begin{pmatrix}-1 & 0\\ 0 & -1 \end{pmatrix}$. For an element to have order \(2\) then \(\mathbf{A}^2 = \mathbf{I}\) ie, \(\mathbf{A} = \mathbf{A}^{-1}\) and \(\mathbf{A} \neq \mathbf{I}\) therefore the only element of order \(2\) is $\begin{pmatrix}-1 & 0\\ 0 & -1 \end{pmatrix}$. For an element to have order \(3\) we must have \(\mathbf{A}^2 = \mathbf{A}^{-1}\), ie $\begin{pmatrix}w^2 + xy & x(w+z)\\ y(w+z) & z^2 + xy \end{pmatrix} = \begin{pmatrix}z & -x\\ -y & w \end{pmatrix}$. Therefore \(w^2 + xy = z, x(w+z) = -x, y(w+z) = -y, z^2+xy = w\). The second and third equations are satisfied iff \(w+z+1 = 0\) or \(x = 0\) and \(y = 0\), but if \(x = 0\) and \(y = 0\) then we aren't order \(3\), so we just need to check this is sufficient for the first and last equations. Since \(\det(\mathbf{A}) = 1\) we have \(wz =xy +1\), so the first and last equations are equivalent to \(w^2 + wz - 1 = z\) and \(x^2 + wz-1 = w\) which are equivalent to \(w(w+z) = z+1\) or \(w + z+ 1 = 0\) as required

1991 Paper 2 Q9
D: 1616.2 B: 1500.0

Let \(G\) be the set of all matrices of the form \[ \begin{pmatrix}a & b\\ 0 & c \end{pmatrix}, \] where \(a,b\) and \(c\) are integers modulo 5, and \(a\neq0\neq c\). Show that \(G\) forms a group under matrix multiplication (which may be assumed to be associative). What is the order of \(G\)? Determine whether or not \(G\) is commutative. Determine whether or not the set consisting of all elements in \(G\) of order \(1\) or \(2\) is a subgroup of \(G\).


Solution: Claim \(G\) is a group under matrix multiplication

  • (Closure) Suppose \(\mathbf{A}\) and \(\mathbf{B}\) are matrices of that form, then \(\begin{pmatrix} a_1 & b_1 \\ 0 & c_1 \end{pmatrix} \begin{pmatrix} a_2 & b_2 \\ 0 & c_2 \end{pmatrix} = \begin{pmatrix} a_1a_2 & a_1b_2 + b_1c_2 \\ 0 & c_1c_2 \end{pmatrix}\), this is clearly of the required form since if \(a_1, a_2, c_1, c_2 \neq 0\) then \(a_1a_2, c_1c_2 \neq 0\)
  • (Associative) By inheritance from matrix multiplication
  • (Identity) Consider \(\begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix}\) also clearly of the required form.
  • (Inverse) Consider \((ac)^{-1}\begin{pmatrix} c & -b \\ 0 & a \end{pmatrix}\), since \(ac \neq 0\) we can assume it has an inverse mod \(5\). therefore we have another matrix of the required form.
There are \(4\) possible values for \(a\) and \(c\) and \(5\) possible values for \(b\), so \(4 \times 4 \times 5 = 80\) elements, so the group is order \(80\). \(G\) is not commutative, consider \(\begin{pmatrix} 1 & 1 \\ 0 & 2 \end{pmatrix} \begin{pmatrix} 1 & 1\\ 0 & 1 \end{pmatrix} = \begin{pmatrix} 1 & 2 \\ 0 & 2 \end{pmatrix}\) \(\begin{pmatrix} 1 & 1 \\ 0 & 1 \end{pmatrix} \begin{pmatrix} 1 & 1 \\ 0 & 2 \end{pmatrix} = \begin{pmatrix} 1 & 3 \\ 0 & 2 \end{pmatrix}\) The elements of order \(1\) or \(2\) satisfy \(\begin{pmatrix} a & b \\ 0 & c \end{pmatrix} = \begin{pmatrix} a^{-1} & -ba^{-1}c^{-1} \\ 0 & c^{-1} \end{pmatrix}\) Therefore \(a^2 = 1, c^2 = 1 \Rightarrow a, c = 1, 4\) and \(b = -ba^{-1}c^{-1} \Rightarrow b = 0\) or , \(ac = -1\), so we have \((a,b,c) = (1,0,1), (4,0,4), (1, *, 4), (4, *, 1)\) So there are \(12\) elements of order \(1\) or \(2\). But this can't be a subgroup since \(12 \not \mid 80\)

1991 Paper 3 Q6
D: 1700.0 B: 1516.0

The transformation \(T\) from \(\binom{x}{y}\) to \(\binom{x'}{y'}\) in two-dimensional space is given by \[ \begin{pmatrix}x'\\ y' \end{pmatrix}=\begin{pmatrix}\cosh u & \sinh u\\ \sinh u & \cosh u \end{pmatrix}\begin{pmatrix}x\\ y \end{pmatrix}, \] where \(u\) is a positive real constant. Show that the curve with equation \(x^{2}-y^{2}=1\) is transformed into itself. Find the equations of two straight lines through the origin which transform into themselves. A line, not necessary through the origin, which has gradient \(\tanh v\) transforms under \(T\) into a line with gradient \(\tanh v'\). Show that \(v'=v+u\). The lines \(\ell_{1}\) and \(\ell_{2}\) with gradients \(\tanh v_{1}\) and \(\tanh v_{2}\) transform under \(T\) into lines with gradients \(\tanh v_{1}'\) and \(\tanh v_{2}'\) respectively. Find the relation satisfied by \(v_{1}\) and \(v_{2}\) that is the necessary and sufficient for \(\ell_{1}\) and \(\ell_{2}\) to intersect at the same angle as their transforms. In the case when \(\ell_{1}\) and \(\ell_{2}\) meet at the origin, illustrate in a diagram the relation between \(\ell_{1}\), \(\ell_{2}\) and their transforms.

1990 Paper 3 Q6
D: 1700.0 B: 1500.0

The transformation \(T\) from \(\begin{pmatrix} x \\ y \end{pmatrix}\) to \(\begin{pmatrix} X \\ Y \end{pmatrix}\) is given by \[ \begin{pmatrix}X\\ Y \end{pmatrix}=\frac{2}{5}\begin{pmatrix}9 & -2\\ -2 & 6 \end{pmatrix}\begin{pmatrix}x\\ y \end{pmatrix}. \] Show that \(T\) leaves the vector \(\begin{pmatrix} 1\\ 2 \end{pmatrix}\) unchanged in direction but multiplied by a scalar, and that \(\begin{pmatrix} 2\\ -1 \end{pmatrix}\) is similarly transformed. The circle \(C\) whose equation is \(x^{2}+y^{2}=1\) transforms under \(T\) to a curve \(E\). Show that \(E\) has equation \[ 8X^{2}+12XY+17Y^{2}=80, \] and state the area of the region bounded by \(E\). Show also that the greatest value of \(X\) on \(E\) is \(2\sqrt{17/5}.\) Find the equation of the tangent to \(E\) at the point which corresponds to the point \(\frac{1}{5}(3,4)\) on \(C\).


Solution: \begin{align*} T\begin{pmatrix}1\\ 2 \end{pmatrix} &= \frac{2}{5}\begin{pmatrix}9 & -2\\ -2 & 6 \end{pmatrix}\begin{pmatrix}1\\ 2 \end{pmatrix} \\ &= \frac25\begin{pmatrix}9 - 4\\ -2+12 \end{pmatrix} \\ &= \begin{pmatrix}2\\ 4 \end{pmatrix} \\ &= 2 \begin{pmatrix}1\\ 2 \end{pmatrix} \end{align*} \begin{align*} T\begin{pmatrix}1\\ 2 \end{pmatrix} &= \frac{2}{5}\begin{pmatrix}9 & -2\\ -2 & 6 \end{pmatrix}\begin{pmatrix}2\\ -1 \end{pmatrix} \\ &= \frac25\begin{pmatrix}18+2\\ -4-6 \end{pmatrix} \\ &= \begin{pmatrix}8\\ -4 \end{pmatrix} \\ &= 4 \begin{pmatrix}2\\ -1 \end{pmatrix} \end{align*} Consider $T^{-1} = \frac{5}{2} \frac{1}{50}\begin{pmatrix}6 & 2\\ 2 & 9 \end{pmatrix}\(, so \)T^{-1} \begin{pmatrix}X\\ Y \end{pmatrix} = \begin{pmatrix}x\\ y \end{pmatrix}$ and so: \begin{align*} x^2 + y^2 & = \begin{pmatrix}x& y \end{pmatrix}\begin{pmatrix}x\\ y \end{pmatrix} \\ &= \begin{pmatrix}X& Y \end{pmatrix} (T^{-1})^T T^{-1} \begin{pmatrix}X\\ Y \end{pmatrix} \\ &= \frac{1}{400}\begin{pmatrix}X& Y \end{pmatrix}\begin{pmatrix}6 & 2\\ 2 & 9 \end{pmatrix}\begin{pmatrix}6 & 2\\ 2 & 9 \end{pmatrix} \begin{pmatrix}X\\ Y \end{pmatrix} \\ &= \frac{1}{400}\begin{pmatrix}X& Y \end{pmatrix}\begin{pmatrix}6 & 2\\ 2 & 9 \end{pmatrix} \begin{pmatrix}6X+2Y\\ 2X+9Y \end{pmatrix} \\ &= \frac{1}{400}\begin{pmatrix}X& Y \end{pmatrix} \begin{pmatrix}6(6X+2Y)+2(2X+9Y)\\ 2(6X+2Y)+9(2X+9Y) \end{pmatrix} \\ &= \frac{1}{400}\begin{pmatrix}X& Y \end{pmatrix} \begin{pmatrix}40X+30Y\\ 30X +85Y \end{pmatrix} \\ &= \frac{1}{80}\begin{pmatrix}X& Y \end{pmatrix} \begin{pmatrix}8X+6Y\\ 6X +17Y \end{pmatrix} \\ &= \frac{1}{80} \l 8X^2 + 12XY + 17Y^2\r \end{align*} Therefore \(8X^2 + 12XY + 17Y^2 = 80\). The area will be \(\det T \cdot \pi = \frac{4}{25} \cdot 50 \cdot \pi = 8 \pi\). Differentiating we obtain \(2 \cdot 8 \cdot X \cdot \frac{dX}{dY} + 2 \cdot 6 \cdot X + 2 \cdot 6 \cdot Y \cdot \frac{dX}{dY} + 2 \cdot 17 Y \Rightarrow \frac{dX}{dY} = -\frac{6X + 17Y}{8X+6Y}\), at a maximum (or minimum, \(6X = -17Y\)). Therefore \begin{align*} \Rightarrow && 8X^2 + 12 \cdot \frac{6}{17}X^2 + 17 ( -\frac{6}{17} X)^2 &= 80 \\ \Rightarrow && \frac{100}{17}X^2 &= 80 \\ \Rightarrow &&X^2 &= \frac{17 \cdot 4}{5} \\ \Rightarrow && |X| = 2 \sqrt {\frac{17}{5}} \end{align*} The point \(\frac15 (3,4)\) maps to \begin{align*} \frac{2}{5}\frac{1}{5}\begin{pmatrix}9 & -2\\ -2 & 6 \end{pmatrix}\begin{pmatrix}3\\ 4 \end{pmatrix} &= \frac{2}{25} \begin{pmatrix}19\\ 18 \end{pmatrix} \end{align*} So the point is \((\frac{38}{25}, \frac{36}{25})\), with gradient \(\frac{dY}{dX} = -\frac{8X+6Y}{6X + 17Y}\) which is \(-\frac{8 \cdot 19+6 \cdot 18}{6\cdot 19 + 17 \cdot 18} = -\frac{13}{21}\) therefore the equation is \(21Y+13X = 50\)

1989 Paper 2 Q9
D: 1600.0 B: 1515.3

The matrix \(\mathbf{F}\) is defined by \[ \mathbf{F}=\mathbf{I}+\sum_{n=1}^{\infty}\frac{1}{n!}t^{n}\mathbf{A}^{n}, \] where $\mathbf{A}=\begin{pmatrix}-3 & -1\\ 8 & 3 \end{pmatrix} \( and \) t \( is a variable scalar. Evaluate \)\mathbf{A}^{2}$, and show that \[ \mathbf{F}=\mathbf{I}\cosh t+\mathbf{A}\sinh t. \] Show also that \(\mathbf{F}^{-1}=\mathbf{I}\cosh t-\mathbf{A}\sinh t\), and that \(\dfrac{\mathrm{d}\mathbf{F}}{\mathrm{d}t}=\mathbf{FA}\). The vector $\mathbf{r}=\begin{pmatrix}x(t)\\ y(t) \end{pmatrix}$ satisfies the differential equation \[ \frac{\mathrm{d}\mathbf{r}}{\mathrm{d}t}+\mathbf{A}\mathbf{r}=\mathbf{0}, \] with \(x=\alpha\) and \(y=\beta\) at \(t=0.\) Solve this equation by means of a suitable matrix integrating factor, and hence show that \begin{alignat*}{1} x(t) & =\alpha\cosh t+(3\alpha+\beta)\sinh t\\ y(t) & =\beta\cosh t-(8\alpha+3\beta)\sinh t. \end{alignat*}


Solution: \begin{align*} \begin{pmatrix} -3 & -1 \\ 8 & 3 \end{pmatrix}^2 &= \begin{pmatrix} 9-8 & 3-3 \\ -24+24 & -8+9 \end{pmatrix} \\ &= \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} \\ &= \mathbf{I} \end{align*} Therefore: \begin{align*} \mathbf{F} &= \mathbf{I}+\sum_{n=1}^{\infty}\frac{1}{n!}t^{n}\mathbf{A}^{n} \\ &= \mathbf{I} + \sum_{n=1}^{\infty} \frac{1}{(2n)!}t^{2n} \mathbf{I} + \sum_{n=0}^{\infty} \frac{1}{(2n+1)!}t^{2n+1} \mathbf{A} \\ &= \cosh t \mathbf{I} + \sinh t \mathbf{A} \end{align*} Notice that \begin{align*} \mathbf{F} (\mathbf{I}\cosh t-\mathbf{A}\sinh t) &= (\mathbf{I}\cosh t+\mathbf{A}\sinh t)(\mathbf{I}\cosh t-\mathbf{A}\sinh t) \\ &= \mathbf{I}^2 \cosh^2 t+\mathbf{A}(\sinh t \cosh t - \cosh t \sinh t) - \mathbf{A}^2\sinh^2 t \\ &= \mathbf{I} \cosh^2 t - \mathbf{I} \sinh^2 t \\ &= \mathbf{I} \end{align*} Therefore \(\mathbf{F}^{-1} = \mathbf{I}\cosh t-\mathbf{A}\sinh t\) \begin{align*} \frac{\d \mathbf{F}}{\d t} &= \frac{\d }{\d t} \left [ \mathbf{I}+\sum_{n=1}^{\infty}\frac{1}{n!}t^{n}\mathbf{A}^{n}\right] \\ &= \sum_{n=1}^{\infty} \frac{1}{(n-1)!}t^{n-1} \mathbf{A}^n \\ &= \left ( \sum_{n=1}^{\infty} \frac{1}{(n-1)!}t^{n-1} \mathbf{A}^{n-1} \right) \mathbf{A} \\ &= \mathbf{F}\mathbf{A} \end{align*} \begin{align*} && \frac{\mathrm{d}\mathbf{r}}{\mathrm{d}t}+\mathbf{A}\mathbf{r}&=\mathbf{0} \\ \Rightarrow && \mathbf{F} \frac{\d \mathbf{r}}{\d t} + \mathbf{FAr} &= \mathbf{0} \\ && \frac{\d }{\d t} \left ( \mathbf{F} \mathbf{r}\right) &= 0 \\ \Rightarrow && \mathbf{Fr} &= \mathbf{c} \\ \Rightarrow && \mathbf{r} &= \mathbf{F}^{-1}\mathbf{c} \\ &&&= ( \mathbf{I}\cosh t-\mathbf{A}\sinh t)\mathbf{c} \\ t = 0: && \binom{\alpha}{\beta} &= \mathbf{c} \\ \Rightarrow && \mathbf{r} &= ( \mathbf{I}\cosh t-\mathbf{A}\sinh t)\binom{\alpha}{\beta} \\ &&&= \binom{\alpha \cosh t}{\beta \cosh t} - \binom{-3\alpha-\beta}{8\alpha + 3\beta}\sinh t \\ &&&= \binom{\alpha \cosh t + (3\alpha + \beta) \sinh t}{\beta \cosh t -(8\alpha + 3\beta)\sinh t} \end{align*} as required

1989 Paper 3 Q3
D: 1675.2 B: 1469.0

The matrix \(\mathbf{M}\) is given by \[ \mathbf{M}=\begin{pmatrix}\cos(2\pi/m) & -\sin(2\pi/m)\\ \sin(2\pi/m) & \cos(2\pi/m) \end{pmatrix}, \] where \(m\) is an integer greater than \(1.\) Prove that \[ \mathbf{M}^{m-1}+\mathbf{M}^{m-2}+\cdots+\mathbf{M}^{2}+\mathbf{M}+\mathbf{I}=\mathbf{O}, \] where $\mathbf{I}=\begin{pmatrix}1 & 0\\ 0 & 1 \end{pmatrix}\( and \)\mathbf{O}=\begin{pmatrix}0 & 0\\ 0 & 0 \end{pmatrix}.$ The sequence \(\mathbf{X}_{0},\mathbf{X}_{1},\mathbf{X}_{2},\ldots\) is defined by \[ \mathbf{X}_{k+1}=\mathbf{PX}_{k}+\mathbf{Q}, \] where \(\mathbf{P,Q}\) and \(\mathbf{X}_{0}\) are given \(2\times2\) matrices. Suggest a suitable expression for \(\mathbf{X}_{k}\) in terms of \(\mathbf{P},\) \(\mathbf{Q}\) and \(\mathbf{X}_{0},\) and justify it by induction. The binary operation \(*\) is defined as follows: \[ \mathbf{X}_{i}*\mathbf{X}_{j}\mbox{ is the result of substituting \ensuremath{\mathbf{X}_{j}}for \ensuremath{\mathbf{X}_{0}}in the expression for \ensuremath{\mathbf{X}_{i}}. } \] Show that if \(\mathbf{P=M},\) the set \(\{\mathbf{X}_{1},\mathbf{X}_{2},\mathbf{X}_{3},\ldots\}\) forms a finite group under the operation \(*\).


Solution: \(\mathbf{M}^m = \mathbf{I}\), we also have \(\mathrm{det}(\mathbf{M - I}) = \cos^2(2\pi/m) - 2\cos(2\pi/m) + 1 + \sin^2(2\pi/m) = 2(1-\cos(2\pi/m))\) therefore \(\mathbf{M-I}\) is invertible. Therefore since \(\mathbf{(M-I)(M^{m-1} + M^{m-1} + \cdots + M^2 + M + I)= M^m-I = 0}\) we can cancel the \(\mathbf{M-I}\) to obtain the desired result. \(\mathbf{X_0 = X_0}\) \(\mathbf{X_1 = PX_0+Q}\) \(\mathbf{X_2 = P(PX_0+Q)+Q = P^2X_0 + PQ + Q}\) Claim: \(\mathbf{X_k = P^k X_0 + (P^{k-1} + P^{k-2} + \cdots + I)Q}\) Proof: (By induction on \(k\)). Base case \(k = 0\) is true. Assume it's true for some \(k = l\), then consider \(k = l+1\) \(\mathbf{X_{l+1} = PX_l + Q = P( P^l X_0 + (P^{l-1} + P^{l-2} + \cdots + I)Q) + Q = P^{l+1}X_0 + (P^l + P^{l-1} + \cdots + P)Q + Q = P^{l+1}X_0 + (P^l + P^{l-1} + \cdots + P + I)Q}\) Suppose \(\mathbf{P} = \mathbf{M}\), then consider the set \(\{\mathbf{X_1, X_2}, \ldots\}\) with the operation \(*\) as defined. \(\mathbf{X_i * X_j} = M^{i}(X_j) + (M^{i-1} + M^{i-2} + \cdots + M + I)Q = M^{i}(M^jX_0 + (M^{j-1} + M^{j-2} + \cdots + M + I)Q) + (M^{i-1} + M^{i-2} + \cdots + M + I)Q = M^{i+j}X_0 + (M^{i+j-1}+\cdots + M + I)Q = X_{i+j}\) Since \(X_m = X_0\) we can check all the requirements of the group, but this is going to be isomorphic to the cyclic group with \(m\) elements.