Advertisements
Advertisements
Question
If\[A = \begin{bmatrix}1 & 2 \\ 2 & 1\end{bmatrix}\] f (x) = x2 − 2x − 3, show that f (A) = 0
Sum
Advertisements
Solution
Here,
`f (x) =x^2−2x−3`
`⇒f(A)=A^2−2A−3I_2`
Now,
`A^2=A A`
`⇒A^2=[[1 2],[2 1]] ,[[1 2],[2 1]]`
`⇒A^2= [[1+4 2+2],[2+2 4+1]]`
`⇒A^2= [[5 4],[4 5]]`
`f(A)=A^2−2A−3I_2`
⇒f(A)= `[[5 4],[4 5]]-2[[1 2],[2 1]]-3[[1 0],[0 1]]`
⇒f(A)=`[[5 4],[4 5]]-[[2 4],[4 2]]-[[3 0],[0 3]]`
⇒f(A)=`[[5-2-3 4-4-0],[4-4-0 5-2-3]]`
⇒f(A)= `[[5-5 0],[0 5 -5]]`
⇒f(A)=0
shaalaa.com
Is there an error in this question or solution?
