Evaluate the determinants:
We know that determinant of A is calculated as
Now,
= 2(-1) – 4(-5)
= -2 – (-20)
= -2 + 20
= 18
The determinant of the above matrix is 18.
Evaluate the determinants:
We know that determinant of A is calculated as
Now,
= cosθ(cosθ) - (-sinθ)(sinθ)
= cos2θ + sin2θ
= 1 [∵ cos2θ + sin2θ = 1]
The determinant of the above matrix is 1.
Evaluate the determinants:
We know that determinant of A is calculated as
Now,
= (x2 – x + 1)(x + 1) - (x - 1)(x + 1)
= (x3 - x2 + x + x2 – x + 1) - (x2 - 1)
= x3 + 1 - x2 + 1
= x3 - x2 + 2
Ans. The determinant of the above matrix is x3 - x2 + 2.
If then show that |2A| = 4|A|.
|A| =
We know that determinant of A is calculated as
= 1(2) - 2(4)
= 2 - 8
|A| = -6
LHS: |2A|
= 2(4) - 4(8)
= 8 - 32 = -24
|2A| = -24 …LHS
RHS: 4|A|
4|A|= 4(-6)
= -24
4|A| = -24 …RHS
LHS = RHS
Hence proved.
If then show that |3A| = 27|A|
We know that a determinant of a 3 x 3 matrix is calculated as
= 1[1(4) - 2(0)] – 0 + 1[0-0]
= 1[4 - 0] – 0 + 0
= 4
|A|= 4
LHS: |3A|
= 3[3(12) - 0(6)] – 0 + 3[0 – 0]
= 3(36) – 0 + 0
= 108
|3A| = 108 ----LHS
RHS: 27|A|
27|A| = 27(4)
= 108
27|A| = 108 ----RHS
LHS=RHS
Hence proved.
Evaluate the determinants
Now,
We know that a determinant of a 3x3 matrix is calculated as
= 3[0 – (-1)(-5)] +1[0 – (-1)(3)] – 2[0 – 0]
= 3(-5) + 1(3) – 0
= -15 + 3
= -12
The determinant of the above matrix is -12
Evaluate the determinants
Now,
We know that a determinant of a 3 x 3 matrix is calculated as
= 3[1 – (-2)(3)] + 4[1 – (-2)(2)] + 5[3-2]
= 3[1 + 6] + 4[1 + 4] + 5[1]
= 3[7] + 4[5] + 5
= 21 + 20 + 5
= 46
The determinant of the above matrix is 46.
Evaluate the determinants
Now,
We know that a determinant of a 3 x 3 matrix is calculated as
= 0 – 1[0 – (-3)(-2)] + 2[(-1)(3) – 0]
= 0 – 1[0 - 6] + 2[-3 – 0]
= 0 – 1[-6] + 2[-3]
= 0 + 6 – 6
= 0
The determinant of the above matrix is 0.
Evaluate the determinants
Now,
We know that a determinant of a 3 x 3 matrix is calculated as
= 2[0 – (-1)(-5)] + 1[0 – (-1)(3)] – 2[0 – 3(2)]
= 2[0 – 5] + 1[0 + 3] – 2[-6]
= 2[-5] + 1[3] -2[-6]
= -10 + 3 + 12
= 5
The determinant of the above matrix is 5.
If A = , find |A|.
GIVEN:
Now,
We know that a determinant of a 3 x 3 matrix is calculated as
.
= 1[-9 – (-3)(4)] – 1[2(-9) – (-3)(5)] – 2[2(4) – 1(5)]
= 1[-9 + 12] – 1[-18 + 15] – 2[8 – 5]
= 1[3] – 1[-3] – 2[3]
= 3 + 3 – 6
= 0
Ans. |A| = 0
Find values of x, if
We have
We know that determinant of A is calculated as
⇒ 2(1) – 4(5) = 2x(x) – 4(6)
⇒ 2 – 20 = 2x2 - 24
⇒ -18 = 2x2 – 24
⇒ 2x2 = -24 + 18
⇒ 2x2 = 6
⇒ x2 = 6/2
⇒ x2 = 3
x = �√3
Ans. The value of x is �√3
Evaluate the determinants
We know that determinant of A is calculated as
We have
⇒ 2(5) – 3(4) = x(5) – 3(2x)
⇒ 10 – 12 = 5x – 6x
⇒ -2 = -x
⇒ x = 2
Ans. The value of x is 2.
If , then x is equal to
A. 6
B. ±6
C. –6
D. 0
We have
We know that determinant of A is calculated as
⇒ x(x) – 2(18) = 6(6) – 2(18)
⇒ x2 - 36 = 36 – 36
⇒ x2 =36 – 36 + 36
⇒ x2 = 36
⇒ x = ±6
Using the property of determinants and without expanding in prove that:
Applying Operations C1→ C1 + C2 (i.e. Replacing 1st column by addition of 1st and 2nd column)
C1→ C1 - C3 (i.e. Replacing 1st column by subtraction of 1st and 3rd column)
If any one of the rows or columns of a determinant is 0 then the value of that determinant is 0.
∴ LHS = 0 = RHS
Using the property of determinants and without expanding in prove that:
Applying Operation C1→ C1 + C2 (i.e. Replacing 1st column by addition of 1st and 2nd column)
C1→ C1 + C3 (i.e. Replacing 1st column by addition of 1st and 3rd column)
If any one of the rows or columns of a determinant is 0 then the value of that determinant is 0.
∴ LHS = 0 = RHS
Using the property of determinants and without expanding in prove that:
Applying Operation C1→ C1 + 9C2 (i.e. Replacing 1st column by addition of 1st column and 9 times second column)
C1→ C1 - C3 (i.e. Replacing 1st column by subtraction of 1st and 3rd column)
If any one of the rows or columns of a determinant is 0 then the value of that determinant is 0.
∴ LHS = 0 = RHS
Using the property of determinants and without expanding in prove that:
C3→ C2 + C3 (i.e. replace 3rd column by addition of 2nd and 3rd column)
Taking ab + bc + ac outside determinant
C1→ C1 - C3 (i.e. replace 1st column by subtraction of 1st and 3rd column)
If any one of the rows or columns of a determinant is 0 then the value of that determinant is 0.
∴ LHS = 0 = RHS
∴
Using the property of determinants and without expanding in prove that:
When two determinants are added each of the corresponding elements gets added.
Here we can split the LHS determinant as
For the determinant, u perform the following transformation R1↔ R3 (i.e. interchange 1st row with 3rd row)
When two particular rows/columns of a determinant are interchanged the value becomes negative 1 times the original value.
R1↔ R2 (i.e. interchange 1st row with 2nd row)
R1 ↔ R3 (i.e. interchange 1st row with 3rd row)
R1 ↔ R2 (i.e. interchange 1st row with 2nd row)
∴ LHS = RHS
∴
Using the property of determinants and without expanding in prove that:
To Prove:
R1→ cR1 (i.e. replace 1st row by multiplying it with c)
As we are multiplying we should also divide c so that the original given determinant is not changed
R1→ R1 - bR2 (i.e. replace 1st row by subtraction of 1st row and b times 2nd row)
Taking a outside the determinant from 1st row
If any two rows or columns of a determinant are identical then the value of that determinant is 0 because we get a row or column with all elements 0 when we when we subtract those particular rows/columns here the transformation is R1 → R1 - R3
∴LHS = 0 = RHS
∴
Using the property of determinants and without expanding in prove that:
Taking ‘a’, ‘b’ and ‘c’ outside the determinant from 1st,2nd and 3rd column respectively
Taking ‘a’, ‘b’ and ‘c’ outside the determinant from 1st,2nd and 3rd row respectively
.
R1 → R1 + R2 (i.e. Replacing 1st row by addition of 1st and 2nd row)
panding the determinant along 1st row
∴ LHS = a2 b2 c2 × 2(1 - (-1)) = 4a2b2c2 = RHS
∴
By using properties of determinants, show that:
R1 → R1 - R2 (i.e. Replacing 1st row by subtraction of 1st and 2nd row)
R2 → R2 - R3 (i.e. Replacing 2nd row by subtraction of 2nd and 3rd row)
Since we know a2 - b2 = (a + b)(a - b)
Therefore taking (a - b) and (b - c) outside the determinant from 1st and 2nd row respectively
R1 → R1 - R2 (i.e. Replacing 1st row by subtraction of 1st and 2nd row)
Expanding the determinant along 1st column
∴
∴LHS = (a - b)(b - c)(0 - (a - c))
∴LHS = (a - b)(b - c)(c - a) = RHS
∴
By using properties of determinants, show that:
C1 → C1 - C2 (i.e. Replacing 1st column by subtraction of 1st and 2nd column)
C2 → C2 - C3 (i.e. Replacing 2nd column by subtraction of 2nd and 3rd column)
We have a3 - b3 = (a - b)(a2 + ab + b2) and b3 - c3 = (b - c)(b2 + bc + c2)
Therefore taking (a - b) and (b - c) outside the determinant from 1st and 2nd column respectively
C1 → C1 - C2 (i.e. Replacing 1st column by subtraction of 1st and 2nd column)
As a2 - c2) = (a + c)(a - c) therefore taking (a - c) outside the determinant from 1st column we get
.
Expanding the determinant along 1st row
∴ LHS = (a - b)(b - c)(a - c)( - (a + b + c))
Adjusting the minus sign with (a - c)
∴LHS = (a - b)(b - c)(c - a)(a + b + c) = RHS
∴
By using properties of determinants, show that:
R1 → R1 - R2 (i.e. Replacing 1st row by subtraction of 1st and 2nd row)
R2 → R2 - R3 (i.e. Replacing 2nd row by subtraction of 2nd and 3rd row)
We know x2 - y2 = (x + y)(x - y) and y2 - z2 = (y + z)(y - z)
Therefore taking (x - y) and (y - z) outside the determinant from 1st and 2nd row respectively
R1 → R1 - R2 (i.e. Replacing 1st row by subtraction of 1st and 2nd row)
Taking (x - z) outside determinant from 1st row
C2 → C2 - C3 (i.e. Replacing 2nd column by subtraction of 2nd and 3rd column)
Expanding the determinant along 1st row
∴ LHS = (x - y)(y - z)(x - z)(z2 - xy - xz - yz - z2)
Cancelling z2 and adjusting the negative sign with (x - z)
∴LHS = (x - y)(y - z)(z - x)(xy + yz + zx) = RHS
∴
By using properties of determinants, show that:
R1 → R1 + R2 + R3 (i.e. replace 1st row by addition of 1st, 2nd and 3rd row)
Taking 5x + 4 outside the determinant from 1st row
C2 → C2 - C1 (i.e. replace 2nd column by subtraction of 2nd and 1st column)
C3 → C3 - C1 (i.e. replace 3rd column by subtraction of 3rd and 1st column)
Expanding the determinant along 1st row
∴ LHS = (5x - 4) (4 - x)2 = RHS
∴
By using properties of determinants, show that:
R1 → R1 + R2 + R3 (i.e. replace 1st row by addition of 1st, 2nd and 3rd row)
Taking 3y + k outside the determinant from 1st row
C2 → C2 - C1 (i.e. replace 2nd column by subtraction of 2nd and 1st column)
C3 → C3 - C1 (i.e. replace 3rd column by subtraction of 3rd and 1st column)
Expanding the determinant along 1st row
∴ LHS = (3y + k) k2 = RHS
∴
By using properties of determinants, show that:
R1 → R1 + R2 + R3 (i.e. replace 1st row by addition of 1st, 2nd and 3rd row)
Taking a + b + c outside the determinant from 1st row
C2 → C2 - C1 (i.e. replace 2nd column by subtraction of 2nd and 1st column)
C3 → C3 - C1 (i.e. replace 3rd column by subtraction of 3rd and 1st column)
Expanding the determinant along 1st row
∴ LHS = (a + b + c)3 = RHS
∴
By using properties of determinants, show that:
C1 → C1 + C2 + C3 (i.e. replace 1st column by addition of 1st, 2nd and 3rd column)
Taking 2(x + y + z) outside the determinant from 1st column
R2 → R2 - R1 (i.e. replace 2nd row by subtraction of 2nd and 1st row)
R3 → R3 - R1 (i.e. replace 3rd row by subtraction of 3rd and 1st row)
Expanding the determinant along 1st column
∴ LHS = 2(x + y + z)3 = RHS
∴
By using properties of determinants, show that:
R1 → R1 - xR2 (i.e. replace 1st row by subtraction of 1st row and ‘x’ times 2nd row)
Taking (1 - x3) outside the determinant from 1st row
.
Expanding the determinant along 1st row
HS = (1 - x3)2 = RHS
∴
By using properties of determinants, show that:
R1 → R1 + bR3 (i.e. replace 1st row by addition of 1st row and b times 3rd row)
R2 → R2 - aR3 (i.e. replace 2nd row by subtraction of 2nd row and a times 3rd row)
.
Taking both (1 + a2 + b2) outside the determinant from 1st and 2nd row
Expanding the determinant along 1st row
∴ LHS = (1 + a2 + b2)2 [1 + a2 - b2 - (-2b2)]
∴ LHS = (1 + a2 + b2 )2 (1 + a2 + b2)
∴ LHS = (1 + a2 + b2 )3 = RHS
By using properties of determinants, show that:
Taking out a, b and c from the determinant from 1st, 2nd and 3rd row respectively.
R2 → R2 - R1 (i.e. replace 2nd row by subtraction of 2nd and 1st row)
R3 → R3 - R1 (i.e. replace 3rd row by subtraction of 3rd and 1st row)
.
C1 → aC1 (i.e. replace 1st column by ‘a’ times 1st column)
C2 → bC2 (i.e. replace 2nd column by ‘b’ times 2nd column)
→ cC3 (i.e. replace 3rd column by ‘c’ times 3rd column)
As we are multiplying by a, b and c we should also divide by a, b and c to keep the original determinant value unchanged.
.
C1 → C1 + C2 + C3 (i.e. replace 1st column by addition of 1st, 2nd and 3rd column)
Expanding determinant along 1st column
∴ LHS = (1 + a2 + b2 + c2) = RHS
∴
Let A be a square matrix of order 3 × 3, then | kA| is equal to
A. k|A|
B. k2 |A|
C. k3 |A|
D. 3k |A|
Let A be any 3×3 matrix
∴
Taking out k from the determinant from 1st, 2nd and 3rd row
∴
∴ |kA| = k3|A|
Therefore, answer is option (C) k3|A|
Which of the following is correct
A. Determinant is a square matrix.
B. Determinant is a number associated to a matrix.
C. Determinant is a number associated to a square matrix.
D. None of these
Determinant is an operation which we perform on arranged numbers. A square matrix is set of arranged numbers. We perform some operations on a matrix and we get a value that value is called as determinant of that matrix hence determinant is a number associated to square matrix.
Find area of the triangle with vertices at the point given in each of the following:
(1, 0), (6, 0), (4, 3)
Given vertices of the triangle are (1, 0), (6, 0), (4, 3)
Let the vertices of the triangle be given by (x1, y1), (x2, y2), (x3, y3)
Area of triangle is given by Δ =
Area of triangle = Δ =
Expanding the determinant along Row 1
Δ = 1/2 × [1 × (0 × 1 – 3 × 1) – 0 × (6 × 1 – 4 × 1) + 1 × (6 × 3 – 4 × 0)]
Δ = 1/2 × [1 × (0 – 3) – 0 + 1 × (18 – 0)]
Δ = 1/2 × (-3 + 18) = 15/2 sq. units
∴ Δ = 15/2 sq. units = 7.5 sq. units
Find area of the triangle with vertices at the point given in each of the following:
(2, 7), (1, 1), (10, 8)
Given vertices of the triangle are (2, 7), (1, 1), (10, 8)
Let the vertices of the triangle be given by (x1, y1), (x2, y2), (x3, y3)
Area of triangle is given by Δ =
Area of triangle = Δ =
Expanding the determinant along Row 1
Δ = 1/2 × [2 × (1 × 1 – 8 × 1) – 7 × (1 × 1 – 10 × 1) + 1 × (8 × 1 – 1 × 10)]
Δ = 1/2 × [2 × (1 – 8) – 7 × (1 – 10) + 1 × (8 – 10)]
Δ = 1/2 × [2 × (-7) – 7 × (-9) + 1 × (-2)] = 1/2 × (-14 + 63 – 2) sq. units
Δ = 1/2 × 47 sq. units = 47/2 sq. units
∴ Δ = 47/2 sq. units = 23.5 sq. units
Find area of the triangle with vertices at the point given in each of the following:
(–2, –3), (3, 2), (–1, –8)
Given vertices of the triangle are (–2, –3), (3, 2), (–1, –8)
Let the vertices of the triangle be given by (x1, y1), (x2, y2), (x3, y3)
Area of triangle is given by Δ =
Area of triangle = Δ =
Expanding the determinant along Row 1
Δ = 1/2 × |[(-2) × (2 × 1 – (-8) × 1) – (-3) × (3 × 1 – (-1) × 1) + 1 × (3 × (-8) – 2 × (-1))]|
Δ = 1/2 × |[(-2) × (2 + 8) + 3 × (3 + 1) + 1 × (-24 + 2)]|
Δ = 1/2 × |[(-2) × 10 + 3 × 4 + 1 × (-22)]| = 1/2 ×| (-20 + 12 – 22)| sq. units
Δ = 1/2 × |-30| sq. units = 30/2 sq. units
∴ Δ = 30/2 sq. units = 15 sq. units
Show that points
A (a, b + c), B (b, c + a), C (c, a + b) are collinear.
Given vertices of the triangle are A (a, b + c), B (b, c + a), C (c, a + b)
Let the vertices of the triangle be given by (x1, y1), (x2, y2), (x3, y3)
Area of triangle is given by Δ =
For points to be collinear area of triangle = Δ = 0
So, we have to show that area of triangle formed by ABC is 0
Area of triangle = Δ =
Expanding the determinant along Row 1
Δ = 1/2 × [a × {(c + a) × 1 – (a + b) × 1) – (b + c) × {b × 1 – c × 1} + 1 × {b × (a + b) – c × (c +a)}]
Δ = 1/2 × [a × (c + a – a – b) – (b + c) × (b – c) + 1 × (ab + b2 – c2 - ca)]
Δ = 1/2 × [a × (c – b) – (b2 – c2) + 1 × (ab + b2 – c2 – ca)]
Δ = 1/2 × (ac – ab – b2 + c2 + ab + b2 – c2 – ca) sq units
Δ = 1/2 × 0 sq units
∴ Δ = 0
∴ Given vertices of the triangle are A (a, b + c), B (b, c + a), C (c, a + b) are collinear
Find values of k if area of triangle is 4 sq. units and vertices are
(k, 0), (4, 0), (0, 2)
Given vertices of the triangle are (k, 0), (4, 0), (0, 2)
Let the vertices of the triangle be given by (x1, y1), (x2, y2), (x3, y3)
Area of triangle is given by Δ =
Given, Area of triangle = Δ = 4 sq. units
= 4
⇒ 4 = 1/2 |[k × (0 × 1 – 2 × 1) – 0 × (4 × 1 – 0 × 1) + 1 × (4 × 2 – 0 × 0)]|
⇒ 4 = 1/2 × |[k × (0 – 2) – 0 + 1 × (8 – 0)]|
⇒ � 4 × 2 = -2k + 8
⇒ 8 = -2k + 8 and -8 = -2k + 8
⇒ 8 – 8 = -2k and 8 + 8 = 2k
⇒ 2k = 0 and 16 = 2k
⇒ k = 0 and k = 8
Find values of k if area of triangle is 4 sq. units and vertices are
(–2, 0), (0, 4), (0, k)
Given vertices of the triangle are (–2, 0), (0, 4), (0, k)
Let the vertices of the triangle be given by (x1, y1), (x2, y2), (x3, y3)
Area of triangle is given by Δ =
Given, Area of triangle = Δ = 4 sq. units
= 4
⇒ 4 = 1/2 |[(-2) × (4 × 1 – k × 1) – 0 × (0 × 1 – 0 × 1) + 1 × (0 × k – 0 × 4)]|
⇒ 4 = 1/2 × |[(-2) × (4 – k) – 0 + 1 × (0 – 0)]|
⇒ 4 × 2 = |(-8 + 2k)
⇒ � 8= 2k - 8
⇒ 8 = 2k - 8 and ⇒ -8 = 2k - 8
⇒ 8 + 8 = 2k and ⇒ 8 - 8 = 2k
⇒ k = 16/2 and ⇒ k = 0/2
⇒ k = 8 and ⇒ k = 0
Find equation of line joining (1, 2) and (3, 6) using determinants.
Equation of line joining points (x1, y1) & (x2, y2) is given by = 0
Given points are (1, 2) and (3, 6)
Equation of line is given by = 0
⇒ 1/2 × [1 × (6 × 1 – y × 1) – 2 × (3 × 1 – x × 1) + 1 × (3 × y – x × 6)] = 0
⇒ [(6 – y) – 2 × (3 – x) + (3y – 6x)] = 0 × 2
⇒ (6 – y – 6 + 2x + 3y – 6x) = 0
⇒ 2y – 4x = 0
⇒ y – 2x = 0 ⇒ y = 2x
∴ Required Equation of line is y = 2x
Find equation of line joining (3, 1) and (9, 3) using determinants.
Equation of line joining points (x1, y1) & (x2, y2) is given by = 0
Given points are (3, 1) and (9, 3)
Equation of line is given by = 0
⇒ 1/2 × [3 × (3 × 1 – y × 1) – 1 × (9 × 1 – x × 1) + 1 × (9 × y – x × 3)] = 0
⇒ [3 × (3 – y) – 1 × (9 – x) + (9y – 3x)] = 0 × 2
⇒ (9 – 3y – 9 + x + 9y – 3x) = 0
⇒ 6y – 2x = 0
⇒ 2x – 6y = 0 ⇒ x – 3y = 0
∴ Required Equation of line is x – 3y = 0
If area of triangle is 35 sq units with vertices (2, –6), (5, 4) and (k, 4). Then k is
A. 12
B. –2
C. –12, –2
D. 12, –2
Given vertices of the triangle are (2, – 6), (5, 4) and (k, 4).
Let the vertices of the triangle be given by (x1, y1), (x2, y2), (x3, y3)
Area of triangle is given by Δ =
Given, Area of triangle = Δ = 35 sq. units
= 35
⇒ � 35 = 1/2 × [2 × (4 × 1 – 4 × 1) – (-6) × (5 × 1 – k × 1) + 1 × (5 × 4 – k × 4)]
⇒ � 35 = 1/2 × [2 × (4 – 4) + 6 × (5 – k) + 1 × (20 – 4k)]
⇒ � 35 × 2 = (2 × 0 + 30 – 6k + 20 – 4k)
⇒ � 70 = 30 – 6k + 20 – 4k
⇒ � 70 = 50 – 10k
⇒ 70 – 50 = -10k and ⇒ -70 – 50 = -10k
⇒ 20 = -10k and ⇒ -120 = -10k
⇒ k = -20/10 and ⇒ k = 120/10
⇒ k = -2 and ⇒ k = 12
Write Minors and Cofactors of the elements of following determinants:
Minor: Minor of an element aij of a determinant is the determinant obtained by removing ith row and jth column in which element aij lies. It is denoted by Mij.
Cofactor: Cofactor of an element aij, Aij = (-1)i+j Mij.
Minor of element aij = Mij
a11 = 2, Minor of element a11 = M11 = 3
Here removing 1st row and 1st column from the determinant we are left out with 3 so M11 = 3.
Similarly, finding other Minors of the determinant
a12 = -4, Minor of element a12 = M12 = 0
a21 = 0, Minor of element a21 = M21 = -4
a22 = 3, Minor of element a22 = M22 = 2
Cofactor of an element aij, Aij = (-1)i+j × Mij
A11 = (-1)1+1 × M11 = 1 × 3 = 3
A12 = (-1)1+2 × M12 = (-1) × 0 = 0
A21 = (-1)2+1 × M11 = (-1) × (-4) = 4
A22 = (-1)2+2 × M22 = 1 × 2 = 2
Write Minors and Cofactors of the elements of following determinants:
Minor of an element aij = Mij
a11 = a, Minor of element a11 = M11 = d
Here removing 1st row and 1st column from the determinant we are left out with d so M11 = d.
Similarly, finding other Minors of the determinant
a12 = c, Minor of element a12 = M12 = b
a21 = b, Minor of element a21 = M21 = c
a22 = d, Minor of element a22 = M22 = a
Cofactor of an element aij, Aij = (-1)i+j × Mij
A11 = (-1)1+1 × M11 = 1 × d = d
A12 = (-1)1+2 × M12 = (-1) × b = -b
A21 = (-1)2+1 × M11 = (-1) × c = -c
A22 = (-1)2+2 × M22 = 1 × a = a
Write Minors and Cofactors of the elements of following determinants:
Minor of an element aij = Mij
a11 = 1, Minor of element a11 = M11 = = (1 × 1) – (0 × 0) = 1
Here removing 1st row and 1st column from the determinant we are left out with the determinant. Solving this we get M11 = 1
Similarly, finding other Minors of the determinant
a12 = 0, Minor of element a12 = M12 = = (0 × 1) – (0 × 0) = 0
a13 = 0, Minor of element a13 = M13 = = (0 × 0) - (1 × 0) = 0
a21 = 0, Minor of element a21 = M21 = = (0 × 1) – (0 × 0) = 0
a22 = 1, Minor of element a22 = M22 = = (1 × 1) – (0 × 0) = 1
a23 = 0, Minor of element a23 = M23 = = (1 × 0) – (0 × 0) = 0
a31 = 0, Minor of element a31 = M31 = = (0 × 0) – (0 × 1) = 0
a32 = 0, Minor of element a32 = M32 = = (1 × 0) – (0 × 0) = 0
a33 = 1, Minor of element a33 = M33 = = (1 × 1) – (0 × 0) = 1
Cofactor of an element aij, Aij = (-1)i+j × Mij
A11 = (-1)1+1 × M11 = 1 × 1 = 1
A12 = (-1)1+2 × M12 = (-1) × 0 = 0
A13 = (-1)1+3 × M13 = 1 × 0 = 0
A21 = (-1)2+1 × M21 = (-1) × 0 = 0
A22 = (-1)2+2 × M22 = 1 × 1 = 1
A23 = (-1)2+3 × M23 = (-1) × 0 = 0
A31 = (-1)3+1 × M31 = 1 × 0 = 0
A32 = (-1)3+2 × M32 = (-1) × 0 = 0
A33 = (-1)3+3 × M33 = 1 × 1 = 1
Write Minors and Cofactors of the elements of following determinants:
Minor of an element aij = Mij
a11 = 1, Minor of element a11 = M11 = = (5 × 2) – ((-1) × 1) = 10 + 1 = 11
Here removing 1st row and 1st column from the determinant we are left out with the determinant. Solving this we get M11 = 11
Similarly, finding other Minors of the determinant
a12 = 0, Minor of element a12 = M12 = = (3 × 2) – ((-1) × 0) = (6 - 0) = 6
a13 = 4, Minor of element a13 = M13 = = (3 × 1) – (5 × 0) = 3 - 0 = 3
a21 = 3, Minor of element a21 = M21 = = (0 × 2) – (4 × 1) = 0 – 4 = -4
a22 = 5, Minor of element a22 = M22 = = (1 × 2) – (4 × 0) = 2 – 0 = 2
a23 = -1, Minor of element a23 = M23 = = (1 × 1) – (0 × 0) = 1
a31 = 0, Minor of element a31 = M31 = = (0 × (-1)) – (4 × 5) = 0 – 20 = -20
a32 = 1, Minor of element a32 = M32 = = (1 × (-1)) – (4 × 3) = -1 – 12 = -13
a33 = 2, Minor of element a33 = M33 = = (1 × 5) – (0 × 3) = (5 – 0) = 5
Cofactor of an element aij, Aij = (-1)i+j × Mij
A11 = (-1)1+1 × M11 = 1 × 11 = 11
A12 = (-1)1+2 × M12 = (-1) × 6 = -6
A13 = (-1)1+3 × M13 = 1 × 3 = 3
A21 = (-1)2+1 × M21 = (-1) × (-4) = 4
A22 = (-1)2+2 × M22 = 1 × 2 = 2
A23 = (-1)2+3 × M23 = (-1) × 1 = -1
A31 = (-1)3+1 × M31 = 1 × (-20) = -20
A32 = (-1)3+2 × M32 = (-1) × (-13) = 13
A33 = (-1)3+3 × M33 = 1 × 5 = 5
Using Cofactors of elements of second row, evaluate
To evaluate a determinant using cofactors, Let
B =
Expanding along Row 1
B =
B = a11 A11 + a12 A12 + a13 A13
[Where Aij represents cofactors of aij of determinant B.]
B = Sum of product of elements of R1 with their corresponding cofactors
Similarly, the determinant can be solved by expanding along column
So, B = sum of product of elements of any row or column with their corresponding cofactors
Cofactors of second row
A21 = (-1)2+1 × M21 = (-1) × = (-1) × (3 × 3 – 8 × 2) = (-1) × (-7) = 7
A22 = (-1)2+2 × M22 = 1 × = (5 × 3 – 8 × 1) = 7
A23 = (-1)2+3 × M23 = (-1) × = (-1) × (5 × 2 – 3 × 1) = (-1) × 7 = -7
[Where Aij = (-1)i+j × Mij, Mij = Minor of ith row & jth column]
Therefore,
Δ = a21A21 + a22A22 + a23A23
Δ = 2 × 7 + 1 × (-7) = 14 - 7 = 7
Ans: Δ = 7
Using Cofactors of elements of third column, evaluate.
To evaluate a determinant using cofactors, Let
B =
Expanding along Row 1
B =
B = a11 A11 + a12 A12 + a13 A13
[Where Aij represents cofactors of aij of determinant B.]
B = Sum of product of elements of R1 with their corresponding cofactors
Similarly, the determinant can be solved by expanding along column
So, B = sum of product of elements of any row or column with their corresponding cofactors
Cofactors of third column
A13 = (-1)1+3 × M13 = 1 × = 1 × (1 × z – 1 × y) = (z – y)
A23 = (-1)2+3 × M23 = (-1) × = (-1) × (1 × z – 1 × x) = - (z - x) = (x - z)
A33 = (-1)3+3 × M33 = 1 × = 1 × (1 × y – 1 × x) = (y – x)
[Where Aij = (-1)i+j × Mij, Mij = Minor of ith row & jth column]
Therefore,
Δ = a13A13 + a23A23 + a33A33
Δ = yz (z - y) + zx (x - z) + xy (y - x) = z [y (z - y) + x (x - z)] + xy (y - x)
Δ = z (yz - y2 + x2 - xz) + xy (y - x) = z [(yz - xz) + (x2 - y2)] + xy (y - x)
Δ = z [z × (y - x) + (x + y) × (x - y)] + xy (y - x)
Δ = z × (y - x) × (z – x - y) + xy (y - x)
Δ = (y - x) × (z2 – xz – yz + xy)
Δ = (y - x) × [z (z - x) – y (z - x)] = (y - x) × (z - y) × (z - x)
Δ = (x - y) (y - z) (z - x)
Ans: Δ = (x - y) (y - z) (z - x)
If and Aij is Cofactors of aij, then value of Δ is given by
A. a11 A31+ a12 A32 + a13 A33
B. a11 A11+ a12 A21 + a13 A31
C. a21 A11+ a22 A12 + a23 A13
D. a11 A11+ a21 A21 + a31 A31
Δ =
Expanding along Column 1
Δ =
Δ = a11A11 + a21A21 + a31A31
Find adjoint of each of the matrices.
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = 4, A12 = -3, A21 = -2, A22 = 1.
∴ Adj A =
=
Find adjoint of each of the matrices.
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = 1{(3×1-0×5)} = 3
Similarly,
A12 = -12, A13 = 6, A21 = 1, A22 = 5, A23 = 2, A31 = -11, A32 = -1, A33 = 5.
Verify A (adj A) = (adj A) A = |A|
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = -6, A12 = 4, A21 = -3, A22 = 2.
∴ Adj A =
=
So LHS = A(AdjA) =
Also AdjA(A) =
Determinant of A = |A| = 2(-6)-(3)(-4) = 0
So RHS = |A|I = 0
Hence A(AdjA) = AdjA(A) = |A|I = 0 {hence proved}
Verify A (adj A) = (adj A) A = |A|
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = 0, A12 = -11, A13 = 0, A21 = 3, A22 = 1, A23 = -1, A31 = 2, A32 = 8, A33 = 3.
∴ Adj A =
=
So, LHS = A(AdjA) =
Also AdjA(A) =
Determinant of A = |A| = 11
So RHS = |A|I = .
Hence A(AdjA) = AdjA(A) = |A|I = {hence proved}
Find the inverse of each of the matrices (if it exists)
We know that
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = 3, A12 = -4, A21 = 2, A22 = 2.
∴ Adj A =
=
And |A| = 2(3)-(-2)(4) = 14
So .
Find the inverse of each of the matrices (if it exists)
We know that
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = 2, A12 = 3, A21 = -5, A22 = -1.
∴ Adj A =
=
And |A| = -1(2)-(-3)(5) = 13
So
Find the inverse of each of the matrices (if it exists)
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = 10, A12 = 0, A13 = 0, A21 = -10, A22 = 5, A23 = 0, A31 = 2, A32 = -4, A33 = 2.
∴ Adj A =
And |A| = 10.
Find the inverse of each of the matrices (if it exists)
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = -3, A12 = 3, A13 = -9, A21 = 0, A22 = -1, A23 = -2, A31 = 0, A32 = 0, A33 = 3.
∴ Adj A =
And |A| = -3.
.
Find the inverse of each of the matrices (if it exists)
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = -1, A12 = -4, A13 = 1, A21 = 5, A22 = 23, A23 = -11, A31 = 3, A32 = 12, A33 = -6.
∴ Adj A =
= .
And |A| = -3.
.
Find the inverse of each of the matrices (if it exists)
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = 2, A12 = -9, A13 = -6, A21 = 0, A22 = -2, A23 = -1, A31 = -1, A32 = 3, A33 = 2.
∴ Adj A =
=
And |A| = -1.
Find the inverse of each of the matrices (if it exists)
Adjoint of the matrix A = [aij]n×n is defined as the transpose of the matrix [Aij]n×n where Aij is the co-factor of the element aij.
Let’s find the cofactors for all the positions first-
Here, A11 = -1, A12 = 0, A13 = 0, A21 = 0, A22 = -cosα, A23 = -sinα, A31 = 0, A32 = -sinα, A33 = cosα.
∴ Adj A =
And |A| = 1.
.
Let . Verify that (AB)–1 = B–1 A–1.
We have AB = = (61)(67)-(47)(87) = -2
Here determinant of matrix = |AB|≠ 0 hence (AB)-1 exists.
Also |A| = 1 ≠ 0 and |B| = -2 ≠ 0.
∴ A-1 and B-1 will also exist and are given by-
And hence,
{Hence proved}
If , show that A2 – 5A + 7I = O. Hence find A–1.
We have A2 = A.A = .
So A2 – 5A + 7I =
Hence A2 – 5A + 7I = 0
∴ A.A – 5A = -7I
Now post multiply with A-1
So A.A.A-1-5A.A-1 = -7I.A-1
→ A.I – 5I = -7I.A-1 {since A.A-1 = I}
A – 5I = -7A-1 {since X.I = X}
For the matrix , find the numbers a and b such that A2 + aA + bI = O.
We have A2 = A.A =
Since A2 + aA + bI =
So A2 + aA + bI =
Hence 10+3a+b = 0 …(i)
5+a = 0 …(ii)
5+2a+b = 0 …(iii)
From (ii) a = -5
Putting a in (iii) we get b = 5
So a = -5 and b = 5 satisfy the equation.
For the matrix
Show that A3– 6A2 + 5A + 11 I = O. Hence, find A–1.
Here A2 = A.A =
And hence A3 = A. A2 =
∴ A3– 6A2 + 5A + 11 I =
Thus, A3– 6A2 + 5A + 11 I = 0
Now, A3– 6A2 + 5A + 11 I = 0,
→ (A.A.A)- 6 (A.A) +5A = -11I
Post-multiply with A-1 on both sides-
→ (A.A.A.A-1)- 6 (A.A.A-1) +5A.A-1 = -11I. A-1
→ (A.A.I) – 6(A.I) + 5I = -11I. A-1 {since A.A-1 = I}
→ (A.A) – 6A +5I = -11A-1 {since X.I = X}
Hence
If Verify that A3 – 6A2 + 9A – 4I = O and hence find A-1.
Here A2 = A.A =
And hence A3 = A. A2 =
∴ A3– 6A2 + 9A -4I
Thus, A3– 6A2 + 9A -4I = 0
Now, A3– 6A2 + 9A -4I = 0,
→ (A.A.A)- 6 (A.A) +9A = 4I
Post-multiply with A-1 on both sides-
→ (A.A.A.A-1)- 6 (A.A.A-1) +9A.A-1 = 4I. A-1
→ (A.A.I) – 6(A.I) + 9I = 4I. A-1 {since A.A-1 = I}
→ (A.A) – 6A +9I = 4A-1 {since X.I = X}
Hence
Let A be a non-singular square matrix of order 3 × 3. Then |adj A| is equal to
A. |A |
B. | A|2
C. | A|3
D. 3|A|
For a square matrix of order n×n,
We know that |AdjA| = |A|n-1
So, |AdjA| = |A|(3-1) = |A|2
If A is an invertible matrix of order 2, then det (A–1) is equal to
A. det (A)
B.
C. 1
D. 0
{since adj(A) is of order n and |Adj(A)| = |A|n-1}
Alternative-
We know that AA-1 = I
So |A||A-1| = |I| = 1
Examine the consistency of the system of equations.
x + 2y = 2
2x + 3y = 3
The given system of equations is:
x + 2y = 2
2x + 3y = 3
The given system of equations can be written in the form of AX = B, where
Now |A| = 3(1)-2(2) = -1 ≠ 0.
∴ A is a non-singular matrix and hence A-1 exists.
So, The system of equations will be consistent.
Examine the consistency of the system of equations.
2x – y = 5
x + y = 4
The given system of equations is:
2x - y = 5
x + y = 4
The given system of equations can be written in the form of AX = B, where
Now |A| = 2(1)-1(-1) = 3 ≠ 0.
∴ A is a non-singular matrix and hence A-1 exists.
So, The system of equations will be consistent.
Examine the consistency of the system of equations.
x + 3y = 5
2x + 6y = 8
The given system of equations is:
x + 3y = 5
2x + 6y = 8
The given system of equations can be written in the form of AX = B, where
Now |A| = 1(6)-3(2) = 0
∴ A is a singular matrix and hence A-1 doesn’t exist.
So, The system of equations will be inconsistent.
Examine the consistency of the system of equations.
x + y + z = 1
2x + 3y + 2z = 2
ax + ay + 2az = 4
The given system of equations is:
x + y + z = 1
2x + 3y + 2z = 2
ax + ay + 2az = 4
The given system of equations can be written in the form of AX = B, where
Now |A| = 1(6a-2a)-1(4a-2a) +1(2a-3a) = a ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
So the system of equations will be consistent.
Examine the consistency of the system of equations.
3x–y – 2z = 2
2y – z = –1
3x – 5y = 3
The given system of equations is:
3x - y - 2z = 2
0x + 2y - z = -1
3x - 5y +0z = 3
The given system of equations can be written in the form of AX = B, where
Now |A| = 3(-5) + 3(5) = 0
∴ A is a singular matrix and hence A-1 doesn’t exist.
So the system of equations will be inconsistent.
Examine the consistency of the system of equations.
5x – y + 4z = 5
2x + 3y + 5z = 2
5x – 2y + 6z = –1
The given system of equations is:
5x - y + 4z = 5
2x + 3y + 5z = 2
5x - 2y + 6z = -1
The given system of equations can be written in the form of AX = B, where
Now |A| = 5(18+10) + 1(12-25) + 4(-4-15) = 51 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
So, The system of equations will be consistent.
Solve system of linear equations, using matrix method.
5x + 2y = 4
7x + 3y = 5
The given system of equations is:
5x + 2y = 4
7x + 3y = 5
The given system of equations can be written in the form of AX = B, where
Now |A| = 3(5)-2(7) = 1 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
Now
So
And
So
Hence x = 2 and y = -3.
Solve system of linear equations, using matrix method.
2x – y = –2
3x + 4y = 3
The given system of equations is:
2x - y = -2
3x + 4y = 3
The given system of equations can be written in the form of AX = B, where
Now |A| = 2(4)-3(-1) = 11 ≠ 0.
∴ A is a non-singular matrix and hence A-1 exists.
Now
So
And
So
Hence
Solve system of linear equations, using matrix method.
4x – 3y = 3
3x – 5y = 7
The given system of equations is:
4x - 3y = 3
3x - 5y = 7
The given system of equations can be written in the form of AX = B, where
Now |A| = 4(-5)-3(-3) = -11 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
Now
AdjA =
So
And
So
Hence
Solve system of linear equations, using matrix method.
5x + 2y = 3
3x + 2y = 5
The given system of equations is:
5x + 2y = 3
3x + 2y = 5
The given system of equations can be written in the form of AX = B, where
Now |A| = 5(2)-2(3) = 4 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
Now
AdjA =
So
And
So
Hence x = -1 and y = 4
Solve system of linear equations, using matrix method.
The given system of equations is:
2x + y + z = 1
0x + 3y - 5z = 9
The given system of equations can be written in the form of AX = B, where
Now |A| = 2(10+3)-1(-5) +1(3) = 34 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
Now A11 = 13, A12 = 5, A13 = 3, A21 = 8, A22 = -10, A23 = -6, A31 = 1, A32 = 3, A33 = -5
So AdjA =
∴
And hence X = A-1B
So
Hence
Solve system of linear equations, using matrix method.
x – y + z = 4
2x + y – 3z = 0
x + y + z = 2
The given system of equations is:
x - y + z = 4
2x + y -3z = 0
x + y + z = 2
The given system of equations can be written in the form of AX = B, where
Now |A| = 1(1+3) + 1(5) + 1(1) = 10 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
Now A11 = 4, A12 = -5, A13 = 1, A21 = 2, A22 = 0, A23 = -2, A31 = 1, A32 = -2, A33 = 3
So AdjA =
And hence X = A-1B
So
Hence x = 2, y = -1 and z = 1.
Solve system of linear equations, using matrix method.
2x + 3y +3 z = 5
x – 2y + z = –4
3x – y – 2z = 3
The given system of equations is:
2x + 3y + 3z = 5
x – 2y + z = -4
3x - y - 2z = 3
The given system of equations can be written in the form of AX = B, where
Now |A| = 2(4+1)-3(-5) +3(5) = 40 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
Now A11 = 5, A12 = 5, A13 = 5, A21 = 3, A22 = -13, A23 = 11, A31 = 9, A32 = 1, A33 = -7
So AdjA =
∴
And hence X = A-1B
So
Hence x = 1, y = 2 and z = -1
Solve system of linear equations, using matrix method.
x – y + 2z = 7
3x + 4y – 5z = –5
2x – y + 3z = 12
The given system of equations is:
x - y + 2z = 7
3x + 4y - 5z = -5
2x – y + 3z = 12
The given system of equations can be written in the form of AX = B, where
Now |A| = 1(12-5) + 1(9+10) + 2(-3-8) = 4 ≠ 0
∴ A is a non-singular matrix and hence A-1 exists.
Now A11 = 7, A12 = -19, A13 = -11, A21 = 1, A22 = -1, A23 = -1, A31 = -3, A32 = 11, A33 = 7
So AdjA =
∴
So
Hence
If , find A–1. Using A–1 solve the system of equations
2x – 3y + 5z = 11
3x + 2y – 4z = – 5
x + y – 2z = – 3
∴ |A| = 2(-4+4) + 3(-6+4) + 5(3-2)
= -1 ≠ 0
So inverse of A exists,
Now A11 = 0, A12 = 2, A13 = 1, A21 = -1, A22 = -9, A23 = -5, A31 = 2, A32 = 23, A33 = 13
So AdjA =
∴
So .
Hence x = 1, y = 2 and z = 3
The cost of 4 kg onion, 3 kg wheat and 2 kg rice is Rs. 60. The cost of 2 kg onion, 4 kg wheat and 6 kg rice is Rs. 90. The cost of 6 kg onion 2 kg wheat and 3 kg rice is Rs. 70. Find cost of each item per kg by matrix method.
Let the cost of onions, wheat and rice per kg be Rs. x, Rs. y and Rs. z respectively.
Then according to the giving situation, it can be represented by a system of equations as-
4x + 3y + 2z = 60
2x + 4y + 6z = 90
6x + 2y + 3z = 70
The system of equations can be written in the form of AX = B, where
Here |A| = 4(12-12)-3(6-36) + 2(4-24) = 50 ≠ 0
Hence A-1 will exist.
Now, A11 = 0, A12 = 30, A13 = -20, A21 = -5, A22 = 0, A23 = 10, A31 = 10, A32 = -20, A33 = 10
Hence x = 5, y = 8 & z = 8.
Hence, the cost of onions is Rs. 5 per kg, the cost of wheat is Rs. 8 per kg, and the cost of rice is Rs. 8 per kg
Prove that the determinant is independent of θ.
Expanding the above determinant along R1 i.e. Row 1
Δ = x (-x2 – 1) – sinθ (-x × sinθ – cosθ × 1) + cosθ (-sinθ × 1 + x cosθ)
Δ = -x3 – x + xsin2θ + sinθ × cos θ – sinθ × cosθ + x cos2θ
Δ = -x3 – x + x (sin2θ + cos2θ)
Since, sin2θ + cos2θ = 1
∴ Δ = -x3 – x + x
Hence Δ is independent of θ.
Without expanding the determinant, prove that
Multiplying row 1 with a, row 2 with b and row 3 with c
[R1→ aR1, R2→ bR2, R3→ cR3]
[Taking common abc from C3]
[Applying C1↔ C3]
= [Applying C2 ↔ C3]
= RHS
Since, LHS = RHS
∴ the given result is proved.
Evaluate
Let Δ =
Expanding along C3 we get
Δ = -sinα (-sinβ × sinα sinβ – cosβ × sinα cosβ) – 0 (sinα cosβ × cosα sinβ – cosα cosβ × sinα sinβ) + cosα (cosα cosβ × cosβ – cosα sinβ × (-sinβ))
Δ = -sinα (-sinα sin2β – sinα cos2β) – 0 + cosα (cosα cos2β + cosα sin2β)
Δ = sinα × sinα (sin2β + cos2β) + cosα × cosα (cos2β + sin2β)
[Taking –sinα and cosα common]
Since, sin2β + cos2β = 1
∴ Δ = sin2α + cos2α
Δ = 1 [sin2α + cos2α = 1]
If a, b and c are real numbers, and
Show that either a + b + c = 0 or a = b = c.
Given, Δ =
Applying Elementary transformations, we get
R1→ R1 + R2 + R3 we have,
Δ = 2 (a + b + c)
Now applying C2→ C2 – C1 and C3→ C3 – C1
Δ = 2 (a + b + c)
Expanding along R1
Δ = 2 (a + b + c) [1 {(b – c)(c – b) – (b – a)(c – a)} – 0 + 0]
Δ = 2 (a + b + c) [- b2 – c2 + 2bc – bc + ba + ac – a2]
Δ = 2 (a + b + c) [ab + bc + ca – a2 – b2 – c2]
Given that Δ = 0
∴ 2 (a + b + c) [ab + bc + ca – a2 – b2 – c2] = 0
⇒ Either a + b + c = 0, or ab + bc + ca – a2 – b2 – c2 = 0
Now
ab + bc + ca – a2 – b2 – c2 = 0
Multiplying both sides by -2
⇒ - 2ab – 2bc – 2ca + 2a2 + 2b2 + 2c2 = 0
⇒ a2 – 2ab + b2 + b2 – 2bc + c2 + c2 – 2ca + a2 = 0
⇒ (a – b)2 + (b – c)2 + (c – a)2 = 0
Since, (a – b)2, (b – c)2, (c – a)2 are non-negative
∴ (a – b)2 = (b – c)2 = (c – a)2
⇒ (a – b) = (b – c) = (c – a)
⇒ a = b = c
Hence, if ∆ = 0, then either a + b + c = 0 or a = b = c
Solve the equation
Given,
Applying elementary transformations we get,
R1→ R1 + R2 + R3
⇒ (3x + a) = 0
C2→ C2 – C1 and C3→ C3 – C1, we get
(3x + a) = 0
Expanding along R1 we have
(3x + a) [1 (a × a – 0 × 0) – 0 (x × a – 0 × a) + 0 (x × 0 – a × x)] = 0
⇒ (3x + a) [1 × a2] = 0
⇒ a2 (3x + a) = 0
Since, a ≠ 0
∴ 3x + a = 0
⇒ x = -a/3
Prove that
Given,
LHS =
RHS = 4a2b2c2
LHS = Δ =
Taking out common factors a, b and c from C1, C2 and C3, we have
Δ = abc
Applying Elementary Transformations
R2→ R2 – R1 and R3→ R3 – R1
Δ = abc
R2→ R2 + R1
Δ = abc
R3→ R3 + R2
Δ = abc
Δ = 2ab2c
C2→ C2 – C1
Δ = 2ab2c
Expanding along R3 we get,
Δ = 2ab2c [a (c – a) + a (a + c)]
= 2ab2c [ac – a2 + a2 + ac]
= 2ab2c (2ac)
= 4a2b2c2
Δ = RHS
∴ LHS = RHS
Hence, Proved
If , find (AB)-1
B =
We need to find B-1
To find the inverse of a matrix we need to find the Adjoint of that matrix
For finding the adjoint of the matrix we need to find its cofactors
Let Bij denote the cofactors of Matrix B
Minor of an element bij = Mij �
b11 = 1, Minor of element b11 = M11 = = (3 × 1) – ((-2) × 0) = 3
b12 = 2, Minor of element b12 = M12 = = ((-1) × 1) – (0 × 0) = -1
b13 = -2, Minor of element b13 = M13 = = ((-1) × (-2)) – (3 × 0) = 2
b21 = -1, Minor of element b21 = M21 = = (2 × 1) – ((-2) × (-2)) = -2
b22 = 3, Minor of element b22 = M22 = = (1 × 1) – ((-2) × 0) = 1
b23 = 0, Minor of element b23 = M23 = = (1 × (-2)) – (2 × 0) = -2
b31 = 0, Minor of element b31 = M31 = = (2 × 0) – ((-2) × 3) = 6
b32 = -2, Minor of element b32 = M32 = = (1 × 0) – ((-2) × (-1)) = -2
b33 = 1, Minor of element b33 = M33 = = (1 × 3) – (2 × (-1)) = 5
Cofactor of an element bij, Bij = (-1)i+j × Mij
B11 = (-1)1+1× M11 = 1 × 3 = 3
B12 = (-1)1+2× M12 = (-1) × (-1) = 1
B13 = (-1)1+3× M13 = 1 × 2 = 2
B21 = (-1)2+1× M21 = (-1) × (-2) = 2
B22 = (-1)2+2 × M22 = 1 × 1 = 1
B23 = (-1)2+3 × M23 = (-1) × (-2) = 2
B31 = (-1)3+1 × M31 = 1 × 6 = 6
B32 = (-1)3+2 × M32 = (-1) × (-2) = 2
B33 = (-1)3+3 × M33 = 1 × 5 = 5
Adj B = =
|B| = 1 (3 × 1 – (-2) × 0) – 2 ((-1) × 1 – 0 × 0) + (-2) ((-1) × (-2) – 3 × 0)
|B| = 3 – 2 ( -1 – 0) – 2 (2 – 0)
|B| = 3 + 2 – 4 = 1
∴ B-1 = (Adj B)/|B| = /1 =
We know (AB)-1 = B-1A-1
(AB)-1 = ×
Solving the above matrix we get
(AB)-1 =
Let . Verify that
[adj A]-1 = adj (A-1)
A =
|A| = 1 (3 × 5 – 1 × 1) – (-2) ((-2) × 5 – 1 × 1) + 1 ((-2) × 1 – 3 × 1)
|A| = (15 – 1) + 2 (-10 – 1) + (-2 – 3)
|A| = 14 – 22 – 5 = -13
To find the inverse of a matrix we need to find the Adjoint of that matrix
For finding the adjoint of the matrix we need to find its cofactors
Let Aij denote the cofactors of Matrix A
Minor of an element aij = Mij �
a11 = 1, Minor of element a11 = M11 = = (3 × 5) – (1 × 1) = 14
a12 = -2, Minor of element a12 = M12 = = (-2 × 5) – (1 × 1) = -11
a13 = 1, Minor of element a13 = M13 = = (-2 × 1) – (3 × 1) = -5
a21 = -2, Minor of element a21 = M21 = = ((-2) × 5) – (1 × 1) = -11
a22 = 3, Minor of element a22 = M22 = = (1 × 5) – (1 × 1) = 4
a23 = 1, Minor of element a23 = M23 = = (1 × 1) – ((-2) × 1) = 3
a31 = 1, Minor of element a31 = M31 = = (-2 × 1) – (3 × 1) = -5
a32 = 1, Minor of element a32 = M32 = = (1 × 1) – (1 × (-2)) = 3
a33 = 5, Minor of element a33 = M33 = = (1 × 3) – ((-2) × (-2)) = -1
Cofactor of an element aij = Aij
A11 = (-1)1+1× 14 = 1 × 14 = 14
A12 = (-1)1+2× (-11) = (-1) × (-11) = 11
A13 = (-1)1+3× (-5) = 1 × (-5) = -5
A21 = (-1)2+1× (-11) = (-1) × (-11) = 11
A22 = (-1)2+2 × 4 = 1 × 4 = 4
A23 = (-1)2+3 × 3 = (-1) × 3 = -3
A31 = (-1)3+1 × (-5) = 1 × (-5) = -5
A32 = (-1)3+2 × 3 = (-1) × 3 = -3
A33 = (-1)3+3 × (-1) = 1 × (-1) = -1
Adj A = =
A-1 = (Adj A)/|A|
A-1 = =
(i) |Adj A| = 14(-4 – 9) – 11 (-11 – 15) – 5 (-33 + 20)
= 14 × (-13) – 11 × (-26) – 5 (-13)
= -182 + 286 + 65 = 169
Similarly Finding the Adj (Adj A) as found above
Adj (Adj A) =
[Adj A]-1 = Adj (Adj A)/|Adj A|
=
=
A-1 = =
Similarly Finding the Adj (A-1) as found above
Adj (A-1) = =
Hence, [Adj A]-1 = Adj (A-1)
Let . Verify that
(A-1)-1 = A
A =
|A| = 1 (3 × 5 – 1 × 1) – (-2) ((-2) × 5 – 1 × 1) + 1 ((-2) × 1 – 3 × 1)
|A| = (15 – 1) + 2 (-10 – 1) + (-2 – 3)
|A| = 14 – 22 – 5 = -13
To find the inverse of a matrix we need to find the Adjoint of that matrix
For finding the adjoint of the matrix we need to find its cofactors
Let Aij denote the cofactors of Matrix A
Minor of an element aij = Mij �
a11 = 1, Minor of element a11 = M11 = = (3 × 5) – (1 × 1) = 14
a12 = -2, Minor of element a12 = M12 = = (-2 × 5) – (1 × 1) = -11
a13 = 1, Minor of element a13 = M13 = = (-2 × 1) – (3 × 1) = -5
a21 = -2, Minor of element a21 = M21 = = ((-2) × 5) – (1 × 1) = -11
a22 = 3, Minor of element a22 = M22 = = (1 × 5) – (1 × 1) = 4
a23 = 1, Minor of element a23 = M23 = = (1 × 1) – ((-2) × 1) = 3
a31 = 1, Minor of element a31 = M31 = = (-2 × 1) – (3 × 1) = -5
a32 = 1, Minor of element a32 = M32 = = (1 × 1) – (1 × (-2)) = 3
a33 = 5, Minor of element a33 = M33 = = (1 × 3) – ((-2) × (-2)) = -1
Cofactor of an element aij = Aij
A11 = (-1)1+1× 14 = 1 × 14 = 14
A12 = (-1)1+2× (-11) = (-1) × (-11) = 11
A13 = (-1)1+3× (-5) = 1 × (-5) = -5
A21 = (-1)2+1× (-11) = (-1) × (-11) = 11
A22 = (-1)2+2 × 4 = 1 × 4 = 4
A23 = (-1)2+3 × 3 = (-1) × 3 = -3
A31 = (-1)3+1 × (-5) = 1 × (-5) = -5
A32 = (-1)3+2 × 3 = (-1) × 3 = -3
A33 = (-1)3+3 × (-1) = 1 × (-1) = -1
Adj A = =
A-1 = (Adj A)/|A|
A-1 = =
(ii) To find (A-1)-1 we have to find out Adj(A-1)
A-1 =
|A-1| = (-1/13)3 [14 (4 × (-1) – (-3) × (-3)) – 11 (11 × (-1) – (-3) × (-5)) + (-5) (11 × (-3) – 4 × (-5))]
|A| = (-1/13)3 [14 (-4 – 9) – 11 (-11 – 15) – 5 (-33 + 20)]
|A| = (-1/13)3 [14 × (-13) – 11 × (-26) – 5 × (-13)]
|A| = (-1/13)3 × 169 = -1/13
Cofactor of an element aij = Aij
A11 = (-1)1+1× (-1/13) = 1 × (-1/13) = -1/13
A12 = (-1)1+2× (-2/13) = (-1) × (-2/13) = 2/13
A13 = (-1)1+3× (-1/13) = 1 × (-1/13) = -1/13
A21 = (-1)2+1× (-2/13) = (-1) × (-2/13) = 2/13
A22 = (-1)2+2 × (3/13) = 1 × (-3/13) = -3/13
A23 = (-1)2+3 × (1/13) = (-1) × (1/13) = -1/13
A31 = (-1)3+1 × (-1/13) = 1 × (-1/13) = -1/13
A32 = (-1)3+2 × (1/13) = (-1) × 1/13 = -1/13
A33 = (-1)3+3 × (-5/13) = 1 × (-5/13) = -5/13
Adj (A-1) = =
(A-1)-1 = Adj(A-1)/|A-1|
(A-1)-1 = = = A
∴ (A-1)-1 = A
Evaluate
Let Δ =
Applying Elementary Transformations.
Applying R1→ R1 + R2 + R3, we have
Δ =
Δ = 2 (x + y)
Applying C2→ C2 – C1 and C3→ C3 – C1, we have
Δ = 2 (x + y)
Expanding along R1, we have
Δ = 2 (x + y) [1 (x × (-x) – (-y) × (x – y)) – 0 + 0]
Δ = 2 (x + y) [-x2 + y (x – y)]
Δ = 2 (x + y) [-x2 + xy – y2]
Δ = -2 (x + y) [x2 – xy + y2]
Δ = -2 (x3 + y3)
Evaluate
Let Δ =
Applying Row transformations
R2→ R2 – R1; R3→ R3 – R1
Δ =
Δ =
Now, Expanding along C1
Δ = 1 (y × x – 0 × 0) – 0 (x × x – 0 × y) + 0 (x × 0 – y × y)
Δ = 1 (xy) – 0 + 0
Δ = xy
Prove that
Let Δ =
Applying Row Transformations
R2→ R2 – R1
Δ =
R3→ R3 – R1
Δ =
Taking (β – α)(γ – α) from R2 and R3 respectively
Δ = (β – α) (γ – α)
Applying R3→ R3 – R2, we have
Δ = (β – α) (γ – α)
Expanding along R3, we have
Δ = (β – α) (γ – α) [0 (α2 × (-1) – (β + γ) × (β + γ) – (γ – β)((-1) × α – 1 × (β + γ) + 0 (α × (β + γ) – 1 × α2)
Δ = (β – α) (γ – α) [0 – (γ – β)( - α - β – γ) + 0]
Δ = (β – α) (γ – α) (γ – β) (α + β + γ)
Δ = (α – β) (β – γ) (γ – α) (α + β + γ)
Prove that where p is any scalar.
Let Δ =
Applying Elementary Row Transformations
R2→ R2 – R1 and R3→ R3 – R1
Δ =
Taking (y – x) and (z – x) common from R2 and R3 respectively
Δ = (y – x) (z – x)
Applying R3→ R3 – R2
Δ = (y – x) (z – x)
Taking (z – y) common from R3
Δ = (y – x) (z – x) (z – y)
Expanding along R3, we have
Δ = (x – y) (y – z) (z – x) [0 – 1 {x × p(y2 + x2 + xy) – 1 × (1 + px3)} + p (x + y + z) {x × (y + x) – 1 × x2}
Δ = (x – y) (y – z) (z – x) (-px3 – pxy2 – px2y + 1 + px3 + px2y + pxy2 + pxyz)
Δ = (x – y) (y – z) (z – x) (1 + pxyz)
Hence, the given result is proved.
Prove that
Let Δ =
Applying Elementary Transformations
C1→ C1 + C2 + C3, we have
Δ =
Taking (a + b + c) common from C1, we get
Δ = (a + b + c)
Applying R2→ R2 – R1 and R3→ R3 – R1
Expanding along C1
Δ = (a + b + c) [1 × {(2b + a) (2c + a) – (a – b) (a – c)} – 0 + 0]
Δ = (a + b + c) [4bc + 2ab + 2ac + a2 – a2 + ac + ba – bc]
Δ = (a + b + c) (3ab + 3bc + 3ac)
Δ = 3 (a + b + c) (ab + bc + ca)
Hence, the given result proved.
Prove that
Let Δ =
Applying Elementary Row Transformations
R2→ R2 – 2R1
Δ =
R3→ R3 – 3R1
Δ =
R3→ R3 – 3R2
Δ =
Expanding Along C1, we have
Δ = 1 (1 × 1 – 0 × (2 + p)) – 0 + 0
Δ = 1 – 0
Δ = 1
Hence, the given result is proved
Prove that
Let Δ =
Δ =
Δ =
Applying Elementary Column Transformations
C1→ C1 + C3
Δ =
Since, the two columns are identical
[In a determinant if two columns are identical the the value of determinant is 0]
So, the value of given determinant is 0
∴ Δ = 0
Hence, the given result is proved.
Solve the system of equations
Given System of equations are
Let
∴ Given system of equation becomes
2p + 3q + 10r = 4
4p – 6q + 5r = 1
6p + 9q – 20r = 2
The given System of Equations can be written in the form of AX = B
Here A = and B = , X =
Now we need to find |A|
∴ |A| = = 2 × (120 – 45) – 3 × (- 80 – 30) + 10 × (36 + 36)
= 150 + 330 + 720
= 1200
Since, |A| ≠ 0
∴ A is non-singular. So, it’s inverse exists
Cofactor of an element aij = Aij
A11 = (-1)1+1× 75 = 1 × 75 = 75
A12 = (-1)1+2× (-110) = (-1) × (-110) = 110
A13 = (-1)1+3× 72 = 1 × (72) = 72
A21 = (-1)2+1× (-150) = (-1) × (-150) = 150
A22 = (-1)2+2 × (-100) = 1 × (-100) = -100
A23 = (-1)2+3 × 0 = (-1) × 0 = 0
A31 = (-1)3+1 × 75 = 1 × 75 = 75
A32 = (-1)3+2 × (-30) = (-1) × (-30) = 30
A33 = (-1)3+3 × (-24) = 1 × (-24) = -24
Adj A = =
A-1 = (Adj A)/|A|
A-1 =
Now,
Since, AX = B
∴ X = A-1B
∴ p = �, q = 1/3, r = 1/5
∴ x = 1/p = 2; y = 1/q = 3; z = 1/r = 5
So, x = 2; y = 3; z = 5
If a, b, c, are in A.P, then the determinant is
A. 0
B. 1
C. x
D. 2x
Let Δ =
Since, a, b, c are in A.P.
∴ 2b = a + c
Δ =
Applying Elementary Row Transformations
R1→ R1 – R2 and R3→ R3 – R2
Δ =
R1→ R1 + R3, we have
Δ =
[In a determinant if all elements of a row is 0 then the value of determinant is 0.]
So, here all the elements of first row (R1) are zero.
∴ Δ = 0
If x, y, z are nonzero real numbers, then the inverse of matrix is
A.
B.
C.
D.
A =
|A| = x × (y × z) = xyz
Since, |A| ≠ 0
A-1 exists
To find the inverse of a matrix we need to find the Adjoint of that matrix
For finding the adjoint of the matrix we need to find its cofactors
Let Aij denote the cofactors of Matrix A
Minor of an element aij = Mij �
a11 = x, Minor of element a11 = M11 = = (y × z) – (0 × 0) = yz
a12 = 0, Minor of element a12 = M12 = = (0 × z) – (0 × 0) = 0
a13 = 0, Minor of element a13 = M13 = = (0 × 0) – (0 × y) = 0
a21 = 0, Minor of element a21 = M21 = = (0 × z) – (0 × 0) = 0
a22 = y, Minor of element a22 = M22 = = (x × z) – (0 × 0) = xz
a23 = 0, Minor of element a23 = M23 = = (x × 0) – (0 × 0) = 0
a31 = 0, Minor of element a31 = M31 = = (z × 0) – (0 × 0) = 0
a32 = 0, Minor of element a32 = M32 = = (x × 0) – (0 × 0) = 0
a33 = z, Minor of element a33 = M33 = = (x × y) – (0 × 0) = xy
Cofactor of an element aij, Aij = (-1)i+j × Mij
A11 = (-1)1+1× M11 = 1 × yz = yz
A12 = (-1)1+2× M12 = (-1) × 0 = 0
A13 = (-1)1+3× M13 = 1 × 0 = 0
A21 = (-1)2+1× M21 = (-1) × 0 = 0
A22 = (-1)2+2 × M22 = 1 × xz = xz
A23 = (-1)2+3 × M23 = (-1) × 0 = 0
A31 = (-1)3+1 × M31 = 1 × 0 = 0
A32 = (-1)3+2 × M32 = (-1) × 0 = 0
A33 = (-1)3+3 × M33 = 1 × xy = xy
Adj A = =
A-1 = adj A / |A|
A-1 = /xyz
A-1 =
A-1 = =
The correct answer is A
Let , where 0 ≤ θ ≤ 2π. Then
A. Det (A) = 0
B. Det (A) є (2, ∞)
C. Det (A) є (2, 4)
D. Det (A) є [2, 4]
A =
|A| = 1 (1 × 1 – sin θ × (-sin θ)) – sin θ ((-sin θ) × 1 – (-1) × sin θ) + 1 ((-sin θ) × (-sin θ) – (-1) × 1)
|A| = 1 + sin2θ + sin2θ – sin2θ + sin2θ + 1
|A| = 2 + 2sin2θ
|A| = 2(1 + sin2θ)
Now, 0 ≤ θ ≤ 2π
⇒ sin 0 ≤ sin θ ≤ sin 2π
⇒ 0 ≤ sin2θ ≤ 1
⇒ 1 + 0 ≤ 1 + sin2θ ≤ 1 + 1
⇒ 2 ≤ 2(1 + sin2θ) ≤ 4
∴ Det (A) є [2, 4]