Math

2x2 Matrix Determinant and Inverse Calculator

Calculator Free · Private
Was this calculator helpful?

A 2×2 matrix is the simplest square matrix that carries full linear-algebraic meaning. Its determinant is a single scalar that tells you whether the matrix is invertible, how it scales areas, and what sign it assigns to orientation. For a matrix A = [[a, b], [c, d]], the determinant is det(A) = ad − bc. If det(A) ≠ 0, the inverse exists and equals (1/det(A)) × [[d, −b], [−c, a]]. These two quantities appear in solving 2×2 linear systems, computing 2D transformations (rotation, shear, scaling), finding eigenvalues, and verifying linear independence of two vectors.

Last reviewed: May 12, 2026 Verified by Hacé Cuentas Team Source: Wikipedia – Determinant (mathematics), Wikipedia – Invertible Matrix, NIST Digital Library of Mathematical Functions – Linear Algebra 100% private

When to use this calculator

  • Solving a 2×2 linear system (e.g., 3x + y = 7, 5x + 2y = 12) by computing the coefficient matrix inverse and multiplying by the constants vector.
  • Checking whether two 2D vectors are linearly independent: arrange them as rows of a 2×2 matrix — if det ≠ 0, they form a basis for ℝ².
  • Computing the area of a parallelogram spanned by vectors u = (a, c) and v = (b, d): Area = |ad − bc|.
  • Verifying a 2D affine or linear transformation (rotation by θ, scaling, shear) is non-singular before applying it in computer graphics pipelines.
  • Finding the inverse of a 2×2 covariance matrix in a bivariate normal distribution calculation for statistics or machine learning.

Example Calculation

  1. [[1,2],[3,4]]
  2. det = −2
Result: det = −2

How it works

3 min read

How It Is Calculated

Given the 2×2 matrix:

A = | a  b |
    | c  d |

Step 1 — Determinant:

det(A) = a·d − b·c

Step 2 — Check invertibility:

If det(A) = 0  →  A is singular (no inverse exists)
If det(A) ≠ 0  →  A is invertible (non-singular)

Step 3 — Inverse (only if det ≠ 0):

A⁻¹ = (1 / det(A)) × |  d  −b |
                       | −c   a |

This formula comes from the adjugate (classical adjoint) method: swap the main-diagonal entries, negate the off-diagonal entries, then divide every entry by the determinant.

Verification: A · A⁻¹ must equal the 2×2 identity matrix I = [[1,0],[0,1]].

---

Reference Table

The table below shows common 2×2 matrices, their determinants, and inverse entries rounded to 4 decimal places.

Matrix [[a,b],[c,d]]det = ad−bcInvertible?A⁻¹ (exact)
[[1,0],[0,1]] (Identity)1✅ Yes[[1,0],[0,1]]
[[2,0],[0,3]] (Diagonal)6✅ Yes[[1/6·6? → 1/2, 0],[0, 1/3]] = [[0.5,0],[0,0.3333]]
[[1,2],[3,4]]−2✅ Yes[[-2, 1],[1.5, -0.5]]
[[3,1],[5,2]]1✅ Yes[[2,−1],[−5,3]]
[[cos θ, −sin θ],[sin θ, cos θ]] (Rotation)1✅ Yes[[cos θ, sin θ],[−sin θ, cos θ]]
[[1,2],[2,4]]0❌ NoDoes not exist
[[0,0],[0,0]] (Zero)0❌ NoDoes not exist
[[−1,0],[0,1]] (Reflection)−1✅ Yes[[−1,0],[0,1]] (self-inverse)

> Key pattern: A rotation matrix always has det = cos²θ + sin²θ = 1, so it is always invertible, and its inverse equals its transpose.

---

Typical Cases (Worked Examples)

Example 1 — Classic textbook matrix


A = [[1, 2],
     [3, 4]]

det(A) = 1·4 − 2·3 = 4 − 6 = −2

A⁻¹ = (1/−2) × [[ 4, −2],
                  [−3,  1]]
     = [[ −2,   1  ],
        [  1.5, −0.5]]

Check: [[1,2],[3,4]] × [[−2,1],[1.5,−0.5]] = [[1·(−2)+2·1.5, 1·1+2·(−0.5)],[3·(−2)+4·1.5, 3·1+4·(−0.5)]] = [[1,0],[0,1]] ✅

---

Example 2 — Solving a linear system with the inverse


System: 2x + 5y = 1 and 1x + 3y = 0

Coefficient matrix: A = [[2, 5], [1, 3]]
det(A) = 2·3 − 5·1 = 6 − 5 = 1

A⁻¹ = (1/1) × [[ 3, −5],
                [−1,  2]]
     = [[ 3, −5],
        [−1,  2]]

Solution vector: A⁻¹ × [1, 0]ᵀ
  x = 3·1 + (−5)·0 = 3
  y = (−1)·1 + 2·0 = −1

Verify: 2(3)+5(−1)=1 ✅ and 1(3)+3(−1)=0 ✅

---

Example 3 — Singular matrix (no inverse)


A = [[4, 6],
     [2, 3]]

det(A) = 4·3 − 6·2 = 12 − 12 = 0

Row 2 is exactly ½ of Row 1 → linearly dependent rows → the system has either infinitely many solutions or no solution, never a unique one. No inverse exists.

---

Common Errors

1. Swapping instead of negating off-diagonal entries. The adjugate formula requires swapping a and d (main diagonal) AND negating b and c (off-diagonal). Students often negate all four entries or only negate one.

2. Forgetting to divide by det(A). Computing the adjugate [[d, −b],[−c, a]] without multiplying by 1/det(A) gives the adjugate matrix, not the inverse. The result will not satisfy A · A⁻¹ = I.

3. Applying the formula when det = 0. Division by zero is undefined. A singular matrix has no inverse — attempting to compute one yields meaningless entries or ±∞.

4. Sign error in the determinant formula. det = ad − bc, NOT ad + bc or ab − cd. Mixing up which product is subtracted is the most common arithmetic mistake, especially when entries are negative.

5. Confusing the determinant with the trace. The trace is a + d (sum of main-diagonal entries). The determinant is ad − bc. Both appear in the characteristic polynomial λ² − tr(A)λ + det(A) = 0, but they serve different roles.

6. Row vs. column ordering. When constructing the matrix from a word problem or a set of equations, always confirm that entries map correctly: a=row1/col1, b=row1/col2, c=row2/col1, d=row2/col2. A transposition error changes the determinant sign and completely changes the inverse.

---

Related Calculators

  • Percentage Calculator

  • Scientific Calculator

  • Ratio & Proportion Calculator

  • Frequently asked questions

    What does it mean geometrically when det(A) = 0?

    A determinant of zero means the two row vectors (or column vectors) of the matrix are linearly dependent — one is a scalar multiple of the other. Geometrically, the linear transformation collapses the plane onto a line (or point), reducing area to zero. The transformation is not reversible, which is why no inverse exists.

    Can the determinant be negative, and what does that mean?

    Yes. A negative determinant means the transformation reverses orientation (flips the plane, like a reflection). Its absolute value |det(A)| still gives the area scale factor. For example, a reflection matrix [[−1,0],[0,1]] has det = −1: it preserves area but mirrors the x-axis, reversing the handedness of the coordinate system.

    How do I use the inverse matrix to solve Ax = b?

    If A is invertible, the unique solution to the system Ax = b is x = A⁻¹b. Compute A⁻¹ using (1/det(A))×[[d,−b],[−c,a]], then multiply it by the constants vector b = [b₁, b₂]ᵀ. This gives x = [x₁, x₂]ᵀ directly, without row reduction. This approach is efficient for 2×2 systems but is computationally expensive for large n×n systems.

    What is Cramer's Rule and how does it relate to the determinant?

    Cramer's Rule is an explicit formula for each variable in a linear system using determinants. For Ax = b (2×2), x₁ = det(A₁)/det(A) and x₂ = det(A₂)/det(A), where A₁ replaces column 1 with b, and A₂ replaces column 2 with b. It is mathematically equivalent to using A⁻¹, and requires det(A) ≠ 0.

    Is the inverse of a 2×2 matrix unique?

    Yes — if an inverse exists, it is unique. This follows from the cancellation law for matrices: if AB = I and AC = I, then B = IB = (CA)B = C(AB) = CI = C. For 2×2 matrices, the formula A⁻¹ = (1/det)×[[d,−b],[−c,a]] always produces the one and only inverse when det ≠ 0.

    What is a rotation matrix and why is its determinant always 1?

    A 2D rotation matrix is R(θ) = [[cos θ, −sin θ],[sin θ, cos θ]]. Its determinant is cos²θ − (−sin θ)(sin θ) = cos²θ + sin²θ = 1 for all θ, by the Pythagorean identity. This means rotations preserve area and orientation. The inverse of R(θ) is R(−θ), which is simply its transpose: [[cos θ, sin θ],[−sin θ, cos θ]].

    How does the determinant relate to eigenvalues?

    For a 2×2 matrix, det(A) equals the product of its two eigenvalues (λ₁ · λ₂). This comes from the characteristic polynomial: det(A − λI) = λ² − tr(A)λ + det(A) = 0. If either eigenvalue is zero, det = 0 and the matrix is singular. Conversely, the trace equals λ₁ + λ₂.

    What happens to the determinant if I swap the two rows?

    Swapping the rows of A = [[a,b],[c,d]] gives [[c,d],[a,b]], with det = cb − da = −(ad−bc) = −det(A). Every elementary row swap multiplies the determinant by −1. This property is fundamental to Gaussian elimination and the definition of the determinant as an alternating multilinear form.

    When is a 2×2 matrix its own inverse (involutory)?

    A matrix A is involutory (self-inverse) when A² = I, which requires A⁻¹ = A. For a 2×2 matrix, this happens when a + d = 0 (trace = 0) and det(A) = −1. Classic examples include reflection matrices like [[1,0],[0,−1]] and [[0,1],[1,0]], and more generally any matrix of the form [[a, b],[c, −a]] with a² + bc = 1.

    Sources and references