4  Linear Algebra for Advanced Applications

4.1 Introduction

Matrix decomposition plays a pivotal role in computational linear algebra, forming the backbone of numerous modern applications in fields such as data science, machine learning, computer vision, and signal processing. The core idea behind matrix decomposition is to break down complex matrices into simpler, structured components that allow for more efficient computation. Techniques such as LU, QR, Singular Value Decomposition (SVD), and Eigenvalue decompositions not only reduce computational complexity but also provide deep insights into the geometry and structure of data. These methods are essential in solving systems of linear equations, performing dimensionality reduction, and extracting meaningful features from data. For instance, LU decomposition is widely used to solve large linear systems, while QR decomposition plays a key role in solving least squares problems—a fundamental task in machine learning models.

In emerging fields like big data analytics and artificial intelligence, matrix decomposition techniques are indispensable for processing and analyzing high-dimensional datasets. SVD and Principal Component Analysis (PCA), for example, are extensively used for data compression and noise reduction, making machine learning algorithms more efficient by reducing the number of variables while retaining key information. Additionally, sparse matrix decompositions allow for the handling of enormous datasets where most entries are zero, optimizing memory usage and computation time. As data science and machine learning continue to evolve, mastering these matrix decomposition techniques provides not only a computational advantage but also deeper insights into the structure and relationships within data, enhancing the performance of algorithms in real-world applications.

4.2 LU Decomposition

LU decomposition is a powerful tool in linear algebra that elegantly unravels the complexity of solving systems of linear equations. At its core, LU decomposition expresses a matrix \(A\) as the product of two distinct matrices: \(L\) (a lower triangular matrix with ones on the diagonal) and \(U\) (an upper triangular matrix). This decomposition transforms the problem of solving \(Ax = b\) into a two-step process: first, solving \(Ly = b\) for \(y\), followed by \(Ux = y\) for \(x\). This systematic approach not only simplifies computations but also provides insightful perspectives on the relationships between the equations involved.

The magic of LU decomposition lies in its utilization of elementary transformations—operations that allow us to manipulate the rows of a matrix to achieve a row-reduced echelon form. These transformations include row swaps, scaling, and adding multiples of one row to another. By applying these operations, we can gradually transform the original matrix \(A\) into the upper triangular matrix \(U\), while simultaneously capturing the essence of these transformations in the lower triangular matrix \(L\). This interplay of \(L\) and \(U\) not only enhances computational efficiency but also unveils the deeper structural relationships within the matrix.

Moreover, the beauty of matrix multiplication shines through in LU decomposition. The product \(A = LU\) showcases how two simpler matrices can combine to reconstruct a more complex one, demonstrating the power of linear combinations in solving equations. As we delve into LU decomposition, we embark on a journey that highlights the synergy between algebraic manipulation and geometric interpretation, empowering us to tackle intricate problems with grace and precision. Given a square matrix \(A\), the LU decomposition expresses \(A\) as a product of a lower triangular matrix \(L\) and an upper triangular matrix \(U\): \[A = LU\]

Where: - \(L\) is a lower triangular matrix with 1’s on the diagonal and other elements like \(l_{21}, l_{31}, l_{32}, \dots\), - \(U\) is an upper triangular matrix with elements \(u_{11}, u_{12}, u_{13}, u_{22}, u_{23}, u_{33}, \dots\).

4.2.1 Step-by-Step Procedure

Let’s assume \(A\) is a \(3 \times 3\) matrix for simplicity: \[A = \begin{pmatrix}a_{11} & a_{12} & a_{13} \\a_{21} & a_{22} & a_{23} \\a_{31} & a_{32} & a_{33}\end{pmatrix}\]

We need to find matrices \(L\) and \(U\), where:

  • \(L = \begin{pmatrix} 1 & 0 & 0 \\ l_{21} & 1 & 0 \\ l_{31} & l_{32} & 1 \end{pmatrix}\)

  • \(U = \begin{pmatrix} u_{11} & u_{12} & u_{13} \\ 0 & u_{22} & u_{23} \\ 0 & 0 & u_{33} \end{pmatrix}\)

The product of \(L\) and \(U\) gives: \[LU = \begin{pmatrix} 1 & 0 & 0 \\l_{21} & 1 & 0 \\l_{31} & l_{32} & 1\end{pmatrix}\begin{pmatrix} u_{11} & u_{12} & u_{13} \\0 & u_{22} & u_{23} \\0 & 0 & u_{33}\end{pmatrix}=\begin{pmatrix} u_{11} & u_{12} & u_{13} \\l_{21}u_{11} & l_{21}u_{12} + u_{22} & l_{21}u_{13} + u_{23} \\l_{31}u_{11} & l_{31}u_{12} + l_{32}u_{22} & l_{31}u_{13} + l_{32}u_{23} + u_{33}\end{pmatrix}\]

By equating this with \(A\), we can set up a system of equations to solve for \(l_{ij}\) and \(u_{ij}\).

Step 1: Solve for \(u_{11}, u_{12}, u_{13}\)

From the first row of \(A = LU\), we have: \[u_{11} = a_{11}\] \[u_{12} = a_{12}\] \[u_{13} = a_{13}\]

Step 2: Solve for \(l_{21}\) and \(u_{22}, u_{23}\)

From the second row, we get: \[l_{21}u_{11} = a_{21} \quad \Rightarrow \quad l_{21} = \frac{a_{21}}{u_{11}}\] \[l_{21}u_{12} + u_{22} = a_{22} \quad \Rightarrow \quad u_{22} = a_{22} - l_{21}u_{12}\] \[l_{21}u_{13} + u_{23} = a_{23} \quad \Rightarrow \quad u_{23} = a_{23} - l_{21}u_{13}\]

Step 3: Solve for \(l_{31}, l_{32}\) and \(u_{33}\)

From the third row, we get: \[l_{31}u_{11} = a_{31} \quad \Rightarrow \quad l_{31} = \frac{a_{31}}{u_{11}}\] \[l_{31}u_{12} + l_{32}u_{22} = a_{32} \quad \Rightarrow \quad l_{32} = \frac{a_{32} - l_{31}u_{12}}{u_{22}}\] \[l_{31}u_{13} + l_{32}u_{23} + u_{33} = a_{33} \quad \Rightarrow \quad u_{33} = a_{33} - l_{31}u_{13} - l_{32}u_{23}\]

Final Result

Thus, the LU decomposition is given by the matrices: - \(L = \begin{pmatrix} 1 & 0 & 0 \\ l_{21} & 1 & 0 \\ l_{31} & l_{32} & 1 \end{pmatrix}\) - \(U = \begin{pmatrix} u_{11} & u_{12} & u_{13} \\ 0 & u_{22} & u_{23} \\ 0 & 0 & u_{33} \end{pmatrix}\)

Where: - \(u_{11} = a_{11}, u_{12} = a_{12}, u_{13} = a_{13}\) - \(l_{21} = \frac{a_{21}}{u_{11}}, u_{22} = a_{22} - l_{21}u_{12}, u_{23} = a_{23} - l_{21}u_{13}\) - \(l_{31} = \frac{a_{31}}{u_{11}}, l_{32} = \frac{a_{32} - l_{31}u_{12}}{u_{22}}, u_{33} = a_{33} - l_{31}u_{13} - l_{32}u_{23}\)

4.2.2 Example

Let’s decompose the following matrix: \[A = \begin{pmatrix} 4 & 3 & 2 \\6 & 3 & 1 \\2 & 1 & 3\end{pmatrix}\]

Following the steps outlined above:

  • \(u_{11} = 4, u_{12} = 3, u_{13} = 2\)
  • \(l_{21} = \frac{6}{4} = 1.5\), so:
    • \(u_{22} = 3 - 1.5 \times 3 = -1.5\)
    • \(u_{23} = 1 - 1.5 \times 2 = -2\)
  • \(l_{31} = \frac{2}{4} = 0.5\), so:
    • \(l_{32} = \frac{1 - 0.5 \times 3}{-1.5} = 0.67\)
    • \(u_{33} = 3 - 0.5 \times 2 - 0.67 \times (-2) = 2.67\)

Thus, the decomposition is: - \(L = \begin{pmatrix} 1 & 0 & 0 \\ 1.5 & 1 & 0 \\ 0.5 & 0.67 & 1 \end{pmatrix}\) - \(U = \begin{pmatrix} 4 & 3 & 2 \\ 0 & -1.5 & -2 \\ 0 & 0 & 2.67 \end{pmatrix}\)

4.2.3 Python Implementation

import numpy as np
from scipy.linalg import lu

# Define matrix A
A = np.array([[4, 3, 2],
              [6, 3, 1],
              [2, 1, 3]])

# Perform LU decomposition
P, L, U = lu(A)

# Print the results
print("L = \n", L)
print("U = \n", U)
L = 
 [[1.         0.         0.        ]
 [0.66666667 1.         0.        ]
 [0.33333333 0.         1.        ]]
U = 
 [[6.         3.         1.        ]
 [0.         1.         1.33333333]
 [0.         0.         2.66666667]]
Note

Since there are many row transformations that reduce a given matrix into row echelon form. So the LU decomposition is not unique.

4.2.4 LU Decomposition Practice Problems with Solutions

Problem 1: Decompose the matrix \[ A = \begin{pmatrix} 4 & 3 \\ 6 & 3 \end{pmatrix} \] into the product of a lower triangular matrix \(L\) and an upper triangular matrix \(U\).

Solution:

Let \[ L = \begin{pmatrix} 1 & 0 \\ l_{21} & 1 \end{pmatrix}, \quad U = \begin{pmatrix} u_{11} & u_{12} \\ 0 & u_{22} \end{pmatrix}. \]

We have:

  1. From the first row: \(u_{11} = 4\) and \(u_{12} = 3\).
  2. From the second row: \(6 = l_{21} \cdot 4\) gives \(l_{21} = \frac{6}{4} = 1.5\).
  3. Finally, \(3 = 1.5 \cdot 3 + u_{22}\) gives \(u_{22} = 3 - 4.5 = -1.5\).

Thus, we have: \[ L = \begin{pmatrix} 1 & 0 \\ 1.5 & 1 \end{pmatrix}, \quad U = \begin{pmatrix} 4 & 3 \\ 0 & -1.5 \end{pmatrix}. \]

Problem 2: Given the matrix \[ A = \begin{pmatrix} 1 & 2 & 3 \\ 2 & 5 & 8 \\ 4 & 5 & 6 \end{pmatrix}, \] perform LU decomposition to find matrices \(L\) and \(U\).

Solution:

Let \[ L = \begin{pmatrix} 1 & 0 & 0 \\ l_{21} & 1 & 0 \\ l_{31} & l_{32} & 1 \end{pmatrix}, \quad U = \begin{pmatrix} u_{11} & u_{12} & u_{13} \\ 0 & u_{22} & u_{23} \\ 0 & 0 & u_{33} \end{pmatrix}. \]

We have:

  1. From Row 1: \(u_{11} = 1, u_{12} = 2, u_{13} = 3\).
  2. From Row 2: \(2 = l_{21} \cdot 1\) gives \(l_{21} = 2\).
    • For Row 2: \(5 = l_{21} \cdot 2 + u_{22}\) gives \(5 = 4 + u_{22} \Rightarrow u_{22} = 1\).
    • \(8 = l_{21} \cdot 3 + u_{23} \Rightarrow 8 = 6 + u_{23} \Rightarrow u_{23} = 2\).
  3. From Row 3: \(4 = l_{31} \cdot 1 \Rightarrow l_{31} = 4\).
    • \(5 = l_{31} \cdot 2 + l_{32} \cdot 1 \Rightarrow 5 = 8 + l_{32} \Rightarrow l_{32} = -3\).
    • Finally, \(6 = l_{31} \cdot 3 + l_{32} \cdot 2 + u_{33} \Rightarrow 6 = 12 - 6 + u_{33} \Rightarrow u_{33} = 0\).

Thus, \[ L = \begin{pmatrix} 1 & 0 & 0 \\ 2 & 1 & 0 \\ 4 & -3 & 1 \end{pmatrix}, \quad U = \begin{pmatrix} 1 & 2 & 3 \\ 0 & 1 & 2 \\ 0 & 0 & 0 \end{pmatrix}. \]

Problem 3: Perform LU decomposition of the matrix \[ A = \begin{pmatrix} 2 & 1 & 1 \\ 4 & -6 & 0 \\ -2 & 7 & 2 \end{pmatrix}, \] and verify the decomposition by checking \(A = LU\).

Solution:

Let \[ L = \begin{pmatrix} 1 & 0 & 0 \\ l_{21} & 1 & 0 \\ l_{31} & l_{32} & 1 \end{pmatrix}, \quad U = \begin{pmatrix} u_{11} & u_{12} & u_{13} \\ 0 & u_{22} & u_{23} \\ 0 & 0 & u_{33} \end{pmatrix}. \]

We have:

  1. From Row 1: \(u_{11} = 2, u_{12} = 1, u_{13} = 1\).
  2. From Row 2: \(4 = l_{21} \cdot 2 \Rightarrow l_{21} = 2\).
    • \(-6 = 2 \cdot 1 + u_{22} \Rightarrow u_{22} = -8\).
    • \(0 = 2 \cdot 1 + u_{23} \Rightarrow u_{23} = -2\).
  3. From Row 3: \(-2 = l_{31} \cdot 2 \Rightarrow l_{31} = -1\).
    • \(7 = -1 \cdot 1 + l_{32} \cdot -8 \Rightarrow 7 = -1 - 8l_{32} \Rightarrow l_{32} = -1\).
    • Finally, \(2 = -1 \cdot 1 + -1 \cdot -2 + u_{33} \Rightarrow 2 = 1 + u_{33} \Rightarrow u_{33} = 1\).

Thus, \[ L = \begin{pmatrix} 1 & 0 & 0 \\ 2 & 1 & 0 \\ -1 & -1 & 1 \end{pmatrix}, \quad U = \begin{pmatrix} 2 & 1 & 1 \\ 0 & -8 & -2 \\ 0 & 0 & 1 \end{pmatrix}. \]

Problem 4: For the matrix \[ A = \begin{pmatrix} 3 & 1 & 6 \\ 2 & 1 & 1 \\ 1 & 2 & 2 \end{pmatrix}, \] find the LU decomposition and use it to solve the system \(Ax = b\) where \(b = \begin{pmatrix} 9 \\ 5 \\ 4 \end{pmatrix}\).

Solution:

Let \[ L = \begin{pmatrix} 1 & 0 & 0 \\ l_{21} & 1 & 0 \\ l_{31} & l_{32} & 1 \end{pmatrix}, \quad U = \begin{pmatrix} u_{11} & u_{12} & u_{13} \\ 0 & u_{22} & u_{23} \\ 0 & 0 & u_{33} \end{pmatrix}. \]

We have:

  1. From Row 1: \(u_{11} = 3, u_{12} = 1, u_{13} = 6\).
  2. From Row 2: \(2 = l_{21} \cdot 3 \Rightarrow l_{21} = \frac{2}{3}\).
    • \(1 = \frac{2}{3} \cdot 1 + u_{22} \Rightarrow 1 = \frac{2}{3} + u_{22} \Rightarrow u_{22} = \frac{1}{3}\).
    • \(1 = \frac{2}{3} \cdot 6 + u_{23} \Rightarrow 1 = 4 + u_{23} \Rightarrow u_{23} = -3\).
  3. From Row 3: \(1 = l_{31} \cdot 3 \Rightarrow l_{31} = \frac{1}{3}\).
    • \(2 = \frac{1}{3} \cdot 1 + l_{32} \cdot \frac{1}{3} \Rightarrow 2 = \frac{1}{3} + \frac{1}{3} l_{32} \Rightarrow l_{32} = 6\).
    • Finally, \(2 = \frac{1}{3} \cdot 6 + 6 \cdot -3 + u_{33} \Rightarrow 2 = 2 - 18 + u_{33} \Rightarrow u_{33} = 18\).

Thus, \[ L = \begin{pmatrix} 1 & 0 & 0 \\ \frac{2}{3} & 1 & 0 \\ \frac{1}{3} & 6 & 1 \end{pmatrix}, \quad U = \begin{pmatrix} 3 & 1 & 6 \\ 0 & \frac{1}{3} & -3 \\ 0 & 0 & 18 \end{pmatrix}. \]

Now, to solve \(Ax = b\), we first solve \(Ly = b\): \[ \begin{pmatrix} 1 & 0 & 0 \\ \frac{2}{3} & 1 & 0 \\ \frac{1}{3} & 6 & 1 \end{pmatrix} \begin{pmatrix} y_1 \\ y_2 \\ y_3 \end{pmatrix} = \begin{pmatrix} 9 \\ 5 \\ 4 \end{pmatrix} \]

Solving this gives: 1. \(y_1 = 9\) 2. \(\frac{2}{3} \cdot 9 + y_2 = 5 \Rightarrow 6 + y_2 = 5 \Rightarrow y_2 = -1\) 3. \(\frac{1}{3} \cdot 9 + 6 \cdot -1 + y_3 = 4 \Rightarrow 3 - 6 + y_3 = 4 \Rightarrow y_3 = 7\)

Next, solve \(Ux = y\): \[ \begin{pmatrix} 3 & 1 & 6 \\ 0 & \frac{1}{3} & -3 \\ 0 & 0 & 18 \end{pmatrix} \begin{pmatrix} x_1 \\ x_2 \\ x_3 \end{pmatrix} = \begin{pmatrix} 9 \\ -1 \\ 7 \end{pmatrix} \]

  1. From Row 3: \(18x_3 = 7 \Rightarrow x_3 = \frac{7}{18}\)
  2. From Row 2: \(\frac{1}{3}x_2 - 3x_3 = -1 \Rightarrow \frac{1}{3}x_2 - \frac{21}{18} = -1 \Rightarrow \frac{1}{3}x_2 = -\frac{18}{18} + \frac{21}{18} = \frac{3}{18} \Rightarrow x_2 = \frac{1}{3}\)
  3. From Row 1: \(3x_1 + x_2 + 6x_3 = 9 \Rightarrow 3x_1 + \frac{1}{3} + \frac{42}{18} = 9 \Rightarrow 3x_1 + \frac{1}{3} + \frac{7}{3} = 9 \Rightarrow 3x_1 = 9 - \frac{8}{3} = \frac{27 - 8}{3} = \frac{19}{3} \Rightarrow x_1 = \frac{19}{9}\)

Thus, the solution to \(Ax = b\) is \[ x = \begin{pmatrix} \frac{19}{9} \\ \frac{1}{3} \\ \frac{7}{18} \end{pmatrix}. \]

4.3 LU Decomposition Practice Problems

Problem 1: Decompose the matrix \[ A = \begin{pmatrix} 4 & 3 \\ 6 & 3 \end{pmatrix} \] into the product of a lower triangular matrix \(L\) and an upper triangular matrix \(U\).

Problem 2: Given the matrix \[ A = \begin{pmatrix} 1 & 2 & 3 \\ 2 & 5 & 8 \\ 4 & 5 & 6 \end{pmatrix}, \] perform LU decomposition to find matrices \(L\) and \(U\).

Problem 3: Perform LU decomposition of the matrix \[ A = \begin{pmatrix} 2 & 1 & 1 \\ 4 & -6 & 0 \\ -2 & 7 & 2 \end{pmatrix}, \] and verify the decomposition by checking \(A = LU\).

Problem 4: For the matrix \[ A = \begin{pmatrix} 3 & 1 & 6 \\ 2 & 1 & 1 \\ 1 & 2 & 2 \end{pmatrix}, \] find the LU decomposition and use it to solve the system \(Ax = b\) where \(b = \begin{pmatrix} 9 \\ 5 \\ 4 \end{pmatrix}\).

Problem 5: Decompose the matrix \[ A = \begin{pmatrix} 1 & 3 & 1 \\ 2 & 6 & 1 \\ 1 & 1 & 4 \end{pmatrix} \] into \(L\) and \(U\), and solve the system \(Ax = \begin{pmatrix} 5 \\ 9 \\ 6 \end{pmatrix}\).

Problem 6: Given the matrix \[ A = \begin{pmatrix} 7 & 3 \\ 2 & 5 \end{pmatrix}, \] perform LU decomposition and use the result to solve \(Ax = b\) for \(b = \begin{pmatrix} 10 \\ 7 \end{pmatrix}\).

Problem 7: Find the LU decomposition of the matrix \[ A = \begin{pmatrix} 2 & -1 & 1 \\ -2 & 2 & -1 \\ 4 & -1 & 3 \end{pmatrix}, \] and use it to solve \(Ax = b\) where \(b = \begin{pmatrix} 1 \\ -1 \\ 7 \end{pmatrix}\).

Problem 8: Perform LU decomposition of the matrix \[ A = \begin{pmatrix} 5 & 2 & 1 \\ 10 & 4 & 3 \\ 15 & 8 & 6 \end{pmatrix}. \]

Problem 9: Use LU decomposition to find the solution to the system \(Ax = b\) where \[ A = \begin{pmatrix} 1 & 1 & 1 \\ 2 & 3 & 5 \\ 4 & 6 & 8 \end{pmatrix}, \quad b = \begin{pmatrix} 6 \\ 15 \\ 30 \end{pmatrix}. \]

Problem 10: Decompose the matrix \[ A = \begin{pmatrix} 6 & -2 & 2 \\ 12 & -8 & 6 \\ -6 & 3 & -3 \end{pmatrix} \] into \(L\) and \(U\), and verify that \(A = LU\).

4.4 Matrix Approach to Create LU Decomposition

LU decomposition can be performed using elementary matrix operations. In this method, we iteratively apply elementary matrices to reduce the given matrix \(A\) into an upper triangular matrix \(U\), while keeping track of the transformations to form the lower triangular matrix \(L\).

The LU decomposition can be written as: \[A = LU\]

where: - \(L\) is the product of the inverses of the elementary matrices. - \(U\) is the upper triangular matrix obtained after applying the row operations.

Example: LU Decomposition of a 3x3 Matrix

Given the matrix: \[ A = \begin{pmatrix} 2 & 1 & 1 \\ 4 & -6 & 0 \\ -2 & 7 & 2 \end{pmatrix} \]

We will decompose \(A\) into \(L\) and \(U\) using elementary row operations.

Step 1: Applying Elementary Matrices

We want to perform row operations to reduce \(A\) into upper triangular form.

Step 1.1: Eliminate the \(a_{21}\) entry (below the pivot in column 1)

To eliminate the \(4\) in position \(a_{21}\), perform the operation: \[R_2 \rightarrow R_2 - 2R_1\]

This corresponds to multiplying \(A\) by the elementary matrix: \[E_1 = \begin{pmatrix}1 & 0 & 0 \\-2 & 1 & 0 \\0 & 0 & 1\end{pmatrix}\]

After this row operation, the matrix becomes: \[ E_1 A = \begin{pmatrix} 2 & 1 & 1 \\ 0 & -8 & -2 \\ -2 & 7 & 2 \end{pmatrix} \]

Step 1.2: Eliminate the \(a_{31}\) entry

To eliminate the \(-2\) in position \(a_{31}\), perform the operation: \[R_3 \rightarrow R_3 + R_1\]

This corresponds to multiplying the matrix by another elementary matrix: \[ E_2 = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 1 & 0 & 1 \end{pmatrix} \]

Now, the matrix becomes: \[ E_2 E_1 A = \begin{pmatrix} 2 & 1 & 1 \\ 0 & -8 & -2 \\ 0 & 8 & 3 \end{pmatrix} \]

Step 1.3: Eliminate the \(a_{32}\) entry

Finally, to eliminate the \(8\) in position \(a_{32}\), perform the operation: \[ R_3 \rightarrow R_3 + R_2 \]

This corresponds to multiplying the matrix by the third elementary matrix:

\[ E_3 = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 1 & 1 \end{pmatrix} \]

After applying this operation, the matrix becomes: \[ E_3 E_2 E_1 A = \begin{pmatrix} 2 & 1 & 1 \\ 0 & -8 & -2 \\ 0 & 0 & 1 \end{pmatrix} \]

This is the upper triangular matrix \(U\).

Step 2: Construct the Lower Triangular Matrix \(L\)

The lower triangular matrix \(L\) is formed by taking the inverses of the elementary matrices \(E_1, E_2, E_3\). Each inverse corresponds to the inverse of the row operations we applied.

  • \(E_1^{-1}\) corresponds to adding back \(2R_1\) to \(R_2\), so: \[ E_1^{-1} = \begin{pmatrix} 1 & 0 & 0 \\ 2 & 1 & 0 \\ 0 & 0 & 1 \end{pmatrix} \]

  • \(E_2^{-1}\) corresponds to subtracting \(R_1\) from \(R_3\), so: \[ E_2^{-1} = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ -1 & 0 & 1 \end{pmatrix} \]

  • \(E_3^{-1}\) corresponds to subtracting \(R_2\) from \(R_3\), so: \[ E_3^{-1} = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & -1 & 1 \end{pmatrix} \]

Now, the lower triangular matrix \(L\) is obtained by multiplying these inverses in reverse order: \[ L = E_3^{-1} E_2^{-1} E_1^{-1} = \begin{pmatrix} 1 & 0 & 0 \\ 2 & 1 & 0 \\ -1 & -1 & 1 \end{pmatrix} \]

Thus, the LU decomposition of \(A\) is: \[ L = \begin{pmatrix} 1 & 0 & 0 \\ 2 & 1 & 0 \\ -1 & -1 & 1 \end{pmatrix}, \quad U = \begin{pmatrix} 2 & 1 & 1 \\ 0 & -8 & -2 \\ 0 & 0 & 1 \end{pmatrix} \]

Verification

Now, we check if \(A = LU\).

Multiply \(L\) and \(U\):

import numpy as np

L = np.array([[1, 0, 0],
              [2, 1, 0],
              [-1, -1, 1]])

U = np.array([[2, 1, 1],
              [0, -8, -2],
              [0, 0, 1]])

A = L @ U
A
array([[ 2,  1,  1],
       [ 4, -6,  0],
       [-2,  7,  2]])

5 Spectral Decomposition

5.1 Background

Imagine encountering a low-resolution image of a familiar scene. The human brain excels at recognizing familiar objects by relying on essential features, often extracting the most significant details while discarding the less important information. This cognitive process mirrors the power of eigenvalue decomposition, where eigenvectors represent the ``nectar’’ of a matrix, capturing its most important characteristics.

As an example, try to identify this image. If you can do it, then your brain know this place!

Before proceeding further just compare the size of its’ original clean image and the low-quality image shown in Figure

Original image size: 985.69 KB
Reconstructed image size: 1.12 KB

The reconstructed image is just 0.2% of the original in size! This is the core principle of optimizing image storage of CCTV system. This resizing can be done and execute with optimal scaling with the help of Linear Algebra. This module mainly focuses on such engineering applications.

5.2 Introduction

Spectral decomposition, also known as eigenvalue decomposition, is a powerful tool in computational linear algebra that breaks down a matrix into its eigenvalues and eigenvectors. This technique allows matrices to be represented in terms of their fundamental components, making it easier to analyze and manipulate them. It is especially useful for symmetric matrices, which are common in various applications. Spectral decomposition facilitates solving systems of equations, optimizing functions, and performing transformations in a simplified, structured manner, as it allows operations to be performed on the eigenvalues, which often leads to more efficient computations.

The importance of spectral decomposition extends across a wide range of fields, including computer science, engineering, and data science. In machine learning, for instance, it forms the backbone of algorithms like Principal Component Analysis (PCA), which is used for dimensionality reduction. It also plays a vital role in numerical stability when dealing with large matrices and is central to many optimization problems, such as those found in machine learning and physics. Spectral decomposition not only provides a deeper understanding of the properties of matrices but also offers practical benefits in improving the efficiency and accuracy of numerical algorithms.

5.3 Spectral Decomposition: Detailed Concepts

5.3.1 Eigenvalues and Eigenvectors

The core idea behind spectral decomposition is that it expresses a matrix in terms of its eigenvalues and eigenvectors. For a square matrix \(A \in \mathbb{R}^{n \times n}\), an eigenvalue \(\lambda \in \mathbb{R}\) and an eigenvector \(v \in \mathbb{R}^{n}\) satisfy the following equation:

\[ A v = \lambda v \]

This implies that when the matrix \(A\) acts on the vector \(v\), it only scales the vector by \(\lambda\), but does not change its direction. The eigenvector \(v\) represents the direction of this scaling, while the eigenvalue \(\lambda\) represents the magnitude of the scaling.

Properties of Eigen values
  • If \(\lambda\) is an eigenvalue of \(A\), then it satisfies the characteristic polynomial:

    \[ p(\lambda) = \text{det}(A - \lambda I) = 0. \]

  • The sum of the eigenvalues (counted with algebraic multiplicity) is equal to the trace of the matrix:

    \[ \sum_{i=1}^{n} \lambda_i = \text{trace}(A). \]

  • The product of the eigenvalues (counted with algebraic multiplicity) is equal to the determinant of the matrix:

    \[ \prod_{i=1}^{n} \lambda_i = \text{det}(A). \]

  • If\(A\) is symmetric, then:

    • All eigenvalues \(\lambda\) are real.
    • If \(\lambda_i\) and \(\lambda_j\) are distinct eigenvalues, then their corresponding eigenvectors \(\mathbf{v}_i\) and \(\mathbf{v}_j\) satisfy:

    \[ \mathbf{v}_i^T \mathbf{v}_j = 0. \]

  • If \(A\) is a scalar multiple of \(k\), then:

    \[ \lambda_i \text{ of } kA = k \cdot \lambda_i \text{ of } A. \]

  • If \(A\) is invertible, then:

    \[ \lambda_i \text{ of } A^{-1} = \frac{1}{\lambda_i \text{ of } A}. \]

  • If \(A\) and \(B\) are similar, then:

    \[ B = P^{-1} A P \implies \lambda_i \text{ of } B = \lambda_i \text{ of } A. \]

  • If \(\lambda\) is an eigenvalue, it has:

    • Algebraic Multiplicity: The number of times \(\lambda\) appears as a root of \(p(\lambda)\).
    • Geometric Multiplicity: The dimension of the eigenspace \(E_{\lambda} = \{\mathbf{v} : A\mathbf{v} = \lambda \mathbf{v}\}\).
  • If\(A\) is symmetric and all eigenvalues \(\lambda\) are positive, then \(A\) is positive definite:

    \[ \lambda_i > 0 \implies A \text{ is positive definite.} \]

  • A square matrix \(A\) has an eigenvalue \(\lambda = 0\) if and only if \(A\) is singular:

    \[ \text{det}(A) = 0 \iff \lambda = 0. \]

Eigen Vectors

Eigen vectors are the non-trivial solutions of \(det(A-\lambda I)=0\) for distinct \(\lambda\).

Properties of Eigen vectors
  • If \(\mathbf{v}\) is an eigenvector of a square matrix \(A\) corresponding to the eigenvalue \(\lambda\), then:

    \[ A\mathbf{v} = \lambda \mathbf{v}. \]

  • Eigenvectors corresponding to distinct eigenvalues are linearly independent. If \(\lambda_1\) and \(\lambda_2\) are distinct eigenvalues of \(A\), with corresponding eigenvectors \(\mathbf{v}_1\) and \(\mathbf{v}_2\), then:

    \[ c_1 \mathbf{v}_1 + c_2 \mathbf{v}_2 = \mathbf{0} \implies c_1 = 0 \text{ and } c_2 = 0. \]

  • If \(\mathbf{v}\) is an eigenvector corresponding to the eigenvalue \(\lambda\), then any non-zero scalar multiple of \(\mathbf{v}\) is also an eigenvector corresponding to \(\lambda\):

    \[ \text{If } \mathbf{v} \text{ is an eigenvector, then } c\mathbf{v} \text{ is an eigenvector for any non-zero scalar } c. \]

  • The eigenspace \(E_{\lambda}\) associated with an eigenvalue \(\lambda\) is defined as:

    \[ E_{\lambda} = \{ \mathbf{v} : A\mathbf{v} = \lambda \mathbf{v} \} = \text{Null}(A - \lambda I). \]

  • The dimension of the eigenspace \(E_{\lambda}\) is equal to the geometric multiplicity of the eigenvalue \(\lambda\).

  • If \(A\) is a symmetric matrix, then eigenvectors corresponding to distinct eigenvalues are orthogonal:

    \[ \mathbf{v}_i^T \mathbf{v}_j = 0 \text{ for distinct eigenvalues } \lambda_i \text{ and } \lambda_j. \]

  • For any square matrix \(A\), if \(\lambda = 0\) is an eigenvalue, the eigenvectors corresponding to this eigenvalue form the null space of \(A\):

    \[ E_{0} = \{ \mathbf{v} : A\mathbf{v} = \mathbf{0} \} = \text{Null}(A). \]

  • If\(A\) is invertible, then \(A\) has no eigenvalue equal to zero, meaning all eigenvectors correspond to non-zero eigenvalues.

  • For\(A\) as a scalar multiple of \(k\):

    \[ A\mathbf{v} = k \lambda \mathbf{v} \text{ for eigenvalue } \lambda. \]

5.3.2 Eigenvalue Decomposition (Spectral Decomposition)

For matrices that are diagonalizable (including symmetric matrices), spectral decomposition expresses the matrix as a combination of its eigenvalues and eigenvectors. Specifically, for a matrix \(A\), spectral decomposition is represented as:

\[ A = V \Lambda V^{-1} \]

where: - \(V\) is the matrix of eigenvectors of \(A\), - \(\Lambda\) is a diagonal matrix of eigenvalues of \(A\), - \(V^{-1}\) is the inverse of the matrix of eigenvectors (if \(V\) is invertible).

For symmetric matrices \(A\), the decomposition becomes simpler:

\[ A = Q \Lambda Q^\top \]

Here, \(Q\) is an orthogonal matrix of eigenvectors (i.e., \(Q^\top Q = I\)), and \(\Lambda\) is a diagonal matrix of eigenvalues.

5.3.3 Geometric Interpretation

Eigenvalues and eigenvectors provide insights into the geometry of linear transformations represented by matrices. Eigenvectors represent directions that remain invariant under the transformation, while eigenvalues indicate how these directions are stretched or compressed.

For example, in the case of a transformation matrix that scales or rotates data points, eigenvalues show the magnitude of scaling along the principal axes (directions defined by eigenvectors).

5.3.4 Importance of Diagonalization

The key advantage of spectral decomposition is that it simplifies matrix operations. When a matrix is diagonalized as \(A = Q \Lambda Q^\top\), any function of the matrix \(A\) (such as powers, exponentials, or inverses) can be easily computed by operating on the diagonal matrix \(\Lambda\). For example:

\[ A^k = Q \Lambda^k Q^\top \]

Since \(\Lambda\) is diagonal, raising \(\Lambda\) to any power \(k\) is straightforward, involving only raising each eigenvalue to the power \(k\).

5.3.5 Properties of Symmetric Matrices

Spectral decomposition applies particularly well to symmetric matrices, which satisfy \(A = A^\top\). Symmetric matrices have the following key properties:

  • Real eigenvalues: The eigenvalues of a symmetric matrix are always real numbers.

  • Orthogonal eigenvectors: The eigenvectors corresponding to distinct eigenvalues of a symmetric matrix are orthogonal to each other.

  • Diagonalizability: Every symmetric matrix can be diagonalized by an orthogonal matrix.

These properties make symmetric matrices highly desirable in computational applications.

5.4 Mathematical Requirements for Spectral Decomposition

5.4.1 Determining Eigenvalues and Eigenvectors

The eigenvalues of a matrix \(A\) are the solutions to the characteristic equation:

\[ \text{det}(A - \lambda I) = 0 \]

Here, \(I\) is the identity matrix, and \(\lambda\) represents the eigenvalues. Solving this polynomial equation provides the eigenvalues \(\lambda_1, \lambda_2, \dots, \lambda_n\). Once the eigenvalues are determined, the eigenvectors can be computed by solving the equation \((A - \lambda I)v = 0\) for each eigenvalue.

5.4.2 Characteristic Polynomial of \(2 \times 2\) Matrices

For a \(2 \times 2\) matrix: \[ A = \begin{pmatrix} a & b \\ c & d \end{pmatrix} \] the characteristic polynomial is derived from the determinant of \(A - \lambda I\), where \(I\) is the identity matrix:

\[ \det(A - \lambda I) = 0 \]

This leads to: \[ \det\begin{pmatrix} a - \lambda & b \\ c & d - \lambda \end{pmatrix} = (a - \lambda)(d - \lambda) - bc = 0 \]

The characteristic polynomial can be simplified to: \[ \lambda^2 - (a + d)\lambda + (ad - bc) = 0 \]

This polynomial can be solved using the quadratic formula: \[ \lambda = \frac{(a + d) \pm \sqrt{(a + d)^2 - 4(ad - bc)}}{2} \]

Shortcut to write Characteristic polynomial of a \(2\times 2\) matrix

If \(A=\begin{bmatrix} a & b\\ c& d\end{bmatrix}\), then the characteristic polynomial is \[\lambda^2-(\text{Trace}(A))\lambda+det(A)=0\]

Eigen vectors can be found by using the formula: \[\begin{equation*} EV(\lambda=\lambda_1)=\begin{bmatrix}\lambda_1-d\\ c\end{bmatrix} \end{equation*}\]

5.4.3 Problems

Example 1: Find Eigenvalues and Eigenvectors of the matrix, \[A = \begin{pmatrix} 3 & 2 \\ 4 & 1 \end{pmatrix}\]

Solution:

The characteristic equation is given by \[det(A-\lambda I)=0\]

\[\begin{align*} \lambda^2 - 4\lambda - 5 &= 0\\ (\lambda-5)(\lambda+1)&=0\\ \end{align*}\]

Hence the eigen values are \(\lambda_1=5,\quad \lambda_2=-1\).

So the eigen vectors are: \[\begin{align*} EV(\lambda=\lambda_1)&=\begin{bmatrix}\lambda_1-d\\ c\end{bmatrix}\\ \therefore EV(\lambda=5)&=\begin{bmatrix}4\\ 4\end{bmatrix}=\begin{bmatrix}1\\ 1\end{bmatrix}\\ \therefore EV(\lambda=-1)&=\begin{bmatrix}-2\\ 4\end{bmatrix}=\begin{bmatrix}-1\\ 2\end{bmatrix} \end{align*}\]

Problem 2: Calculate the eigenvalues and eigenvectors of the matrix: \(A = \begin{pmatrix} 2 & 1 \\ 1 & 2 \end{pmatrix}\)

Solution:

To find the eigenvalues and eigenvectors of a \(2 \times 2\) matrix, we can use the shortcut formula for the characteristic polynomial:

\[ \lambda^2 - \text{trace}(A)\lambda + \det(A) = 0, \]

where \(A\) is the matrix. Let’s apply this to the matrix

\[ A = \begin{pmatrix} 2 & 1 \\ 1 & 2 \end{pmatrix}. \]

First, we calculate the trace and determinant of \(A\):

  • The trace is the sum of the diagonal elements:

\[ \text{trace}(A) = 2 + 2 = 4. \]

  • The determinant is calculated as follows:

\[ \det(A) = (2)(2) - (1)(1) = 4 - 1 = 3. \]

Next, substituting the trace and determinant into the characteristic polynomial gives:

\[ \lambda^2 - (4)\lambda + 3 = 0, \]

which simplifies to:

\[ \lambda^2 - 4\lambda + 3 = 0. \]

We can factor this quadratic equation:

\[ (\lambda - 1)(\lambda - 3) = 0. \]

Setting each factor to zero gives the eigenvalues:

\[ \lambda_1 = 1, \quad \lambda_2 = 3. \]

To find the eigenvectors corresponding to each eigenvalue, we use the shortcut for the eigenvector of a \(2 \times 2\) matrix \(A = \begin{pmatrix} a & b \\ c & d \end{pmatrix}\):

\[ EV(\lambda) = \begin{pmatrix} \lambda - d \\ c \end{pmatrix}. \]

For the eigenvalue \(\lambda_1 = 1\):

\[ EV(1) = \begin{pmatrix} 1 - 2 \\ 1 \end{pmatrix} = \begin{pmatrix} -1 \\ 1 \end{pmatrix}. \]

This eigenvector can be simplified (up to a scalar multiple) to:

\[ \mathbf{v_1} = \begin{pmatrix} 1 \\ -1 \end{pmatrix}. \]

For the eigenvalue \(\lambda_2 = 3\):

\[ EV(3) = \begin{pmatrix} 3 - 2 \\ 1 \end{pmatrix} = \begin{pmatrix} 1 \\ 1 \end{pmatrix}. \]

This eigenvector is already in a simple form:

\[ \mathbf{v_2} = \begin{pmatrix} 1 \\ 1 \end{pmatrix}. \]

Problem 3: For the matrix: \(A = \begin{pmatrix} 1 & 2 & 1 \\ 0 & 1 & 0 \\ 1 & 0 & 1 \end{pmatrix}\), find the eigenvalues and eigenvectors.

Solution:

We are given the matrix \[ A = \begin{pmatrix} 1 & 2 & 1 \\ 0 & 1 & 0 \\ 1 & 0 & 1 \end{pmatrix} \]

and we aim to find its eigenvalues using the characteristic polynomial.

The shortcut formula for the characteristic polynomial of a \(3 \times 3\) matrix is given by: \[ \lambda^3 - \text{tr}(A)\lambda^2 + (\text{sum of principal minors of } A)\lambda - \det(A) = 0. \]

The trace of a matrix is the sum of its diagonal elements. For matrix \(A\), we have: \[ \text{tr}(A) = 1 + 1 + 1 = 3. \]

The principal minors are the determinants of the \(2 \times 2\) submatrices obtained by deleting one row and one column of \(A\).

The first minor is obtained by deleting the third row and third column: \[ \det\begin{pmatrix} 1 & 2 \\ 0 & 1 \end{pmatrix} = (1)(1) - (2)(0) = 1. \]

The second minor is obtained by deleting the second row and second column: \[ \det\begin{pmatrix} 1 & 1 \\ 1 & 1 \end{pmatrix} = (1)(1) - (1)(1) = 0. \]

The third minor is obtained by deleting the first row and first column: \[ \det\begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} = (1)(1) - (0)(0) = 1. \]

Thus, the sum of the principal minors is: \[ 1 + 0 + 1 = 2. \]

The determinant of \(A\) can be calculated using cofactor expansion along the first row: \[ \det(A) = 1 \cdot \det\begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} - 2 \cdot \det\begin{pmatrix} 0 & 0 \\ 1 & 1 \end{pmatrix} + 1 \cdot \det\begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix} \] \[ = 1 \cdot (1) - 2 \cdot (0) + 1 \cdot (-1) = 1 - 0 - 1 = 0. \]

Now, we substitute these values into the characteristic polynomial formula: \[ \lambda^3 - \text{tr}(A)\lambda^2 + (\text{sum of principal minors})\lambda - \det(A) = 0 \] \[ \lambda^3 - 3\lambda^2 + 2\lambda - 0 = 0. \]

We now solve the equation: \[ \lambda^3 - 3\lambda^2 + 2\lambda = 0. \] Factoring out \(\lambda\) and apply factor theorem, we get:

\[\begin{align*} \lambda(\lambda^2 - 3\lambda + 2) &= 0\\ \lambda(\lambda-2)(\lambda-1)&=0 \end{align*}\]

This gives one eigenvalue: \[ \lambda_1 = 0;\quad \lambda_2=2;\quad \lambda_3=1 \]

Now we find the eigenvectors corresponding to each eigenvalue.

For \(\lambda_1 = 0\), solve \((A - 0I)\mathbf{v} = 0\): \[ \begin{pmatrix} 1 & 2 & 1 \\ 0 & 1 & 0 \\ 1 & 0 & 1 \end{pmatrix} \begin{pmatrix} x \\ y \\ z \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \\ 0 \end{pmatrix}. \] This gives the system: \[ x + 2y + z = 0, \quad y = 0, \quad x + z = 0. \] Thus, \(x = -z\), and the eigenvector is: \[ \mathbf{v}_1 = \begin{pmatrix} -1 \\ 0 \\ 1 \end{pmatrix}. \]

For \(\lambda_2 = 2\), solve \((A - 2I)\mathbf{v} = 0\): \[ \begin{pmatrix} -1 & 2 & 1 \\ 0 & -1 & 0 \\ 1 & 0 & -1 \end{pmatrix} \begin{pmatrix} x \\ y \\ z \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \\ 0 \end{pmatrix}. \] This gives the system: \[ -x + 2y + z = 0, \quad -y = 0, \quad x - z = 0. \] Thus, \(x = z\), and the eigenvector is: \[ \mathbf{v}_2 = \begin{pmatrix} 1 \\ 0 \\ 1 \end{pmatrix}. \]

For \(\lambda_3 = 1\), solve \((A - I)\mathbf{v} = 0\): \[ \begin{pmatrix} 0 & 2 & 1 \\ 0 & 0 & 0 \\ 1 & 0 & 0 \end{pmatrix} \begin{pmatrix} x \\ y \\ z \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \\ 0 \end{pmatrix}. \] This gives the system: \[ 2y + z = 0, \quad x = 0. \] Thus, \(z = -2y\), and the eigenvector is: \[ \mathbf{v}_3 = \begin{pmatrix} 0 \\ 1 \\ -2 \end{pmatrix}. \]

Problem 3: If \(A=\begin{bmatrix}1&2&4\\ 0&3&4\\ 1&-1&-1 \end{bmatrix}\), compute the eigen values and eigen vectors and left eigen vectors of \(A\).

Solution:

We are given the matrix \[ A = \begin{pmatrix} 1 & 2 & 4 \\ 0 & 3 & 4 \\ 1 & -1 & -1 \end{pmatrix} \]

and need to find its eigenvalues and eigenvectors.

The characteristic polynomial for a \(3 \times 3\) matrix is given by: \[ \lambda^3 - \text{tr}(A)\lambda^2 + (\text{sum of principal minors})\lambda - \det(A) = 0. \]

The trace is the sum of the diagonal elements: \[ \text{tr}(A) = 1 + 3 + (-1) = 3. \]

We now compute the \(2 \times 2\) principal minors:

  • Minor by removing the third row and third column: \[ \det\begin{pmatrix} 1 & 2 \\ 0 & 3 \end{pmatrix} = (1)(3) - (2)(0) = 3. \]

  • Minor by removing the second row and second column: \[ \det\begin{pmatrix} 1 & 4 \\ 1 & -1 \end{pmatrix} = (1)(-1) - (4)(1) = -1 - 4 = -5. \]

  • Minor by removing the first row and first column: \[ \det\begin{pmatrix} 3 & 4 \\ -1 & -1 \end{pmatrix} = (3)(-1) - (4)(-1) = -3 + 4 = 1. \]

Thus, the sum of the principal minors is: \[ 3 + (-5) + 1 = -1. \]

We calculate the determinant of \(A\) by cofactor expansion along the first row: \[ \det(A) = 1 \cdot \det\begin{pmatrix} 3 & 4 \\ -1 & -1 \end{pmatrix} - 2 \cdot \det\begin{pmatrix} 0 & 4 \\ 1 & -1 \end{pmatrix} + 4 \cdot \det\begin{pmatrix} 0 & 3 \\ 1 & -1 \end{pmatrix}. \] The \(2 \times 2\) determinants are: \[ \det\begin{pmatrix} 3 & 4 \\ -1 & -1 \end{pmatrix} = -3 + 4 = 1, \quad \det\begin{pmatrix} 0 & 4 \\ 1 & -1 \end{pmatrix} = -4, \] \[ \det\begin{pmatrix} 0 & 3 \\ 1 & -1 \end{pmatrix} = -3. \]

Thus: \[ \det(A) = 1 \cdot 1 - 2 \cdot (-4) + 4 \cdot (-3) = 1 + 8 - 12 = -3. \]

Substituting into the characteristic polynomial: \[ \lambda^3 - \text{tr}(A)\lambda^2 + (\text{sum of principal minors})\lambda - \det(A) = 0, \]

we get: \[ \lambda^3 - 3\lambda^2 - \lambda + 3 = 0. \]

We now solve the cubic equation: \[\begin{align*} \lambda^3 - 3\lambda^2 - \lambda + 3& = 0. \\ (\lambda-1)(\lambda+1)(\lambda -3)&=0 \end{align*}\]

\[\lambda_1 = 1, \quad \lambda_2 = -1, \quad \lambda_3 = 3.\]

To find the eigenvector corresponding to \(\lambda_1 = 3\), solve \((A - 3I)\mathbf{v} = 0\): \[ A - 3I = \begin{pmatrix} 1 & 2 & 4 \\ 0 & 3 & 4 \\ 1 & -1 & -1 \end{pmatrix} - 3\begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{pmatrix} = \begin{pmatrix} -2 & 2 & 4 \\ 0 & 0 & 4 \\ 1 & -1 & -4 \end{pmatrix}. \]

Solving this system gives the eigenvector: \[ \mathbf{v}_1 = \begin{pmatrix} 1 \\ 1 \\ 0 \end{pmatrix}. \]

For \(\lambda_2 = -1\), solve \((A +I)\mathbf{v} = 0\): \[ A +I = \begin{pmatrix} 1 & 2 & 4 \\ 0 & 3 & 4 \\ 1 & -1 & -1 \end{pmatrix} +\begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{pmatrix} = \begin{pmatrix} 2 & 2 & 4 \\ 0 & 4 & 4 \\ 1 & -1 & 0 \end{pmatrix}. \]

Note that the third row is depending on first and second rows. So by finding the cross product of first two rows,

\[ \mathbf{v}_2 = \begin{pmatrix} -1 \\ -1 \\ 1 \end{pmatrix}. \]

For \(\lambda_3 = 1\), solve \((A -I)\mathbf{v} = 0\): \[ A - I = \begin{pmatrix} 1 & 2 & 4 \\ 0 & 3 & 4 \\ 1 & -1 & -1 \end{pmatrix} -\begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{pmatrix} = \begin{pmatrix} 0 & 2 & 4 \\ 0 & 2 & 4 \\ 1 & -1 & -2\end{pmatrix}. \]

Note that the second row is same as first row. So by finding the cross product of first and third rows, \[ \mathbf{v}_3 = \begin{pmatrix} 0 \\ -2 \\ 1 \end{pmatrix}. \]

Thus, the eigenvalues of the matrix are: \[ \lambda_1 = 3, \quad \lambda_2 = -1, \quad \lambda_3 = 1 \]

with corresponding eigenvectors \(\mathbf{v}_1=\begin{pmatrix} 1 \\ 1 \\ 0 \end{pmatrix}\), \(\mathbf{v}_2=\begin{pmatrix} -1 \\ -1 \\ 1 \end{pmatrix}\), and \(\mathbf{v}_3=\begin{pmatrix} 0 \\ -2 \\ 1 \end{pmatrix}\).

Left eigen vectors of the matrix \(A\) are eigen vectors of \(A^T\).

Here \(A^T=\begin{bmatrix} 1&0&1\\ 2&3&-1\\ 4&4&-1 \end{bmatrix}\).

Since \(A\) and \(A^T\) have same eigen values, it is enough to find corresponding eigen vectors. When \(\lambda=3\), the coefficient matrix of \((A-\lambda I)X=0\) reduced into \(\begin{bmatrix} -2&0&1\\ 2&0&-1\\ 4&4&-4 \end{bmatrix}\)

Here the only independent rows are first and last. So the eigen vector can be found as the cross product of these two rows. \(\therefore v_1=\begin{bmatrix} 1\\1\\2 \end{bmatrix}\).

When \(\lambda=-1\), the coefficient matrix of \((A-\lambda I)X=0\) reduced into \(\begin{bmatrix} 2&0&1\\ 2&4&-1\\ 4&4&0 \end{bmatrix}\)

Here the only independent rows are first and second. So the eigen vector can be found as the cross product of these two rows. \(\therefore v_2=\begin{bmatrix} -1\\1\\2 \end{bmatrix}\). When \(\lambda=1\), the coefficient matrix of \((A-\lambda I)X=0\) reduced into \(\begin{bmatrix} 0&0&1\\ 2&2&-1\\ 4&4&-2 \end{bmatrix}\)

Here the only independent rows are first and second. So the eigen vector can be found as the cross product of these two rows. \(\therefore v_2=\begin{bmatrix} -1\\1\\0 \end{bmatrix}\).

5.4.4 Python code to find eigen values and eigen vectors

  1. Find eigen values and eigen vectors of \(A=\begin{bmatrix} 2&1\\ 1&2\end{bmatrix}\).
import numpy as np
from scipy.linalg import null_space

# Define matrix A
A = np.array([[2, 1], 
              [1, 2]])

# Find eigenvalues
eigenvalues, _ = np.linalg.eig(A)

# Define identity matrix I
I = np.eye(A.shape[0])

# Iterate over eigenvalues to find corresponding eigenvectors
for i, eigenvalue in enumerate(eigenvalues):
    # Compute A - lambda * I
    A_lambda_I = A - eigenvalue * I
    
    # Find the null space (which gives the eigenvector)
    eig_vector = null_space(A_lambda_I)
    
    print(f"Eigenvalue {i+1}: {eigenvalue}")
    print(f"Eigenvector {i+1}:\n{eig_vector}\n")
Eigenvalue 1: 3.0
Eigenvector 1:
[[0.70710678]
 [0.70710678]]

Eigenvalue 2: 1.0
Eigenvector 2:
[[-0.70710678]
 [ 0.70710678]]

Same can be done using direct approach. Code for this task is given below.

import numpy as np

# Define matrix A
A = np.array([[2, 1], 
              [1, 2]])

# Find eigenvalues and eigenvectors
eigenvalues, eigenvectors = np.linalg.eig(A)

# Display the results
print("Eigenvalues:", eigenvalues)
print("Eigenvectors:\n", eigenvectors)
Eigenvalues: [3. 1.]
Eigenvectors:
 [[ 0.70710678 -0.70710678]
 [ 0.70710678  0.70710678]]

5.4.5 Diagonalization of Symmetric Matrices

For a symmetric matrix \(A\), the process of diagonalization can be summarized as follows:

  1. Compute eigenvalues: Solve the characteristic equation \(\text{det}(A - \lambda I) = 0\) to find the eigenvalues.

  2. Find eigenvectors: For each eigenvalue \(\lambda_i\), solve \((A - \lambda_i I)v_i = 0\) to find the corresponding eigenvector \(v_i\).

  3. Form the eigenvector matrix: Arrange the eigenvectors into a matrix \(Q\), with each eigenvector as a column.

  4. Form the diagonal matrix of eigenvalues: Construct \(\Lambda\) by placing the eigenvalues along the diagonal of the matrix.

Thus, the matrix can be expressed as \(A = Q \Lambda Q^\top\).

  1. Diagonalize the matrix \(A=\begin{bmatrix} 2&1\\ 1&2\end{bmatrix}\).

Python code for this task is given below.

import numpy as np

# Define matrix A
A = np.array([[2, 1], 
              [1, 2]])

# Step 1: Find eigenvalues and eigenvectors
eigenvalues, eigenvectors = np.linalg.eig(A)

# Step 2: Construct the diagonal matrix D (eigenvalues)
D = np.diag(eigenvalues)

# Step 3: Construct the matrix P (eigenvectors)
P = eigenvectors

# Step 4: Calculate the inverse of P
P_inv = np.linalg.inv(P)

# Verify the diagonalization: A = P D P_inv
A_reconstructed = P @ D @ P_inv

print("Matrix A:")
print(A)

print("\nEigenvalues (Diagonal matrix D):")
print(D)

print("\nEigenvectors (Matrix P):")
print(P)

print("\nInverse of P:")
print(P_inv)

print("\nReconstructed matrix A (P D P^(-1)):")
print(A_reconstructed)
Matrix A:
[[2 1]
 [1 2]]

Eigenvalues (Diagonal matrix D):
[[3. 0.]
 [0. 1.]]

Eigenvectors (Matrix P):
[[ 0.70710678 -0.70710678]
 [ 0.70710678  0.70710678]]

Inverse of P:
[[ 0.70710678  0.70710678]
 [-0.70710678  0.70710678]]

Reconstructed matrix A (P D P^(-1)):
[[2. 1.]
 [1. 2.]]

5.4.6 Matrix Functions and Spectral Theorem

Once a matrix is diagonalized, various matrix functions become easier to compute. For a function \(f(A)\), such as the exponential of a matrix or any power, the function can be applied to the diagonal matrix of eigenvalues:

\[ f(A) = Q f(\Lambda) Q^\top \]

where \(f(\Lambda)\) is the function applied element-wise to the eigenvalues in the diagonal matrix \(\Lambda\).

6 QR Decomposition

The QR decomposition of a matrix is a factorization technique that expresses a matrix \(A\) as the product of an orthogonal matrix \(Q\) and an upper triangular matrix \(R\):

\[ A = QR \]

where: - \(Q\) is an orthogonal matrix (\(Q^T Q = I\)), meaning its columns are orthonormal vectors. - \(R\) is an upper triangular matrix.

Properties
  1. Orthogonality: The columns of \(Q\) are orthonormal, which implies that \(Q^T = Q^{-1}\).
  2. Uniqueness: The QR decomposition is unique if the columns of \(A\) are linearly independent.
Relation with Other Decompositions
  1. LU Decomposition:
    • LU decomposition factors a matrix \(A\) into a lower triangular matrix \(L\) and an upper triangular matrix \(U\): \[ A = LU \]
    • Unlike QR, LU decomposition does not require the columns of \(A\) to be orthogonal.
    • QR decomposition is often used when \(A\) is not square or when numerical stability is a concern, as it can be computed using Gram-Schmidt or Householder reflections.
  2. Cholesky Decomposition (CR):
    • The Cholesky decomposition is a specific case of LU decomposition applicable to symmetric positive-definite matrices: \[ A = LL^T \]
    • It is more efficient than LU decomposition for suitable matrices but does not provide orthogonality like QR.
  3. Spectral Decomposition:
    • The spectral decomposition expresses a symmetric matrix \(A\) in terms of its eigenvalues and eigenvectors: \[ A = Q \Lambda Q^T \]
    • While QR decomposition provides an orthogonal basis for any matrix, spectral decomposition is specifically used for symmetric matrices, providing insights into the matrix’s properties through its eigenvalues and eigenvectors.
  4. Singular Value Decomposition (SVD):
    • The SVD decomposes a matrix \(A\) into three matrices: \[ A = U \Sigma V^T \]
    • \(U\) and \(V\) are orthogonal matrices, and \(\Sigma\) is a diagonal matrix of singular values.
    • SVD is more general than QR and is particularly useful in applications involving rank-deficient matrices, dimensionality reduction, and noise reduction.

6.0.1 Practical Uses of QR Decomposition

  1. Solving Linear Systems:
    • QR decomposition is used to solve linear systems of equations, especially over-determined systems where there are more equations than unknowns. The least squares solution can be efficiently obtained via QR.
  2. Eigenvalue Problems:
    • QR algorithms are often used in iterative methods for finding eigenvalues and eigenvectors of matrices, especially for large matrices.
  3. Numerical Stability:
    • QR decomposition is numerically stable, making it suitable for computations involving floating-point arithmetic, particularly when dealing with ill-conditioned matrices.
  4. Computer Graphics:
    • In computer graphics, QR decomposition can be used in perspective projection, where 3D points are projected onto a 2D plane, often needing orthogonal transformations.
  5. Signal Processing:
    • In signal processing, QR decomposition is utilized in adaptive filtering algorithms and for solving problems related to estimation theory.

6.0.2 Python method for DR decomposition

  1. Find the QR decomposition of the matrix, \(A=\begin{bmatrix}12&-51&4\\ 6&167&-68\\ -4&24&-41\end{bmatrix}\). Verify the decomposition using the reconstruction.
import numpy as np

# Define  A
A = np.array([[12, -51, 4],
              [6, 167, -68],
              [-4, 24, -41]])

# Perform QR decomposition
Q, R = np.linalg.qr(A)

# Display the results
print("Matrix A:")
print(A)

print("\nOrthogonal Matrix Q:")
print(Q)

print("\nUpper Triangular Matrix R:")
print(R)

# Verify the decomposition
print("\nVerification (Q @ R):")
print(np.dot(Q, R))

# Check if Q is orthogonal (Q^T @ Q should be the identity matrix)
print("\nQ^T @ Q (should be identity):")
print(np.dot(Q.T, Q))
Matrix A:
[[ 12 -51   4]
 [  6 167 -68]
 [ -4  24 -41]]

Orthogonal Matrix Q:
[[-0.85714286  0.39428571  0.33142857]
 [-0.42857143 -0.90285714 -0.03428571]
 [ 0.28571429 -0.17142857  0.94285714]]

Upper Triangular Matrix R:
[[ -14.  -21.   14.]
 [   0. -175.   70.]
 [   0.    0.  -35.]]

Verification (Q @ R):
[[ 12. -51.   4.]
 [  6. 167. -68.]
 [ -4.  24. -41.]]

Q^T @ Q (should be identity):
[[ 1.00000000e+00 -5.04131884e-17 -3.39864191e-17]
 [-5.04131884e-17  1.00000000e+00  2.30881074e-17]
 [-3.39864191e-17  2.30881074e-17  1.00000000e+00]]

6.1 Overdetermined Systems

An overdetermined system of linear equations is a system in which there are more equations than unknowns. Mathematically, if we have a matrix \(A\) of size \(m \times n\) where \(m > n\), the system can be represented as:

\[ A \mathbf{x} = \mathbf{b} \]

where: - \(A\) is the coefficient matrix, - \(\mathbf{x}\) is the vector of unknowns (with size \(n\)), - \(\mathbf{b}\) is the vector of constants (with size \(m\)).

6.1.1 Example of an Overdetermined System

Consider the following system of equations:

\[\begin{align*} 2x_1 + 3x_2 &= 5 \\ 4x_1 + 6x_2 &= 10 \\ 1x_1 + 2x_2 &= 3 \\ \end{align*}\]

Here, we have three equations but only two unknowns (\(x_1\) and \(x_2\)). This system is overdetermined.

6.1.2 Challenges in Solving Overdetermined Systems

  1. No Exact Solutions: In most cases, an overdetermined system does not have an exact solution because the equations may be inconsistent. For example, if one equation contradicts another, no value of \(\mathbf{x}\) can satisfy all equations simultaneously.

  2. Finding Best Approximation: When the system is consistent, it may still be that no single solution satisfies all equations perfectly. Therefore, the goal is often to find an approximate solution that minimizes the error.

6.1.3 Why We Need QR Decomposition

QR decomposition is particularly useful for solving overdetermined systems for the following reasons:

  1. Least Squares Solution:
    • The primary goal in solving an overdetermined system is to find the least squares solution, which minimizes the sum of the squared residuals (the differences between the left and right sides of the equations). QR decomposition allows us to efficiently compute this solution.
  2. Orthogonality:
    • The QR decomposition expresses the matrix \(A\) as the product of an orthogonal matrix \(Q\) and an upper triangular matrix \(R\): \[ A = QR \]
    • The orthogonality of \(Q\) ensures numerical stability and helps in reducing the problem to solving triangular systems.
  3. Stability:
    • QR decomposition is more stable than other methods, such as Gaussian elimination, especially when dealing with ill-conditioned matrices. This is crucial in applications where precision is important.
  4. Computational Efficiency:
    • The process of obtaining the QR decomposition can be performed using efficient algorithms, such as Gram-Schmidt orthogonalization or Householder reflections, which makes it suitable for large systems.

6.1.4 Solving an Overdetermined System using QR Decomposition

Given an overdetermined system represented as \(A \mathbf{x} = \mathbf{b}\), the steps to find the least squares solution using QR decomposition are as follows:

  1. Compute QR Decomposition:
    • Decompose the matrix \(A\) into \(Q\) and \(R\).
  2. Formulate the Normal Equations:
    • The least squares solution can be found from the equation: \[ R \mathbf{x} = Q^T \mathbf{b} \]
  3. Solve the Triangular System:
    • Solve for \(\mathbf{x}\) using back substitution, as \(R\) is an upper triangular matrix.

Python code for solving the above system of equations is given below.

import numpy as np

# Define the coefficient matrix A and the constant vector b
A = np.array([[2, 3],
              [4, 6],
              [1, 2]])

b = np.array([5, 10, 3])

# Perform QR decomposition
Q, R = np.linalg.qr(A)

# Calculate the least squares solution
# Solve the equation R * x = Q^T * b
Q_b = np.dot(Q.T, b)
x = np.linalg.solve(R, Q_b)

print("The least squares solution is:")
print(x)
The least squares solution is:
[1. 1.]

6.1.5 Problems

Problem 1: Simple Overdetermined System

Problem Statement:
Solve the overdetermined system given by the equations:

\[\begin{align*} 2x + 3y &= 5 \\ 4x + 6y &= 10 \\ 1x + 2y &= 2 \end{align*}\]

# Problem 1: Simple Overdetermined System

import numpy as np

# Define the coefficient matrix A and the vector b
A1 = np.array([[2, 3],
               [4, 6],
               [1, 2]])

b1 = np.array([5, 10, 2])

# QR decomposition
Q1, R1 = np.linalg.qr(A1)
x_qr1 = np.linalg.solve(R1, Q1.T @ b1)

# Ordinary Least Squares solution using np.linalg.lstsq
x_ols1, residuals1, rank1, s1 = np.linalg.lstsq(A1, b1, rcond=None)

print("Problem 1 - QR Decomposition Solution:", x_qr1)
print("Problem 1 - Ordinary Least Squares Solution:", x_ols1)
Problem 1 - QR Decomposition Solution: [ 4. -1.]
Problem 1 - Ordinary Least Squares Solution: [ 4. -1.]

Problem 2: Overdetermined System with No Exact Solution

Problem Statement:
Solve the following system:

\[\begin{align*} x + 2y &= 3 \\ 2x + 4y &= 6 \\ 3x + 1y &= 5 \end{align*}\]

# Problem 2: Overdetermined System with No Exact Solution

# Define the coefficient matrix A and the vector b
A2 = np.array([[1, 2],
               [2, 4],
               [3, 1]])

b2 = np.array([3, 6, 5])

# QR decomposition
Q2, R2 = np.linalg.qr(A2)
x_qr2 = np.linalg.solve(R2, Q2.T @ b2)

# Ordinary Least Squares solution using np.linalg.lstsq
x_ols2, residuals2, rank2, s2 = np.linalg.lstsq(A2, b2, rcond=None)

print("Problem 2 - QR Decomposition Solution:", x_qr2)
print("Problem 2 - Ordinary Least Squares Solution:", x_ols2)
Problem 2 - QR Decomposition Solution: [1.4 0.8]
Problem 2 - Ordinary Least Squares Solution: [1.4 0.8]

Problem 3: Overdetermined System with Random Data

Problem Statement:
Generate a random overdetermined system and solve it:

\[ Ax = b \]

Where \(A\) is a random \(6 \times 3\) matrix and \(b\) is generated accordingly.

# Problem 3: Overdetermined System with Random Data

# Generate a random overdetermined system
np.random.seed(0)  # For reproducibility
A3 = np.random.rand(6, 3)
x_true = np.array([1, 2, 3])  # True solution
b3 = A3 @ x_true + np.random.normal(0, 0.1, 6)  # Adding some noise

# QR decomposition
Q3, R3 = np.linalg.qr(A3)
x_qr3 = np.linalg.solve(R3, Q3.T @ b3)

# Ordinary Least Squares solution using np.linalg.lstsq
x_ols3, residuals3, rank3, s3 = np.linalg.lstsq(A3, b3, rcond=None)

print("Problem 3 - QR Decomposition Solution:", x_qr3)
print("Problem 3 - Ordinary Least Squares Solution:", x_ols3)
Problem 3 - QR Decomposition Solution: [0.94791379 2.10331498 2.98999875]
Problem 3 - Ordinary Least Squares Solution: [0.94791379 2.10331498 2.98999875]

Problem 4: Real-World Data Fitting

Problem Statement:
Fit a linear model to the following data points:

\[\begin{align*} (1, 2), (2, 3), (3, 5), (4, 7), (5, 11) \end{align*}\]

# Problem 4: Real-World Data Fitting

# Data points
x_data = np.array([1, 2, 3, 4, 5])
y_data = np.array([2, 3, 5, 7, 11])

# Create the design matrix A
A4 = np.vstack([x_data, np.ones(len(x_data))]).T  # Add intercept

# QR decomposition
Q4, R4 = np.linalg.qr(A4)
x_qr4 = np.linalg.solve(R4, Q4.T @ y_data)

# Ordinary Least Squares solution using np.linalg.lstsq
x_ols4, residuals4, rank4, s4 = np.linalg.lstsq(A4, y_data, rcond=None)

print("Problem 4 - QR Decomposition Solution:", x_qr4)
print("Problem 4 - Ordinary Least Squares Solution:", x_ols4)
Problem 4 - QR Decomposition Solution: [ 2.2 -1. ]
Problem 4 - Ordinary Least Squares Solution: [ 2.2 -1. ]

Problem 5: Polynomial Fit (Higher Degree)

Problem Statement:
Fit a quadratic polynomial to the following data points:

\[\begin{align*} (1, 1), (2, 4), (3, 9), (4, 16), (5, 25) \end{align*}\]

# Problem 5: Polynomial Fit (Higher Degree)

# Data points for polynomial fitting
x_data_poly = np.array([1, 2, 3, 4, 5])
y_data_poly = np.array([1, 4, 9, 16, 25])

# Create the design matrix for a quadratic polynomial
A5 = np.vstack([x_data_poly**2, x_data_poly, np.ones(len(x_data_poly))]).T

# QR decomposition
Q5, R5 = np.linalg.qr(A5)
x_qr5 = np.linalg.solve(R5, Q5.T @ y_data_poly)

# Ordinary Least Squares solution using np.linalg.lstsq
x_ols5, residuals5, rank5, s5 = np.linalg.lstsq(A5, y_data_poly, rcond=None)

print("Problem 5 - QR Decomposition Solution:", x_qr5)
print("Problem 5 - Ordinary Least Squares Solution:", x_ols5)
Problem 5 - QR Decomposition Solution: [ 1.00000000e+00 -2.73316113e-15  3.80986098e-15]
Problem 5 - Ordinary Least Squares Solution: [ 1.00000000e+00 -6.77505297e-15  1.06049542e-14]

6.2 Module review

  1. Diagonalize the matrix \(A = \begin{bmatrix} 1 & 1 & 3 \\ 1 & 5 & 1 \\ 3 & 1 & 1 \end{bmatrix}\) and write the spectral decomposition of \(A\).

Hint:
- To diagonalize a matrix, find the eigenvalues and eigenvectors of the matrix. - Compute the eigenvectors of \(A\) to form the matrix \(P\), and diagonalize \(A\) using the formula \(A = P D P^{-1}\) where \(D\) is the diagonal matrix of eigenvalues. - Spectral decomposition expresses the matrix as a sum of eigenvalue-weighted outer products of eigenvectors.


  1. Find the eigenvalues and eigenvectors of the matrix \(A = \begin{bmatrix} 2 & 1 \\ 1 & 2 \end{bmatrix}\) and write the spectral decomposition.

Hint:
- To find the eigenvalues, solve the characteristic equation \(det(A - \lambda I) = 0\). - Once you have the eigenvalues \(\lambda\), find the eigenvectors by solving \((A - \lambda I)v = 0\) for each eigenvalue. - Express \(A\) as a sum of eigenvalue-weighted outer products of its eigenvectors.


  1. Compute the QR decomposition of the matrix \(A = \begin{bmatrix} 4 & 3 \\ 3 & 2 \end{bmatrix}\).

Hint:
- Use Gram-Schmidt process or the numpy.linalg.qr() function to compute the QR decomposition. - The matrix \(A\) is factored into an orthogonal matrix \(Q\) and an upper triangular matrix \(R\) such that \(A = QR\).


  1. Derive the LU decomposition of the matrix \(A = \begin{bmatrix} 4 & 3 & 1 \\ 6 & 3 & 2 \\ 3 & 1 & 2 \end{bmatrix}\).

Hint:
- Use Gaussian elimination to decompose the matrix into lower and upper triangular matrices. - The matrix \(A\) is written as \(A = LU\), where \(L\) is a lower triangular matrix with 1s on the diagonal, and \(U\) is an upper triangular matrix.


  1. Verify the rank-nullity theorem for the matrix \(A = \begin{bmatrix} 7 & -3 & 5 \\ 9 & 11 & 2 \\ 16 & 8 & 7 \end{bmatrix}\).

Hint:
- The rank-nullity theorem states that the sum of the rank and nullity (dimension of the null space) of a matrix equals the number of its columns. - Compute the rank of the matrix \(A\) by finding the number of linearly independent rows or columns, and use this to determine the nullity.


  1. What are the five important matrix decompositions in linear algebra? Compare them in terms of application.

Hint:
- The five important matrix decompositions include: - LU Decomposition: Used for solving systems of linear equations, especially in computational methods. - QR Decomposition: Primarily used in solving least squares problems and orthogonalization. - Eigenvalue Decomposition: Useful in spectral analysis and principal component analysis (PCA). - Singular Value Decomposition (SVD): Crucial for data compression, noise reduction, and dimensionality reduction. - Cholesky Decomposition: Applied in numerical optimization, especially for solving linear systems with symmetric, positive-definite matrices. - Compare their efficiency, use cases, and computational costs in different applications such as image processing, machine learning, etc.


  1. If \(A = \begin{bmatrix} 1 & 2 & 4 \\ 0 & 3 & 4 \\ 1 & -1 & -1 \end{bmatrix}\), find the eigenvalues and eigenvectors of \(A\).

Hint:
- To find the eigenvalues of \(A\), solve the characteristic equation \(det(A - \lambda I) = 0\). - After finding the eigenvalues \(\lambda\), substitute them back into the equation \((A - \lambda I) v = 0\) to find the corresponding eigenvectors. - Eigenvectors should be normalized if required.


  1. Perform the QR decomposition of the matrix \(A = \begin{bmatrix} 4 & 1 \\ 3 & 2 \end{bmatrix}\).

Hint:
- Use Gram-Schmidt process or the numpy.linalg.qr() function to compute the QR decomposition. - The matrix \(A\) is decomposed into \(A = QR\), where \(Q\) is an orthogonal matrix and \(R\) is an upper triangular matrix.


  1. Compute the LU decomposition of the matrix \(A = \begin{bmatrix} 2 & 3 & 1 \\ 3 & 4 & 2 \\ 1 & 2 & 3 \end{bmatrix}\).

Hint:
- Use Gaussian elimination or the scipy.linalg.lu() function to find the LU decomposition. - The matrix \(A\) is decomposed as \(A = LU\), where \(L\) is a lower triangular matrix with 1s on the diagonal and \(U\) is an upper triangular matrix.


  1. Find the Singular Value Decomposition (SVD) of the matrix \(A = \begin{bmatrix} 1 & 2 & 3 \\ 4 & 5 & 6 \\ 7 & 8 & 9 \end{bmatrix}\).

Hint:
- Use the numpy.linalg.svd() function to compute the SVD of matrix \(A\). - The SVD decomposes \(A\) into \(A = U \Sigma V^T\), where \(U\) and \(V\) are orthogonal matrices and \(\Sigma\) is a diagonal matrix containing the singular values.