In the following we want to study how to use numpy and scipy to solve Linear Systems with Python. Most of the functions in numpy and scipy for Linear Algebra are contained in the sub-packages np.linalg and scipy.linalg, as you will see in the following.
To fix the notation, given a matrix AβRnΓn and a vector yβRn, solving a linear system means finding (when exists) a vector xβRn such that it solves
Ax=y
This is not hard to do in numpy, since it implements a function np.linalg.solve, taking as input a 2-dimensional array A and a 1-dimensional array y, and returns the solution x to the linear system. In particular:
# Define the problemA = np.array([[1, 1, 1], [2, 1, 2], [0, 0, 1]])y = np.array([0, 1, 0])# Solve the systemx_sol = np.linalg.solve(A, y)print(f"The solution is {x_sol}.")
Testing the accuracy
When the matrix A is ill-conditioned, the solution of a linear system wonβt be correct, since the small perturbations on y introduced by the floating point system will be amplified and the corresponding solution will be dramatically distant to the true solution. To check how accurate our computed solution is to the true solution of the system (i.e. to quantify the amplification of the perturbation on the data), it is common to use the relative error, which is defined as
Clearly, the problem is that if our algorithm fails in recovering the true solution due to the ill-conditioning of the system matrix A, how can we compute the true solution xtrueβ, required to compute E(xtrueβ,x)? The solution is to build a test problem.
Creating a Test Problem
Consider a matrix AβRnΓn and assume we want to test the accuracy of an algorithm solving systems involving A. Fix an n-dimensional vector xtrueββRn, and compute y=Axtrueβ. Clearly, this procedure defines a linear system
Ax=y
of which we know that xtrueβ is a solution, since we built the term b accordingly. Now, when we apply our algorithm to that linear system, we get a solution xsolβ, approximately solving Ax=y. Given that, we can always compute the relative error E(xtrueβ,xsolβ) associated to the solution obtained by the algorithm. Using numpy, this can be simply done as
import numpy as np# Setting up the dimensionn = 10# Creating the test problemA = np.random.randn(n, n) # n x n random matrixx_true = np.ones((n, )) # n-dimensional vector of onesy = A @ x_true # Compute the term y s.t. x_true is a sol.# Solving the system with numpyx_sol = np.solve(A, y)# Computing the accuracyE_rel = np.linalg.norm(x_true - x_sol, 2) / np.linalg.norm(x_true, 2)print(f"The relative error is {E_rel}")
Condition number
You should already know that the conditioning of an nΓn matrix A can be quantified by a term called condition number which, whenever A is invertible, is defined as
kpβ(A)=β£β£Aβ£β£pββ£β£Aβ1β£β£pβ
where pβ₯1 identifies the norm with respect to which the condition number is computed (hence the term p-condition number).
An invertible matrix A is said to be ill-conditioned when its condition number grows exponentially with the dimension of the problem, n.
The condition number is related to the accuracy of the computed solution of a linear system by the following inequality
which implies that the relative error on the computed solution is big whenever k(A) is big. Moreover, note that as a consequence of the formula above, the accuracy of a computed solution is partially a property of the condition number of A itself, meaning that no algorithm is able to compute an accurate solution to an ill-conditioned system.
Computing the p-condition number of a matrix A in numpy is trivial: just using the function np.linalg.cond(A, p) to compute kpβ(A) does the job.
Solving Linear System by Matrix Splitting
As you should know, when the matrix A is unstructured, the linear system Ax=y can be efficiently solved by using LU Decomposition. In particular, with Gaussian elimination algorithm, one can factorize any non-singular matrix AβRnΓn into:
A=PLU
where LβRnΓn is a lower-triangular matrix, UβRnΓn is an upper-triangular matrix with all ones on the diagonal and PβRnΓn is a permutation matrix (i.e. a matrix obtained by permutating the rows of the identity matrix). If the decomposition is computed without pivoting, the permutation matrix equals the identity. Note that the assumption that A is non-singular is not restrictive, since it is a necessary condition for the solvability of Ax=y.
Since P is an orthogonal matrix, Pβ1=PT, thus
A=PLUβΊPTA=LU
Since linear systems of the form
Lx=yΒ andΒ Ux=y
can be efficiently solved by the Forward (Backward) substitution, and the computation of the LU factorization by Gaussian elimination is pretty fast (O(n3) floating point operations), we can use that to solve the former linear system. Indeed,
Ax=yβΊPTAx=PTyβΊLUx=PTy
then, by Forward-Backward substitution, this system can be solved by subsequently solve
Lz=PTyΒ thenΒ Ux=z
whose solution is a solution for Ax=y.
Even if this procedure is automatically performed by the np.linalg.solve function, we can unroll it with the functions scipy.linalg.lu(A) and scipy.linalg.solve_triangular(A, b), whose documentation can be found here and here.
Exercise
Write a function solve(A,y) that takes as input a non-singular matrix AβRnΓn and a vector yβRn and returns the solution xβRn of Ax=y without using np.linalg.lu.
Visualize solution
import numpy as npimport scipy# Define a function that solves the systemdef solve(A, y):# LU factorization of A P, L, U = scipy.linalg.lu(A) # Solve Lz = P.Ty z = scipy.linalg.solve_triangular(L, P.T@y, lower=True) # Solve Ux = z x = scipy.linalg.solve_triangular(U, z) return x