*This preview shows
page 1. Sign up
to
view the full content.*

**Unformatted text preview: **Chapter 2 Linear Equations One of the problems encountered most frequently in scientific computation is the solution of systems of simultaneous linear equations. This chapter covers the solution of linear systems by Gaussian elimination and the sensitivity of the solution to errors in the data and roundoff errors in the computation. 2.1 Solving Linear Systems
Ax = b. With matrix notation, a system of simultaneous linear equations is written In the most frequent case, when there are as many equations as unknowns, A is a given square matrix of order n, b is a given column vector of n components, and x is an unknown column vector of n components. Students of linear algebra learn that the solution to Ax = b can be written x = A-1 b, where A-1 is the inverse of A. However, in the vast majority of practical computational problems, it is unnecessary and inadvisable to actually compute A-1 . As an extreme but illustrative example, consider a system consisting of just one equation, such as 7x = 21. The best way to solve such a system is by division: x= Use of the matrix inverse would lead to x = 7-1 21 = 0.142857 21 = 2.99997. The inverse requires more arithmetic--a division and a multiplication instead of just a division--and produces a less accurate answer. Similar considerations apply
February 15, 2008 21 = 3. 7 1 2 Chapter 2. Linear Equations to systems of more than one equation. This is even true in the common situation where there are several systems of equations with the same matrix A but different right-hand sides b. Consequently, we shall concentrate on the direct solution of systems of equations rather than the computation of the inverse. 2.2 The MATLAB Backslash Operator To emphasize the distinction between solving linear equations and computing inverses, Matlab has introduced nonstandard notation using backward slash and forward slash operators, "\" and "/". If A is a matrix of any size and shape and B is a matrix with as many rows as A, then the solution to the system of simultaneous equations AX = B is denoted by X = A\B. Think of this as dividing both sides of the equation by the coefficient matrix A. Because matrix multiplication is not commutative and A occurs on the left in the original equation, this is left division. Similarly, the solution to a system with A on the right and B with as many columns as A, XA = B, is obtained by right division, X = B/A. This notation applies even if A is not square, so that the number of equations is not the same as the number of unknowns. However, in this chapter, we limit ourselves to systems with square coefficient matrices. 2.3 A 3-by-3 Example To illustrate the general linear equation solution algorithm, consider an example of order three: 10 -7 0 x1 7 -3 2 6 x2 = 4 . 5 -1 5 x3 6 This, of course, represents the three simultaneous equations -3x1 + 2x2 + 6x3 = 4, 5x1 - x2 + 5x3 = 6. The first step of the solution algorithm uses the first equation to eliminate x1 from the other equations. This is accomplished by adding 0.3 times the first equation 10x1 - 7x2 = 7, 2.3. A 3-by-3 Example 3 to the second equation and subtracting 0.5 times the first equation from the third equation. The coefficient 10 of x1 in the first equation is called the first pivot and the quantities -0.3 and 0.5, obtained by dividing the coefficients of x1 in the other equations by the pivot, are called the multipliers. The first step changes the equations to 10 -7 0 x1 7 0 -0.1 6 x2 = 6.1 . 0 2.5 5 x3 2.5 The second step might use the second equation to eliminate x2 from the third equation. However, the second pivot, which is the coefficient of x2 in the second equation, would be -0.1, which is smaller than the other coefficients. Consequently, the last two equations are interchanged. This is called pivoting. It is not actually necessary in this example because there are no roundoff errors, but it is crucial in general: 10 -7 0 x1 7 0 2.5 5 x2 = 2.5 . 0 -0.1 6 x3 6.1 Now the second pivot is 2.5 and the second equation can be used to eliminate x2 from the third equation. This is accomplished by adding 0.04 times the second equation to the third equation. (What would the multiplier have been if the equations had not been interchanged?) 10 -7 0 x1 7 0 2.5 5 x2 = 2.5 . 0 0 6.2 x3 6.2 The last equation is now 6.2x3 = 6.2. This can be solved to give x3 = 1. This value is substituted into the second equation: 2.5x2 + (5)(1) = 2.5. Hence x2 = -1. Finally, the values of x2 and x3 are substituted into the first equation: 10x1 + (-7)(-1) = 7. Hence x1 = 0. The solution is 0 x = -1 . 1 This solution can be easily checked 10 -7 -3 2 5 -1 using the original equations: 0 0 7 6 -1 = 4 . 5 1 6 4 Chapter 2. Linear Equations The entire algorithm can be compactly expressed in matrix notation. For this example, let 1 0 0 10 -7 0 1 0 0 L = 0.5 1 0 , U = 0 2.5 5 , P = 0 0 1 . -0.3 -0.04 1 0 0 6.2 0 1 0 The matrix L contains the multipliers used during the elimination, the matrix U is the final coefficient matrix, and the matrix P describes the pivoting. With these three matrices, we have LU = P A. In other words, the original coefficient matrix can be expressed in terms of products involving matrices with simpler structure. 2.4 Permutation and Triangular Matrices A permutation matrix is an identity matrix with the rows and columns interchanged. It has exactly one 1 in each row and column; all the other elements are 0. For example, 0 0 0 1 1 0 0 0 P = . 0 0 1 0 0 1 0 0 Multiplying a matrix A on the left by a permutation matrix to give P A permutes the rows of A. Multiplying on the right, AP , permutes the columns of A. Matlab can also use a permutation vector as a row or column index to rearrange the rows or columns of a matrix. Continuing with the P above, let p be the vector p = [4 1 3 2] Then P*A and A(p,:) are equal. The resulting matrix has the fourth row of A as its first row, the first row of A as its second row, and so on. Similarly, A*P and A(:,p) both produce the same permutation of the columns of A. The P*A notation is closer to traditional mathematics, P A, while the A(p,:) notation is faster and uses less memory. Linear equations involving permutation matrices are trivial to solve. The solution to Px = b is simply a rearrangement of the components of b: x = P T b. An upper triangular matrix has all its nonzero elements above or on the main diagonal. A unit lower triangular matrix has ones on the main diagonal and all the 2.5. LU Factorization rest of its nonzero elements below the main 1 2 0 5 U = 0 0 0 0 is upper triangular, and 1 2 L= 3 4 0 1 5 6 diagonal. For example, 3 4 6 7 8 9 0 10 0 0 1 7 0 0 0 1 5 is unit lower triangular. Linear equations involving triangular matrices are also easily solved. There are two variants of the algorithm for solving an n-by-n upper triangular system U x = b. Both begin by solving the last equation for the last variable, then the next-to-last equation for the next-to-last variable, and so on. One subtracts multiples of the columns of U from b. x = zeros(n,1); for k = n:-1:1 x(k) = b(k)/U(k,k); i = (1:k-1)'; b(i) = b(i) - x(k)*U(i,k); end The other uses inner products between the rows of U and portions of the emerging solution x. x = zeros(n,1); for k = n:-1:1 j = k+1:n; x(k) = (b(k) - U(k,j)*x(j))/U(k,k); end 2.5 LU Factorization The algorithm that is almost universally used to solve square systems of simultaneous linear equations is one of the oldest numerical methods, the systematic elimination method, generally named after C. F. Gauss. Research in the period 1955 to 1965 revealed the importance of two aspects of Gaussian elimination that were not emphasized in earlier work: the search for pivots and the proper interpretation of the effect of rounding errors. In general, Gaussian elimination has two stages, the forward elimination and the back substitution. The forward elimination consists of n - 1 steps. At the kth step, multiples of the kth equation are subtracted from the remaining equations to eliminate the kth variable. If the coefficient of xk is "small," it is advisable to 6 Chapter 2. Linear Equations interchange equations before this is done. The elimination steps can be simultaneously applied to the right-hand side, or the interchanges and multipliers saved and applied to the right-hand side later. The back substitution consists of solving the last equation for xn , then the next-to-last equation for xn-1 , and so on, until x1 is computed from the first equation. Let Pk , k = 1, . . . , n - 1, denote the permutation matrix obtained by interchanging the rows of the identity matrix in the same way the rows of A are interchanged at the kth step of the elimination. Let Mk denote the unit lower triangular matrix obtained by inserting the negatives of the multipliers used at the kth step below the diagonal in the kth column of the identity matrix. Let U be the final upper triangular matrix obtained after the n - 1 steps. The entire process can be described by one matrix equation, U = Mn-1 Pn-1 M2 P2 M1 P1 A. It turns out that this equation can be rewritten L1 L2 Ln-1 U = Pn-1 P2 P1 A, where Lk is obtained from Mk by permuting and changing the signs of the multipliers below the diagonal. So, if we let L = L1 L2 Ln-1 , P = Pn-1 P2 P1 , then we have LU = P A. The unit lower triangular matrix L contains all the multipliers used during the elimination and the permutation matrix P accounts for all the interchanges. For our example 10 -7 0 A = -3 2 6, 5 -1 5 the matrices defined during the elimination are 1 0 0 1 0 0 P1 = 0 1 0 , M1 = 0.3 1 0 , 0 0 1 -0.5 0 1 1 0 P2 = 0 0 0 1 0 1 0 0 1 , M2 = 0 1 0. 0 0 0.04 1 The corresponding L's are 1 0 0 1 L1 = 0.5 1 0 , L2 = 0 -0.3 0 1 0 0 0 1 0. -0.04 1 2.6. Why Is Pivoting Necessary? 7 The relation LU = P A is called the LU factorization or the triangular decomposition of A. It should be emphasized that nothing new has been introduced. Computationally, elimination is done by row operations on the coefficient matrix, not by actual matrix multiplication. LU factorization is simply Gaussian elimination expressed in matrix notation. With this factorization, a general system of equations Ax = b becomes a pair of triangular systems Ly = P b, U x = y. 2.6 Why Is Pivoting Necessary? The diagonal elements of U are called pivots. The kth pivot is the coefficient of the kth variable in the kth equation at the kth step of the elimination. In our 3-by-3 example, the pivots are 10, 2.5, and 6.2. Both the computation of the multipliers and the back substitution require divisions by the pivots. Consequently, the algorithm cannot be carried out if any of the pivots are zero. Intuition should tell us that it is a bad idea to complete the computation if any of the pivots are nearly zero. To see this, let us change our example slightly to 10 -7 0 x1 7 -3 2.099 6 x2 = 3.901 . 5 -1 5 x3 6 The (2, 2) element of the matrix has been changed from 2.000 to 2.099, and the right-hand side has also been changed so that the exact answer is still (0, -1, 1)T . Let us assume that the solution is to be computed on a hypothetical machine that does decimal floating-point arithmetic with five significant digits. The first step of the elimination produces 10 -7 0 x1 7 0 -0.001 6 x2 = 6.001 . 0 2.5 5 x3 2.5 The (2, 2) element is now quite small compared with the other elements in the matrix. Nevertheless, let us complete the elimination without using any interchanges. The next step requires adding 2.5 103 times the second equation to the third: On the right-hand side, this involves multiplying 6.001 by 2.5 103 . The result is 1.50025 104 , which cannot be exactly represented in our hypothetical floating-point number system. It must be rounded to 1.5002 104 . The result is then added to 2.5 and rounded again. In other words, both of the 5's shown in italics in (5 + 1.5000 104 )x3 = (2.5 + 1.50025 104 ) (5 + (2.5 103 )(6))x3 = (2.5 + (2.5 103 )(6.001)). 8 Chapter 2. Linear Equations are lost in roundoff errors. On this hypothetical machine, the last equation becomes 1.5005 104 x3 = 1.5004 104 . The back substitution begins with x3 = 1.5004 104 = 0.99993. 1.5005 104 Because the exact answer is x3 = 1, it does not appear that the error is too serious. Unfortunately, x2 must be determined from the equation -0.001x2 + (6)(0.99993) = 6.001, which gives 1.5 10-3 = -1.5. -1.0 10-3 Finally, x1 is determined from the first equation, x2 = 10x1 + (-7)(-1.5) = 7, which gives
T Instead of (0, -1, 1) , we have obtained (-0.35, -1.5, 0.99993)T . Where did things go wrong? There was no "accumulation of rounding error" caused by doing thousands of arithmetic operations. The matrix is not close to singular. The difficulty comes from choosing a small pivot at the second step of the elimination. As a result, the multiplier is 2.5 103 , and the final equation involves coefficients that are 103 times as large as those in the original problem. Roundoff errors that are small if compared to these large coefficients are unacceptable in terms of the original matrix and the actual solution. We leave it to the reader to verify that if the second and third equations are interchanged, then no large multipliers are necessary and the final result is accurate. This turns out to be true in general: If the multipliers are all less than or equal to one in magnitude, then the computed solution can be proved to be satisfactory. Keeping the multipliers less than one in absolute value can be ensured by a process known as partial pivoting. At the kth step of the forward elimination, the pivot is taken to be the largest (in absolute value) element in the unreduced part of the kth column. The row containing this pivot is interchanged with the kth row to bring the pivot element into the (k, k) position. The same interchanges must be done with the elements of the right-hand side b. The unknowns in x are not reordered because the columns of A are not interchanged. x1 = -0.35. 2.7 lutx, bslashtx, lugui We have three functions implementing the algorithms discussed in this chapter. The first function, lutx, is a readable version of the built-in Matlab function lu. 2.7. lutx, bslashtx, lugui 9 There is one outer for loop on k that counts the elimination steps. The inner loops on i and j are implemented with vector and matrix operations, so that the overall function is reasonably efficient. function [L,U,p] = lutx(A) %LU Triangular factorization % [L,U,p] = lutx(A) produces a unit lower triangular % matrix L, an upper triangular matrix U, and a % permutation vector p, so that L*U = A(p,:). [n,n] = size(A); p = (1:n)' for k = 1:n-1 % Find largest element below diagonal in k-th column [r,m] = max(abs(A(k:n,k))); m = m+k-1; % Skip elimination if column is zero if (A(m,k) ~= 0) % Swap pivot row if (m ~= k) A([k m],:) = A([m k],:); p([k m]) = p([m k]); end % Compute multipliers i = k+1:n; A(i,k) = A(i,k)/A(k,k); % Update the remainder of the matrix j = k+1:n; A(i,j) = A(i,j) - A(i,k)*A(k,j); end end % Separate result L = tril(A,-1) + eye(n,n); U = triu(A); Study this function carefully. Almost all the execution time is spent in the statement A(i,j) = A(i,j) - A(i,k)*A(k,j); 10 Chapter 2. Linear Equations At the kth step of the elimination, i and j are index vectors of length n-k. The operation A(i,k)*A(k,j) multiplies a column vector by a row vector to produce a square, rank one matrix of order n-k. This matrix is then subtracted from the submatrix of the same size in the bottom right corner of A. In a programming language without vector and matrix operations, this update of a portion of A would be done with doubly nested loops on i and j. The second function, bslashtx, is a simplified version of the built-in Matlab backslash operator. It begins by checking for three important special cases: lower triangular, upper triangular, and symmetric positive definite. Linear systems with these properties can be solved in less time than a general system. function x = bslashtx(A,b) % BSLASHTX Solve linear system (backslash) % x = bslashtx(A,b) solves A*x = b [n,n] = size(A); if isequal(triu(A,1),zeros(n,n)) % Lower triangular x = forward(A,b); return elseif isequal(tril(A,-1),zeros(n,n)) % Upper triangular x = backsubs(A,b); return elseif isequal(A,A') [R,fail] = chol(A); if ~fail % Positive definite y = forward(R',b); x = backsubs(R,y); return end end If none of the special cases is detected, bslashtx calls lutx to permute and factor the coefficient matrix, then uses the permutation and factors to complete the solution of a linear system. % Triangular factorization [L,U,p] = lutx(A); % Permutation and forward elimination y = forward(L,b(p)); % Back substitution x = backsubs(U,y); 2.8. Effect of Roundoff Errors 11 The bslashtx function employs subfunctions to carry out the solution of lower and upper triangular systems. function x = forward(L,x) % FORWARD. Forward elimination. % For lower triangular L, x = forward(L,b) solves L*x = b. [n,n] = size(L); for k = 1:n j = 1:k-1; x(k) = (x(k) - L(k,j)*x(j))/L(k,k); end function x = backsubs(U,x) % BACKSUBS. Back substitution. % For upper triangular U, x = backsubs(U,b) solves U*x = b. [n,n] = size(U); for k = n:-1:1 j = k+1:n; x(k) = (x(k) - U(k,j)*x(j))/U(k,k); end A third function, lugui, shows the steps in LU decomposition by Gaussian elimination. It is a version of lutx that allows you to experiment with various pivot selection strategies. At the kth step of the elimination, the largest element in the unreduced portion of the kth column is shown in magenta. This is the element that partial pivoting would ordinarily select as the pivot. You can then choose among four different pivoting strategies: Pick a pivot. Use the mouse to pick the magenta element, or any other element, as pivot. Diagonal pivoting. Use the diagonal element as the pivot. Partial pivoting. Same strategy as lu and lutx. Complete pivoting. Use the largest element in the unfactored submatrix as the pivot. The chosen pivot is shown in red and the resulting elimination step is taken. As the process proceeds, the emerging columns of L are shown in green and the emerging rows of U in blue. 2.8 Effect of Roundoff Errors The rounding errors introduced during the solution of a linear system of equations almost always cause the computed solution--which we now denote by x --to differ somewhat from the theoretical solution x = A-1 b. In fact, if the elements of x 12 Chapter 2. Linear Equations are not floating-point numbers, then x cannot equal x. There are two common measures of the discrepancy in x : the error, e = x - x , and the residual, r = b - Ax . Matrix theory tells us that, because A is nonsingular, if one of these is zero, the other must also be zero. But they are not necessarily both "small" at the same time. Consider the following example: 0.780 0.563 x1 0.217 = . 0.913 0.659 x2 0.254 What happens if we carry out Gaussian elimination with partial pivoting on a hypothetical three-digit decimal computer? First, the two rows (equations) are interchanged so that 0.913 becomes the pivot. Then the multiplier 0.780 = 0.854 (to three places) 0.913 is computed. Next, 0.854 times the new first row is subtracted from the new second row to produce the system 0.913 0.659 x1 0.254 = . 0 0.001 x2 0.001 Finally, the back substitution is carried out: 0.001 = 1.00 (exactly), 0.001 0.254 - 0.659x2 x1 = 0.913 = -0.443 (to three places). x2 = Thus the computed solution is x = -0.443 1.000 . To assess the accuracy without knowing the exact answer, we compute the residuals (exactly): 0.217 - ((0.780)(-0.443) + (0.563)(1.00)) r = b - Ax = 0.254 - ((0.913)(-0.443) + (0.659)(1.00)) -0.000460 = . -0.000541 2.8. Effect of Roundoff Errors 13 The residuals are less than 10-3 . We could hardly expect better on a three-digit machine. However, it is easy to see that the exact solution to this system is 1.000 x= . -1.000 So the components of our computed solution actually have the wrong signs; the error is larger than the solution itself. Were the small residuals just a lucky fluke? You should realize that this example is highly contrived. The matrix is very close to being singular and is not typical of most problems encountered in practice. Nevertheless, let us track down the reason for the small residuals. If Gaussian elimination with partial pivoting is carried out for this example on a computer with six or more digits, the forward elimination will produce a system something like 0.913000 0.659000 x1 0.254000 = . 0 -0.000001 x2 0.000001 Notice that the sign of U2,2 differs from that obtained with three-digit computation. Now the back substitution produces 0.000001 = -1.00000, -0.000001 0.254 - 0.659x2 x1 = 0.913 = 1.00000, x2 = the exact answer. On our three-digit machine, x2 was computed by dividing two quantities, both of which were on the order of rounding errors and one of which did not even have the correct sign. Hence x2 can turn out to be almost anything. Then this arbitrary value of x2 was substituted into the first equation to obtain x1 . We can reasonably expect the residual from the first equation to be small-- x1 was computed in such a way as to make this certain. Now comes a subtle but crucial point. We can also expect the residual from the second equation to be small, precisely because the matrix is so close to being singular. The two equations are very nearly multiples of one another, so any pair (x1 , x2 ) that nearly satisfies the first equation will also nearly satisfy the second. If the matrix were known to be exactly singular, we would not need the second equation at all--any solution of the first would automatically satisfy the second. In Figure 2.1, the exact solution is marked with a circle and the computed solution with an asterisk. Even though the computed solution is far from the exact intersection, it is close to both lines because they are nearly parallel. Although this example is contrived and atypical, the conclusion we reached is not. It is probably the single most important fact that we have learned about matrix computation since the invention of the digital computer: Gaussian elimination with partial pivoting is guaranteed to produce small residuals. 14 Chapter 2. Linear Equations 1.5 1 0.5 0 -0.5 -1 -1.5 -1.5 -1 -0.5 0 0.5 1 1.5 2 Figure 2.1. The computed solution, marked by an asterisk, shows a large error, but a small residual. Now that we have stated it so strongly, we must make a couple of qualifying remarks. By "guaranteed" we mean it is possible to prove a precise theorem that assumes certain technical details about how the floating-point arithmetic system works and that establishes certain inequalities that the components of the residual must satisfy. If the arithmetic units work some other way or if there is a bug in the particular program, then the "guarantee" is void. Furthermore, by "small" we mean on the order of roundoff error relative to three quantities: the size of the elements of the original coefficient matrix, the size of the elements of the coefficient matrix at intermediate steps of the elimination process, and the size of the elements of the computed solution. If any of these are "large," then the residual will not necessarily be small in an absolute sense. Finally, even if the residual is small, we have made no claims that the error will be small. The relationship between the size of the residual and the size of the error is determined in part by a quantity known as the condition number of the matrix, which is the subject of the next section. 2.9 Norms and Condition Numbers The coefficients in the matrix and right-hand side of a system of simultaneous linear equations are rarely known exactly. Some systems arise from experiments, and so the coefficients are subject to observational errors. Other systems have coefficients given by formulas that involve roundoff error in their evaluation. Even if the system can be stored exactly in the computer, it is almost inevitabe that roundoff errors will be introduced during its solution. It can be shown that roundoff errors in Gaussian elimination have the same effect on the answer as errors in the original coefficients. 2.9. Norms and Condition Numbers 15 Consequently, we are led to a fundamental question. If perturbations are made in the coefficients of a system of linear equations, how much is the solution altered? In other words, if Ax = b, how can we measure the sensitivity of x to changes in A and b? The answer to this question lies in making the idea of nearly singular precise. If A is a singular matrix, then for some b's a solution x will not exist, while for others it will not be unique. So if A is nearly singular, we can expect small changes in A and b to cause very large changes in x. On the other hand, if A is the identity matrix, then b and x are the same vector. So if A is nearly the identity, small changes in A and b should result in correspondingly small changes in x. At first glance, it might appear that there is some connection between the size of the pivots encountered in Gaussian elimination with partial pivoting and nearness to singularity, because if the arithmetic could be done exactly, all the pivots would be nonzero if and only if the matrix is nonsingular. To some extent, it is also true that if the pivots are small, then the matrix is close to singular. However, when roundoff errors are encountered, the converse is no longer true--a matrix might be close to singular even though none of the pivots are small. To get a more precise, and reliable, measure of nearness to singularity than the size of the pivots, we need to introduce the concept of a norm of a vector. This is a single number that measures the general size of the elements of the vector. The family of vector norms known as lp depends on a parameter p in the range 1 p : n 1/p p xp = |xi | .
i=1 We almost always use p = 1, p = 2, or lim p : x1 = x2 =
n i=1 |xi |, |xi |
2 x = max |xi |.
i n i=1 1/2 , The l1 -norm is also known as the Manhattan norm because it corresponds to the distance traveled on a grid of city streets. The l2 -norm is the familiar Euclidean distance. The l -norm is also known as the Chebyshev norm. The particular value of p is often unimportant and we simply use x. All vector norms have the following basic properties associated with the notion of distance: 0 = 0, cx = |c|x for all scalars c, x + y x + y (the triangle inequality). x > 0 if x = 0, 16 Chapter 2. Linear Equations In Matlab, xp is computed by norm(x,p), and norm(x) is the same as norm(x,2). For example, x = (1:4)/5 norm1 = norm(x,1) norm2 = norm(x) norminf = norm(x,inf) produces x = 0.2000 norm1 = 2.0000 norm2 = 1.0954 norminf = 0.8000 Multiplication of a vector x by a matrix A results in a new vector Ax that can have a very different norm from x. This change in norm is directly related to the sensitivity we want to measure. The range of the possible change can be expressed by two numbers: M = max Ax , x Ax m = min . x 0.4000 0.6000 0.8000 The max and min are taken over all nonzero vectors x. Note that if A is singular, then m = 0. The ratio M/m is called the condition number of A: (A) = max Ax x min Ax x . The actual numerical value of (A) depends on the vector norm being used, but we are usually only interested in order of magnitude estimates of the condition number, so the particular norm is usually not very important. Consider a system of equations Ax = b and a second system obtained by altering the right-hand side: A(x + x) = b + b. 2.9. Norms and Condition Numbers 17 We think of b as being the error in b and x as being the resulting error in x, although we need not make any assumptions that the errors are small. Because A(x) = b, the definitions of M and m immediately lead to b M x and Consequently, if m = 0, b mx. x b (A) . x b The quantity b/b is the relative change in the right-hand side, and the quantity x/x is the relative error caused by this change. The advantage of using relative changes is that they are dimensionless, that is, they are not affected by overall scale factors. This shows that the condition number is a relative error magnification factor. Changes in the right-hand side can cause changes (A) times as large in the solution. It turns out that the same is true of changes in the coefficient matrix itself. The condition number is also a measure of nearness to singularity. Although we have not yet developed the mathematical tools necessary to make the idea precise, the condition number can be thought of as the reciprocal of the relative distance from the matrix to the set of singular matrices. So, if (A) is large, A is close to singular. Some of the basic properties of the condition number are easily derived. Clearly, M m, and so (A) 1. If P is a permutation matrix, then the components of P x are simply a rearrangement of the components of x. It follows that P x = x for all x, and so (P ) = 1. In particular, (I) = 1. If A is multiplied by a scalar c, then M and m are both multiplied by the same scalar, and so (cA) = (A). If D is a diagonal matrix, then (D) = max |dii | . min |dii | These last two properties are two of the reasons that (A) is a better measure of nearness to singularity than the determinant of A. As an extreme example, consider a 100-by-100 diagonal matrix with 0.1 on the diagonal. Then det(A) = 10-100 , which is usually regarded as a small number. But (A) = 1, and the components of Ax are simply 0.1 times the corresponding components of x. For linear systems of equations, such a matrix behaves more like the identity than like a singular matrix. 18 The following example uses the l1 -norm: 4.1 2.8 A= , 9.7 6.6 4.1 b= , 9.7 1 x= . 0 Clearly, Ax = b, and b = 13.8, x = 1. ~= b the solution becomes x= ~ Let b = b - ~ and x = x - x. Then b ~ 4.11 9.70 0.34 0.97 Chapter 2. Linear Equations If the right-hand side is changed to , . x = 1.63. We have made a fairly small perturbation in b that completely changes x. In fact, the relative changes are b = 0.0007246, b x = 1.63. x Because (A) is the maximum magnification factor, (A) 1.63 = 2249.4. 0.0007246 b = 0.01, We have actually chosen the b and b that give the maximum, and so, for this example with the l1 -norm, (A) = 2249.4. It is important to realize that this example is concerned with the exact solutions to two slightly different systems of equations and that the method used to obtain the solutions is irrelevant. The example is constructed to have a fairly large condition number so that the effect of changes in b is quite pronounced, but similar behavior can be expected in any problem with a large condition number. The condition number also plays a fundamental role in the analysis of the roundoff errors introduced during the solution by Gaussian elimination. Let us 2.9. Norms and Condition Numbers 19 assume that A and b have elements that are exact floating-point numbers, and let x be the vector of floating-point numbers obtained from a linear equation solver such as the function we shall present in the next section. We also assume that exact singularity is not detected and that there are no underflows or overflows. Then it is possible to establish the following inequalities: b - Ax , Ax x - x (A). x Here is the relative machine precision eps and is defined more carefully later, but it usually has a value no larger than about 10. The first inequality says that the relative residual can usually be expected to be about the size of roundoff error, no matter how badly conditioned the matrix is. This was illustrated by the example in the previous section. The second inequality requires that A be nonsingular and involves the exact solution x. It follows directly from the first inequality and the definition of (A) and says that the relative error will also be small if (A) is small but might be quite large if the matrix is nearly singular. In the extreme case where A is singular but the singularity is not detected, the first inequality still holds but the second has no meaning. To be more precise about the quantity , it is necessary to introduce the idea of a matrix norm and establish some further inequalities. Readers who are not interested in such details can skip the remainder of this section. The quantity M defined earlier is known as the norm of the matrix. The notation for the matrix norm is the same as for the vector norm: Ax A = max . x It is not hard to see that A-1 = 1/m, so an equivalent definition of the condition number is (A) = AA-1 . Again, the actual numerical values of the matrix norm and condition number depend on the underlying vector norm. It is easy to compute the matrix norms corresponding to the l1 and l vector norms. In fact, it is not hard to show that A1 = max |ai,j |,
j A = max
i Computing the matrix norm corresponding to the l2 vector norm involves the singular value decomposition (SVD), which is discussed in a later chapter. Matlab computes matrix norms with norm(A,p) for p = 1, 2, or inf. The basic result in the study of roundoff error in Gaussian elimination is due to J. H. Wilkinson. He proved that the computed solution x exactly satisfies (A + E)x = b, j i |ai,j |. 20 Chapter 2. Linear Equations where E is a matrix whose elements are about the size of roundoff errors in the elements of A. There are some rare situations where the intermediate matrices obtained during Gaussian elimination have elements that are larger than those of A, and there is some effect from accumulation of rounding errors in large matrices, but it can be expected that if is defined by E = , A then will rarely be bigger than about 10. From this basic result, we can immediately derive inequalities involving the residual and the error in the computed solution. The residual is given by b - Ax = Ex , and hence b - Ax = Ex Ex . The residual involves the product Ax , so it is appropriate to consider the relative residual, which compares the norm of b - Ax to the norms of A and x . It follows directly from the above inequalities that b - Ax . Ax If A is nonsingular, the error can be expressed using the inverse of A by x - x = A-1 (b - Ax ), and so x - x A-1 Ex . It is simplest to compare the norm of the error with the norm of the computed solution. Thus the relative error satisfies x - x AA-1 . x Hence x - x (A). x The actual computation of (A) requires knowing A-1 . But computing A-1 requires roughly three times as much work as solving a single linear system. Computing the l2 condition number requires the SVD and even more work. Fortunately, the exact value of (A) is rarely required. Any reasonably good estimate of it is satisfactory. Matlab has several functions for computing or estimating condition numbers. cond(A) or cond(A,2) computes 2 (A). Uses svd(A). Suitable for smaller matrices where the geometric properties of the l2 -norm are important. 2.10. Sparse Matrices and Band Matrices cond(A,1) computes 1 (A). Uses inv(A). Less work than cond(A,2). cond(A,inf) computes (A). Uses inv(A). Same as cond(A',1). 21 condest(A) estimates 1 (A). Uses lu(A) and a recent algorithm of Higham and Tisseur [9]. Especially suitable for large, sparse matrices. rcond(A) estimates 1/1 (A). Uses lu(A) and an older algorithm developed by the LINPACK and LAPACK projects. Primarily of historical interest. 2.10 Sparse Matrices and Band Matrices Sparse matrices and band matrices occur frequently in technical computing. The sparsity of a matrix is the fraction of its elements that are zero. The Matlab function nnz counts the number of nonzeros in a matrix, so the sparsity of A is given by density = nnz(A)/prod(size(A)) sparsity = 1 - density A sparse matrix is a matrix whose sparsity is nearly equal to 1. The bandwidth of a matrix is the maximum distance of the nonzero elements from the main diagonal. [i,j] = find(A) bandwidth = max(abs(i-j)) A band matrix is a matrix whose bandwidth is small. As you can see, both sparsity and bandwidth are matters of degree. An n-by-n diagonal matrix with no zeros on the diagonal has sparsity 1 - 1/n and bandwidth 0, so it is an extreme example of both a sparse matrix and a band matrix. On the other hand, an n-by-n matrix with no zero elements, such as the one created by rand(n,n), has sparsity equal to zero and bandwidth equal to n - 1, and so is far from qualifying for either category. The Matlab sparse data structure stores the nonzero elements together with information about their indices. The sparse data structure also provides efficient handling of band matrices, so Matlab does not have a separate band matrix storage class. The statement S = sparse(A) converts a matrix to its sparse representation. The statement A = full(S) reverses the process. However, most sparse matrices have orders so large that it is impractical to store the full representation. More frequently, sparse matrices are created by S = sparse(i,j,x,m,n) 22 This produces a matrix S with [i,j,x] = find(S) [m,n] = size(S) Chapter 2. Linear Equations Most Matlab matrix operations and functions can be applied to both full and sparse matrices. The dominant factor in determining the execution time and memory requirements for sparse matrix operations is the number of nonzeros, nnz(S), in the various matrices involved. A matrix with bandwidth equal to 1 is known as a tridiagonal matrix. It is worthwhile to have a specialized function for one particular band matrix operation, the solution of a tridiagonal system of simultaneous linear equations: b1 a1 c1 b2 a2 c2 b3 .. . . bn-1 an-1 .. cn-1 bn x x1 x2 x3 . . .
n-1 c3 .. . an-2 xn The function tridisolve is included in the NCM directory. The statement x = tridisolve(a,b,c,d) solves the tridiagonal system with subdiagonal a, diagonal b, superdiagonal c, and right-hand side d. We have already seen the algorithm that tridisolve uses; it is Gaussian elimination. In many situations involving tridiagonal matrices, the diagonal elements dominate the off-diagonal elements, so pivoting is unnecessary. Furthermore, the right-hand side is processed at the same time as the matrix itself. In this context, Gaussian elimination without pivoting is also known as the Thomas algorithm. The body of tridisolve begins by copying the right-hand side to a vector that will become the solution. x = d; n = length(x); The forward elimination step is a simple for loop. for j = 1:n-1 mu = a(j)/b(j); b(j+1) = b(j+1) - mu*c(j); x(j+1) = x(j+1) - mu*x(j); end The mu's would be the multipliers on the subdiagonal of L if we were saving the LU factorization. Instead, the right-hand side is processed in the same loop. The back substitution step is another simple loop. = d d1 d2 d3 . . .
n-1 dn . 2.11. PageRank and Markov Chains x(n) = x(n)/b(n); for j = n-1:-1:1 x(j) = (x(j)-c(j)*x(j+1))/b(j); end 23 Because tridisolve does not use pivoting, the results might be inaccurate if abs(b) is much smaller than abs(a)+abs(c). More robust, but slower, alternatives that do use pivoting include generating a full matrix with diag: T = diag(a,-1) + diag(b,0) + diag(c,1); x = T\d or generating a sparse matrix with spdiags: S = spdiags([a b c],[-1 0 1],n,n); x = S\d 2.11 PageRank and Markov Chains One of the reasons why GoogleTM is such an effective search engine is the PageRankTM algorithm developed by Google's founders, Larry Page and Sergey Brin, when they were graduate students at Stanford University. PageRank is determined entirely by the link structure of the World Wide Web. It is recomputed about once a month and does not involve the actual content of any Web pages or individual queries. Then, for any particular query, Google finds the pages on the Web that match that query and lists those pages in the order of their PageRank. Imagine surfing the Web, going from page to page by randomly choosing an outgoing link from one page to get to the next. This can lead to dead ends at pages with no outgoing links, or cycles around cliques of interconnected pages. So, a certain fraction of the time, simply choose a random page from the Web. This theoretical random walk is known as a Markov chain or Markov process. The limiting probability that an infinitely dedicated random surfer visits any particular page is its PageRank. A page has high rank if other pages with high rank link to it. Let W be the set of Web pages that can be reached by following a chain of hyperlinks starting at some root page, and let n be the number of pages in W . For Google, the set W actually varies with time, In June 2004, their value of n was over 4 billion. Today, Google does not reveal how many pages they reach. Let G be the n-by-n connectivity matrix of a portion of the Web, that is, gij = 1 if there is a hyperlink to page i from page j and gij = 0 otherwise. The matrix G can be huge, but it is very sparse. Its jth column shows the links on the jth page. The number of nonzeros in G is the total number of hyperlinks in W . Let ri and cj be the row and column sums of G: ri = gij , cj = gij .
j i The quantities rj and cj are the in-degree and out-degree of the jth page. Let p be the probability that the random walk follows a link. A typical value is p = 0.85. 24 Chapter 2. Linear Equations Then 1 - p is the probability that some arbitrary page is chosen and = (1 - p)/n is the probability that a particular random page is chosen. Let A be the n-by-n matrix whose elements are pgij /cj + : cj = 0 aij = 1/n : cj = 0. Notice that A comes from scaling the connectivity matrix by its column sums. The jth column is the probability of jumping from the jth page to the other pages on the Web. If the jth page is a dead end, that is has no out-links, then we assign a uniform probability of 1/n to all the elements in its column. Most of the elements of A are equal to , the probability of jumping from one page to another without following a link. If n = 4 109 and p = 0.85, then = 3.75 10-11 . The matrix A is the transition probability matrix of the Markov chain. Its elements are all strictly between zero and one and its column sums are all equal to one. An important result in matrix theory known as the PerronFrobenius theorem applies to such matrices. It concludes that a nonzero solution of the equation x = Ax exists and is unique to within a scaling factor. If this scaling factor is chosen so that xi = 1,
i then x is the state vector of the Markov chain and is Google's PageRank. The elements of x are all positive and less than one. The vector x is the solution to the singular, homogeneous linear system (I - A)x = 0. For modest n, an easy way to compute x in Matlab is to start with some approximate solution, such as the PageRanks from the previous month, or x = ones(n,1)/n Then simply repeat the assignment statement x = A*x until successive vectors agree to within a specified tolerance. This is known as the power method and is about the only possible approach for very large n. In practice, the matrices G and A are never actually formed. One step of the power method would be done by one pass over a database of Web pages, updating weighted reference counts generated by the hyperlinks between pages. The best way to compute PageRank in Matlab is to take advantage of the particular structure of the Markov matrix. Here is an approach that preserves the sparsity of G. The transition matrix can be written A = pGD + ez T 2.11. PageRank and Markov Chains where D is the diagonal matrix formed from the reciprocals of the outdegrees, 1/cj : cj = 0 djj = 0 : cj = 0, e is the n-vector of all ones, and z is the vector with components : cj = 0 zj = 1/n : cj = 0. 25 The rank-one matrix ez T accounts for the random choices of Web pages that do not follow links. The equation x = Ax can be written where (I - pGD)x = e = z T x. We do not know the value of because it depends upon the unknown vector x, but we can temporarily take = 1. As long as p is strictly less than one, the coefficient matrix I - pGD is nonsingular and the equation (I - pGD)x = e can be solved for x. Then the resulting x can be rescaled so that xi = 1.
i Notice that the vector z is not actually involved in this calculation. The following Matlab statements implement this approach c k D e I x x = = = = = = = sum(G,1); find(c~=0); sparse(k,k,1./c(k),n,n); ones(n,1); speye(n,n); (I - p*G*D)\e; x/sum(x); The power method can also be implemented in a way that does not actually form the Markov matrix and so preserves sparsity. Compute G = p*G*D; z = ((1-p)*(c~=0) + (c==0))/n; Start with x = e/n 26 Then repeat the statement x = G*x + e*(z*x) Chapter 2. Linear Equations until x settles down to several decimal places. It is also possible to use an algorithm known as inverse iteration. A = p*G*D + delta x = (I - A)\e x = x/sum(x) At first glance, this appears to be a very dangerous idea. Because I - A is theoretically singular, with exact computation some diagonal element of the upper triangular factor of I - A should be zero and this computation should fail. But with roundoff error, the computed matrix I - A is probably not exactly singular. Even if it is singular, roundoff during Gaussian elimination will most likely prevent any exact zero diagonal elements. We know that Gaussian elimination with partial pivoting always produces a solution with a small residual, relative to the computed solution, even if the matrix is badly conditioned. The vector obtained with the backslash operation, (I - A)\e, usually has very large components. If it is rescaled by its sum, the residual is scaled by the same factor and becomes very small. Consequently, the two vectors x and A*x equal each other to within roundoff error. In this setting, solving the singular system with Gaussian elimination blows up, but it blows up in exactly the right direction.
alpha delta beta gamma sigma rho Figure 2.2. A tiny Web. Figure 2.2 is the graph for a tiny example, with n = 6 instead of n = 4 109 . Pages on the Web are identified by strings known as uniform resource locators, or URLs. Most URLs begin with http because they use the hypertext transfer protocol. In Matlab , we can store the URLs as an array of strings in a cell array. This example involves a 6-by-1 cell array. 2.11. PageRank and Markov Chains U = {'http://www.alpha.com' 'http://www.beta.com' 'http://www.gamma.com' 'http://www.delta.com' 'http://www.rho.com' 'http://www.sigma.com'} 27 Two different kinds of indexing into cell arrays are possible. Parentheses denote subarrays, including individual cells, and curly braces denote the contents of the cells. If k is a scalar, then U(k) is a 1-by-1 cell array consisting of the kth cell in U, while U{k} is the string in that cell. Thus U(1) is a single cell and U{1} is the string 'http://www.alpha.com'. Think of mail boxes with addresses on a city street. B(502) is the box at number 502, while B{502} is the mail in that box. We can generate the connectivity matrix by specifying the pairs of indices (i,j) of the nonzero elements. Because there is a link to beta.com from alpha.com, the (2,1) element of G is nonzero. The nine connections are described by i = [ 2 6 3 4 4 5 6 1 1] j = [ 1 1 2 2 3 3 3 4 6] A sparse matrix is stored in a data structure that requires memory only for the nonzero elements and their indices. This is hardly necessary for a 6-by-6 matrix with only 27 zero entries, but it becomes crucially important for larger problems. The statements n = 6 G = sparse(i,j,1,n,n); full(G) generate the sparse representation of an n-by-n matrix with ones in the positions specified by the vectors i and j and display its full representation. 0 1 0 0 0 1 0 0 1 1 0 0 0 0 0 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 The statement c = full(sum(G)) computes the column sums c = 2 2 3 1 0 1 Notice that c(5) = 0 because the 5th page, labeled rho, has no out-links. The statements 28 x = (I - p*G*D)\e x = x/sum(x) solve the sparse linear system to produce x = 0.3210 0.1705 0.1066 0.1368 0.0643 0.2007 Chapter 2. Linear Equations Page Rank 0.35 0.3 0.25 0.2 0.15 0.1 0.05 0 1 2 3 4 5 6 Figure 2.3. Page Rank for the tiny Web The bar graph of x is shown in figure 2.3. If the URLs are sorted in PageRank order and listed along with their in- and out-degrees, the result is 1 6 2 4 3 5 page-rank 0.3210 0.2007 0.1705 0.1368 0.1066 0.0643 in 2 2 1 2 1 1 out 2 1 2 1 3 0 url http://www.alpha.com http://www.sigma.com http://www.beta.com http://www.delta.com http://www.gamma.com http://www.rho.com We see that alpha has a higher PageRank than delta or sigma, even though they all have the same number of in-links. A random surfer will visit alpha over 32% of the time and rho only about 6% of the time. For this tiny example with p = .85, the smallest element of the Markov transition matrix is = .15/6 = .0250. 2.11. PageRank and Markov Chains A = 0.0250 0.4500 0.0250 0.0250 0.0250 0.4500 29 0.0250 0.0250 0.4500 0.4500 0.0250 0.0250 0.0250 0.0250 0.0250 0.3083 0.3083 0.3083 0.8750 0.0250 0.0250 0.0250 0.0250 0.0250 0.1667 0.1667 0.1667 0.1667 0.1667 0.1667 0.8750 0.0250 0.0250 0.0250 0.0250 0.0250 Notice that the column sums of A are all equal to one. Our collection of NCM programs includes surfer.m. A statement like [U,G] = surfer('http://www.xxx.zzz',n) starts at a specified URL and tries to surf the Web until it has visited n pages. If successful, it returns an n-by-1 cell array of URLs and an n-by-n sparse connectivity matrix. The function uses urlread, which was introduced in Matlab 6.5, along with underlying Java utilities to access the Web. Surfing the Web automatically is a dangerous undertaking and this function must be used with care. Some URLs contain typographical errors and illegal characters. There is a list of URLs to avoid that includes .gif files and Web sites known to cause difficulties. Most importantly, surfer can get completely bogged down trying to read a page from a site that appears to be responding, but that never delivers the complete page. When this happens, it may be necessary to have the computer's operating system ruthlessly terminate Matlab. With these precautions in mind, you can use surfer to generate your own PageRank examples.
0 50 100 150 200 250 300 350 400 450 500 0 100 200 300 nz = 2636 400 500 Figure 2.4. Spy plot of the harvard500 graph. 30 The statement Chapter 2. Linear Equations [U,G] = surfer('http://www.harvard.edu',500) accesses the home page of Harvard University and generates a 500-by-500 test case. The graph generated in August 2003 is available in the NCM directory. The statements load harvard500 spy(G) produce a spy plot (Figure 2.4) that shows the nonzero structure of the connectivity matrix. The statement pagerank(U,G) computes page ranks, produces a bar graph (Figure 2.5) of the ranks, and prints the most highly ranked URLs in PageRank order. For the harvard500 data, the dozen most highly ranked pages are 1 10 42 130 18 15 9 17 46 13 260 19 page-rank in 0.0843 195 0.0167 21 0.0166 42 0.0163 0.0139 0.0131 0.0114 0.0111 0.0100 0.0086 0.0086 0.0084 24 45 16 21 13 18 9 26 23 out 26 18 0 12 46 49 27 6 21 1 1 21 url http://www.harvard.edu http://www.hbs.edu http://search.harvard.edu:8765/ custom/query.html http://www.med.harvard.edu http://www.gse.harvard.edu http://www.hms.harvard.edu http://www.ksg.harvard.edu http://www.hsph.harvard.edu http://www.gocrimson.com http://www.hsdm.med.harvard.edu http://search.harvard.edu:8765/ query.html http://www.radcliffe.edu The URL where the search began, www.harvard.edu, dominates. Like most universities, Harvard is organized into various colleges and institutes, including the Kennedy School of Government, the Harvard Medical School, the Harvard Business School, and the Radcliffe Institute. You can see that the home pages of these schools have high PageRank. With a different sample, such as the one generated by Google itself, the ranks would be different. 2.12 Further Reading Further reading on matrix computation includes books by Demmel [2], Golub and Van Loan [3], Stewart [4, 5], and Trefethen and Bau [6]. The definitive references on Fortran matrix computation software are the LAPACK Users' Guide and Web site [1]. The Matlab sparse matrix data structure and operations are described Exercises
Page Rank 0.02 0.018 0.016 0.014 0.012 0.01 0.008 0.006 0.004 0.002 0 31 0 50 100 150 200 250 300 350 400 450 500 Figure 2.5. PageRank of the harvard500 graph. in [8]. Information available on Web sites about PageRank includes a brief explanation at Google [7], a technical report by Page, Brin, and colleagues [11], and a comprehensive survey by Langville and Meyer [10]. Exercises
2.1. Alice buys three apples, a dozen bananas, and one cantaloupe for $2.36. Bob buys a dozen apples and two cantaloupes for $5.26. Carol buys two bananas and three cantaloupes for $2.77. How much do single pieces of each fruit cost? (You might want to set format bank.) 2.2. What Matlab function computes the reduced row echelon form of a matrix? What Matlab function generates magic square matrices? What is the reduced row echelon form of the magic square of order six? 2.3. Figure 2.6 depicts a plane truss having 13 members (the numbered lines) connecting 8 joints (the numbered circles). The indicated loads, in tons, are applied at joints 2, 5, and 6, and we want to determine the resulting force on each member of the truss. For the truss to be in static equilibrium, there must be no net force, horizontally or vertically, at any joint. Thus, we can determine the member forces by equating the horizontal forces to the left and right at each joint, and similarly equating the vertical forces upward and downward at each joint. For the eight joints, this would give 16 equations, which is more than the 13 unknown factors to be determined. For the truss to be statically determinate, that is, for there to be a unique solution, we assume that joint 1 is rigidly fixed both horizontally and vertically and that joint 8 is fixed verti- 32 Chapter 2. Linear Equations 3 4 4 8 7 1 3 5 7 9 11 12 1 2 2 6 5 10 6 13 8 10 15 20 Figure 2.6. A plane truss. cally. Resolving the member forces into horizontal and vertical components and defining = 1/ 2, we obtain the following system of equations for the member forces fi : Joint 2: f2 = f6 , f3 = 10; Joint 3: f1 = f4 + f5 , f1 + f3 + f5 = 0; Joint 4: f4 = f8 , f7 = 0; Joint 5: f5 + f6 = f9 + f10 , f5 + f7 + f9 = 15; Joint 6: f10 = f13 , Joint 7: Joint 8: f11 = 20; f8 + f9 = f12 , f9 + f11 + f12 = 0; f13 + f12 = 0. Solve this system of equations to find the vector f of member forces. 2.4. Figure 2.7 is the circuit diagram for a small network of resistors. There are five nodes, eight resistors, and one constant voltage source. We want to compute the voltage drops between the nodes and the currents around each of the loops. Several different linear systems of equations can be formed to describe this circuit. Let vk , k = 1, . . . , 4, denote the voltage difference between each of the first four nodes and node number 5 and let ik , k = 1, . . . , 4, denote the clockwise current around each of the loops in the diagram. Ohm's law says that the voltage drop across a resistor is the resistance times the current. For Exercises
3 33 r23 2 r25 r13 i2 r12 1 i1 r34 i3 r14 r45 r35 4 i4 vs 5 Figure 2.7. A resistor network. example, the branch between nodes 1 and 2 gives v1 - v2 = r12 (i2 - i1 ). Using the conductance, which is the reciprocal of the resistance, gkj = 1/rkj , Ohm's law becomes i2 - i1 = g12 (v1 - v2 ). The voltage source is included in the equation v3 - vs = r35 i4 . Kirchhoff 's voltage law says that the sum of the voltage differences around each loop is zero. For example, around loop 1, (v1 - v4 ) + (v4 - v5 ) + (v5 - v2 ) + (v2 - v1 ) = 0. Combining the voltage law with Ohm's law leads to the loop equations for the currents: Ri = b. Here i is the current vector, i1 i i = 2, i3 i4 0 0 b = , 0 vs b is the source voltage vector, 34 and R is the resistance matrix, r25 + r12 + r14 + r45 -r12 -r14 -r45 -r12 r23 + r12 + r13 -r13 0 Chapter 2. Linear Equations -r14 -r13 r14 + r13 + r34 -r34 Kirchhoff 's current law says that the sum of the currents at each node is zero. For example, at node 1, (i1 - i2 ) + (i2 - i3 ) + (i3 - i1 ) = 0. Combining the current law with the conductance version of Ohm's law leads to the nodal equations for the voltages: Gv = c. Here v is the voltage vector, v1 v v = 2 , v3 v4 0 0 c= , g35 vs 0 -g13 -g23 g13 + g23 + g34 + g35 -g34 -r45 0 . -r34 r35 + r45 + r34 c is the source current vector, and G is the conductance matrix, g12 + g13 + g14 -g12 -g13 -g14 -g12 g12 + g23 + g25 -g23 0 You can solve the linear system obtained from the loop equations to compute the currents and then use Ohm's law to recover the voltages. Or you can solve the linear system obtained from the node equations to compute the voltages and then use Ohm's law to recover the currents. Your assignment is to verify that these two approaches produce the same results for this circuit. You can choose your own numerical values for the resistances and the voltage source. 2.5. The Cholesky algorithm factors an important class of matrices known as positive definite matrices. Andre-Louis Cholesky (18751918) was a French military officer involved in geodesy and surveying in Crete and North Africa just before World War I. He developed the method now named after him to compute solutions to the normal equations for some least squares data-fitting -g14 0 . -g34 g14 + g34 + g45 Exercises 35 problems arising in geodesy. His work was posthumously published on his behalf in 1924 by a fellow officer, Benoit, in the Bulletin Geodesique. A real symmetric matrix A = AT is positive definite if any of the following equivalent conditions hold: The quadratic form xT Ax is positive for all nonzero vectors x. All determinants formed from symmetric submatrices of any order centered on the diagonal of A are positive. All eigenvalues (A) are positive. There is a real matrix R such that A = RT R. These conditions are difficult or expensive to use as the basis for checking if a particular matrix is positive definite. In Matlab , the best way to check positive definiteness is with the chol function. See help chol (a) Which of the following families of matrices are positive definite? M H P I R R R R = = = = = = = = magic(n) hilb(n) pascal(n) eye(n,n) randn(n,n) randn(n,n); A = R' * R randn(n,n); A = R' + R randn(n,n); I = eye(n,n); A = R' + R + n*I (b) If the matrix R is upper triangular, then equating individual elements in the equation A = RT R gives akj =
k i=1 rik rij , k j. Using these equations in different orders yields different variants of the Cholesky algorithm for computing the elements of R. What is one such algorithm? 2.6. This example shows that a badly conditioned matrix does not necessarily lead to small pivots in Gaussian elimination. The matrix is the n-by-n upper triangular matrix A with elements -1, i < j, 1, i = j, aij = 0, i > j. 36 Chapter 2. Linear Equations Show how to generate this matrix in Matlab with eye, ones, and triu. Show that 1 (A) = n2n-1 . For what n does 1 (A) exceed 1/eps? This matrix is not singular, so Ax cannot be zero unless x is zero. However, there are vectors x for which Ax is much smaller than x. Find one such x. Because this matrix is already upper triangular, Gaussian elimination with partial pivoting has no work to do. What are the pivots? Use lugui to design a pivot strategy that will produce smaller pivots than partial pivoting. (Even these pivots do not completely reveal the large condition number.) 2.7. The matrix factorization LU = P A can be used to compute the determinant of A. We have det(L)det(U ) = det(P )det(A). Because L is triangular with ones on the diagonal, det(L) = 1. Because U is triangular, det(U ) = u11 u22 unn . Because P is a permutation, det(P ) = +1 if the number of interchanges is even and -1 if it is odd. So det(A) = u11 u22 unn . Modify the lutx function so that it returns four outputs. function [L,U,p,sig] = lutx(A) %LU Triangular factorization % [L,U,p,sig] = lutx(A) computes a unit lower triangular % matrix L, an upper triangular matrix U, a permutation % vector p, and a scalar sig, so that L*U = A(p,:) and % sig = +1 or -1 if p is an even or odd permutation. Write a function mydet(A) that uses your modified lutx to compute the determinant of A. In Matlab, the product u11 u22 unn can be computed by the expression prod(diag(U)). 2.8. Modify the lutx function so that it uses explicit for loops instead of Matlab vector notation. For example, one section of your modified program will read % Compute the multipliers for i = k+1:n A(i,k) = A(i,k)/A(k,k); end Compare the execution time of your modified lutx program with the original lutx program and with the built-in lu function by finding the order of the matrix for which each of the three programs takes about 10 s on your computer. Exercises 2.9. Let 37 (a) Show that the set of linear equations Ax = b has infinitely many solutions. Describe the set of possible solutions. (b) Suppose Gaussian elimination is used to solve Ax = b using exact arithmetic. Because there are infinitely many solutions, it is unreasonable to expect one particular solution to be computed. What does happen? (c) Use bslashtx to solve Ax = b on an actual computer with floating-point arithmetic. What solution is obtained? Why? In what sense is it a "good" solution? In what sense is it a "bad" solution? (d) Explain why the built-in backslash operator x = A\b gives a different solution from x = bslashtx(A,b). 2.10. Section 2.4 gives two algorithms for solving triangular systems. One subtracts columns of the triangular matrix from the right-hand side; the other uses inner products between the rows of the triangular matrix and the emerging solution. (a) Which of these two algorithms does bslashtx use? (b) Write another function, bslashtx2, that uses the other algorithm. 2.11. The inverse of a matrix A can be defined as the matrix X whose columns xj solve the equations Axj = ej , where ej is the jth column of the identity matrix. (a) Starting with the function bslashtx, write a Matlab function X = myinv(A) that computes the inverse of A. Your function should call lutx only once and should not use the built-in Matlab backslash operator or inv function. (b) Test your function by comparing the inverses it computes with the inverses obtained from the built-in inv(A) on a few test matrices. 2.12. If the built-in Matlab lu function is called with only two output arguments [L,U] = lu(A) the permutations are incorporated into the output matrix L. The help entry for lu describes L as "psychologically lower triangular." Modify lutx so that it does the same thing. You can use if nargout == 2, ... to test the number of output arguments. 2.13. (a) Why is M = magic(8) lugui(M) 1 2 A = 4 5 7 8 3 1 6, b = 3. 9 5 38 Chapter 2. Linear Equations an interesting example? (b) How is the behavior of lugui(M) related to rank(M)? (c) Can you pick a sequence of pivots so that no roundoff error occurs in lugui(M)? 2.14. The pivot selection strategy known as complete pivoting is one of the options available in lugui. It has some slight numerical advantages over partial pivoting. At each stage of the elimination, the element of largest magnitude in the entire unreduced matrix is selected as the pivot. This involves both row and column interchanges and produces two permutation vectors p and q so that L*U = A(p,q) Modify lutx and bslashtx so that they use complete pivoting. 2.15. The function golub in the NCM directory is named after Stanford professor Gene Golub. The function generates test matrices with random integer entries. The matrices are very badly conditioned, but Gaussian elimination without pivoting fails to produce the small pivots that would reveal the large condition number. (a) How does condest(golub(n)) grow with increasing order n? Because these are random matrices you can't be very precise here, but you can give some qualitative description. (b) What atypical behavior do you observe with the diagonal pivoting option in lugui(golub(n))? (c) What is det(golub(n))? Why? 2.16. The function pascal generates symmetric test matrices based on Pascal's triangle. (a) How are the elements of pascal(n+1) related to the binomial coefficients generated by nchoosek(n,k)? (b) How is chol(pascal(n)) related to pascal(n)? (c) How does condest(pascal(n)) grow with increasing order n? (d) What is det(pascal(n))? Why? (e) Let Q be the matrix generated by Q = pascal(n); Q(n,n) = Q(n,n) - 1; How is chol(Q) related to chol(pascal(n))? Why? (f) What is det(Q)? Why? 2.17. Play "Pivot Pickin' Golf" with pivotgolf. The goal is to use lugui to compute the LU decompositions of nine matrices with as little roundoff error as possible. The score for each hole is R + L + U , where R = LU -P AQ is the residual and L and U are the nonzeros that should be zero in L and U . Exercises 39 (a) Can you beat the scores obtained by partial pivoting on any of the courses? (b) Can you get a perfect score of zero on any of the courses? 2.18. The object of this exercise is to investigate how the condition numbers of random matrices grow with their order. Let Rn denote an n-by-n matrix with normally distributed random elements. You should observe experimentally that there is an exponent p so that 1 (Rn ) = O(np ). In other words, there are constants c1 and c2 so that most values of 1 (Rn ) satisfy c1 np 1 (Rn ) c2 np . Your job is to find p, c1 , and c2 . The NCM M-file randncond.m is the starting point for your experiments. This program generates random matrices with normally distributed elements and plots their l1 condition numbers versus their order on a loglog scale. The program also plots two lines that are intended to enclose most of the observations. (On a loglog scale, power laws like = cnp produce straight lines.) (a) Modify randncond.m so that the two lines have the same slope and enclose most of the observations. (b) Based on this experiment, what is your guess for the exponent p in (Rn ) = O(np )? How confident are you? (c) The program uses ('erasemode','none'), so you cannot print the results. What would you have to change to make printing possible? 2.19. For n = 100, solve this tridiagonal system of equations three different ways: 2x1 - x2 = 1, -xj-1 + 2xj - xj+1 = j, j = 2, . . . , n - 1, -xn-1 + 2xn = n. (a) Use diag three times to form the coefficient matrix and then use lutx and bslashtx to solve the system. (b) Use spdiags once to form a sparse representation of the coefficient matrix and then use the backslash operator to solve the system. (c) Use tridisolve to solve the system. (d) Use condest to estimate the condition of the coefficient matrix. 2.20. Use surfer and pagerank to compute PageRanks for some subset of the Web that you choose. Do you see any interesting structure in the results? 2.21. Suppose that U and G are the URL cell array and the connectivity matrix produced by surfer and that k is an integer. Explain what U{k}, U(k), G(k,:), G(:,k), U(G(k,:)), U(G(:,k)) are. 40 Chapter 2. Linear Equations 2.22. The connectivity matrix for the harvard500 data set has four small, almost entirely nonzero, submatrices that produce dense patches near the diagonal of the spy plot. You can use the zoom button to find their indices. The first submatrix has indices around 170 and the other three have indices in the 200s and 300s. Mathematically, a graph with every node connected to every other node is known as a clique. Identify the organizations within the Harvard community that are responsible for these near cliques. 2.23. A Web connectivity matrix G has gij = 1 if it is possible to get to page i from page j with one click. If you multiply the matrix by itself, the entries of the matrix G2 count the number of different paths of length two to page i from page j. The matrix power Gp shows the number of paths of length p. (a) For the harvard500 data set, find the power p where the number of nonzeros stops increasing. In other words, for any q greater than p, nnz(G^q) is equal to nnz(G^p). (b) What fraction of the entries in Gp are nonzero? (c) Use subplot and spy to show the nonzeros in the successive powers. (d) Is there a set of interconnected pages that do not link to the other pages? 2.24. The function surfer uses a subfunction, hashfun, to speed up the search for a possibly new URL in the list of URLs that have already been processed. Find two different URLs on The MathWorks home page http://www.mathworks.com that have the same hashfun value. 2.25. Figure 2.8 is the graph of another six-node subset of the Web. In this example, there are two disjoint subgraphs.
alpha delta beta gamma sigma rho Figure 2.8. Another tiny Web. (a) What is the connectivity matrix G? (b) What are the PageRanks if the hyperlink transition probability p is the default value 0.85? (c) Describe what happens with this example to both the definition of PageRank and the computation done by pagerank in the limit p 1. Exercises 41 2.26. The function pagerank(U,G) computes PageRanks by solving a sparse linear system. It then plots a bar graph and prints the dominant URLs. (a) Create pagerank1(G) by modifying pagerank so that it just computes the PageRanks, but does not do any plotting or printing. (b) Create pagerank2(G) by modifying pagerank1 to use inverse iteration instead of solving the sparse linear system. The key statements are x = (I - A)\e x = x/sum(x) What should be done in the unlikely event that the backslash operation involves a division by zero? (c) Create pagerank3(G) by modifying pagerank1 to use the power method instead of solving the sparse linear system. The key statements are G = p*G*D z = ((1-p)*(c~=0) + (c==0))/n; while termination_test x = G*x + e*(z*x) end What is an appropriate test for terminating the power iteration? (d) Use your functions to compute the PageRanks of the six-node example discussed in the text. Make sure you get the correct result from each of your three functions. 2.27. Here is yet another function for computing PageRank. This version uses the power method, but does not do any matrix operations. Only the link structure of the connectivity matrix is involved. function [x,cnt] = pagerankpow(G) % PAGERANKPOW PageRank by power method. % x = pagerankpow(G) is the PageRank of the graph G. % [x,cnt] = pagerankpow(G) % counts the number of iterations. % Link structure [n,n] = for j = L{j} c(j) end size(G); 1:n = find(G(:,j)); = length(L{j}); % Power method p = .85; delta = (1-p)/n; x = ones(n,1)/n; 42 Chapter 2. Linear Equations z = zeros(n,1); cnt = 0; while max(abs(x-z)) > .0001 z = x; x = zeros(n,1); for j = 1:n if c(j) == 0 x = x + z(j)/n; else x(L{j}) = x(L{j}) + z(j)/c(j); end end x = p*x + delta; cnt = cnt+1; end (a) How do the storage requirements and execution time of this function compare with the three pagerank functions from the previous exercise? (b) Use this function as a template to write a function that computes PageRank in some other programming language. Bibliography
[1] E. Anderson, Z. Bai, C. Bischof, S. Blackford, J. Demmel, J. Dongarra, J. Du Croz, A. Greenbaum, S. Hammarling, A. McKenney, and D. Sorensen, LAPACK Users' Guide, Third Edition, SIAM, Philadelphia, 1999. http://www.netlib.org/lapack [2] J. W. Demmel, Applied Numerical Linear Algebra, SIAM, Philadelphia, 1997. [3] G. H. Golub and C. F. Van Loan, Matrix Computations, Third Edition, The Johns Hopkins University Press, Baltimore, 1996. [4] G. W. Stewart, Introduction to Matrix Computations, Academic Press, New York, 1973. [5] G. W. Stewart, Matrix Algorithms: Basic Decompositions, SIAM, Philadelphia, 1998. [6] L. N. Trefethen and D. Bau, III, Numerical Linear Algebra, SIAM, Philadelphia, 1997. [7] Google, Google Technology. http://www.google.com/technology/index.html [8] J. R. Gilbert, C. Moler, and R. Schreiber, Sparse matrices in MATLAB: Design and implementation, SIAM Journal on Matrix Analysis and Applications, 13 (1992), pp. 333356. [9] N. J. Higham, and F. Tisseur, A block algorithm for matrix 1-norm estimation, with an application to 1-norm pseudospectra, SIAM Journal on Matrix Analysis and Applications, 21 (2000), pp. 11851201. [10] A. Langville, and C. Meyer, Deeper Inside PageRank, http://meyer.math.ncsu.edu/Meyer/PS_Files/DeeperInsidePR.pdf [11] L. Page, S. Brin, R. Motwani, and T. Winograd, The PageRank Citation Ranking: Bringing Order to the Web. http://dbpubs.stanford.edu:8090/pub/1999-66 43 ...

View
Full
Document