diff options
author | Holden Rohrer <hr@hrhr.dev> | 2021-09-21 17:12:46 -0400 |
---|---|---|
committer | Holden Rohrer <hr@hrhr.dev> | 2021-09-21 17:12:46 -0400 |
commit | 32f4af5f369fa9f0b2988ecad7797f4bec3661c3 (patch) | |
tree | 7ce1c56011914681d6e2ffb5737dcdf1078d3930 /li | |
parent | b8433c9909bc5d29df16fd3011251a0a214d2b1a (diff) |
notes and homework
Diffstat (limited to 'li')
-rw-r--r-- | li/04_vector_space | 30 | ||||
-rw-r--r-- | li/05_linear_span | 17 | ||||
-rw-r--r-- | li/06_basis | 40 | ||||
-rw-r--r-- | li/hw2.tex | 265 | ||||
-rw-r--r-- | li/hw3.tex | 378 | ||||
-rw-r--r-- | li/matlab_hw.py | 183 |
6 files changed, 913 insertions, 0 deletions
diff --git a/li/04_vector_space b/li/04_vector_space new file mode 100644 index 0000000..720f44b --- /dev/null +++ b/li/04_vector_space @@ -0,0 +1,30 @@ +Vector Space is a set V and binary addition operator and a +scalar-by-vector multiplication operator where V is closed under the +binary and multiplication operators. (V, +, *) denotes the space. +V = R^n is one example of a vector space. + +A subspace is a subset of V s.t. (S, +, *) is a closed vector space. + +With A an mxn matrix, + +S = { x in R^n | Ax = 0 } is a subspace in R^n + +It is sufficient proof to check: +(i) \forall x,y \in S, x+y \in S + If Ax = 0, Ay = 0, A(x+y) = Ax + Ay = 0 + 0 = 0. +(ii) \forall x\in S, \alpha\in R, \alpha x \in S. + A(\alpha*x)= \alpha(Ax) = \alpha*0 = 0. +Therefore, S is a subspace of R^n. + +This is called the null space of A, null(A). + +AX = 0 is also a vector space over X\in R^M by R^M. + +If A is invertible, AX = 0 \to A^{-1} AX = A^{-1} 0 \to X = 0. + +S = span{a1, a2, ..., a_n} is a subspace in R^m + Proof of subspace omitted. + +S = column space of A = C(A) = R(A) = span of A. +R(A) stands for range. +Ax = b has a solution, iff b \in R(A) diff --git a/li/05_linear_span b/li/05_linear_span new file mode 100644 index 0000000..ed5fdf2 --- /dev/null +++ b/li/05_linear_span @@ -0,0 +1,17 @@ + Linear Independence and Linear Span +Solution set for Ax=0, Ax=B, and the dimension of null space/rank of a +matrix. + +Def. linear independence over vector space (V, +, *) +Let v = {v1, ..., vk} be vectors in V. We say v is linearly independent +iff whenever a1v1 + a2v2 + ... akvk = 0, a1 = a2 = ... = ak = 0. + +Where v_i in v in R^n, we can find linear independence by writing it as + +[ v1 | v2 | ... | vk ][a1 a2 ... ak]^T = 0 + +Prove: basis of a vector space has the same dimension regardless of the +vectors. + +Def basis: a linearly independent set of vectors which spans a vector +space. diff --git a/li/06_basis b/li/06_basis new file mode 100644 index 0000000..2a103b7 --- /dev/null +++ b/li/06_basis @@ -0,0 +1,40 @@ +(V, +, *) vector space. + +Let v be {v1, ..., vn} where vi \in V. +v is a basis of V if +(1) v is linearly independent +(2) span{v} = V. + +Lemma: Suppose v = {v1, ..., vj} is a linearly independent set in V and +w = {w1, w2, ..., wk} s.t. span{w} = V. Then j \leq k. + +Since w spans V, + +[ v1 | v2 | ... | vj ] = [ w1 | w2 | ... | wk ] A, + B C +where A is a k-by-j matrix. + +Suppose k < j. + Then Ax = 0 has nontrivial solutions because more unknowns (j) than + outputs. + Therefore, there exists x0 s.t. Ax0 = 0 with x0 \neq 0. + +B = CA \to Bx0 = CAx0 = 0, but Bx0 = 0 with nontrivial x0 is not +possible because B is defined to be linearly independent. +Therefore, k \geq j. + +Thm: If v and w are each bases of V, then k = j by application of the +previous lemma (k\geq j and j\geq k) +Let dim V = k = j. + +Suppose that dim V = n. + Any linearly independent set in V will have at most n vectors, and + Any spanning set in V must have at least n vectors. +Converses + Every set with more than n vectors in V is linearly dependent. + Every set with less than n vectors in V does not span V. +(2')For linearly independent set w which doesn't span V, there exists a +vector which may be added to V outside the span, and w \cup {v} is also +linearly independent. + [I think this is equivalent to saying "if w is linearly independent, + then w1 \not\in span{w \setminus {w1}} diff --git a/li/hw2.tex b/li/hw2.tex new file mode 100644 index 0000000..769cad3 --- /dev/null +++ b/li/hw2.tex @@ -0,0 +1,265 @@ +\def\bmatrix#1{\left[\matrix{#1}\right]} +{\bf\noindent Section 1.6} +{\bf\noindent 4.} +{\it (a)} +That means $A^{-1}$ exists s.t. $A^{-1}A = I.$ +$$AB = AC \to A^{-1}AB = A^{-1}AC \to B = C$$ + +{\it (b)} +$$B = \pmatrix{1 & 1\cr 1 & 0}$$ +$$C = \pmatrix{1 & 2\cr 1 & 1}$$ + +$$AB = AC.$$ % double check + +{\bf\noindent 10.} + +$$A_1^{-1} = \bmatrix{0 & 0 & 0 & 1\cr 0 & 0 & 1/2 & 0\cr 0 & 1/3 & 0 & +0\cr 1/4 & 0 & 0 & 0}$$ + +$$A_2^{-1} = \bmatrix{1 & 0 & 0 & 0\cr 1/2 & 1 & 0 & 0\cr 0 & 2/3 & 1 & +0\cr 0 & 0 & 3/4 & 1}$$ + +\def\rowone{{d\over ad-bc}&{-b\over ad-bc}} +\def\rowtwo{{-c\over ad-bc}&{a\over ad-bc}} +$$A_3^{-1} = \bmatrix{\rowone & 0 & 0\cr \rowtwo & 0 & 0\cr 0 & 0 & +\rowone \cr 0 & 0 & \rowtwo}$$ + +{\bf\noindent 11.} + +If $A = I$ and $B = -I,$ both are invertible, but $A + B$ is not. + +If $A = \pmatrix{1&0\cr 0&0}$ and $B = \pmatrix{0&0\cr 0&1},$ $A + B = +I$ is invertible, but neither $A$ nor $B$ are. + +$A = B = I$ gives $A + B,$ $A,$ and $B$ invertibility. +$A^{-1} = B^{-1} = I \to A^{-1}(A+B)B^{-1} = B^{-1} + A^{-1} \to 2I = +2I.$ + +{\bf\noindent 12.} + +{\it (a)} + +This one remains. + +{\it (b)} + +This one remains. + +{\it (c)} + +This one remains. + +{\it (d)} + +This one doesn't remain. + +{\it (e)} + +This one remains. + +% I don't really have any coherent reasons for this. +% But these all make a lot of sense to me. +% This isn't graded, but I would like to understand it better. Office +% hours? + +{\bf\noindent 23.} + +$$\bmatrix{10&20\cr 20&50}\bmatrix{x\cr y} = \bmatrix{1\cr 0} \to +\bmatrix{10&20\cr 0&10}\bmatrix{x \cr y} = \bmatrix{1\cr -2}$$$$\to +\bmatrix{1&2\cr 0&1}\bmatrix{x \cr y} = \bmatrix{1/10\cr -2/10} \to +\bmatrix{1&0\cr 0&1}\bmatrix{x \cr y} = \bmatrix{5/10\cr -2/10}. +$$ + +$$\bmatrix{10&20\cr 20&50}\bmatrix{x\cr y} = \bmatrix{0\cr 1} \to +\bmatrix{1&2\cr 0&1}\bmatrix{x\cr y} = \bmatrix{0\cr 1/10} \to +\bmatrix{1&0\cr 0&1}\bmatrix{x\cr y} = \bmatrix{-2/10\cr 1/10}. +$$ + +$$A^{-1} = {1\over10}\bmatrix{5&-2\cr -2&1}$$ + +{\bf\noindent 25.} + +{\it (a)} +$$\bmatrix{\hbox{row 1}\cr\hbox{row 2}\cr\hbox{row 1 + row +2}}\bmatrix{x\cr y\cr z} = \bmatrix{1\cr0\cr0} +\to \bmatrix{\hbox{row 1}\cr\hbox{row 2}\cr0}\bmatrix{x\cr y\cr z} = +\bmatrix{1\cr0\cr-1}.$$ +$0 \neq -1,$ so this is inconsistent. + +{\it (b)} + +If $b_3 = b_2 + b_1,$ this matrix admits a solution. + +{\it (c)} + +As already shown, row 3 becomes a zero row during elimination. + +{\bf\noindent 27.} + +$B = E_12A,$ +and the elementary matrix which switches the first two rows is by +definition invertible ($E_12E_12 = I$) so if $A$ is invertible, $B$ is +also invertible, and $B^{-1} = A^{-1}E_12$ (I think this is a switch of +columns). + +{\bf\noindent 36.} + +$$\bmatrix{2 & 1 & 0 & 1 & 0 & 0\cr + 1 & 2 & 1 & 0 & 1 & 0\cr + 0 & 1 & 2 & 0 & 0 & 1} +\to +\bmatrix{2 & 1 & 0 & 1 & 0 & 0\cr + 0 & 1.5 & 1 & .5 & 1 & 0\cr + 0 & 1 & 2 & 0 & 0 & 1} +\to +\bmatrix{2 & 1 & 0 & 1 & 0 & 0\cr + 0 & 1 & 0 & .5 & 1 & .5\cr + 0 & 1 & 2 & 0 & 0 & 1} +$$$$\to +\bmatrix{2 & 0 & 0 & .5 & -1 & -.5\cr + 0 & 1 & 0 & .5 & 1 & .5\cr + 0 & 1 & 2 & 0 & 0 & 1} +\to +\bmatrix{2 & 0 & 0 & .5 & -1 & -.5\cr + 0 & 1 & 0 & .5 & 1 & .5\cr + 0 & 0 & 2 & -.5 & -1 & .5} +$$ + +Inverse is +$$ +\bmatrix{.25 & -.5 & -.25\cr + .5 & 1 & .5\cr + -.25 & -.5 & .25} +$$ + +% {\bf\noindent 40.} + +{\bf\noindent Section 2.1} +{\bf\noindent 2.} + +{\it (a)} + +This is. $k(0, b_2, b_3) = (0, kb_2, kb_3),$ and similar for addition. + +{\it (b)} + +This isn't. $2(0, 1, 0) = (0, 2, 0),$ which isn't in the subset. + +{\it (c)} + +This isn't. $(1, 1, 0) + (1, 0, 1) = (2, 1, 1),$ which isn't in the +subset, even though its two starting components are. + +{\it (d)} + +This is, by definition. + +{\it (e)} + +Yes, this is the null space of a homogenous matrix with one row. It is a +subspace. + +{\bf\noindent 7.} + +{\it (a)} + +This is not a subspace. The given sequence plus the given sequence with +a zero prepended would sum to a sequence without a zero, meaning it is +not closed under addition. + +{\it (b)} + +This is a subspace because the sums and multiples are closed. + +{\it (c)} + +$x_{j+1} \leq x_j \land y_{j+1} \leq y_j \to x_{j+1} + y_{j+1} \leq x_j ++ y_j,$ +and similar for multiplication. + +{\it (d)} + +Yes, the sum of two convergent sequences will tend to the sum of their +limits, and each can be multiplied by a real. + +{\it (e)} + +Yes, the sum of two arithmetic series will be an arithmetic series, and +they work with multiplications by a real. + +{\it (f)} + +This isn't because $x_1 = 1,$ $k = 2,$ and $x_1 = 1,$ $k = 3$ sum to +$$2 + 5 + 14 + \cdots,$$ which is not a geometric sequence. + +{\bf\noindent 8.} % required + +The two rows are linearly independent, so with 3 unknows, this forms a +line (intersection of two equations/planes). + +{\bf\noindent 18.} + +{\it (a)} + +It's probably a line, but it could be a plane. + +{\it (b)} + +It's probably $(0, 0, 0)$, but it could be a line. + +% {\it (c)} + +% I don't know. + +{\bf\noindent 22.} % required + +{\it (a)} + +This matrix has a column space with basis $(1, 2, -1)^T,$ so it is only +solvable if $b = (b_1, b_2, b_3)^T$ is a multiple of that vector. + +{\it (b)} + +These two columns are linearly independent, so any b within the span of +them has a solution. + +{\bf\noindent 25.} + +Unless $b$ is in the span of $A.$ If A is the zero $3\times 1$ matrix, +and $b$ is the first column of $I_3,$ then the column space increases. +If $A = b,$ then it would not extend the column space. Iff $b$ is +already in the column space, then $Ax = b$ has a solution. + +{\bf\noindent 26.} + +They are not equal only if $B$ is not invertible (it is a necessary but +not sufficient condition). If $A$ is invertible, like $A = I,$ then +non-invertible $B$ gives $AB = B$ with smaller column space. + +{\bf\noindent 28.} % required + +{\it (a)} + +False. Counterexample: The complement of $(0, 0, 0)$ does not pass the +test $0v \in V.$ + +{\it (b)} + +True. The space with only the zero vector has a basis of only zero +vectors. + +{\it (c)} + +True. Each of $2A$'s columns can be transformed into the columns of $A$ +by division by 2. + +{\it (d)} + +False. The column space of $-I$ is the full space, unlike the zero +matrix it's ``based on.'' + +{\bf\noindent 30.} % required + +$${\bf C}(A) = {\bf R}^9.$$ + +\bye diff --git a/li/hw3.tex b/li/hw3.tex new file mode 100644 index 0000000..870daf0 --- /dev/null +++ b/li/hw3.tex @@ -0,0 +1,378 @@ +\def\bmatrix#1{\left[\matrix{#1}\right]} + + {\noindent\bf Section 2.2} + +{\noindent\bf 12.} + +{\it (a)} + +This is correct. It's equal to the number of linearly independent +rows/dimension of the row space/rank because if any of the non-zero rows +were linearly dependent, they would have been eliminated to a zero row +when forming $R.$ + +{\it (b)} + +This is false. A zero matrix has rank zero but can have a nonzero number +in this property if it has more columns than rows. + +{\it (c)} + +This is true. All columns are either pivot columns or free columns, and +the rank is the number of pivot columns. + +{\it (d)} + +No. The following matrix has four ones but rank one: +$$\bmatrix{1&1&1&1}$$ + +{\noindent\bf 26.} + +The maximum rank of a matrix is the smaller of its number of rows and +its number of columns because the pivot columns and rows are strictly +less than the total number of each. Therefore, $C$ and $A$ have at most +rank 2, and $CA$ also has at most rank 2 (column space of $A$ is a +superset of the column space of $CA,$ which becomes obvious if they're +treated like functions). $CA$ is a $3\times 3$ matrix, and $I_3$ has +rank 3, so $CA \neq I.$ + +$AC = I$ if +$$A = \bmatrix{1 & 0 & 0\cr + 0 & 1 & 0}$$ + and +$$C = \bmatrix{1 & 0\cr + 0 & 1\cr + 0 & 0}$$ + +{\noindent\bf 42.} + +If $Ax = b$ has infinitely many solutions, then there exists infinitely +many solutions to $Ay = 0$ if $y = x - x_0$ where $x_0$ is a particular +solution to $Ax_0 = b.$ If there exists one particular solution $x_1$ to +$Ax_1 = B,$ then there must be an infinite number $A(x_1+y) = B$ where +$y$ is in the null space of $A$ as noted earlier. + +However, $Ax = B$ could have zero solutions. The matrix +$$A = \bmatrix{1&0\cr 0&0}$$ +does not include $b_0 = \bmatrix{0\cr 1}$ in its column space, so $Ax = +b_0$ would have zero solutions even though $Ax = \bmatrix{1\cr 0}$ has +an infinite number of solutions. + +\iffalse % practice problems + +{\noindent\bf 7.} + +$$R_3 = R_2 + R_1 \to c = 5 + 2.$$ + +{\noindent\bf 9.} + +{\it (a)} + +$$\bmatrix{1&2&3&4\cr 0&0&1&2\cr 0&0&0&0}\bmatrix{x_1\cr x_2\cr x_3\cr +x_4} = \bmatrix{0\cr 0\cr 0} \to x = \bmatrix{-4\cr 0\cr -2\cr 1}x_4 + +\bmatrix{-2\cr 1\cr 0\cr 0}x_2$$ +$$R = \bmatrix{1&2&0&-2\cr 0&0&0&1&2\cr 0&0&0&0}.$$ +$$Rx = 0 \to x = \bmatrix{2&0&-2&1}x_4 + \bmatrix{-2&1&0&0}x_2.$$ + +{\it (b)} + +If the right-hand side is $(a, b, 0),$ the solution set will be the null +space plus a particular solution. In the case of $U,$ a particular +solution would be $(a, 0, b, 0).$ + +{\noindent\bf 10.} + +$$\bmatrix{0&1&-1\cr 1&0&-1}x = \bmatrix{1\cr -2\cr 0}.$$ +$$\bmatrix{0&1&-1\cr 1&0&-1\cr 1&1&-2}x = \bmatrix{1\cr -2\cr 0}.$$ + +{\noindent\bf 14.} + +$$R_A = \bmatrix{1&2&0\cr 0&0&1\cr 0&0&0}.$$ +$$R_B = \bmatrix{1&2&0&1&2&0\cr 0&0&1&0&0&1\cr 0&0&0&0&0&0}.$$ +$$R_C = \bmatrix{1&2&0&0&0&0\cr 0&0&1&0&0&0\cr 0&0&0&0&0&0\cr +0&0&0&1&2&0\cr 0&0&0&0&0&1\cr 0&0&0&0&0&0}.$$ + +{\noindent\bf 21.} + +The rank $r$ is the number of pivot rows and the number of pivot +columns, so the subset of these rows and columns would be an $r\times r$ +matrix. They are by definition linearly independent, so each spans/forms +a basis for $R^r,$ giving them invertibility. + +{\noindent\bf 24.} + +The rank of $A$ is the same as the rank of $A^T,$ so +$${\rm rank}(AB) \leq {\rm rank}(A) \to {\rm rank}((AB)^T) \leq +{\rm rank}(A^T) \to {\rm rank}(B^TA^T) \leq {\rm rank}(A^T) \to +{\rm rank}(AB) \leq {\rm rank}(B).$$ + +{\noindent\bf 25.} + + + +{\noindent\bf 36.} + +{\it (a)} + +All vectors in $R^3$ are in the column space, so only the trivial +combination of the rows of $A$ gives zero. + +{\it (b)} + +Only vectors where $b_3 = 2b_2$ are within the column space. This means +that $2x_2 = -x_3$ gives a zero combination. % double check this. + +{\noindent\bf 40.} + +$x_5$ is a free variable, the zero vector isn't the only solution to +$Ax=0,$ and if $Ax=b$ has a solution, then it has infinite solutions. + +{\noindent\bf 43.} + +{\it (a)} + +$q=6$ gives a rank of 1 for $B,$ and $q=3$ gives a rank of 1 for the +frist matrix. + +{\it (b)} + +$q = 7$ gives a rank of 2 for both matrices. + +{\it (c)} + +A rank of 3 is impossible for both matrices. + +{\noindent\bf 45.} +% idk come back to this. + +{\it (a)} + +$r < n.$ + +{\it (b)} + +$r > m.$ $r\geq n.$ % ??? + +{\it (c)} + +$r < n.$ + +{\it (d)} + +{\noindent\bf 53.} + +{\it (a)} + +False. The zero matrix has $n$ free variables. + +{\it (b)} + +True. If the linear function corresponding to the matrix can be +inverted, it must not have a non-zero null-space (i.e. a non-injective +relation). + +{\noindent\bf 60.} + +$$\bmatrix{1&0&-2&-3\cr0&1&-2&-1}$$ has this nullspace. + +{\noindent\bf 61.} + +% simple enough to construct + +{\noindent\bf 62.} + + + +{\noindent\bf 63.} +{\noindent\bf 64.} + +{\noindent\bf 65.} + +$$\bmatrix{0&0\cr 1&0}$$ + +{\noindent\bf 66.} + +Dimension of null space is $n-r = 3-r,$ and dimension of column space is +$r,$ so they cannot have the same dimension and therefore cannot be +equal. + +\fi + + {\noindent\bf Section 2.3} + +{\noindent\bf 22.} + +{\it (a)} + +They might not span ${\bf R}^4$ if, for example, they are all the zero +vector, but they could span it, like if the first four were elementary +vectors $e_1$ to $e_4.$ + +{\it (b)} + +They are not linearly independent because 4 is the maximal independent +set. + +{\it (c)} + +Any four might be a basis for ${\bf R}^4,$ because they could be +linearly independent and four vectors in ${\bf R}^4$ could span it. + +{\it (d)} + +$Ax = b$ might not have a solution. It could have a solution depending +on the $b,$ but $0x = e_1,$ where $0$ refers to the zero vector for $A$ +has zero solutions. + +{\noindent\bf 27.} + +The column space of $A$ has basis in $\{(1, 0, 1)^T, (3, 1, 3)^T\}$ and +the column space of $U$ has basis in $\{(1, 0, 0)^T, (3, 1, 0)^T\}.$ +The two matrices have the same row space, based in $\{(1, 3, 2), (0, 1, +1)\}.$ +They also have the same null space, based in $\{(-1, 1, -1)\}.$ + +{\noindent\bf 32.} + +{\it (a)} + +The dimension is 3 because this is the set of vectors on ${\bf R}^4$ +under one linear constraint: $v_4 = -(v_3 + v_2 + v_1).$ + +{\it (b)} + +The dimension is 0 because the identity matrix, by definition only +returns 0 if given 0. + +{\it (c)} + +The dimension is 16 because there are 16 unconstrained components. + +{\noindent\bf 36.} + +6 independent vectors satisfy $Ax=0$ by the rank theorem. $A^T$ has the +same rank, so 53 independent vectors satisfy $A^Ty = 0.$ + +{\noindent\bf 42.} + +$\{x^3, x^2, x, 1\}$ form a basis of the polynomials of degree up to 3, +and this set restricted to those where $p(1) = 0$ has basis $\{x^3-1, +x^2-1, x-1\}.$ + +\iffalse % practice problems + +{\noindent\bf 7.} + +$$v_1 - v_2 + v_3 = w_2 - w_3 - w_1 + w_3 + w_1 - w_2 = 0,$$ +proving dependence of these vectors. + +{\noindent\bf 8.} % not an actual problem + +$$c_1v_1 + c_2v_2 + c_3v_3 = c_1(w_2 + w_3) + c_2(w_1 + w_3) + c_3(w_1 + +w_2) = (c_2+c_3)w_1 + (c_1+c_3)w_2 + (c_1+c_2)w_3 = 0.$$ +Because the set of $w$ vectors are independent, this sum is only equal +to zero if $c_2 + c_3 = 0 \to c_3 = -c_2,$ $c_1 + c_3 = 0 \to c_1 = -c_3 += +c_2,$ and $c_1+c_2 = 0 \to c_2 = c_1 = 0 \to c_3 = 0.$ + +{\noindent\bf 9.} + +{\it (a)} + +If $v_1$ to $v_3$ are linearly independent, the dimension of their +spanning set must be 3 (and the set equal to $R^3$), so $v_4 \in R^3$ +can be written as a combination of the other three. + +{\it (b)} + +$v_2 = kv_1$ where $k\in\bf R$ + +{\it (c)} + +$0v_1 + k(0,0,0) = 0,$ giving a non-trivial combination with the value +0. + +{\noindent\bf 12.} + +The vector $b$ is in the subspace spanned by the columns of $A$ when +there is a solution to $Ax = b.$ The vector $c$ is in the row space of +$A$ when there is a solution to $A^Tx = c$ or $x^TA = c.$ + +The zero vector is in every space, so the rows may still be independent. +(False) + +{\noindent\bf 13.} + +The dimensions of the column spaces and of the row spaces of $A$ and $U$ +are the same (2), and the row spaces are the same between the two (and +conversely, the null space) + +{\noindent\bf 21.} + +% easy + +{\noindent\bf 23.} + +If they are linearly independent, the rank of $A$ is $n.$ If they span +$R^m,$ the rank is $m.$ If they are a basis for $R^m,$ then both are +true and $n = m.$ + +{\noindent\bf 25.} + +{\it (a)} + +The columns are linearly independent, so there is no nontrivial linear +combination equal to 0. + +{\it (b)} + +The columns of $A$ span $R^5,$ so there must be a linear combination +(value of $x$) equal to $b.$ + +{\noindent\bf 26.} + +{\it (a)} + +True. Thm in the book. + +{\it (b)} + +False. See 31. + +{\noindent\bf 31.} + +If we let $v_k = e_k,$ the subspace with basis $(0, 0, 1, 1)$ does not +have a basis in the elementary vectors. + +{\noindent\bf 34.} + +% seems simple enough, don't know if I can do it. + +{\noindent\bf 35.} + +{\it (a)} + +False. The unit vector $e_1$'s single column is linearly independent, +but except in $R,$ it doesn't span $R^k,$ and $e_1x = e_2$ has no +solution. + +{\it (b)} + +True. The rank is at most $5,$ meaning there must be two free variables. + +{\noindent\bf 41.} + +{\it (a)} + +For dimension 1, $y_k = kx.$ + +{\it (b)} + +For dimension 2, $y_1 = x^2,$ $y_2 = 2x,$ and $y_3 = 3x.$ + +{\it (c)} + +For dimension 3, $y_k = x^k.$ + +\fi + +\bye diff --git a/li/matlab_hw.py b/li/matlab_hw.py new file mode 100644 index 0000000..71b89b6 --- /dev/null +++ b/li/matlab_hw.py @@ -0,0 +1,183 @@ +import numpy as np +import scipy.linalg + +print("Section 1.3: #30") + +coeff = ( + [[ 1, 1, 1 ], + [ 1, 2, 2 ], + [ 2, 3, -4]] + ) +print( +np.linalg.solve( + coeff, + [6, 11, 3]) +) + +print( +np.linalg.solve( + coeff, + [7, 10, 3]) +) + +print("Section 1.3: #32") + +dimension = 3 +tot = [0]*3 +ct = 10000 +for x in range(ct): + P, L, U = scipy.linalg.lu(np.random.rand(dimension, dimension)) + for i in range(dimension): + tot[i] += U[i][i] +print([x/ct for x in tot]); + +print("Section 1.4: #21") + +A = [[.5, .5], [.5, .5]] +B = [[1, 0], [0, -1]] +C = np.matmul(A, B) +print("A^2") +print(np.linalg.matrix_power(A, 2)) +print("A^3") +print(np.linalg.matrix_power(A, 3)) +print("A^k = A"); +print("B^2") +print(np.linalg.matrix_power(B, 2)) +print("B^3") +print(np.linalg.matrix_power(B, 3)) +print("B^{2k} = B^2. B^{2k+1} = B.") +print("C^2") +print(np.linalg.matrix_power(C, 2)) +print("C^3") +print(np.linalg.matrix_power(C, 3)) +print("C^k = 0") + +print("Section 1.4: #59") + +print("A * v = [3, 4, 5]") +print("v' * v = 50") +print(f"v * A = {np.matmul([3,4,5], np.identity(3))}") + +print("Section 1.4: #60") + +print("Av = [4. 4. 4. 4.]") +print("Bw = [10. 10. 10. 10.]") + +print("Section 1.6: #12") + +print("Section 1.6: #32") + +print("the inverse of the first matrix is") +print(np.linalg.inv(5*np.identity(4) - np.ones((4,4)))) +print("(meaning a = .4, b = .2)") +print("ones(4)*ones(4) =") +print(np.matmul(np.ones((4,4)),np.ones((4,4)))) +print("so we can solve this like the following equation: (a*eye +\n\ +b*ones)(c*eye + d*ones) = ac*eye + (ad+bc+dimension*bd)*ones = eye") +print("c = 1/a. d = -b/a/(a+dimension*b), giving for the next matrix") +print("a = 1/6, d = 6/(1/6+5*-1)") # WRONG + +print("Section 1.6: #47") + +print("I'm not sure. This is Python. It says (in either case):") +print("numpy.linalg.LinAlgError: Singular matrix") +''' +print(np.linalg.solve(np.ones((4, 4)), np.random.rand(4, 1))) +print(np.linalg.solve(np.ones((4, 4)), np.ones((4, 1)))) +''' + +''' #68 +dimension = 500 +northwest = np.random.rand(dimension, dimension) +for x in range(0, dimension): + for y in range(x+1, dimension): + northwest[y][x] = 0 +print(northwest) +''' +print("Section 1.6: #69") + + +from time import perf_counter +matrix = np.random.rand(500, 500) +start = perf_counter() +np.linalg.inv(matrix) +first_time = perf_counter() - start +print("500x500 inverse:", str(first_time) + "s") +matrix = np.random.rand(1000, 1000) +start = perf_counter() +np.linalg.inv(matrix) +second_time = perf_counter() - start +print("1000x1000 inverse:", str(second_time) + "s") +print("factor:", second_time / first_time) + +print("Section 1.6: #70") + +dimension = 1000 +I = np.identity(dimension) +A = np.random.rand(dimension, dimension) +U = np.random.rand(dimension, dimension) +for x in range(0, dimension): + for y in range(x+1, dimension): + U[y][x] = 0 + +start = perf_counter() +np.linalg.inv(U) +print("inverse of U:", str(perf_counter() - start) + "s") +start = perf_counter() +np.linalg.solve(U, I) +print("U\I:", str(perf_counter() - start) + "s") +start = perf_counter() +np.linalg.inv(A) +print("inverse of A:", str(perf_counter() - start) + "s") +start = perf_counter() +np.linalg.solve(A, I) +print("A\I:", str(perf_counter() - start) + "s") + +#print("Section 1.6: #71") + +print("Section 2.2: #33") + +print("For the first matrix:") +A = [[1, 3, 3], [2, 6, 9], [-1, -3, 3]] +b = [[1, 5, 5]] +# TODO +print(scipy.linalg.null_space(A)) +print("For the second matrix:") +A = [[1, 3, 1, 2], [2, 6, 4, 8], [0, 0, 2, 4]] +b = [[1, 3, 1]] + +print("Section 2.2: #35") + +print("For the first system:") +A = [[1, 2], [2, 4], [2, 5], [3, 9]] +print(sympy.Matrix.rref(A)) # ew is there no automated way to do this ? + +print("Section 2.2: #36") + +print("(a)") +A = np.array([[1, 2, 1], [2, 6, 3], [0, 2, 5]]) +print("Basis of the column space:") +print(scipy.linalg.orth(A)) +print("Multiplying the rows by this gives zero (for this matrix, \ +equivalent to [0 0 0]^T)") +print(scipy.linalg.null_space(A.transpose())) + +print("(b)") +A = np.array([[1, 1, 1], [1, 2, 4], [2, 4, 8]]) +print("Basis of the column space:") +print(scipy.linalg.orth(A)) +print("Multiplying the rows by this gives zero (for this matrix, \ +equivalent to [0 2 -1]^T)") +print(scipy.linalg.null_space(A.transpose())) + +print("Section 2.3: #2") + +print("Section 2.3: #5") + +print("Section 2.3: #13") + +print("Section 2.3: #16") + +print("Section 2.3: #18") + +print("Section 2.3: #24") |