• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/34

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

34 Cards in this Set

  • Front
  • Back

Injective

If f(x) = f(y) x=y (outputs can only hv1 input)

Surjective

If our codomaine is B and y is in B then there exists some x where f(x) = y (all y have an x)

Image of a linear map

The set of outputs

Kernel

The inputs that map to 0

Nullity T

Dim ker T

Rank

Dim Im T

Rank nullity theorem

Dim(T) = rankT + nullityT

Fundamental theorem of algebra

every polynomial equation of degree n with complex number coefficients has n roots, or solutions, in the complex numbers

Isomophism

Linear map thats bijective (a map that preserves the desired structure)

If V and W are vector spaces (both over a field K) of the same dimension n then V and W are
isomorphic
Let V , W be two vector spaces over the same field K. Then a function f : V → W is called a linear map if for all a, b ∈ K and u, v ∈ V
f(au + bv) = af(u) + bf(v).
Theorem 1.28. Let V /= {0} be a vector space and we assume that there exists a finite set S which spans V . Then
(1) There exists a subset of S which is a basis for V . In particular, V is finite-dimensional. (2) If a subset T ⊂ V is linearly independent then T can be extended to a basis of V . (3) If U is a subspace of V then U is finite-dimensional. If furthermore dim U = dim V then U = V
Definition 1.30 (Matrix of a linear map). Let V, W be vector spaces over a field K. Let A = {v1, . . . , vn} be an ordered basis of V and B = {w1, . . . , wm} an ordered basis of W. Given f ∈ L(V, W), the matrix associated to f (with respect to the bases A, B) is the m × n matrix

Let U, V, W be vector spaces over a field K with ordered bases A = {u1, . . . , ul}, B ={v1, . . . , vm} and C = {w1, . . . , wn}, respectively and suppose f : U → V and g : V → W are linear maps.Let MBA(f) = (aij ) be the matrix of f and MCB(g) = (bij ) the matrix of g relative to these bases. Theng ◦ f : U → W has matrix

Definition 1.37. Matrices A and A' are said to be similar, denoted A ~A'

if there exists an invertible matrix C such that A' = C −1AC. Similarity defines an equivalence relation on the space of square n × n matrices

Theorem 1.38. Suppose that A~ A' and A represents a linear operator f relative to some ordered basisB = {v1, . . . , vn}.

Then there is an ordered basis B' = {v' 1 , . . . , v'n} relative to which f has matrix A;

Let A, B ∈be Mn(K) be similar matrices. Then pA(x)

= pB(x).

Lemma 1.44. Let A ∈be Mn(K). Then the constant coefficient of pA(x) equals

det(A).

if A is similar to B, then det(A) =

det(B).

Definition 1.45. Let A be Mn(K). Then the trace of A, denoted trA, is equal to =

Sum of aii

Lemma 1.46. Let A be Mn(K). Then the x^n−1coefficient of pA(x) equals

(−1)^n−1 tr(A).

A is similar to B, then tr(A) =

tr(B)

Definition 1.1. A function f : Rn → Rm is called R-linear if it satisfies two properties:
• f(v + w) = f(v) + f(w) for all v, w ∈ Rn; and • f(λv) = λf(v) for all v ∈ Rn and λ ∈ R.
Definition 1.2. A function f : C n → C m is called C-linear if it satisfies two properties:
• f(v + w) = f(v) + f(w) for all v, w ∈ C n; and • f(λv) = λf(v) for all v ∈ C n and λ ∈ C.
Definition 1.3. Let V and W be vector spaces (technically over a field F). Then a function f : V → W is called F-linear if it satisfies two properties:
• f(v + w) = f(v) + f(w) for all v, w ∈ V ; and • f(λv) = λf(v) for all v ∈ V and λ ∈ F. If, from the context, the field F is clear, then we will talk about the function f being linear and omit the notation of F.
Definition 1.7. Fix an n ∈ {1, 2, 3, . . .} and choose an i ∈ {1, . . . , n}. Then ei
is the vector in Rn with ith entry 1 and all other entries zero. Another way to phrase this is that the jth entry of ei is given by the Kronecker delta function δij .
Lemma 1.10. Let V and W be vector spaces over F and f and g be F-linear functions from V to W such that f(ei) = g(ei) for all valid i. Then
f = g
Definition 2.1. Let V be a vector space over F and T : V → V a linear operator. Then a vector v ∈ V with v /=0 is called an eigenvector of T if
there exists a λ ∈ F such that T(v) = λv . The number λ is the called an eigenvalue of T.
Definition 2.4. Let T : V → V be a linear operator, we call the set of eigenvalues of T
the spectrum of T, denoted spec T.
Lemma 2.5. Let T : V → V be a linear operator and λ be an eigenvalue for T. Then the set of eigenvectors for λ, together with the zero vector,
form a subspace of V
Theorem 2.6. Let T : V → V be a linear operator and dim V < ∞. Then λ ∈ F is an eigenvalue of T if and only if
det(T − λI) = 0.
Definition 2.7. Let V be vector space over F and T : V → V be a linear operator,
• if dim V < ∞ then the characteristic polynomial of T is defined as pT (x) := det(T − xI). • if λ ∈ F is an eigenvalue of T the corresponding eigenspace is defined as E(λ) := ker(T − λI)
Definition 2.17. For a polynomial p(x) we say that λ1 ∈ C is a root of p if
p(λ1) = 0.
We can also talk about the multiplicity of a root. Specifically, if λ1 is a root of multiplicity m1
there exists a polynomial q(x) of degree n-m1 with q(λ1) /= 0 where p(x) = (x − λ1) m1 q(x), Note that the multiplicity of any root is a natural number.