SlideShare una empresa de Scribd logo
1 de 57
Descargar para leer sin conexión
Lesson 16 (S&H, Section 14.6)
      The Spectral Theorem and Applications

                          Math 20


                      October 26, 2007


Announcements
   Welcome parents!
   Problem Set 6 on the website. Due October 31.
   OH: Mondays 1–2, Tuesdays 3–4, Wednesdays 1–3 (SC 323)
   Prob. Sess.: Sundays 6–7 (SC B-10), Tuesdays 1–2 (SC 116)
Outline

   Hatsumon

   Concept Review
      Eigenbusiness
      Diagonalization

   The Spectral Theorem
      The split case
      The symmetric case
      Iterations

   Applications
      Back to Fibonacci
      Markov chains
A famous math problem

 “A certain man had one pair
 of rabbits together in a
 certain enclosed place, and
 one wishes to know how
 many are created from the
 pair in one year when it is the
 nature of them in a single
 month to bear another pair,
 and in the second month
 those born to bear also.
 Because the abovewritten pair        Leonardo of Pisa
 in the first month bore, you       (1170s or 1180s–1250)
 will double it; there will be        a/k/a Fibonacci
 two pairs in one month.”
Diagram of rabbits



                     f (0) = 1
Diagram of rabbits



                     f (0) = 1

                     f (1) = 1
Diagram of rabbits



                     f (0) = 1

                     f (1) = 1

                     f (2) = 2
Diagram of rabbits



                     f (0) = 1

                     f (1) = 1

                     f (2) = 2

                     f (3) = 3
Diagram of rabbits



                     f (0) = 1

                     f (1) = 1

                     f (2) = 2

                     f (3) = 3

                     f (4) = 5
Diagram of rabbits



                     f (0) = 1

                     f (1) = 1

                     f (2) = 2

                     f (3) = 3

                     f (4) = 5

                     f (5) = 8
An equation for the rabbits



   Let f (k) be the number of pairs of rabbits in month k. Each new
   month we have
       The same rabbits as last month
       Every pair of rabbits at least one month old producing a new
       pair of rabbits
An equation for the rabbits



   Let f (k) be the number of pairs of rabbits in month k. Each new
   month we have
        The same rabbits as last month
        Every pair of rabbits at least one month old producing a new
        pair of rabbits
   So
                      f (k) = f (k − 1) + f (k − 2)
Some fibonacci numbers

     k   f (k)
     0       1
     1       1
     2       2
     3       3
     4       5
     5       8     Question
     6      13     Can we find an explicit formula for f (k)?
     7      21
     8      34
     9      55
    10      89
    11    144
    12    233
Outline

   Hatsumon

   Concept Review
      Eigenbusiness
      Diagonalization

   The Spectral Theorem
      The split case
      The symmetric case
      Iterations

   Applications
      Back to Fibonacci
      Markov chains
Concept Review




  Definition
  Let A be an n × n matrix. The number λ is called an eigenvalue
  of A if there exists a nonzero vector x ∈ Rn such that

                               Ax = λx.                           (1)

  Every nonzero vector satisfying (??) is called an eigenvector of A
  associated with the eigenvalue λ.
Diagonalization Procedure



      Find the eigenvalues and eigenvectors.
      Arrange the eigenvectors in a matrix P and the corresponding
      eigenvalues in a diagonal matrix D.
      If you have “enough” eigenvectors (that is, one for each
      column of A), the original matrix is diagonalizable and equal
      to PDP−1 .
      Pitfalls:
          Repeated eigenvalues
          Nonreal eigenvalues
Outline

   Hatsumon

   Concept Review
      Eigenbusiness
      Diagonalization

   The Spectral Theorem
      The split case
      The symmetric case
      Iterations

   Applications
      Back to Fibonacci
      Markov chains
Question
Under what conditions on A would you be able to guarantee that
A is diagonalizable?
Theorem (Baby Spectral Theorem)
Suppose An×n has n distinct real eigenvalues. Then A is
diagonalizable.
Theorem (Spectral Theorem for Symmetric Matrices)
Suppose An×n is symmetric, that is, A = A. Then A is
diagonalizable. In fact, the eigenvectors can be chosen to be
pairwise orthogonal with length one, which means that P−1 = P .
Thus a symmetric matrix can be diagonalized as

                          A = PDP ,
Powers of diagonalizable matrices




   Remember if A is diagonalizable then

       Ak = (PDP−1 )k = (PDP−1 )(PDP−1 ) · · · (PDP−1 )
                                                k
                   −1       −1                  −1
          = PD(P        P)D(P    P) · · · D(P        P)DP−1 = PDk P−1
Another way to look at it


      If v is an eigenvector corresponding to eigenvalue λ, then

                               Ak v =
Another way to look at it


      If v is an eigenvector corresponding to eigenvalue λ, then

                               Ak v = λ k v
Another way to look at it


      If v is an eigenvector corresponding to eigenvalue λ, then

                                    Ak v = λ k v


      If v1 , . . . vn are eigenvectors with eigenvalues λ1 , . . . , λn , then

              Ak (c1 v1 + · · · + cn vn )
Another way to look at it


      If v is an eigenvector corresponding to eigenvalue λ, then

                                    Ak v = λ k v


      If v1 , . . . vn are eigenvectors with eigenvalues λ1 , . . . , λn , then

              Ak (c1 v1 + · · · + cn vn ) = c1 λk v1 + · · · + cn λk vn
                                                1                  n
Another way to look at it


      If v is an eigenvector corresponding to eigenvalue λ, then

                                    Ak v = λ k v


      If v1 , . . . vn are eigenvectors with eigenvalues λ1 , . . . , λn , then

              Ak (c1 v1 + · · · + cn vn ) = c1 λk v1 + · · · + cn λk vn
                                                1                  n



      If A is diagonalizable, there are n linearly independent
      eigenvectors, so any v can be written as a linear combination
      of them.
Outline

   Hatsumon

   Concept Review
      Eigenbusiness
      Diagonalization

   The Spectral Theorem
      The split case
      The symmetric case
      Iterations

   Applications
      Back to Fibonacci
      Markov chains
Setting up the Fibonacci sequence
   Recall the Fibonacci sequence defined by

          f (k + 2) = f (k) + f (k + 1),   f (0) = 1,   f (1) = 1
Setting up the Fibonacci sequence
   Recall the Fibonacci sequence defined by

          f (k + 2) = f (k) + f (k + 1),   f (0) = 1,   f (1) = 1

   Let’s let g (k) = f (k + 1). Then

        g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k).
Setting up the Fibonacci sequence
   Recall the Fibonacci sequence defined by

          f (k + 2) = f (k) + f (k + 1),     f (0) = 1,    f (1) = 1

   Let’s let g (k) = f (k + 1). Then

        g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k).

                  f (k)
   So if y(k) =         , we have
                  g (k)

                    f (k + 1)              g (k)            0 1
       y(k + 1) =               =                      =        y(k)
                    g (k + 1)          f (k) + g (k)        1 1
Setting up the Fibonacci sequence
   Recall the Fibonacci sequence defined by

          f (k + 2) = f (k) + f (k + 1),     f (0) = 1,    f (1) = 1

   Let’s let g (k) = f (k + 1). Then

        g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k).

                  f (k)
   So if y(k) =         , we have
                  g (k)

                     f (k + 1)             g (k)            0 1
       y(k + 1) =                 =                    =        y(k)
                     g (k + 1)         f (k) + g (k)        1 1

   So if A is this matrix, then

                             y(k) =
Setting up the Fibonacci sequence
   Recall the Fibonacci sequence defined by

          f (k + 2) = f (k) + f (k + 1),     f (0) = 1,    f (1) = 1

   Let’s let g (k) = f (k + 1). Then

        g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k).

                  f (k)
   So if y(k) =         , we have
                  g (k)

                     f (k + 1)             g (k)            0 1
       y(k + 1) =                 =                    =        y(k)
                     g (k + 1)         f (k) + g (k)        1 1

   So if A is this matrix, then

                             y(k) = Ak y(0).
Diagonalize

                            0 1
   The eigenvalues of A =         are found by solving
                            1 1

                      −λ  1
                 0=          = (−λ)(1 − λ) − 1
                       1 1−λ
                  = λ2 − λ − 1
Diagonalize

                            0 1
   The eigenvalues of A =         are found by solving
                            1 1

                        −λ  1
                   0=          = (−λ)(1 − λ) − 1
                         1 1−λ
                   = λ2 − λ − 1

   The roots are          √               √
                       1+ 5            1− 5
                    ϕ=              ϕ=
                                    ¯
                         2               2
Diagonalize

                             0 1
   The eigenvalues of A =            are found by solving
                             1 1

                        −λ  1
                   0=          = (−λ)(1 − λ) − 1
                         1 1−λ
                    = λ2 − λ − 1

   The roots are           √                 √
                        1+ 5              1− 5
                     ϕ=                ϕ=
                                       ¯
                          2                 2
   Notice that
                        ϕ + ϕ = 1,
                            ¯           ϕϕ = −1
                                         ¯
   (These facts make later calculations simpler.)
Eigenvectors

   We row reduce to find the eigenvectors:

                   −ϕ  1             −ϕ 1   −ϕ
                                             ¯             −ϕ 1
        A − ϕI =                =
                    1 1−ϕ            1  ϕ ←+
                                        ¯ −                 0 0

         1
   So        is an eigenvector for A corresponding to the eigenvalue ϕ.
         ϕ
Eigenvectors

   We row reduce to find the eigenvectors:

                     −ϕ  1              −ϕ 1   −ϕ
                                                ¯             −ϕ 1
        A − ϕI =                   =
                      1 1−ϕ             1  ϕ ←+
                                           ¯ −                 0 0

         1
   So           is an eigenvector for A corresponding to the eigenvalue ϕ.
         ϕ
              1
   Similarly,    is an eigenvector for A corresponding to the
              ϕ¯
   eigenvalue ϕ.
              ¯
Eigenvectors

   We row reduce to find the eigenvectors:

                     −ϕ  1              −ϕ 1   −ϕ
                                                ¯             −ϕ 1
        A − ϕI =                   =
                      1 1−ϕ             1  ϕ ←+
                                           ¯ −                 0 0

         1
   So           is an eigenvector for A corresponding to the eigenvalue ϕ.
         ϕ
              1
   Similarly,     is an eigenvector for A corresponding to the
              ϕ¯
   eigenvalue ϕ. So now we know that
              ¯

                                       1            1
                        y(k) = c1 ϕk     + c2 ϕk
                                              ¯
                                       ϕ            ϕ
                                                    ¯
What are the constants?


   To find c1 and c2 , we solve

                  1          1      1          1 1   c1
                      = c1     + c2       =
                  1          ϕ      ¯
                                    ϕ          ϕ ϕ
                                                 ¯   c2
                                 −1
                 c1       1 1         1         1   ϕ −1
                                                     ¯
          =⇒          =                   =
                 c2       ϕ ϕ
                            ¯         1       ϕ−ϕ
                                              ¯     −ϕ 1
                          1   ϕ−1
                              ¯           1
                      =
                       ϕ−ϕ ϕ+1
                       ¯                  1
                        1   ϕ
                      =√
                         5  −ϕ
                             ¯
Finally


   Putting this all together we have

                             ϕ     1      ϕ
                                          ¯      1
                      y(k) = √ ϕk      − √ ϕk¯
                               5   ϕ       5     ϕ
                                                 ¯
                   f (k)      1 ϕk+1 − ϕk+1
                                        ¯
                           =√     k+2 − ϕk+2
                   g (k)       5 ϕ      ¯
Finally


   Putting this all together we have

                             ϕ     1      ϕ
                                          ¯            1
                      y(k) = √ ϕk      − √ ϕk¯
                               5   ϕ       5           ϕ
                                                       ¯
                   f (k)      1 ϕk+1 − ϕk+1
                                        ¯
                           =√     k+2 − ϕk+2
                   g (k)       5 ϕ      ¯

   So
                                √                  √
                                                                
                                       k+1                 k+1
                     1  1+ 5                     1− 5
            f (k) = √                       −                   
                      5   2                        2
Markov Chains




      Recall the setup: T is a transition matrix giving the
      probabilities of switching from any state to any of the other
      states.
Markov Chains




      Recall the setup: T is a transition matrix giving the
      probabilities of switching from any state to any of the other
      states.
      We seek a steady-state vector, i.e., a probability vector u
      such that Tu = u.
Markov Chains




      Recall the setup: T is a transition matrix giving the
      probabilities of switching from any state to any of the other
      states.
      We seek a steady-state vector, i.e., a probability vector u
      such that Tu = u.
      This is nothing more than an eigenvector of eigenvalue 1!
Theorem
If T is a regular doubly-stochastic matrix, then
    1 is an eigenvalue for T
    all other eigenvalues of T have absolute value less than 1.
Let u be an eigenvector of eigenvalue 1, scaled so it’s a probability
vector. Let v2 , . . . , vn be eigenvectors corresponding to the other
eigenvalues λ2 , . . . , λn . Then for any initial state x(0), we have

       x(k) = Ak x(0) = Ak (c1 u + c2 λ2 v2 + · · · + cn λn vn )
                        = c1 u + c2 λk v2 + · · · + cn λk vn
                                     2                  n

So
                             x(k) → c1 u
Since each x(k) is a probability vector, c1 = 1. Hence

                             x(k) → c1 u

Más contenido relacionado

La actualidad más candente

Section 10: Lagrange's Theorem
Section 10: Lagrange's TheoremSection 10: Lagrange's Theorem
Section 10: Lagrange's TheoremKevin Johnson
 
Higher order ODE with applications
Higher order ODE with applicationsHigher order ODE with applications
Higher order ODE with applicationsPratik Gadhiya
 
Graph theory
Graph theoryGraph theory
Graph theoryKumar
 
Fourier series Introduction
Fourier series IntroductionFourier series Introduction
Fourier series IntroductionRizwan Kazi
 
Ordinary differential equation
Ordinary differential equationOrdinary differential equation
Ordinary differential equationJUGAL BORAH
 
First Order Differential Equations
First Order Differential EquationsFirst Order Differential Equations
First Order Differential EquationsItishree Dash
 
Diagonalization of matrix
Diagonalization of matrixDiagonalization of matrix
Diagonalization of matrixVANDANASAINI29
 
Application of eigen value eigen vector to design
Application of eigen value eigen vector to designApplication of eigen value eigen vector to design
Application of eigen value eigen vector to designHome
 
Differential Equations
Differential EquationsDifferential Equations
Differential EquationsKrupaSuthar3
 
Eigenvectors & Eigenvalues: The Road to Diagonalisation
Eigenvectors & Eigenvalues: The Road to DiagonalisationEigenvectors & Eigenvalues: The Road to Diagonalisation
Eigenvectors & Eigenvalues: The Road to DiagonalisationChristopher Gratton
 
Fractional Calculus
Fractional CalculusFractional Calculus
Fractional CalculusVRRITC
 
Generating functions solve recurrence
Generating functions solve recurrenceGenerating functions solve recurrence
Generating functions solve recurrenceHae Morgia
 
Convex optimization methods
Convex optimization methodsConvex optimization methods
Convex optimization methodsDong Guo
 

La actualidad más candente (20)

Integral calculus
Integral calculusIntegral calculus
Integral calculus
 
Section 10: Lagrange's Theorem
Section 10: Lagrange's TheoremSection 10: Lagrange's Theorem
Section 10: Lagrange's Theorem
 
Higher order ODE with applications
Higher order ODE with applicationsHigher order ODE with applications
Higher order ODE with applications
 
Graph theory
Graph theoryGraph theory
Graph theory
 
Fourier series Introduction
Fourier series IntroductionFourier series Introduction
Fourier series Introduction
 
Introduction to Graph Theory
Introduction to Graph TheoryIntroduction to Graph Theory
Introduction to Graph Theory
 
Ordinary differential equation
Ordinary differential equationOrdinary differential equation
Ordinary differential equation
 
First Order Differential Equations
First Order Differential EquationsFirst Order Differential Equations
First Order Differential Equations
 
Fourier series
Fourier seriesFourier series
Fourier series
 
Tensor analysis
Tensor analysisTensor analysis
Tensor analysis
 
Diagonalization of matrix
Diagonalization of matrixDiagonalization of matrix
Diagonalization of matrix
 
Application of eigen value eigen vector to design
Application of eigen value eigen vector to designApplication of eigen value eigen vector to design
Application of eigen value eigen vector to design
 
Interpolation
InterpolationInterpolation
Interpolation
 
Differential Equations
Differential EquationsDifferential Equations
Differential Equations
 
Eigenvectors & Eigenvalues: The Road to Diagonalisation
Eigenvectors & Eigenvalues: The Road to DiagonalisationEigenvectors & Eigenvalues: The Road to Diagonalisation
Eigenvectors & Eigenvalues: The Road to Diagonalisation
 
Fractional Calculus
Fractional CalculusFractional Calculus
Fractional Calculus
 
Limits and derivatives
Limits and derivativesLimits and derivatives
Limits and derivatives
 
derivatives math
derivatives mathderivatives math
derivatives math
 
Generating functions solve recurrence
Generating functions solve recurrenceGenerating functions solve recurrence
Generating functions solve recurrence
 
Convex optimization methods
Convex optimization methodsConvex optimization methods
Convex optimization methods
 

Destacado

Lesson 18: Geometric Representations of Functions
Lesson 18: Geometric Representations of FunctionsLesson 18: Geometric Representations of Functions
Lesson 18: Geometric Representations of FunctionsMatthew Leingang
 
Lesson05 Continuity Slides+Notes
Lesson05    Continuity Slides+NotesLesson05    Continuity Slides+Notes
Lesson05 Continuity Slides+NotesMatthew Leingang
 
Lesson 8: Tangents, Velocity, the Derivative
Lesson 8: Tangents, Velocity, the DerivativeLesson 8: Tangents, Velocity, the Derivative
Lesson 8: Tangents, Velocity, the DerivativeMatthew Leingang
 
Lesson05 Continuity Slides+Notes
Lesson05    Continuity Slides+NotesLesson05    Continuity Slides+Notes
Lesson05 Continuity Slides+NotesMatthew Leingang
 
Lesson 6: Continuity II, Infinite Limits
Lesson 6: Continuity II, Infinite LimitsLesson 6: Continuity II, Infinite Limits
Lesson 6: Continuity II, Infinite LimitsMatthew Leingang
 
Lesson 13: Rates of Change in Science
Lesson 13: Rates of Change in ScienceLesson 13: Rates of Change in Science
Lesson 13: Rates of Change in ScienceMatthew Leingang
 
Lesson 12: The Product and Quotient Rule
Lesson 12: The Product and Quotient RuleLesson 12: The Product and Quotient Rule
Lesson 12: The Product and Quotient RuleMatthew Leingang
 
Lesson 10: What does f' say about f?
Lesson 10: What does f' say about f?Lesson 10: What does f' say about f?
Lesson 10: What does f' say about f?Matthew Leingang
 
Lesson 9: The Derivative as a Function
Lesson 9: The Derivative  as  a FunctionLesson 9: The Derivative  as  a Function
Lesson 9: The Derivative as a FunctionMatthew Leingang
 
Lesson 15: Diagonalization
Lesson 15: DiagonalizationLesson 15: Diagonalization
Lesson 15: DiagonalizationMatthew Leingang
 
Lesson17: Functions Of Several Variables
Lesson17: Functions Of  Several  VariablesLesson17: Functions Of  Several  Variables
Lesson17: Functions Of Several VariablesMatthew Leingang
 
Lesson14: Derivatives of Trigonometric Functions
Lesson14: Derivatives of Trigonometric FunctionsLesson14: Derivatives of Trigonometric Functions
Lesson14: Derivatives of Trigonometric FunctionsMatthew Leingang
 
Lesson 7: Limits at Infinity
Lesson 7: Limits at InfinityLesson 7: Limits at Infinity
Lesson 7: Limits at InfinityMatthew Leingang
 
Lesson 12: Linear Independence
Lesson 12: Linear IndependenceLesson 12: Linear Independence
Lesson 12: Linear IndependenceMatthew Leingang
 
Lesson 16: Implicit Differentiation
Lesson 16: Implicit DifferentiationLesson 16: Implicit Differentiation
Lesson 16: Implicit DifferentiationMatthew Leingang
 
Lesson 13: Rank and Solutions to Systems of Linear Equations
Lesson 13: Rank and Solutions to Systems of Linear EquationsLesson 13: Rank and Solutions to Systems of Linear Equations
Lesson 13: Rank and Solutions to Systems of Linear EquationsMatthew Leingang
 

Destacado (20)

Lesson 18: Geometric Representations of Functions
Lesson 18: Geometric Representations of FunctionsLesson 18: Geometric Representations of Functions
Lesson 18: Geometric Representations of Functions
 
Lesson05 Continuity Slides+Notes
Lesson05    Continuity Slides+NotesLesson05    Continuity Slides+Notes
Lesson05 Continuity Slides+Notes
 
Lesson 5: Continuity
Lesson 5: ContinuityLesson 5: Continuity
Lesson 5: Continuity
 
Lesson 15: The Chain Rule
Lesson 15: The Chain RuleLesson 15: The Chain Rule
Lesson 15: The Chain Rule
 
Lesson 8: Tangents, Velocity, the Derivative
Lesson 8: Tangents, Velocity, the DerivativeLesson 8: Tangents, Velocity, the Derivative
Lesson 8: Tangents, Velocity, the Derivative
 
Lesson05 Continuity Slides+Notes
Lesson05    Continuity Slides+NotesLesson05    Continuity Slides+Notes
Lesson05 Continuity Slides+Notes
 
Lesson 6: Continuity II, Infinite Limits
Lesson 6: Continuity II, Infinite LimitsLesson 6: Continuity II, Infinite Limits
Lesson 6: Continuity II, Infinite Limits
 
Lesson 13: Rates of Change in Science
Lesson 13: Rates of Change in ScienceLesson 13: Rates of Change in Science
Lesson 13: Rates of Change in Science
 
Lesson 10: Inverses
Lesson 10: InversesLesson 10: Inverses
Lesson 10: Inverses
 
Lesson 12: The Product and Quotient Rule
Lesson 12: The Product and Quotient RuleLesson 12: The Product and Quotient Rule
Lesson 12: The Product and Quotient Rule
 
Lesson 10: What does f' say about f?
Lesson 10: What does f' say about f?Lesson 10: What does f' say about f?
Lesson 10: What does f' say about f?
 
Lesson 9: The Derivative as a Function
Lesson 9: The Derivative  as  a FunctionLesson 9: The Derivative  as  a Function
Lesson 9: The Derivative as a Function
 
Lesson 15: Diagonalization
Lesson 15: DiagonalizationLesson 15: Diagonalization
Lesson 15: Diagonalization
 
Lesson17: Functions Of Several Variables
Lesson17: Functions Of  Several  VariablesLesson17: Functions Of  Several  Variables
Lesson17: Functions Of Several Variables
 
Lesson 19: Related Rates
Lesson 19: Related RatesLesson 19: Related Rates
Lesson 19: Related Rates
 
Lesson14: Derivatives of Trigonometric Functions
Lesson14: Derivatives of Trigonometric FunctionsLesson14: Derivatives of Trigonometric Functions
Lesson14: Derivatives of Trigonometric Functions
 
Lesson 7: Limits at Infinity
Lesson 7: Limits at InfinityLesson 7: Limits at Infinity
Lesson 7: Limits at Infinity
 
Lesson 12: Linear Independence
Lesson 12: Linear IndependenceLesson 12: Linear Independence
Lesson 12: Linear Independence
 
Lesson 16: Implicit Differentiation
Lesson 16: Implicit DifferentiationLesson 16: Implicit Differentiation
Lesson 16: Implicit Differentiation
 
Lesson 13: Rank and Solutions to Systems of Linear Equations
Lesson 13: Rank and Solutions to Systems of Linear EquationsLesson 13: Rank and Solutions to Systems of Linear Equations
Lesson 13: Rank and Solutions to Systems of Linear Equations
 

Similar a Lesson 16 The Spectral Theorem and Applications

Taylor Polynomials and Series
Taylor Polynomials and SeriesTaylor Polynomials and Series
Taylor Polynomials and SeriesMatthew Leingang
 
An introduction to quantum stochastic calculus
An introduction to quantum stochastic calculusAn introduction to quantum stochastic calculus
An introduction to quantum stochastic calculusSpringer
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
Lesson31 Higher Dimensional First Order Difference Equations Slides
Lesson31   Higher Dimensional First Order Difference Equations SlidesLesson31   Higher Dimensional First Order Difference Equations Slides
Lesson31 Higher Dimensional First Order Difference Equations SlidesMatthew Leingang
 
Cs229 cvxopt
Cs229 cvxoptCs229 cvxopt
Cs229 cvxoptcerezaso
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
 
The Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is DiagonalizableThe Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is DiagonalizableJay Liew
 
Omiros' talk on the Bernoulli factory problem
Omiros' talk on the  Bernoulli factory problemOmiros' talk on the  Bernoulli factory problem
Omiros' talk on the Bernoulli factory problemBigMC
 
SMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionSMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionLilyana Vankova
 

Similar a Lesson 16 The Spectral Theorem and Applications (20)

Taylor Polynomials and Series
Taylor Polynomials and SeriesTaylor Polynomials and Series
Taylor Polynomials and Series
 
An introduction to quantum stochastic calculus
An introduction to quantum stochastic calculusAn introduction to quantum stochastic calculus
An introduction to quantum stochastic calculus
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)
 
Taylor series
Taylor seriesTaylor series
Taylor series
 
Imc2016 day2-solutions
Imc2016 day2-solutionsImc2016 day2-solutions
Imc2016 day2-solutions
 
1105 ch 11 day 5
1105 ch 11 day 51105 ch 11 day 5
1105 ch 11 day 5
 
Lesson31 Higher Dimensional First Order Difference Equations Slides
Lesson31   Higher Dimensional First Order Difference Equations SlidesLesson31   Higher Dimensional First Order Difference Equations Slides
Lesson31 Higher Dimensional First Order Difference Equations Slides
 
Imc2016 day1-solutions
Imc2016 day1-solutionsImc2016 day1-solutions
Imc2016 day1-solutions
 
Imc2017 day2-solutions
Imc2017 day2-solutionsImc2017 day2-solutions
Imc2017 day2-solutions
 
Taylor problem
Taylor problemTaylor problem
Taylor problem
 
Cs229 cvxopt
Cs229 cvxoptCs229 cvxopt
Cs229 cvxopt
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
 
Mgm
MgmMgm
Mgm
 
The Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is DiagonalizableThe Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is Diagonalizable
 
AJMS_389_22.pdf
AJMS_389_22.pdfAJMS_389_22.pdf
AJMS_389_22.pdf
 
0708 ch 7 day 8
0708 ch 7 day 80708 ch 7 day 8
0708 ch 7 day 8
 
Thesis 6
Thesis 6Thesis 6
Thesis 6
 
Omiros' talk on the Bernoulli factory problem
Omiros' talk on the  Bernoulli factory problemOmiros' talk on the  Bernoulli factory problem
Omiros' talk on the Bernoulli factory problem
 
Md2521102111
Md2521102111Md2521102111
Md2521102111
 
SMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionSMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last version
 

Más de Matthew Leingang

Streamlining assessment, feedback, and archival with auto-multiple-choice
Streamlining assessment, feedback, and archival with auto-multiple-choiceStreamlining assessment, feedback, and archival with auto-multiple-choice
Streamlining assessment, feedback, and archival with auto-multiple-choiceMatthew Leingang
 
Electronic Grading of Paper Assessments
Electronic Grading of Paper AssessmentsElectronic Grading of Paper Assessments
Electronic Grading of Paper AssessmentsMatthew Leingang
 
Lesson 27: Integration by Substitution (slides)
Lesson 27: Integration by Substitution (slides)Lesson 27: Integration by Substitution (slides)
Lesson 27: Integration by Substitution (slides)Matthew Leingang
 
Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Matthew Leingang
 
Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Matthew Leingang
 
Lesson 27: Integration by Substitution (handout)
Lesson 27: Integration by Substitution (handout)Lesson 27: Integration by Substitution (handout)
Lesson 27: Integration by Substitution (handout)Matthew Leingang
 
Lesson 26: The Fundamental Theorem of Calculus (handout)
Lesson 26: The Fundamental Theorem of Calculus (handout)Lesson 26: The Fundamental Theorem of Calculus (handout)
Lesson 26: The Fundamental Theorem of Calculus (handout)Matthew Leingang
 
Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)Matthew Leingang
 
Lesson 25: Evaluating Definite Integrals (handout)
Lesson 25: Evaluating Definite Integrals (handout)Lesson 25: Evaluating Definite Integrals (handout)
Lesson 25: Evaluating Definite Integrals (handout)Matthew Leingang
 
Lesson 24: Areas and Distances, The Definite Integral (handout)
Lesson 24: Areas and Distances, The Definite Integral (handout)Lesson 24: Areas and Distances, The Definite Integral (handout)
Lesson 24: Areas and Distances, The Definite Integral (handout)Matthew Leingang
 
Lesson 24: Areas and Distances, The Definite Integral (slides)
Lesson 24: Areas and Distances, The Definite Integral (slides)Lesson 24: Areas and Distances, The Definite Integral (slides)
Lesson 24: Areas and Distances, The Definite Integral (slides)Matthew Leingang
 
Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)Matthew Leingang
 
Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)Matthew Leingang
 
Lesson 22: Optimization Problems (slides)
Lesson 22: Optimization Problems (slides)Lesson 22: Optimization Problems (slides)
Lesson 22: Optimization Problems (slides)Matthew Leingang
 
Lesson 22: Optimization Problems (handout)
Lesson 22: Optimization Problems (handout)Lesson 22: Optimization Problems (handout)
Lesson 22: Optimization Problems (handout)Matthew Leingang
 
Lesson 21: Curve Sketching (slides)
Lesson 21: Curve Sketching (slides)Lesson 21: Curve Sketching (slides)
Lesson 21: Curve Sketching (slides)Matthew Leingang
 
Lesson 21: Curve Sketching (handout)
Lesson 21: Curve Sketching (handout)Lesson 21: Curve Sketching (handout)
Lesson 21: Curve Sketching (handout)Matthew Leingang
 
Lesson 20: Derivatives and the Shapes of Curves (slides)
Lesson 20: Derivatives and the Shapes of Curves (slides)Lesson 20: Derivatives and the Shapes of Curves (slides)
Lesson 20: Derivatives and the Shapes of Curves (slides)Matthew Leingang
 
Lesson 20: Derivatives and the Shapes of Curves (handout)
Lesson 20: Derivatives and the Shapes of Curves (handout)Lesson 20: Derivatives and the Shapes of Curves (handout)
Lesson 20: Derivatives and the Shapes of Curves (handout)Matthew Leingang
 

Más de Matthew Leingang (20)

Making Lesson Plans
Making Lesson PlansMaking Lesson Plans
Making Lesson Plans
 
Streamlining assessment, feedback, and archival with auto-multiple-choice
Streamlining assessment, feedback, and archival with auto-multiple-choiceStreamlining assessment, feedback, and archival with auto-multiple-choice
Streamlining assessment, feedback, and archival with auto-multiple-choice
 
Electronic Grading of Paper Assessments
Electronic Grading of Paper AssessmentsElectronic Grading of Paper Assessments
Electronic Grading of Paper Assessments
 
Lesson 27: Integration by Substitution (slides)
Lesson 27: Integration by Substitution (slides)Lesson 27: Integration by Substitution (slides)
Lesson 27: Integration by Substitution (slides)
 
Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)
 
Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)
 
Lesson 27: Integration by Substitution (handout)
Lesson 27: Integration by Substitution (handout)Lesson 27: Integration by Substitution (handout)
Lesson 27: Integration by Substitution (handout)
 
Lesson 26: The Fundamental Theorem of Calculus (handout)
Lesson 26: The Fundamental Theorem of Calculus (handout)Lesson 26: The Fundamental Theorem of Calculus (handout)
Lesson 26: The Fundamental Theorem of Calculus (handout)
 
Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)
 
Lesson 25: Evaluating Definite Integrals (handout)
Lesson 25: Evaluating Definite Integrals (handout)Lesson 25: Evaluating Definite Integrals (handout)
Lesson 25: Evaluating Definite Integrals (handout)
 
Lesson 24: Areas and Distances, The Definite Integral (handout)
Lesson 24: Areas and Distances, The Definite Integral (handout)Lesson 24: Areas and Distances, The Definite Integral (handout)
Lesson 24: Areas and Distances, The Definite Integral (handout)
 
Lesson 24: Areas and Distances, The Definite Integral (slides)
Lesson 24: Areas and Distances, The Definite Integral (slides)Lesson 24: Areas and Distances, The Definite Integral (slides)
Lesson 24: Areas and Distances, The Definite Integral (slides)
 
Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)
 
Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)Lesson 23: Antiderivatives (slides)
Lesson 23: Antiderivatives (slides)
 
Lesson 22: Optimization Problems (slides)
Lesson 22: Optimization Problems (slides)Lesson 22: Optimization Problems (slides)
Lesson 22: Optimization Problems (slides)
 
Lesson 22: Optimization Problems (handout)
Lesson 22: Optimization Problems (handout)Lesson 22: Optimization Problems (handout)
Lesson 22: Optimization Problems (handout)
 
Lesson 21: Curve Sketching (slides)
Lesson 21: Curve Sketching (slides)Lesson 21: Curve Sketching (slides)
Lesson 21: Curve Sketching (slides)
 
Lesson 21: Curve Sketching (handout)
Lesson 21: Curve Sketching (handout)Lesson 21: Curve Sketching (handout)
Lesson 21: Curve Sketching (handout)
 
Lesson 20: Derivatives and the Shapes of Curves (slides)
Lesson 20: Derivatives and the Shapes of Curves (slides)Lesson 20: Derivatives and the Shapes of Curves (slides)
Lesson 20: Derivatives and the Shapes of Curves (slides)
 
Lesson 20: Derivatives and the Shapes of Curves (handout)
Lesson 20: Derivatives and the Shapes of Curves (handout)Lesson 20: Derivatives and the Shapes of Curves (handout)
Lesson 20: Derivatives and the Shapes of Curves (handout)
 

Último

How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...
How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...
How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...Hector Del Castillo, CPM, CPMM
 
GUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdf
GUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdfGUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdf
GUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdfDanny Diep To
 
business environment micro environment macro environment.pptx
business environment micro environment macro environment.pptxbusiness environment micro environment macro environment.pptx
business environment micro environment macro environment.pptxShruti Mittal
 
Cyber Security Training in Office Environment
Cyber Security Training in Office EnvironmentCyber Security Training in Office Environment
Cyber Security Training in Office Environmentelijahj01012
 
NAB Show Exhibitor List 2024 - Exhibitors Data
NAB Show Exhibitor List 2024 - Exhibitors DataNAB Show Exhibitor List 2024 - Exhibitors Data
NAB Show Exhibitor List 2024 - Exhibitors DataExhibitors Data
 
20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdf
20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdf20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdf
20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdfChris Skinner
 
Jewish Resources in the Family Resource Centre
Jewish Resources in the Family Resource CentreJewish Resources in the Family Resource Centre
Jewish Resources in the Family Resource CentreNZSG
 
Healthcare Feb. & Mar. Healthcare Newsletter
Healthcare Feb. & Mar. Healthcare NewsletterHealthcare Feb. & Mar. Healthcare Newsletter
Healthcare Feb. & Mar. Healthcare NewsletterJamesConcepcion7
 
Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...
Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...
Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...ssuserf63bd7
 
1911 Gold Corporate Presentation Apr 2024.pdf
1911 Gold Corporate Presentation Apr 2024.pdf1911 Gold Corporate Presentation Apr 2024.pdf
1911 Gold Corporate Presentation Apr 2024.pdfShaun Heinrichs
 
How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...
How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...
How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...SOFTTECHHUB
 
Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...
Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...
Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...ssuserf63bd7
 
Fordham -How effective decision-making is within the IT department - Analysis...
Fordham -How effective decision-making is within the IT department - Analysis...Fordham -How effective decision-making is within the IT department - Analysis...
Fordham -How effective decision-making is within the IT department - Analysis...Peter Ward
 
20200128 Ethical by Design - Whitepaper.pdf
20200128 Ethical by Design - Whitepaper.pdf20200128 Ethical by Design - Whitepaper.pdf
20200128 Ethical by Design - Whitepaper.pdfChris Skinner
 
BAILMENT & PLEDGE business law notes.pptx
BAILMENT & PLEDGE business law notes.pptxBAILMENT & PLEDGE business law notes.pptx
BAILMENT & PLEDGE business law notes.pptxran17april2001
 
EUDR Info Meeting Ethiopian coffee exporters
EUDR Info Meeting Ethiopian coffee exportersEUDR Info Meeting Ethiopian coffee exporters
EUDR Info Meeting Ethiopian coffee exportersPeter Horsten
 
Church Building Grants To Assist With New Construction, Additions, And Restor...
Church Building Grants To Assist With New Construction, Additions, And Restor...Church Building Grants To Assist With New Construction, Additions, And Restor...
Church Building Grants To Assist With New Construction, Additions, And Restor...Americas Got Grants
 
trending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdf
trending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdftrending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdf
trending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdfMintel Group
 
WSMM Technology February.March Newsletter_vF.pdf
WSMM Technology February.March Newsletter_vF.pdfWSMM Technology February.March Newsletter_vF.pdf
WSMM Technology February.March Newsletter_vF.pdfJamesConcepcion7
 
Cybersecurity Awareness Training Presentation v2024.03
Cybersecurity Awareness Training Presentation v2024.03Cybersecurity Awareness Training Presentation v2024.03
Cybersecurity Awareness Training Presentation v2024.03DallasHaselhorst
 

Último (20)

How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...
How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...
How Generative AI Is Transforming Your Business | Byond Growth Insights | Apr...
 
GUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdf
GUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdfGUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdf
GUIDELINES ON USEFUL FORMS IN FREIGHT FORWARDING (F) Danny Diep Toh MBA.pdf
 
business environment micro environment macro environment.pptx
business environment micro environment macro environment.pptxbusiness environment micro environment macro environment.pptx
business environment micro environment macro environment.pptx
 
Cyber Security Training in Office Environment
Cyber Security Training in Office EnvironmentCyber Security Training in Office Environment
Cyber Security Training in Office Environment
 
NAB Show Exhibitor List 2024 - Exhibitors Data
NAB Show Exhibitor List 2024 - Exhibitors DataNAB Show Exhibitor List 2024 - Exhibitors Data
NAB Show Exhibitor List 2024 - Exhibitors Data
 
20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdf
20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdf20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdf
20220816-EthicsGrade_Scorecard-JP_Morgan_Chase-Q2-63_57.pdf
 
Jewish Resources in the Family Resource Centre
Jewish Resources in the Family Resource CentreJewish Resources in the Family Resource Centre
Jewish Resources in the Family Resource Centre
 
Healthcare Feb. & Mar. Healthcare Newsletter
Healthcare Feb. & Mar. Healthcare NewsletterHealthcare Feb. & Mar. Healthcare Newsletter
Healthcare Feb. & Mar. Healthcare Newsletter
 
Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...
Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...
Intermediate Accounting, Volume 2, 13th Canadian Edition by Donald E. Kieso t...
 
1911 Gold Corporate Presentation Apr 2024.pdf
1911 Gold Corporate Presentation Apr 2024.pdf1911 Gold Corporate Presentation Apr 2024.pdf
1911 Gold Corporate Presentation Apr 2024.pdf
 
How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...
How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...
How To Simplify Your Scheduling with AI Calendarfly The Hassle-Free Online Bo...
 
Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...
Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...
Horngren’s Financial & Managerial Accounting, 7th edition by Miller-Nobles so...
 
Fordham -How effective decision-making is within the IT department - Analysis...
Fordham -How effective decision-making is within the IT department - Analysis...Fordham -How effective decision-making is within the IT department - Analysis...
Fordham -How effective decision-making is within the IT department - Analysis...
 
20200128 Ethical by Design - Whitepaper.pdf
20200128 Ethical by Design - Whitepaper.pdf20200128 Ethical by Design - Whitepaper.pdf
20200128 Ethical by Design - Whitepaper.pdf
 
BAILMENT & PLEDGE business law notes.pptx
BAILMENT & PLEDGE business law notes.pptxBAILMENT & PLEDGE business law notes.pptx
BAILMENT & PLEDGE business law notes.pptx
 
EUDR Info Meeting Ethiopian coffee exporters
EUDR Info Meeting Ethiopian coffee exportersEUDR Info Meeting Ethiopian coffee exporters
EUDR Info Meeting Ethiopian coffee exporters
 
Church Building Grants To Assist With New Construction, Additions, And Restor...
Church Building Grants To Assist With New Construction, Additions, And Restor...Church Building Grants To Assist With New Construction, Additions, And Restor...
Church Building Grants To Assist With New Construction, Additions, And Restor...
 
trending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdf
trending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdftrending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdf
trending-flavors-and-ingredients-in-salty-snacks-us-2024_Redacted-V2.pdf
 
WSMM Technology February.March Newsletter_vF.pdf
WSMM Technology February.March Newsletter_vF.pdfWSMM Technology February.March Newsletter_vF.pdf
WSMM Technology February.March Newsletter_vF.pdf
 
Cybersecurity Awareness Training Presentation v2024.03
Cybersecurity Awareness Training Presentation v2024.03Cybersecurity Awareness Training Presentation v2024.03
Cybersecurity Awareness Training Presentation v2024.03
 

Lesson 16 The Spectral Theorem and Applications

  • 1. Lesson 16 (S&H, Section 14.6) The Spectral Theorem and Applications Math 20 October 26, 2007 Announcements Welcome parents! Problem Set 6 on the website. Due October 31. OH: Mondays 1–2, Tuesdays 3–4, Wednesdays 1–3 (SC 323) Prob. Sess.: Sundays 6–7 (SC B-10), Tuesdays 1–2 (SC 116)
  • 2. Outline Hatsumon Concept Review Eigenbusiness Diagonalization The Spectral Theorem The split case The symmetric case Iterations Applications Back to Fibonacci Markov chains
  • 3. A famous math problem “A certain man had one pair of rabbits together in a certain enclosed place, and one wishes to know how many are created from the pair in one year when it is the nature of them in a single month to bear another pair, and in the second month those born to bear also. Because the abovewritten pair Leonardo of Pisa in the first month bore, you (1170s or 1180s–1250) will double it; there will be a/k/a Fibonacci two pairs in one month.”
  • 5. Diagram of rabbits f (0) = 1 f (1) = 1
  • 6. Diagram of rabbits f (0) = 1 f (1) = 1 f (2) = 2
  • 7. Diagram of rabbits f (0) = 1 f (1) = 1 f (2) = 2 f (3) = 3
  • 8. Diagram of rabbits f (0) = 1 f (1) = 1 f (2) = 2 f (3) = 3 f (4) = 5
  • 9. Diagram of rabbits f (0) = 1 f (1) = 1 f (2) = 2 f (3) = 3 f (4) = 5 f (5) = 8
  • 10. An equation for the rabbits Let f (k) be the number of pairs of rabbits in month k. Each new month we have The same rabbits as last month Every pair of rabbits at least one month old producing a new pair of rabbits
  • 11. An equation for the rabbits Let f (k) be the number of pairs of rabbits in month k. Each new month we have The same rabbits as last month Every pair of rabbits at least one month old producing a new pair of rabbits So f (k) = f (k − 1) + f (k − 2)
  • 12.
  • 13. Some fibonacci numbers k f (k) 0 1 1 1 2 2 3 3 4 5 5 8 Question 6 13 Can we find an explicit formula for f (k)? 7 21 8 34 9 55 10 89 11 144 12 233
  • 14. Outline Hatsumon Concept Review Eigenbusiness Diagonalization The Spectral Theorem The split case The symmetric case Iterations Applications Back to Fibonacci Markov chains
  • 15. Concept Review Definition Let A be an n × n matrix. The number λ is called an eigenvalue of A if there exists a nonzero vector x ∈ Rn such that Ax = λx. (1) Every nonzero vector satisfying (??) is called an eigenvector of A associated with the eigenvalue λ.
  • 16. Diagonalization Procedure Find the eigenvalues and eigenvectors. Arrange the eigenvectors in a matrix P and the corresponding eigenvalues in a diagonal matrix D. If you have “enough” eigenvectors (that is, one for each column of A), the original matrix is diagonalizable and equal to PDP−1 . Pitfalls: Repeated eigenvalues Nonreal eigenvalues
  • 17.
  • 18. Outline Hatsumon Concept Review Eigenbusiness Diagonalization The Spectral Theorem The split case The symmetric case Iterations Applications Back to Fibonacci Markov chains
  • 19. Question Under what conditions on A would you be able to guarantee that A is diagonalizable?
  • 20. Theorem (Baby Spectral Theorem) Suppose An×n has n distinct real eigenvalues. Then A is diagonalizable.
  • 21.
  • 22. Theorem (Spectral Theorem for Symmetric Matrices) Suppose An×n is symmetric, that is, A = A. Then A is diagonalizable. In fact, the eigenvectors can be chosen to be pairwise orthogonal with length one, which means that P−1 = P . Thus a symmetric matrix can be diagonalized as A = PDP ,
  • 23.
  • 24. Powers of diagonalizable matrices Remember if A is diagonalizable then Ak = (PDP−1 )k = (PDP−1 )(PDP−1 ) · · · (PDP−1 ) k −1 −1 −1 = PD(P P)D(P P) · · · D(P P)DP−1 = PDk P−1
  • 25.
  • 26. Another way to look at it If v is an eigenvector corresponding to eigenvalue λ, then Ak v =
  • 27.
  • 28. Another way to look at it If v is an eigenvector corresponding to eigenvalue λ, then Ak v = λ k v
  • 29.
  • 30. Another way to look at it If v is an eigenvector corresponding to eigenvalue λ, then Ak v = λ k v If v1 , . . . vn are eigenvectors with eigenvalues λ1 , . . . , λn , then Ak (c1 v1 + · · · + cn vn )
  • 31. Another way to look at it If v is an eigenvector corresponding to eigenvalue λ, then Ak v = λ k v If v1 , . . . vn are eigenvectors with eigenvalues λ1 , . . . , λn , then Ak (c1 v1 + · · · + cn vn ) = c1 λk v1 + · · · + cn λk vn 1 n
  • 32. Another way to look at it If v is an eigenvector corresponding to eigenvalue λ, then Ak v = λ k v If v1 , . . . vn are eigenvectors with eigenvalues λ1 , . . . , λn , then Ak (c1 v1 + · · · + cn vn ) = c1 λk v1 + · · · + cn λk vn 1 n If A is diagonalizable, there are n linearly independent eigenvectors, so any v can be written as a linear combination of them.
  • 33. Outline Hatsumon Concept Review Eigenbusiness Diagonalization The Spectral Theorem The split case The symmetric case Iterations Applications Back to Fibonacci Markov chains
  • 34.
  • 35. Setting up the Fibonacci sequence Recall the Fibonacci sequence defined by f (k + 2) = f (k) + f (k + 1), f (0) = 1, f (1) = 1
  • 36. Setting up the Fibonacci sequence Recall the Fibonacci sequence defined by f (k + 2) = f (k) + f (k + 1), f (0) = 1, f (1) = 1 Let’s let g (k) = f (k + 1). Then g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k).
  • 37.
  • 38. Setting up the Fibonacci sequence Recall the Fibonacci sequence defined by f (k + 2) = f (k) + f (k + 1), f (0) = 1, f (1) = 1 Let’s let g (k) = f (k + 1). Then g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k). f (k) So if y(k) = , we have g (k) f (k + 1) g (k) 0 1 y(k + 1) = = = y(k) g (k + 1) f (k) + g (k) 1 1
  • 39.
  • 40. Setting up the Fibonacci sequence Recall the Fibonacci sequence defined by f (k + 2) = f (k) + f (k + 1), f (0) = 1, f (1) = 1 Let’s let g (k) = f (k + 1). Then g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k). f (k) So if y(k) = , we have g (k) f (k + 1) g (k) 0 1 y(k + 1) = = = y(k) g (k + 1) f (k) + g (k) 1 1 So if A is this matrix, then y(k) =
  • 41. Setting up the Fibonacci sequence Recall the Fibonacci sequence defined by f (k + 2) = f (k) + f (k + 1), f (0) = 1, f (1) = 1 Let’s let g (k) = f (k + 1). Then g (k + 1) = f (k + 2) = f (k) + f (k + 1) = f (k) + g (k). f (k) So if y(k) = , we have g (k) f (k + 1) g (k) 0 1 y(k + 1) = = = y(k) g (k + 1) f (k) + g (k) 1 1 So if A is this matrix, then y(k) = Ak y(0).
  • 42.
  • 43. Diagonalize 0 1 The eigenvalues of A = are found by solving 1 1 −λ 1 0= = (−λ)(1 − λ) − 1 1 1−λ = λ2 − λ − 1
  • 44. Diagonalize 0 1 The eigenvalues of A = are found by solving 1 1 −λ 1 0= = (−λ)(1 − λ) − 1 1 1−λ = λ2 − λ − 1 The roots are √ √ 1+ 5 1− 5 ϕ= ϕ= ¯ 2 2
  • 45. Diagonalize 0 1 The eigenvalues of A = are found by solving 1 1 −λ 1 0= = (−λ)(1 − λ) − 1 1 1−λ = λ2 − λ − 1 The roots are √ √ 1+ 5 1− 5 ϕ= ϕ= ¯ 2 2 Notice that ϕ + ϕ = 1, ¯ ϕϕ = −1 ¯ (These facts make later calculations simpler.)
  • 46. Eigenvectors We row reduce to find the eigenvectors: −ϕ 1 −ϕ 1 −ϕ ¯ −ϕ 1 A − ϕI = = 1 1−ϕ 1 ϕ ←+ ¯ − 0 0 1 So is an eigenvector for A corresponding to the eigenvalue ϕ. ϕ
  • 47. Eigenvectors We row reduce to find the eigenvectors: −ϕ 1 −ϕ 1 −ϕ ¯ −ϕ 1 A − ϕI = = 1 1−ϕ 1 ϕ ←+ ¯ − 0 0 1 So is an eigenvector for A corresponding to the eigenvalue ϕ. ϕ 1 Similarly, is an eigenvector for A corresponding to the ϕ¯ eigenvalue ϕ. ¯
  • 48. Eigenvectors We row reduce to find the eigenvectors: −ϕ 1 −ϕ 1 −ϕ ¯ −ϕ 1 A − ϕI = = 1 1−ϕ 1 ϕ ←+ ¯ − 0 0 1 So is an eigenvector for A corresponding to the eigenvalue ϕ. ϕ 1 Similarly, is an eigenvector for A corresponding to the ϕ¯ eigenvalue ϕ. So now we know that ¯ 1 1 y(k) = c1 ϕk + c2 ϕk ¯ ϕ ϕ ¯
  • 49. What are the constants? To find c1 and c2 , we solve 1 1 1 1 1 c1 = c1 + c2 = 1 ϕ ¯ ϕ ϕ ϕ ¯ c2 −1 c1 1 1 1 1 ϕ −1 ¯ =⇒ = = c2 ϕ ϕ ¯ 1 ϕ−ϕ ¯ −ϕ 1 1 ϕ−1 ¯ 1 = ϕ−ϕ ϕ+1 ¯ 1 1 ϕ =√ 5 −ϕ ¯
  • 50.
  • 51. Finally Putting this all together we have ϕ 1 ϕ ¯ 1 y(k) = √ ϕk − √ ϕk¯ 5 ϕ 5 ϕ ¯ f (k) 1 ϕk+1 − ϕk+1 ¯ =√ k+2 − ϕk+2 g (k) 5 ϕ ¯
  • 52. Finally Putting this all together we have ϕ 1 ϕ ¯ 1 y(k) = √ ϕk − √ ϕk¯ 5 ϕ 5 ϕ ¯ f (k) 1 ϕk+1 − ϕk+1 ¯ =√ k+2 − ϕk+2 g (k) 5 ϕ ¯ So √ √   k+1 k+1 1 1+ 5 1− 5 f (k) = √  −  5 2 2
  • 53. Markov Chains Recall the setup: T is a transition matrix giving the probabilities of switching from any state to any of the other states.
  • 54. Markov Chains Recall the setup: T is a transition matrix giving the probabilities of switching from any state to any of the other states. We seek a steady-state vector, i.e., a probability vector u such that Tu = u.
  • 55. Markov Chains Recall the setup: T is a transition matrix giving the probabilities of switching from any state to any of the other states. We seek a steady-state vector, i.e., a probability vector u such that Tu = u. This is nothing more than an eigenvector of eigenvalue 1!
  • 56. Theorem If T is a regular doubly-stochastic matrix, then 1 is an eigenvalue for T all other eigenvalues of T have absolute value less than 1.
  • 57. Let u be an eigenvector of eigenvalue 1, scaled so it’s a probability vector. Let v2 , . . . , vn be eigenvectors corresponding to the other eigenvalues λ2 , . . . , λn . Then for any initial state x(0), we have x(k) = Ak x(0) = Ak (c1 u + c2 λ2 v2 + · · · + cn λn vn ) = c1 u + c2 λk v2 + · · · + cn λk vn 2 n So x(k) → c1 u Since each x(k) is a probability vector, c1 = 1. Hence x(k) → c1 u