SlideShare a Scribd company logo
1 of 26
Download to read offline
The R´nyi entropy and
                   e
the uncertainty relations in quantum mechanics



               Iwo Bialynicki-Birula
           Center for Theoretical Physics
                  Warsaw, Poland


               www.cft.edu.pl/˜ birula



                   November 2008
Heisenberg uncertainty relation

   The standard formulation of the uncertainty relations for the
measurements of the position and momentum has the following form
                                        h
                                        ¯
                             ∆x∆p ≥
                                        2

              ∆x and ∆p are the mean standard deviations

                         ∆a =       (a − a )2

          2                                 2
     (∆x) =       dxρ(x)(x − x )2     (∆p) =    dp˜(p)(p − p )2
                                                  ρ

          This form of the uncertainty is simple to prove

           The proof requires only the Schwartz inequality
  and is usually included in the first course on quantum mechanics
Critique of the standard uncertainty relations

        The mean standard deviation ∆a (dispersion) is used a lot
         in the statistical analysis of experiments that measure a
  It is a reasonable measure of the spread when (as in measurements)
          the distribution in question is of a simple ”hump” type
For example, it is a very good characteristic for a Gaussian distribution
           since it determines the half-width of this distribution
However, when the distribution of the values has more than one hump
             the standard deviation loses some of its usefulness
          especially in connection with the notion of uncertainty

         Examples of two states of a particle in one dimension
First state describes a particle with a uniformly distributed probability
                        in a box of total length L
  Second state describes a particle localized with equal probabilities
                in two smaller boxes each of length L/4
The probability distributions corresponding to these two states
  F=flat and C=clustered, are graphically represented as follows:


                      I             II          III           IV
State F


                      II                                      IV

                      I                                       III
State C
                                1
                                     inside the box,
            State F        ρ=   L
                                0    outside the box,

                                2
            State C        ρ=   L    inside the box,
                                0    outside the box.
  In which case, F or C, is the uncertainty in the position greater?
According to our intuition, the uncertainty is greater in the case F
  In the case C we know more about the position; we know that the
particle is not in the regions II and III. However, when we calculate the
          standard deviation ∆x we obtain the opposite result:

                                Case F
                                  L
                           ∆xF = √
                                   12




                                Case C

                                    7 L
                         ∆xC =       √
                                    4 12
The second, somewhat more dramatic example of the situation where
the standard deviation does not give a sensible measure of uncertainty
        is provided by the following distribution of probability.
                                State D




      L(1-1/N)        ————— Distance NL —————–                 L/N




     The probability density is constant in two regions I and II
      separated by a large distance N L (N is a large number)
    The region I has the size L(1 − 1/N ) and the distant region II
                          has the size L/N
1
                                  L    in region I,
               State D    ρ=      1
                                       in region II,
                                  L
                                  0    elsewhere.
 For large N the standard deviation ∆x is approximately equal to:

                                 Case D
                             √            L
                     ∆xD ∼       1 + 12N √
                                           12
   This expression tends to infinity with N in spite of the fact that
   the probability of finding the particle in the region I tends to 1

This example shows clearly what is wrong with the standard deviation
         It gets very high contributions from distant regions
        because these contribution enter with a large weight:
              namely, the distance from the mean value.
Information entropy as a measure of uncertainty

The best measure of information was introduced in 1948 by Shannon
    in his paper “The mathematical theory of communication”
      Shannon entropy closely resembles the physical entropy


                        H=−          pi ln pi
                                 i

     where pi is the probability of the occurrence of the i-th event
             or the a priori probability of the i-th message
    In information theory one uses the logarithms to the base of 2
                       H is then measured in bits
     but a change of the base results only in a change of the scale

The information entropy may serve as a precise measure of uncertainty
   It has even been described already by Shannon as a measure of
                “information, choice or uncertainty”
H may indeed serve at the same time as a measure of information and
as a measure of uncertainty, because the two notions are directly related

              The information gained in the measurement
              removes the uncertainty about its outcome

           The information uncertainty is in many ways
  a much better measure of uncertainty than the standard deviation

      The Shannon entropy has a simple intuitive interpretation
                   The average number of questions
needed to discover ”the truth” (that is to remove uncertainty) which is
    hidden in one of the boxes with the probability distribution pi
                     is bounded from below by H
To apply the Shannon measure of uncertainty to quantum mechanics
 we must discretize the distributions of the position and momentum
Histograms

The probability density to find a quantum particle at x is given by
    the modulus squared of the wave function ρ(x) = |ψ(x)|2
  The probability density to find the momentum p is given by
                                                         ˜
  the modulus squared of the Fourier transform ρ(p) = |ψ(p)|2
                                                ˜

 In order to have a discrete set of probabilities we introduce the
    finite resolutions (bins) for the measurements of x and p
 To distinguish them from the statistical dispersions ∆x and ∆p
                they will be denoted by δx and δp

           The probabilities ql and k to find the values
           of x and p in the l-th and the k-th bin are
                     xl +δx                 pk +δp
             ql =         dx ρ(x) pk =           dp ρ(p)
                                                    ˜
                    xl                     pk

 The values of ql and pk represent a histogram of measurements
Following D. Deutsch (1983) and H. M. Partovi (1983) we can
formulate the uncertainty relations between the positions and momenta
 that restrict the available information about x and p measured by the
    Shannon entropies H (x) = − ql ln ql and H (p) = − pk ln pk
           The best known (but not sharp) bound expressing
         the uncertainty relations (I. Bialynicki-Birula 1984) is

                    (x)        (p)                           δxδp
                H         +H         ≥ 1 − ln(2) − ln
                                                              h


                    RHS
                     3



                     2



                     1


                                                             ∆x∆p
                               0.5      1.    1.5       2.
                                                              h
The R´nyi entropy
                             e

Is there an uncertainty relation expressed in term of the R´nyi entropy?
                                                           e


 R´nyi entropy is a one parameter extension of the Shannon entropy
  e
                              1
                        Hα =     ln(       pα )
                                            k
                             1−α
      The limit α → 1 produces the Shannon entropy (      pk = 1)

                       d
                         ln(    pα ) =
                                 k        pk ln pk
                      dα
R´nyi entropy has many applications in information theory (coding and
 e
cryptography), physics (particle production, diffusion), signal analysis,
      nonlinear dynamics, fractals, etc. (150 000 hits in Google)
R´nyi entropy shares many properties with Shannon entropy
   e

• Hα is nonnegative
• Hα vanishes when the uncertainty is the smallest (only one pk = 0)
• Hα takes on the maximal value ln n when all n probabilities are equal
• Hα is extensive — satisfies the following condition If pij is a product of
  two probability distributions pij = pi pj then

              1                       1             α           α
      Hα =       ln(        pα ) =
                             ij          ln(       pi       p   j)   = Hα + Hα
             1−α       ij
                                     1−α       i        j

   R´nyi entropies for the position and momentum are easily defined using
    e
  the same probabilities pk and ql as for the Shannon entropy

          (x)            α           (p)
         Hα = ln        ql /(1 − α) Hα = ln                     pα /(1 − α)
                                                                 k

                                                            (x)          (p)
         What are the limitations on the values of Hα             and Hα ?
Brief mathematical interlude
  The (q, q )-norm of T is the smallest k(q, q ) such that


       Tψ    q   ≤ k(q, q ) ψ        q        1/q + 1/q = 1             q≥q
                                                            1
                                                            q
                      ψ     q   =             dx|ψ(x)|q

     K. I. Babenko (1961) and W. Beckner (1975)
      found k(q, q ) for the Fourier transformation
                                            ˜
   Choosing the physical normalization of ψ we obtain

                  ˜        1
                  ψ(p) = √                     dx e−ipx/¯ ψ(x)
                                                        h
                           2π¯
                             h
                                                        1
                                          1          − 2q
         ˜            1 1
                      q−2
                                 q       2q     q
         ψ   q   ≤¯
                  h                                             ψ   q    (I)
                                2π              2π
This inequality is saturated for all Gaussian wave functions
It is convenient to change q → 2α and q → 2β
          1  1                  α     β
            + =2                   =                 α > 1 =⇒ β < 1
          α β                  α−1   1−β

         and rewrite the norms in the inequality (I) as follows
                                     1                                      1
                                    2α                                     2β
         ˜
         ψ   2α   =     dp|˜(p)|α
                           ρ             ψ     2β    =        dx|ρ(x)|β

                                                 ˜
                      ρ(x) = |ψ(x)|2 and ρ(p) = |ψ(p)|2
                                         ˜

              are quantum mechanical probability densities

After squaring both sides of the inequality (I) we obtain (for α > β)
                                                                       1
                                                              1
                                              1
                                              α −1
                                                         α   2α   β   2β
     ρ
     ˜   α   ≤ n(α, β) ρ   β    n(α, β) = ¯
                                          h                                (II)
                                                         π        π
To make use of this mathematical inequality in the proof
           of the physical uncertainty relations we have to
       introduce into this analysis the probabilities pk and ql
           This is done by splitting the integrals into bins
                                         xk +δp
                   dp(˜(p))α =
                      ρ                           dp (˜(p))α
                                                      ρ
                                        xk

        and using the famous property of convex functions:
           The value of the function at the average point
          cannot exceed the average value of the function

 In the present case I need a somewhat more general version of this
       statement in the form of an integral Jensen inequality

             Φ        dx h(x)w(x)   ≤        dx Φ (h(x)) w(x)
                  D                      D

which is valid for a convex function Φ, an arbitrary function h(x) and
a nonnegative weight function w(x) normalized to 1 ( D dx w(x) = 1)
The power xα is a convex function for α > 1 hence
                               xk +δp        α        xk +δp
           pk
                  α
                               xk
                                     dp ρ(p)
                                        ˜             xk
                                                            dp (˜(p))α
                                                                ρ
                      =                          ≤
           δp                       δp                      δp

                                                            α
           δp1−α           pα ≤
                            k           dp (˜(p))α = ( ρ α )
                                            ρ          ˜          (A)
  For concave functions we have: The average value of the function
   cannot exceed the value of the function at the average point
                       β                                     β
            ( ρ   β)       =    dx (ρ(x))β ≤ δx1−β          ql    (B)

     Raising (A) to the power 1/α and (B) to the power of 1/β
       and we the use of the basic inequality (II) we obtain
                                    1                                 1
                 1−α                α                 1−β             β
                                                                  β
            δp    α            pα
                                k       ≤ n(α, β)δx    β         ql

1−α         1                         1−β                      1           β
    ln(δp) + ln                pα
                                k   ≤     ln(δx) + ln n(α, β) + ln        ql
 α          α                          β                       β
Final form of the uncertainty relation
              From the last inequality with the use of
      α/(α − 1) = β/(1 − β) after some work we obtain finally

       (x) (p)             1      ln α   ln β                      δxδp
     Hβ + Hα ≥ −                       +            − ln(2) − ln
                           2     1−α 1−β                            h

                 In the limit when α → 1 also β → 1
             With the help of the l’Hˆpital rule one obtains
                                     o
                                      ln α
                                 lim       = −1
                                 α→1 1 − α

The inequality for R´nyi entropies reduces to the uncertainty relation
                    e
      for the Shannon entropies proved by me twenty years ago

                     (x)        (p)                      δxδp
                 H         +H         ≥ 1 − ln(2) − ln
                                                          h
Uncertainty relations for N -level systems
For quantum systems described by vectors in the N -dimensional Hilbert
 space the analog of the uncertainty relation for the R´nyi entropies is
                                                       e
                     N                          N
           1                      1
              ln          ρα
                          ˜k   +     ln              ρβ
                                                      l    ≥ ln N,
          1−α                    1−β
                    k=1                        l=1

     where ρk = |˜k |2 , ρl = |al |2 and the amplitudes ak and al are
           ˜     a                                      ˜
          connected by the discrete Fourier transformation
                               N
                        1
                  ak = √
                  ˜                  exp(2πik l/N ) al .
                         N     l=1

       This uncertainty relation is saturated for the states that
                are either localized in “position space”
        exactly one of the amplitudes al is different from zero
                or are localized in “momentum space”
        exactly one of the amplitudes ak is different from zero
                                        ˜
               The bound does not depend on α and β
Uncertainty relations for mixed states
         The uncertainty relations for the R´nyi entropies also
                                            e
                       hold for all mixed states

        This result is not obvious because the R´nyi entropy is
                                                 e
not a convex function of the probability distributions for all values of α

      The terms on the left hand side of the uncertainty relation
                 may decrease as a result of mixing

             In order to extend our results to mixed states
        one has to go back to the original inequalities that were
            used in the proof and show that they still hold
Uncertainty relations for continuous distributions
         There also exist purely mathematical versions of
            the uncertainty relations that do not involve
  the experimental resolutions δx and δp of the measuring devices

                ∞                          ∞
     1                            1
        ln        dp (˜(p))α +
                      ρ              ln      dx (ρ(x))β
    1−α        −∞              1−β        −∞
                                 1      α       1       β
                          ≥−          ln −           ln
                             2(1 − α) π     2(1 − β) π
On the left hand side of this inequality we have, what might be called,
       the continuous or integral versions of the R´nyi entropies
                                                   e
 In order to derive this purely mathematical inequality that includes
  no reference to the finite resolution of the physical measurements
      I have dropped ¯ in the definition of the Fourier transform
                       h
This inequality has been independently proven by Zozor and Vignat
        Analogous relations for the continuous Tsallis entropies
                for x and p were obtained by Rajagopal
Uncertainty relations for continuous Shannon entropies
 Assuming that the distributions ρ(x) and ρ(p) are normalized
                                            ˜
     in the limit, when α → 1, β → 1 from the formula for
the entropic uncertainty relation in terms of the R´nyi entropies
                                                    e
          we obtain the classic uncertainty relation for
               the continuous Shannon entropies


        ∞                        ∞
    −        dp ρ(p) ln ρ(p) −
                ˜       ˜             dx ρ(x) ln ρ(x) ≥ ln(eπ)
        −∞                       −∞



  This inequality had been conjectured by Hirschman (1957)
  and later proved by Bialynicki-Birula and Mycielski (1975)
                    and by Beckner (1975)
References


    1. I. Bialynicki-Birula, “Renyi entropy and the uncertainty rela-
tions”, Foundations of Probability and Physics, Ed. G. Adenier, C. A.
Fuchs, and A. Yu. Khrennikov, American Institute of Physics, Melville,
2007.


   2. I. Bialynicki-Birula, “Formulation of the uncertainty relations in
terms of the Renyi entropies” Phys. Rev. A 74, 052101 (2006).


   Both references can be found on my home page: www.cft.edu.pl/˜ birula
Summary
   I have shown that the R´nyi entropies for canonically conjugate
                            e
           physical variables obey the inequalities that have
            the interpretation of the uncertainty relations

      In order to use the R´nyi entropies in their standard form
                             e
     (i.e. using a discrete set of probabilities), I had to introduce
        the finite resolutions δx and δp of the measuring devices

The resulting new uncertainty relations for the position and momentum
  acquire an additional meaning not found in their standard version

           The more accurately we wish to measure x and p
   the stronger the limitations imposed by the uncertainty relation

                      1     ln α   ln β             δxδp
    The lower limit −            +           − ln            increases
                      2     1−α 1−β                  π¯
                                                      h

More Related Content

What's hot

Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image ijcsa
 
NIACFDS2014-12-16_Nishikawa_BoundaryFormula
NIACFDS2014-12-16_Nishikawa_BoundaryFormulaNIACFDS2014-12-16_Nishikawa_BoundaryFormula
NIACFDS2014-12-16_Nishikawa_BoundaryFormulaHiroaki Nishikawa
 
Decreasing of quantity of radiation de fects in
Decreasing of quantity of radiation de fects inDecreasing of quantity of radiation de fects in
Decreasing of quantity of radiation de fects inijcsa
 
Geometric properties for parabolic and elliptic pde
Geometric properties for parabolic and elliptic pdeGeometric properties for parabolic and elliptic pde
Geometric properties for parabolic and elliptic pdeSpringer
 
Robust wavelet denoising
Robust wavelet denoisingRobust wavelet denoising
Robust wavelet denoisinggulanglhs
 
Stability criterion of periodic oscillations in a (9)
Stability criterion of periodic oscillations in a (9)Stability criterion of periodic oscillations in a (9)
Stability criterion of periodic oscillations in a (9)Alexander Decker
 
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?Sebastian De Haro
 
Math Conference Poster
Math Conference PosterMath Conference Poster
Math Conference Postermary41679
 
Nonparametric Density Estimation
Nonparametric Density EstimationNonparametric Density Estimation
Nonparametric Density Estimationjachno
 
Biased normalized cuts
Biased normalized cutsBiased normalized cuts
Biased normalized cutsirisshicat
 
Primer for ordinary differential equations
Primer for ordinary differential equationsPrimer for ordinary differential equations
Primer for ordinary differential equationsTarun Gehlot
 
Logistics. Terminology
Logistics. TerminologyLogistics. Terminology
Logistics. TerminologySSA KPI
 
A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...
A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...
A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...IJMER
 
The Graph Minor Theorem: a walk on the wild side of graphs
The Graph Minor Theorem: a walk on the wild side of graphsThe Graph Minor Theorem: a walk on the wild side of graphs
The Graph Minor Theorem: a walk on the wild side of graphsMarco Benini
 
Polynomial functions modelllings
Polynomial functions modelllingsPolynomial functions modelllings
Polynomial functions modelllingsTarun Gehlot
 

What's hot (20)

Puy chosuantai2
Puy chosuantai2Puy chosuantai2
Puy chosuantai2
 
Astaño 4
Astaño 4Astaño 4
Astaño 4
 
Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image
 
NIACFDS2014-12-16_Nishikawa_BoundaryFormula
NIACFDS2014-12-16_Nishikawa_BoundaryFormulaNIACFDS2014-12-16_Nishikawa_BoundaryFormula
NIACFDS2014-12-16_Nishikawa_BoundaryFormula
 
Decreasing of quantity of radiation de fects in
Decreasing of quantity of radiation de fects inDecreasing of quantity of radiation de fects in
Decreasing of quantity of radiation de fects in
 
Geometric properties for parabolic and elliptic pde
Geometric properties for parabolic and elliptic pdeGeometric properties for parabolic and elliptic pde
Geometric properties for parabolic and elliptic pde
 
Robust wavelet denoising
Robust wavelet denoisingRobust wavelet denoising
Robust wavelet denoising
 
Stability criterion of periodic oscillations in a (9)
Stability criterion of periodic oscillations in a (9)Stability criterion of periodic oscillations in a (9)
Stability criterion of periodic oscillations in a (9)
 
Lecture 2
Lecture 2Lecture 2
Lecture 2
 
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
 
Math Conference Poster
Math Conference PosterMath Conference Poster
Math Conference Poster
 
Nonparametric Density Estimation
Nonparametric Density EstimationNonparametric Density Estimation
Nonparametric Density Estimation
 
Biased normalized cuts
Biased normalized cutsBiased normalized cuts
Biased normalized cuts
 
Primer for ordinary differential equations
Primer for ordinary differential equationsPrimer for ordinary differential equations
Primer for ordinary differential equations
 
New version
New versionNew version
New version
 
Logistics. Terminology
Logistics. TerminologyLogistics. Terminology
Logistics. Terminology
 
A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...
A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...
A Study of Periodic Points and Their Stability on a One-Dimensional Chaotic S...
 
Paper06
Paper06Paper06
Paper06
 
The Graph Minor Theorem: a walk on the wild side of graphs
The Graph Minor Theorem: a walk on the wild side of graphsThe Graph Minor Theorem: a walk on the wild side of graphs
The Graph Minor Theorem: a walk on the wild side of graphs
 
Polynomial functions modelllings
Polynomial functions modelllingsPolynomial functions modelllings
Polynomial functions modelllings
 

Viewers also liked

Experimental demonstration of continuous variable quantum key distribution ov...
Experimental demonstration of continuous variable quantum key distribution ov...Experimental demonstration of continuous variable quantum key distribution ov...
Experimental demonstration of continuous variable quantum key distribution ov...wtyru1989
 
Completely positive maps in quantum information
Completely positive maps in quantum informationCompletely positive maps in quantum information
Completely positive maps in quantum informationwtyru1989
 
Smooth entropies a tutorial
Smooth entropies a tutorialSmooth entropies a tutorial
Smooth entropies a tutorialwtyru1989
 
Lattices, sphere packings, spherical codes
Lattices, sphere packings, spherical codesLattices, sphere packings, spherical codes
Lattices, sphere packings, spherical codeswtyru1989
 
Postselection technique for quantum channels and applications for qkd
Postselection technique for quantum channels and applications for qkdPostselection technique for quantum channels and applications for qkd
Postselection technique for quantum channels and applications for qkdwtyru1989
 
Renyis entropy
Renyis entropyRenyis entropy
Renyis entropywtyru1989
 
Atmospheric aberrations in coherent laser systems
Atmospheric aberrations in coherent laser systemsAtmospheric aberrations in coherent laser systems
Atmospheric aberrations in coherent laser systemswtyru1989
 
Qkd and de finetti theorem
Qkd and de finetti theoremQkd and de finetti theorem
Qkd and de finetti theoremwtyru1989
 
Continuous variable quantum key distribution finite key analysis of composabl...
Continuous variable quantum key distribution finite key analysis of composabl...Continuous variable quantum key distribution finite key analysis of composabl...
Continuous variable quantum key distribution finite key analysis of composabl...wtyru1989
 

Viewers also liked (9)

Experimental demonstration of continuous variable quantum key distribution ov...
Experimental demonstration of continuous variable quantum key distribution ov...Experimental demonstration of continuous variable quantum key distribution ov...
Experimental demonstration of continuous variable quantum key distribution ov...
 
Completely positive maps in quantum information
Completely positive maps in quantum informationCompletely positive maps in quantum information
Completely positive maps in quantum information
 
Smooth entropies a tutorial
Smooth entropies a tutorialSmooth entropies a tutorial
Smooth entropies a tutorial
 
Lattices, sphere packings, spherical codes
Lattices, sphere packings, spherical codesLattices, sphere packings, spherical codes
Lattices, sphere packings, spherical codes
 
Postselection technique for quantum channels and applications for qkd
Postselection technique for quantum channels and applications for qkdPostselection technique for quantum channels and applications for qkd
Postselection technique for quantum channels and applications for qkd
 
Renyis entropy
Renyis entropyRenyis entropy
Renyis entropy
 
Atmospheric aberrations in coherent laser systems
Atmospheric aberrations in coherent laser systemsAtmospheric aberrations in coherent laser systems
Atmospheric aberrations in coherent laser systems
 
Qkd and de finetti theorem
Qkd and de finetti theoremQkd and de finetti theorem
Qkd and de finetti theorem
 
Continuous variable quantum key distribution finite key analysis of composabl...
Continuous variable quantum key distribution finite key analysis of composabl...Continuous variable quantum key distribution finite key analysis of composabl...
Continuous variable quantum key distribution finite key analysis of composabl...
 

Similar to The renyi entropy and the uncertainty relations in quantum mechanics

Further discriminatory signature of inflation
Further discriminatory signature of inflationFurther discriminatory signature of inflation
Further discriminatory signature of inflationLaila A
 
Bubbles of True Vacuum and Duality in M-theory
Bubbles of True Vacuum and Duality in M-theoryBubbles of True Vacuum and Duality in M-theory
Bubbles of True Vacuum and Duality in M-theorySebastian De Haro
 
A jensen-shannon
A    jensen-shannonA    jensen-shannon
A jensen-shannonUnirio
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
2Solution of NonlinearEquations2.1 IntroductionMan.docx
2Solution of NonlinearEquations2.1 IntroductionMan.docx2Solution of NonlinearEquations2.1 IntroductionMan.docx
2Solution of NonlinearEquations2.1 IntroductionMan.docxgilbertkpeters11344
 
Threshold network models
Threshold network modelsThreshold network models
Threshold network modelsNaoki Masuda
 
Heesterbeek 90 definition_computation
Heesterbeek 90 definition_computationHeesterbeek 90 definition_computation
Heesterbeek 90 definition_computationDilruk Gallage
 
Topological Inference via Meshing
Topological Inference via MeshingTopological Inference via Meshing
Topological Inference via MeshingDon Sheehy
 
A note on arithmetic progressions in sets of integers
A note on arithmetic progressions in sets of integersA note on arithmetic progressions in sets of integers
A note on arithmetic progressions in sets of integersLukas Nabergall
 
Small Sampling Theory Presentation1
Small Sampling Theory Presentation1Small Sampling Theory Presentation1
Small Sampling Theory Presentation1jravish
 
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...GLOBAL HEAVYLIFT HOLDINGS
 
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...Global HeavyLift Holdings, LLC
 
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...Global HeavyLift Holdings, LLC
 
THE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDS
THE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDSTHE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDS
THE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDSMichaelRabinovich
 

Similar to The renyi entropy and the uncertainty relations in quantum mechanics (20)

Further discriminatory signature of inflation
Further discriminatory signature of inflationFurther discriminatory signature of inflation
Further discriminatory signature of inflation
 
Bubbles of True Vacuum and Duality in M-theory
Bubbles of True Vacuum and Duality in M-theoryBubbles of True Vacuum and Duality in M-theory
Bubbles of True Vacuum and Duality in M-theory
 
Galichon jds
Galichon jdsGalichon jds
Galichon jds
 
A jensen-shannon
A    jensen-shannonA    jensen-shannon
A jensen-shannon
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)
 
2Solution of NonlinearEquations2.1 IntroductionMan.docx
2Solution of NonlinearEquations2.1 IntroductionMan.docx2Solution of NonlinearEquations2.1 IntroductionMan.docx
2Solution of NonlinearEquations2.1 IntroductionMan.docx
 
Serie de dyson
Serie de dysonSerie de dyson
Serie de dyson
 
Threshold network models
Threshold network modelsThreshold network models
Threshold network models
 
Quantum chaos of generic systems - Marko Robnik
Quantum chaos of generic systems - Marko RobnikQuantum chaos of generic systems - Marko Robnik
Quantum chaos of generic systems - Marko Robnik
 
Wave functions
Wave functionsWave functions
Wave functions
 
Heesterbeek 90 definition_computation
Heesterbeek 90 definition_computationHeesterbeek 90 definition_computation
Heesterbeek 90 definition_computation
 
Topological Inference via Meshing
Topological Inference via MeshingTopological Inference via Meshing
Topological Inference via Meshing
 
Sm08a10
Sm08a10Sm08a10
Sm08a10
 
A note on arithmetic progressions in sets of integers
A note on arithmetic progressions in sets of integersA note on arithmetic progressions in sets of integers
A note on arithmetic progressions in sets of integers
 
Berans qm overview
Berans qm overviewBerans qm overview
Berans qm overview
 
Small Sampling Theory Presentation1
Small Sampling Theory Presentation1Small Sampling Theory Presentation1
Small Sampling Theory Presentation1
 
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
 
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
 
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
New Presentation From Astrophysicist Dr. Andrew Beckwith: "Detailing Coherent...
 
THE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDS
THE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDSTHE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDS
THE MASS OF ASYMPTOTICALLY HYPERBOLIC MANIFOLDS
 

More from wtyru1989

Quantum optical measurement
Quantum optical measurementQuantum optical measurement
Quantum optical measurementwtyru1989
 
Gaussian discord imperial
Gaussian discord imperialGaussian discord imperial
Gaussian discord imperialwtyru1989
 
Entropic characteristics of quantum channels and the additivity problem
Entropic characteristics of quantum channels and the additivity problemEntropic characteristics of quantum channels and the additivity problem
Entropic characteristics of quantum channels and the additivity problemwtyru1989
 
Manipulating continuous variable photonic entanglement
Manipulating continuous variable photonic entanglementManipulating continuous variable photonic entanglement
Manipulating continuous variable photonic entanglementwtyru1989
 
The gaussian minimum entropy conjecture
The gaussian minimum entropy conjectureThe gaussian minimum entropy conjecture
The gaussian minimum entropy conjecturewtyru1989
 
The security of quantum cryptography
The security of quantum cryptographyThe security of quantum cryptography
The security of quantum cryptographywtyru1989
 
Entanglement of formation
Entanglement of formationEntanglement of formation
Entanglement of formationwtyru1989
 
Bound entanglement is not rare
Bound entanglement is not rareBound entanglement is not rare
Bound entanglement is not rarewtyru1989
 
Continuous variable quantum entanglement and its applications
Continuous variable quantum entanglement and its applicationsContinuous variable quantum entanglement and its applications
Continuous variable quantum entanglement and its applicationswtyru1989
 
Relative entropy and_squahed_entanglement
Relative entropy and_squahed_entanglementRelative entropy and_squahed_entanglement
Relative entropy and_squahed_entanglementwtyru1989
 
Lect12 photodiode detectors
Lect12 photodiode detectorsLect12 photodiode detectors
Lect12 photodiode detectorswtyru1989
 
Towards a one shot entanglement theory
Towards a one shot entanglement theoryTowards a one shot entanglement theory
Towards a one shot entanglement theorywtyru1989
 
Encrypting with entanglement matthias christandl
Encrypting with entanglement matthias christandlEncrypting with entanglement matthias christandl
Encrypting with entanglement matthias christandlwtyru1989
 
Dic rd theory_quantization_07
Dic rd theory_quantization_07Dic rd theory_quantization_07
Dic rd theory_quantization_07wtyru1989
 
标量量化
标量量化标量量化
标量量化wtyru1989
 
Fully understanding cmrr taiwan-2012
Fully understanding cmrr taiwan-2012Fully understanding cmrr taiwan-2012
Fully understanding cmrr taiwan-2012wtyru1989
 
Op amp tutorial-1
Op amp tutorial-1Op amp tutorial-1
Op amp tutorial-1wtyru1989
 
Tsinghua visit
Tsinghua visitTsinghua visit
Tsinghua visitwtyru1989
 
Quantum conditional states, bayes' rule, and state compatibility
Quantum conditional states, bayes' rule, and state compatibilityQuantum conditional states, bayes' rule, and state compatibility
Quantum conditional states, bayes' rule, and state compatibilitywtyru1989
 

More from wtyru1989 (20)

Quantum optical measurement
Quantum optical measurementQuantum optical measurement
Quantum optical measurement
 
Gaussian discord imperial
Gaussian discord imperialGaussian discord imperial
Gaussian discord imperial
 
Entropic characteristics of quantum channels and the additivity problem
Entropic characteristics of quantum channels and the additivity problemEntropic characteristics of quantum channels and the additivity problem
Entropic characteristics of quantum channels and the additivity problem
 
Manipulating continuous variable photonic entanglement
Manipulating continuous variable photonic entanglementManipulating continuous variable photonic entanglement
Manipulating continuous variable photonic entanglement
 
The gaussian minimum entropy conjecture
The gaussian minimum entropy conjectureThe gaussian minimum entropy conjecture
The gaussian minimum entropy conjecture
 
The security of quantum cryptography
The security of quantum cryptographyThe security of quantum cryptography
The security of quantum cryptography
 
Entanglement of formation
Entanglement of formationEntanglement of formation
Entanglement of formation
 
Bound entanglement is not rare
Bound entanglement is not rareBound entanglement is not rare
Bound entanglement is not rare
 
Continuous variable quantum entanglement and its applications
Continuous variable quantum entanglement and its applicationsContinuous variable quantum entanglement and its applications
Continuous variable quantum entanglement and its applications
 
Relative entropy and_squahed_entanglement
Relative entropy and_squahed_entanglementRelative entropy and_squahed_entanglement
Relative entropy and_squahed_entanglement
 
Lect12 photodiode detectors
Lect12 photodiode detectorsLect12 photodiode detectors
Lect12 photodiode detectors
 
Towards a one shot entanglement theory
Towards a one shot entanglement theoryTowards a one shot entanglement theory
Towards a one shot entanglement theory
 
Encrypting with entanglement matthias christandl
Encrypting with entanglement matthias christandlEncrypting with entanglement matthias christandl
Encrypting with entanglement matthias christandl
 
Dic rd theory_quantization_07
Dic rd theory_quantization_07Dic rd theory_quantization_07
Dic rd theory_quantization_07
 
Em method
Em methodEm method
Em method
 
标量量化
标量量化标量量化
标量量化
 
Fully understanding cmrr taiwan-2012
Fully understanding cmrr taiwan-2012Fully understanding cmrr taiwan-2012
Fully understanding cmrr taiwan-2012
 
Op amp tutorial-1
Op amp tutorial-1Op amp tutorial-1
Op amp tutorial-1
 
Tsinghua visit
Tsinghua visitTsinghua visit
Tsinghua visit
 
Quantum conditional states, bayes' rule, and state compatibility
Quantum conditional states, bayes' rule, and state compatibilityQuantum conditional states, bayes' rule, and state compatibility
Quantum conditional states, bayes' rule, and state compatibility
 

Recently uploaded

Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfPrecisely
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionDilum Bandara
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningLars Bell
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 

Recently uploaded (20)

Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An Introduction
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine Tuning
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 

The renyi entropy and the uncertainty relations in quantum mechanics

  • 1. The R´nyi entropy and e the uncertainty relations in quantum mechanics Iwo Bialynicki-Birula Center for Theoretical Physics Warsaw, Poland www.cft.edu.pl/˜ birula November 2008
  • 2. Heisenberg uncertainty relation The standard formulation of the uncertainty relations for the measurements of the position and momentum has the following form h ¯ ∆x∆p ≥ 2 ∆x and ∆p are the mean standard deviations ∆a = (a − a )2 2 2 (∆x) = dxρ(x)(x − x )2 (∆p) = dp˜(p)(p − p )2 ρ This form of the uncertainty is simple to prove The proof requires only the Schwartz inequality and is usually included in the first course on quantum mechanics
  • 3. Critique of the standard uncertainty relations The mean standard deviation ∆a (dispersion) is used a lot in the statistical analysis of experiments that measure a It is a reasonable measure of the spread when (as in measurements) the distribution in question is of a simple ”hump” type For example, it is a very good characteristic for a Gaussian distribution since it determines the half-width of this distribution However, when the distribution of the values has more than one hump the standard deviation loses some of its usefulness especially in connection with the notion of uncertainty Examples of two states of a particle in one dimension First state describes a particle with a uniformly distributed probability in a box of total length L Second state describes a particle localized with equal probabilities in two smaller boxes each of length L/4
  • 4. The probability distributions corresponding to these two states F=flat and C=clustered, are graphically represented as follows: I II III IV State F II IV I III State C 1 inside the box, State F ρ= L 0 outside the box, 2 State C ρ= L inside the box, 0 outside the box. In which case, F or C, is the uncertainty in the position greater?
  • 5. According to our intuition, the uncertainty is greater in the case F In the case C we know more about the position; we know that the particle is not in the regions II and III. However, when we calculate the standard deviation ∆x we obtain the opposite result: Case F L ∆xF = √ 12 Case C 7 L ∆xC = √ 4 12
  • 6. The second, somewhat more dramatic example of the situation where the standard deviation does not give a sensible measure of uncertainty is provided by the following distribution of probability. State D L(1-1/N) ————— Distance NL —————– L/N The probability density is constant in two regions I and II separated by a large distance N L (N is a large number) The region I has the size L(1 − 1/N ) and the distant region II has the size L/N
  • 7. 1 L in region I, State D ρ= 1 in region II, L 0 elsewhere. For large N the standard deviation ∆x is approximately equal to: Case D √ L ∆xD ∼ 1 + 12N √ 12 This expression tends to infinity with N in spite of the fact that the probability of finding the particle in the region I tends to 1 This example shows clearly what is wrong with the standard deviation It gets very high contributions from distant regions because these contribution enter with a large weight: namely, the distance from the mean value.
  • 8. Information entropy as a measure of uncertainty The best measure of information was introduced in 1948 by Shannon in his paper “The mathematical theory of communication” Shannon entropy closely resembles the physical entropy H=− pi ln pi i where pi is the probability of the occurrence of the i-th event or the a priori probability of the i-th message In information theory one uses the logarithms to the base of 2 H is then measured in bits but a change of the base results only in a change of the scale The information entropy may serve as a precise measure of uncertainty It has even been described already by Shannon as a measure of “information, choice or uncertainty”
  • 9. H may indeed serve at the same time as a measure of information and as a measure of uncertainty, because the two notions are directly related The information gained in the measurement removes the uncertainty about its outcome The information uncertainty is in many ways a much better measure of uncertainty than the standard deviation The Shannon entropy has a simple intuitive interpretation The average number of questions needed to discover ”the truth” (that is to remove uncertainty) which is hidden in one of the boxes with the probability distribution pi is bounded from below by H
  • 10.
  • 11. To apply the Shannon measure of uncertainty to quantum mechanics we must discretize the distributions of the position and momentum
  • 12. Histograms The probability density to find a quantum particle at x is given by the modulus squared of the wave function ρ(x) = |ψ(x)|2 The probability density to find the momentum p is given by ˜ the modulus squared of the Fourier transform ρ(p) = |ψ(p)|2 ˜ In order to have a discrete set of probabilities we introduce the finite resolutions (bins) for the measurements of x and p To distinguish them from the statistical dispersions ∆x and ∆p they will be denoted by δx and δp The probabilities ql and k to find the values of x and p in the l-th and the k-th bin are xl +δx pk +δp ql = dx ρ(x) pk = dp ρ(p) ˜ xl pk The values of ql and pk represent a histogram of measurements
  • 13. Following D. Deutsch (1983) and H. M. Partovi (1983) we can formulate the uncertainty relations between the positions and momenta that restrict the available information about x and p measured by the Shannon entropies H (x) = − ql ln ql and H (p) = − pk ln pk The best known (but not sharp) bound expressing the uncertainty relations (I. Bialynicki-Birula 1984) is (x) (p) δxδp H +H ≥ 1 − ln(2) − ln h RHS 3 2 1 ∆x∆p 0.5 1. 1.5 2. h
  • 14. The R´nyi entropy e Is there an uncertainty relation expressed in term of the R´nyi entropy? e R´nyi entropy is a one parameter extension of the Shannon entropy e 1 Hα = ln( pα ) k 1−α The limit α → 1 produces the Shannon entropy ( pk = 1) d ln( pα ) = k pk ln pk dα R´nyi entropy has many applications in information theory (coding and e cryptography), physics (particle production, diffusion), signal analysis, nonlinear dynamics, fractals, etc. (150 000 hits in Google)
  • 15. R´nyi entropy shares many properties with Shannon entropy e • Hα is nonnegative • Hα vanishes when the uncertainty is the smallest (only one pk = 0) • Hα takes on the maximal value ln n when all n probabilities are equal • Hα is extensive — satisfies the following condition If pij is a product of two probability distributions pij = pi pj then 1 1 α α Hα = ln( pα ) = ij ln( pi p j) = Hα + Hα 1−α ij 1−α i j R´nyi entropies for the position and momentum are easily defined using e the same probabilities pk and ql as for the Shannon entropy (x) α (p) Hα = ln ql /(1 − α) Hα = ln pα /(1 − α) k (x) (p) What are the limitations on the values of Hα and Hα ?
  • 16. Brief mathematical interlude The (q, q )-norm of T is the smallest k(q, q ) such that Tψ q ≤ k(q, q ) ψ q 1/q + 1/q = 1 q≥q 1 q ψ q = dx|ψ(x)|q K. I. Babenko (1961) and W. Beckner (1975) found k(q, q ) for the Fourier transformation ˜ Choosing the physical normalization of ψ we obtain ˜ 1 ψ(p) = √ dx e−ipx/¯ ψ(x) h 2π¯ h 1 1 − 2q ˜ 1 1 q−2 q 2q q ψ q ≤¯ h ψ q (I) 2π 2π This inequality is saturated for all Gaussian wave functions
  • 17. It is convenient to change q → 2α and q → 2β 1 1 α β + =2 = α > 1 =⇒ β < 1 α β α−1 1−β and rewrite the norms in the inequality (I) as follows 1 1 2α 2β ˜ ψ 2α = dp|˜(p)|α ρ ψ 2β = dx|ρ(x)|β ˜ ρ(x) = |ψ(x)|2 and ρ(p) = |ψ(p)|2 ˜ are quantum mechanical probability densities After squaring both sides of the inequality (I) we obtain (for α > β) 1 1 1 α −1 α 2α β 2β ρ ˜ α ≤ n(α, β) ρ β n(α, β) = ¯ h (II) π π
  • 18. To make use of this mathematical inequality in the proof of the physical uncertainty relations we have to introduce into this analysis the probabilities pk and ql This is done by splitting the integrals into bins xk +δp dp(˜(p))α = ρ dp (˜(p))α ρ xk and using the famous property of convex functions: The value of the function at the average point cannot exceed the average value of the function In the present case I need a somewhat more general version of this statement in the form of an integral Jensen inequality Φ dx h(x)w(x) ≤ dx Φ (h(x)) w(x) D D which is valid for a convex function Φ, an arbitrary function h(x) and a nonnegative weight function w(x) normalized to 1 ( D dx w(x) = 1)
  • 19. The power xα is a convex function for α > 1 hence xk +δp α xk +δp pk α xk dp ρ(p) ˜ xk dp (˜(p))α ρ = ≤ δp δp δp α δp1−α pα ≤ k dp (˜(p))α = ( ρ α ) ρ ˜ (A) For concave functions we have: The average value of the function cannot exceed the value of the function at the average point β β ( ρ β) = dx (ρ(x))β ≤ δx1−β ql (B) Raising (A) to the power 1/α and (B) to the power of 1/β and we the use of the basic inequality (II) we obtain 1 1 1−α α 1−β β β δp α pα k ≤ n(α, β)δx β ql 1−α 1 1−β 1 β ln(δp) + ln pα k ≤ ln(δx) + ln n(α, β) + ln ql α α β β
  • 20. Final form of the uncertainty relation From the last inequality with the use of α/(α − 1) = β/(1 − β) after some work we obtain finally (x) (p) 1 ln α ln β δxδp Hβ + Hα ≥ − + − ln(2) − ln 2 1−α 1−β h In the limit when α → 1 also β → 1 With the help of the l’Hˆpital rule one obtains o ln α lim = −1 α→1 1 − α The inequality for R´nyi entropies reduces to the uncertainty relation e for the Shannon entropies proved by me twenty years ago (x) (p) δxδp H +H ≥ 1 − ln(2) − ln h
  • 21. Uncertainty relations for N -level systems For quantum systems described by vectors in the N -dimensional Hilbert space the analog of the uncertainty relation for the R´nyi entropies is e N N 1 1 ln ρα ˜k + ln ρβ l ≥ ln N, 1−α 1−β k=1 l=1 where ρk = |˜k |2 , ρl = |al |2 and the amplitudes ak and al are ˜ a ˜ connected by the discrete Fourier transformation N 1 ak = √ ˜ exp(2πik l/N ) al . N l=1 This uncertainty relation is saturated for the states that are either localized in “position space” exactly one of the amplitudes al is different from zero or are localized in “momentum space” exactly one of the amplitudes ak is different from zero ˜ The bound does not depend on α and β
  • 22. Uncertainty relations for mixed states The uncertainty relations for the R´nyi entropies also e hold for all mixed states This result is not obvious because the R´nyi entropy is e not a convex function of the probability distributions for all values of α The terms on the left hand side of the uncertainty relation may decrease as a result of mixing In order to extend our results to mixed states one has to go back to the original inequalities that were used in the proof and show that they still hold
  • 23. Uncertainty relations for continuous distributions There also exist purely mathematical versions of the uncertainty relations that do not involve the experimental resolutions δx and δp of the measuring devices ∞ ∞ 1 1 ln dp (˜(p))α + ρ ln dx (ρ(x))β 1−α −∞ 1−β −∞ 1 α 1 β ≥− ln − ln 2(1 − α) π 2(1 − β) π On the left hand side of this inequality we have, what might be called, the continuous or integral versions of the R´nyi entropies e In order to derive this purely mathematical inequality that includes no reference to the finite resolution of the physical measurements I have dropped ¯ in the definition of the Fourier transform h This inequality has been independently proven by Zozor and Vignat Analogous relations for the continuous Tsallis entropies for x and p were obtained by Rajagopal
  • 24. Uncertainty relations for continuous Shannon entropies Assuming that the distributions ρ(x) and ρ(p) are normalized ˜ in the limit, when α → 1, β → 1 from the formula for the entropic uncertainty relation in terms of the R´nyi entropies e we obtain the classic uncertainty relation for the continuous Shannon entropies ∞ ∞ − dp ρ(p) ln ρ(p) − ˜ ˜ dx ρ(x) ln ρ(x) ≥ ln(eπ) −∞ −∞ This inequality had been conjectured by Hirschman (1957) and later proved by Bialynicki-Birula and Mycielski (1975) and by Beckner (1975)
  • 25. References 1. I. Bialynicki-Birula, “Renyi entropy and the uncertainty rela- tions”, Foundations of Probability and Physics, Ed. G. Adenier, C. A. Fuchs, and A. Yu. Khrennikov, American Institute of Physics, Melville, 2007. 2. I. Bialynicki-Birula, “Formulation of the uncertainty relations in terms of the Renyi entropies” Phys. Rev. A 74, 052101 (2006). Both references can be found on my home page: www.cft.edu.pl/˜ birula
  • 26. Summary I have shown that the R´nyi entropies for canonically conjugate e physical variables obey the inequalities that have the interpretation of the uncertainty relations In order to use the R´nyi entropies in their standard form e (i.e. using a discrete set of probabilities), I had to introduce the finite resolutions δx and δp of the measuring devices The resulting new uncertainty relations for the position and momentum acquire an additional meaning not found in their standard version The more accurately we wish to measure x and p the stronger the limitations imposed by the uncertainty relation 1 ln α ln β δxδp The lower limit − + − ln increases 2 1−α 1−β π¯ h