Dirac delta function: Difference between revisions

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search
imported>Dedhert.Jr
History: fix sfn
imported>Tito Omburo
 
Line 6: Line 6:
{{Differential equations}}
{{Differential equations}}


In [[mathematical analysis]], the '''Dirac delta function''' (or '''{{mvar|δ}} distribution'''), also known as the '''unit impulse''',{{sfn|atis|2013|loc=unit impulse}} is a [[generalized function]] on the [[real numbers]], whose value is zero everywhere except at zero, and whose [[integral]] over the entire real line is equal to one.{{sfn|Arfken|Weber|2000|p=84}}{{sfn|Dirac|1930|loc=§22 The ''δ'' function}}{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.1}} Thus it can be [[Heuristic|represented heuristically]] as
In [[mathematical analysis]], the '''Dirac delta function''' (or '''{{mvar|δ}} distribution'''), also known as the '''unit impulse''',{{sfn|Jeffrey|1993|p=639}} is a [[generalized function]] on the [[real numbers]], whose value is zero everywhere except at zero, and whose [[integral]] over the entire real line is equal to one.{{sfn|Arfken|Weber|2000|p=84}}{{sfn|Dirac|1930|loc=§22 The ''δ'' function}}{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.1}} Thus it can be [[Heuristic|represented heuristically]] as


<math display="block">\delta (x) = \begin{cases} 0, & x \neq 0 \\ {\infty} , & x = 0 \end{cases}</math>
<math display="block">\delta (x) = \begin{cases} 0, & x \neq 0 \\ {\infty} , & x = 0 \end{cases}</math>
Line 16: Line 16:
Since there is no function having this property, modelling the delta "function" rigorously involves the use of [[limit (mathematics)|limits]] or, as is common in mathematics, [[measure theory]] and the theory of [[distribution (mathematics)|distributions]].
Since there is no function having this property, modelling the delta "function" rigorously involves the use of [[limit (mathematics)|limits]] or, as is common in mathematics, [[measure theory]] and the theory of [[distribution (mathematics)|distributions]].


The delta function was introduced by physicist [[Paul Dirac]], and has since been applied routinely in physics and engineering to model point masses and instantaneous impulses. It is called the delta function because it is a continuous analogue of the [[Kronecker delta]] function, which is usually defined on a discrete domain and takes values 0 and 1. The mathematical rigor of the delta function was disputed until [[Laurent Schwartz]] developed the theory of distributions, where it is defined as a linear form acting on functions.
The delta function was introduced by physicist [[Paul Dirac]], and has since been applied routinely in physics and engineering to model point masses and instantaneous impulses. It is called the delta function because it is a continuous analogue of the [[Kronecker delta]] function, which is usually defined on a discrete domain and takes values 0 and 1. The mathematical rigor of the delta function was disputed until [[Laurent Schwartz]] developed the theory of distributions, where it is defined as a linear form acting on functions.


== Motivation and overview ==
== Motivation and overview ==
The [[graph of a function|graph]] of the Dirac delta is usually thought of as following the whole ''x''-axis and the positive ''y''-axis.{{sfn|Zhao|2011|p=[https://books.google.com/books?id=blZYGDREpk8C&pg=PA174 174]}} The Dirac delta is used to model a tall narrow spike function (an ''impulse''), and other similar [[abstraction]]s such as a [[point charge]], [[point mass]]{{sfn|Snieder|2004|p=[http://books.google.com/books?id=digNulgdDBIC&pg=PA212 212]}} or [[electron]] point.{{cn|date=June 2025}} For example, to calculate the [[dynamics (mechanics)|dynamics]] of a [[billiard ball]] being struck, one can approximate the [[force]] of the impact by a Dirac delta. In doing so, one can simplify the equations and calculate the [[motion (physics)|motion]] of the ball by only considering the total impulse of the collision, without a detailed model of all of the elastic energy transfer at subatomic levels (for instance).
The [[graph of a function|graph]] of the Dirac delta is usually thought of as following the whole ''x''-axis and the positive ''y''-axis.{{sfn|Zhao|2011|p=[https://books.google.com/books?id=blZYGDREpk8C&pg=PA174 174]}} The Dirac delta is used to model a tall narrow spike function (an ''impulse''), and other similar [[abstraction]]s such as a [[point charge]] or [[point mass]].{{sfn|Bracewell|2000|p=74}}{{sfn|Snieder|2004|p=[http://books.google.com/books?id=digNulgdDBIC&pg=PA212 212]}} For example, to calculate the [[dynamics (mechanics)|dynamics]] of a [[billiard ball]] being struck, one can approximate the [[force]] of the impact by a Dirac delta. In doing so, one can simplify the equations and calculate the [[motion (physics)|motion]] of the ball by only considering the total impulse of the collision, without a detailed model of all of the elastic energy transfer at subatomic levels (for instance).


To be specific, suppose that a billiard ball is at rest. At time <math>t=0</math> it is struck by another ball, imparting it with a [[momentum]] {{mvar|P}}, with units kg&sdot;m&sdot;s<sup>&minus;1</sup>. The exchange of momentum is not actually instantaneous, being mediated by elastic processes at the molecular and subatomic level, but for practical purposes it is convenient to consider that energy transfer as effectively instantaneous. The [[force]] therefore is {{math|''P'' ''δ''(''t'')}}; the units of {{math|''δ''(''t'')}} are s<sup>&minus;1</sup>.
To be specific, suppose that a billiard ball is at rest. At time <math>t=0</math> it is struck by another ball, imparting it with a [[momentum]] {{mvar|P}}, with units kg&sdot;m&sdot;s<sup>&minus;1</sup>. The exchange of momentum is not actually instantaneous, being mediated by elastic processes at the molecular and subatomic level, but for practical purposes it is convenient to consider that energy transfer as effectively instantaneous. The [[force]] therefore is {{math|''P'' ''δ''(''t'')}}; the units of {{math|''δ''(''t'')}} are s<sup>&minus;1</sup>.


To model this situation more rigorously, suppose that the force instead is uniformly distributed over a small time interval {{nowrap|<math>\Delta t = [0,T]</math>.}} That is,
To model this situation more rigorously, suppose that the force instead is uniformly distributed over a small time interval {{nowrap|<math>\Delta t = [0,T]</math>.}} That is,
Line 43: Line 43:
which holds for all {{nowrap|<math>\Delta t>0</math>,}} should continue to hold in the limit. So, in the equation {{nowrap|<math display="inline">F(t)=P\,\delta(t)=\lim_{\Delta t\to 0}F_{\Delta t}(t)</math>,}} it is understood that the limit is always taken {{em|outside the integral}}.
which holds for all {{nowrap|<math>\Delta t>0</math>,}} should continue to hold in the limit. So, in the equation {{nowrap|<math display="inline">F(t)=P\,\delta(t)=\lim_{\Delta t\to 0}F_{\Delta t}(t)</math>,}} it is understood that the limit is always taken {{em|outside the integral}}.


In applied mathematics, as we have done here, the delta function is often manipulated as a kind of limit (a [[weak limit]]) of a [[sequence]] of functions, each member of which has a tall spike at the origin: for example, a sequence of [[Gaussian distribution]]s centered at the origin with [[variance]] tending to zero.
In applied mathematics, as we have done here, the delta function is often manipulated as a kind of limit (a [[weak limit]]) of a [[sequence]] of functions, each member of which has a tall spike at the origin: for example, a sequence of [[Gaussian distribution]]s centered at the origin with [[variance]] tending to zero. (However, even in some applications, highly oscillatory functions are used as approximations to the delta function, see [[#Representations|below]].)


The Dirac delta is not truly a function, at least not a usual one with domain and range in [[real number]]s.{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.1}} For example, the objects {{math|1=''f''(''x'') = ''δ''(''x'')}} and {{math|1=''g''(''x'') = 0}} are equal everywhere except at {{math|1=''x'' = 0}} yet have integrals that are different. According to [[Lebesgue integral#Basic theorems of the Lebesgue integral|Lebesgue integration theory]], if {{mvar|f}} and {{mvar|g}} are functions such that {{math|1=''f'' = ''g''}} [[almost everywhere]], then {{mvar|f}} is integrable [[if and only if]] {{mvar|g}} is integrable and the integrals of {{mvar|f}} and {{mvar|g}} are identical.{{sfn|Schwartz|1950|p=19}} A rigorous approach to regarding the Dirac delta function as a [[mathematical object]] in its own right requires [[measure theory]] or the theory of [[distribution (mathematics)|distribution]]s.{{cn|date=June 2025}}
The Dirac delta, given the desired properties outlined above, cannot be a function with domain and range in [[real number]]s.{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.1}} For example, the objects {{math|1=''f''(''x'') = ''δ''(''x'')}} and {{math|1=''g''(''x'') = 0}} are equal everywhere except at {{math|1=''x'' = 0}} yet have integrals that are different. According to [[Lebesgue integral#Basic theorems of the Lebesgue integral|Lebesgue integration theory]], if {{mvar|f}} and {{mvar|g}} are functions such that {{math|1=''f'' = ''g''}} [[almost everywhere]], then {{mvar|f}} is integrable [[if and only if]] {{mvar|g}} is integrable and the integrals of {{mvar|f}} and {{mvar|g}} are identical.{{sfn|Schwartz|1950|p=19}} A rigorous approach to regarding the Dirac delta function as a [[mathematical object]] in its own right uses [[measure theory]] or the theory of [[distribution (mathematics)|distribution]]s.{{sfn|Schwartz|1950|p=5}}


==History==
==History==
In physics, the Dirac delta function was popularized by [[Paul Dirac]] in this book ''[[The Principles of Quantum Mechanics]]'' published in 1930.{{sfn|Dirac|1930|loc=§22 The ''δ'' function}} However, [[Oliver Heaviside]], 35 years before Dirac, described an impulsive function called the [[Heaviside step]] function for purposes and with properties analogous to Dirac's work. Even earlier several mathematicians and physicists used limits of sharply peaked functions in derivations.<ref name=JacksonHistory>{{Cite journal |last=Jackson |first=J. D. |date=2008-08-01 |title=Examples of the zeroth theorem of the history of science |url=https://pubs.aip.org/aapt/ajp/article-abstract/76/8/704/1057888/Examples-of-the-zeroth-theorem-of-the-history-of?redirectedFrom=fulltext |journal=American Journal of Physics |volume=76 |issue=8 |pages=704–719 |doi=10.1119/1.2904468 |issn=0002-9505|arxiv=0708.4249 |bibcode=2008AmJPh..76..704J }}</ref>
In physics, the Dirac delta function was popularized by [[Paul Dirac]] in this book ''[[The Principles of Quantum Mechanics]]'' published in 1930.{{sfn|Dirac|1930|loc=§22 The ''δ'' function}} However, [[Oliver Heaviside]], 35 years before Dirac, described an impulsive function called the [[Heaviside step]] function for purposes and with properties analogous to Dirac's work. Even earlier several mathematicians and physicists used limits of sharply peaked functions in derivations.<ref name=JacksonHistory>{{Cite journal |last=Jackson |first=J. D. |date=2008-08-01 |title=Examples of the zeroth theorem of the history of science |url=https://pubs.aip.org/aapt/ajp/article-abstract/76/8/704/1057888/Examples-of-the-zeroth-theorem-of-the-history-of?redirectedFrom=fulltext |journal=American Journal of Physics |volume=76 |issue=8 |pages=704–719 |doi=10.1119/1.2904468 |issn=0002-9505|arxiv=0708.4249 |bibcode=2008AmJPh..76..704J }}</ref>
An [[infinitesimal]] formula for an infinitely tall, unit impulse delta function (infinitesimal version of [[Cauchy distribution]]) explicitly appears in an 1827 text of [[Augustin-Louis Cauchy]].{{sfn|Laugwitz|1989|p=230}} [[Siméon Denis Poisson]] considered the issue in connection with the study of wave propagation as did [[Gustav Kirchhoff]] somewhat later. Kirchhoff and [[Hermann von Helmholtz]] also introduced the unit impulse as a limit of [[Gaussian distribution|Gaussians]], which also corresponded to [[Lord Kelvin]]'s notion of a point heat source.<ref>A more complete historical account can be found in {{harvnb|van der Pol|Bremmer|1987|loc=§V.4}}.</ref> The Dirac delta function as such was introduced by [[Paul Dirac]] in his 1927 paper ''The Physical Interpretation of the Quantum Dynamics.''<ref>{{Cite journal |date=January 1927 |title=The physical interpretation of the quantum dynamics |journal=Proceedings of the Royal Society of London. Series A, Containing Papers of a Mathematical and Physical Character |language=en |volume=113 |issue=765 |pages=621–641 |doi=10.1098/rspa.1927.0012 |bibcode=1927RSPSA.113..621D |issn=0950-1207|last1=Dirac |first1=P. A. M. |s2cid=122855515 |doi-access=free }}</ref> He called it the "delta function" since he used it as a [[Continuum (set theory)|continuum]] analogue of the discrete [[Kronecker delta]].{{sfn|Dirac|1930|loc=§22 The ''δ'' function}}
An [[infinitesimal]] formula for an infinitely tall, unit impulse delta function (infinitesimal version of [[Cauchy distribution]]) explicitly appears in an 1827 text of [[Augustin-Louis Cauchy]].{{sfn|Laugwitz|1989|p=230}} [[Siméon Denis Poisson]] considered the issue in connection with the study of wave propagation as did [[Gustav Kirchhoff]] somewhat later. Kirchhoff and [[Hermann von Helmholtz]] also introduced the unit impulse as a limit of [[Gaussian distribution|Gaussians]], which also corresponded to [[Lord Kelvin]]'s notion of a point heat source.<ref>A more complete historical account can be found in {{harvnb|van der Pol|Bremmer|1987|loc=§V.4}}.</ref> The Dirac delta function as such was introduced by [[Paul Dirac]] in his 1927 paper ''The Physical Interpretation of the Quantum Dynamics.''<ref>{{Cite journal |date=January 1927 |title=The physical interpretation of the quantum dynamics |journal=Proceedings of the Royal Society of London. Series A, Containing Papers of a Mathematical and Physical Character |language=en |volume=113 |issue=765 |pages=621–641 |doi=10.1098/rspa.1927.0012 |bibcode=1927RSPSA.113..621D |issn=0950-1207|last1=Dirac |first1=P. A. M. |s2cid=122855515 |doi-access=free }}</ref> He called it the "delta function" since he used it as a [[Continuum (set theory)|continuum]] analogue of the discrete [[Kronecker delta]].{{sfn|Dirac|1930|loc=§22 The ''δ'' function}}


Mathematicians refer to the same concept as a [[Distribution (mathematics)|distribution]] rather than a function.<ref>{{Cite book |last=Zee |first=Anthony |title=Einstein Gravity in a Nutshell |date=2013 |publisher=Princeton University Press |isbn=978-0-691-14558-7 |edition=1st |series=In a Nutshell Series |location=Princeton}}</ref>{{rp|33}}
Mathematicians refer to the same concept as a [[Distribution (mathematics)|distribution]] rather than a function.<ref>{{Cite book |last=Zee |first=Anthony |title=Einstein Gravity in a Nutshell |date=2013|page=33 |publisher=Princeton University Press |isbn=978-0-691-14558-7 |edition=1st |series=In a Nutshell Series |location=Princeton}}</ref>
[[Joseph Fourier]] presented what is now called the [[Fourier integral theorem]] in his treatise ''Théorie analytique de la chaleur'' in the form:<ref name=Fourier>{{cite book |title=The Analytical Theory of Heat |first=JB |last=Fourier |author-link=Joseph Fourier |year=1822 |page=[{{google books |plainurl=y |id=-N8EAAAAYAAJ|page=408}}] |edition= English translation by Alexander Freeman, 1878 |publisher=The University Press}}, cf. {{google books |plainurl=y |id=-N8EAAAAYAAJ|page=449
[[Joseph Fourier]] presented what is now called the [[Fourier integral theorem]] in his treatise ''Théorie analytique de la chaleur'' in the form:<ref name=Fourier>{{cite book |title=The Analytical Theory of Heat |first=JB |last=Fourier |author-link=Joseph Fourier |year=1822 |page=[{{google books |plainurl=y |id=-N8EAAAAYAAJ|page=408}}] |edition= English translation by Alexander Freeman, 1878 |publisher=The University Press}}, cf. {{google books |plainurl=y |id=-N8EAAAAYAAJ|page=449
}} and pp. 546–551. [{{google books |plainurl=y |id=TDQJAAAAIAAJ|page=525}} Original French text].</ref>
}} and pp. 546–551. [{{google books |plainurl=y |id=TDQJAAAAIAAJ|page=525}} Original French text].</ref>
Line 90: Line 90:
<math display="block">\delta(x) \simeq \begin{cases} +\infty, & x = 0 \\ 0, & x \ne 0 \end{cases}</math>
<math display="block">\delta(x) \simeq \begin{cases} +\infty, & x = 0 \\ 0, & x \ne 0 \end{cases}</math>


and which is also constrained to satisfy the identity{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.1, p. 1}}
and which is also constrained to satisfy the identity{{sfn|Halperin|Schwartz|1952|p=1}}


<math display="block">\int_{-\infty}^\infty \delta(x) \, dx = 1.</math>
<math display="block">\int_{-\infty}^\infty \delta(x) \, dx = 1.</math>
Line 97: Line 97:


===As a measure===
===As a measure===
One way to rigorously capture the notion of the Dirac delta function is to define a [[Measure (mathematics)|measure]], called [[Dirac measure]], which accepts a subset {{mvar|A}} of the real line {{math|'''R'''}} as an argument, and returns {{math|1=''δ''(''A'') = 1}} if {{math|0 ∈ ''A''}}, and {{math|1=''δ''(''A'') = 0}} otherwise.<ref name="Rudin 1966 loc=§1.20">{{harvnb|Rudin|1966|loc=§1.20}}</ref> If the delta function is conceptualized as modeling an idealized point mass at 0, then {{math|''δ''(''A'')}} represents the mass contained in the set {{mvar|A}}. One may then define the integral against {{mvar|δ}} as the integral of a function against this mass distribution. Formally, the [[Lebesgue integral]] provides the necessary analytic device. The Lebesgue integral with respect to the measure {{mvar|δ}} satisfies
One way to rigorously capture the notion of the Dirac delta function is to define a [[Measure (mathematics)|measure]], called [[Dirac measure]], which accepts a subset {{mvar|A}} of the real line {{math|'''R'''}} as an argument, and returns {{math|1=''δ''(''A'') = 1}} if {{math|0 ∈ ''A''}}, and {{math|1=''δ''(''A'') = 0}} otherwise.<ref name="Rudin 1966 loc=§1.20">{{harvnb|Rudin|1966|loc=§1.20}}</ref> If the delta function is conceptualized as modeling an idealized point mass at 0, then {{math|''δ''(''A'')}} represents the mass contained in the set {{mvar|A}}. One may then define the integral against {{mvar|δ}} as the integral of a function against this mass distribution. Formally, the [[Lebesgue integral]] provides the necessary analytic device. The Lebesgue integral with respect to the measure {{mvar|δ}} satisfies


<math display="block">\int_{-\infty}^\infty f(x) \, \delta(dx) =  f(0)</math>
<math display="block">\int_{-\infty}^\infty f(x) \, \delta(dx) =  f(0)</math>
Line 105: Line 105:
<math display="block">\int_{-\infty}^\infty f(x)\, \delta(x)\, dx = f(0)</math>
<math display="block">\int_{-\infty}^\infty f(x)\, \delta(x)\, dx = f(0)</math>


holds.{{sfn|Hewitt|Stromberg|1963|loc=§19.61}} As a result, the latter notation is a convenient [[abuse of notation]], and not a standard ([[Riemann integral|Riemann]] or [[Lebesgue integral|Lebesgue]]) integral.{{sfn|Gelfand|Shilov|p=1}}
holds.{{sfn|Hewitt|Stromberg|1963|loc=§19.61}} As a result, the latter notation is a convenient [[abuse of notation]], and not a standard ([[Riemann integral|Riemann]] or [[Lebesgue integral|Lebesgue]]) integral.{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.3}}


As a [[probability measure]] on {{math|'''R'''}}, the delta measure is characterized by its [[cumulative distribution function]], which is the [[unit step function]].<ref>{{harvnb|Driggers|2003|p=2321}} See also {{harvnb|Bracewell|1986|loc=Chapter 5}} for a different interpretation. Other conventions for the assigning the value of the Heaviside function at zero exist, and some of these are not consistent with what follows.</ref>
As a [[probability measure]] on {{math|'''R'''}}, the delta measure is characterized by its [[cumulative distribution function]], which is the [[unit step function]].<ref>{{harvnb|Driggers|2003|p=2321}} See also {{harvnb|Bracewell|1986|loc=Chapter 5}} for a different interpretation. Other conventions for the assigning the value of the Heaviside function at zero exist, and some of these are not consistent with what follows.</ref>


<math display="block">H(x) =  
<math display="block">H(x) =  
Line 123: Line 123:
<math display="block">\int_{-\infty}^\infty f(x)\,\delta(dx) = \int_{-\infty}^\infty f(x) \,dH(x).</math>
<math display="block">\int_{-\infty}^\infty f(x)\,\delta(dx) = \int_{-\infty}^\infty f(x) \,dH(x).</math>


All higher [[moment (mathematics)|moments]] of {{mvar|δ}} are zero. In particular, [[characteristic function (probability theory)|characteristic function]] and [[moment generating function]] are both equal to one.{{cn|date=June 2025}}
All higher [[moment (mathematics)|moments]] of {{mvar|δ}} are zero. In particular, [[characteristic function (probability theory)|characteristic function]] and [[moment generating function]] are both equal to one.{{sfn|Billingsley|1986|p=356}}


===As a distribution===
===As a distribution===
In the theory of [[distribution (mathematics)|distributions]], a generalized function is considered not a function in itself but only through how it affects other functions when "integrated" against them.{{sfn|Hazewinkel|2011|p=[{{google books |plainurl=y |id=_YPtCAAAQBAJ|page=41}} 41]}} In keeping with this philosophy, to define the delta function properly, it is enough to say what the "integral" of the delta function is against a sufficiently "good" '''test function''' {{mvar|φ}}.{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.1}} If the delta function is already understood as a measure, then the Lebesgue integral of a test function against that measure supplies the necessary integral.{{sfn|Stein|Shararchi|2007|p=285}}
In the theory of [[distribution (mathematics)|distributions]], a generalized function is considered not a function in itself but only through how it affects other functions when "integrated" against them.{{sfn|Hazewinkel|2011|p=[{{google books |plainurl=y |id=_YPtCAAAQBAJ|page=41}} 41]}} In keeping with this philosophy, to define the delta function properly, it is enough to say what the "integral" of the delta function is against a sufficiently "good" '''test function''' {{mvar|φ}}.{{sfn|Gelfand|Shilov|1966–1968|loc=Volume I, §1.1}} If the delta function is already understood as a measure, then the Lebesgue integral of a test function against that measure supplies the necessary integral.{{sfn|Stein|Shakarchi|2007|p=285}}


A typical space of test functions consists of all [[smooth function]]s on {{math|'''R'''}} with [[compact support]] that have as many derivatives as required. As a distribution, the Dirac delta is a [[linear functional]] on the space of test functions and is defined by{{sfn|Strichartz|1994|loc=§2.2}}
A typical space of test functions consists of all [[smooth function]]s on {{math|'''R'''}} with [[compact support]] that have as many derivatives as required. As a distribution, the Dirac delta is a [[linear functional]] on the space of test functions and is defined by{{sfn|Strichartz|1994|loc=§2.2}}


{{NumBlk2|:| <math>\delta[\varphi] = \varphi(0)</math>|1}}
{{NumBlk2|:| <math>\delta[\varphi] = \varphi(0)</math>|1}}
Line 134: Line 134:
for every test function {{mvar|φ}}.
for every test function {{mvar|φ}}.


For {{mvar|δ}} to be properly a distribution, it must be continuous in a suitable topology on the space of test functions. In general, for a linear functional {{mvar|S}} on the space of test functions to define a distribution, it is necessary and sufficient that, for every positive integer {{mvar|N}} there is an integer {{math|''M''<sub>''N''</sub>}} and a constant {{mvar|''C''<sub>''N''</sub>}} such that for every test function {{mvar|φ}}, one has the inequality{{sfn|Hörmander|1983|loc=Theorem 2.1.5}}
For {{mvar|δ}} to be properly a distribution, it must be continuous in a suitable topology on the space of test functions. In general, for a linear functional {{mvar|S}} on the space of test functions to define a distribution, it is necessary and sufficient that, for every positive integer {{mvar|N}} there is an integer {{math|''M''<sub>''N''</sub>}} and a constant {{mvar|''C''<sub>''N''</sub>}} such that for every test function {{mvar|φ}}, one has the inequality{{sfn|Hörmander|1983|loc=Theorem 2.1.5}}


<math display="block">\left|S[\varphi]\right| \le C_N \sum_{k=0}^{M_N}\sup_{x\in [-N,N]} \left|\varphi^{(k)}(x)\right|</math>
<math display="block">\left|S[\varphi]\right| \le C_N \sum_{k=0}^{M_N}\sup_{x\in [-N,N]} \left|\varphi^{(k)}(x)\right|</math>


where {{math|sup}} represents the [[Infimum and supremum|supremum]]. With the {{mvar|δ}} distribution, one has such an inequality (with {{math|1=''C''<sub>''N''</sub> = 1)}} with {{math|1=''M''<sub>''N''</sub> = 0}} for all {{mvar|N}}. Thus {{mvar|δ}} is a distribution of order zero. It is, furthermore, a distribution with compact support (the [[support (mathematics)|support]] being {{math|{{brace|0}}}}).
where {{math|sup}} represents the [[Infimum and supremum|supremum]]. With the {{mvar|δ}} distribution, one has such an inequality (with {{math|1=''C''<sub>''N''</sub> = 1)}} with {{math|1=''M''<sub>''N''</sub> = 0}} for all {{mvar|N}}. Thus {{mvar|δ}} is a distribution of order zero. It is, furthermore, a distribution with compact support (the [[support (mathematics)|support]] being {{math|{{brace|0}}}}).


The delta distribution can also be defined in several equivalent ways. For instance, it is the [[distributional derivative]] of the [[Heaviside step function]]. This means that for every test function {{mvar|φ}}, one has
The delta distribution can also be defined in several equivalent ways. For instance, it is the [[distributional derivative]] of the [[Heaviside step function]]. This means that for every test function {{mvar|φ}}, one has


<math display="block">\delta[\varphi] = -\int_{-\infty}^\infty \varphi'(x)\,H(x)\,dx.</math>
<math display="block">\delta[\varphi] = -\int_{-\infty}^\infty \varphi'(x)\,H(x)\,dx.</math>
Line 152: Line 152:
<math display="block">-\int_{-\infty}^\infty \varphi'(x)\,H(x)\, dx = \int_{-\infty}^\infty \varphi(x)\,dH(x).</math>
<math display="block">-\int_{-\infty}^\infty \varphi'(x)\,H(x)\, dx = \int_{-\infty}^\infty \varphi(x)\,dH(x).</math>


In the context of measure theory, the Dirac measure gives rise to distribution by integration. Conversely, equation ({{EquationNote|1}}) defines a [[Daniell integral]] on the space of all compactly supported continuous functions {{mvar|φ}} which, by the [[Riesz–Markov–Kakutani representation theorem|Riesz representation theorem]], can be represented as the Lebesgue integral of {{mvar|φ}} with respect to some [[Radon measure]].{{sfn|Schwartz|1950}}}
In the context of measure theory, the Dirac measure gives rise to distribution by integration. Conversely, equation ({{EquationNote|1}}) defines a [[Daniell integral]] on the space of all compactly supported continuous functions {{mvar|φ}} which, by the [[Riesz–Markov–Kakutani representation theorem|Riesz representation theorem]], can be represented as the Lebesgue integral of {{mvar|φ}} with respect to some [[Radon measure]].{{sfn|Schwartz|1950}}}


Generally, when the term ''Dirac delta function'' is used, it is in the sense of distributions rather than measures, the [[Dirac measure]] being among several terms for the corresponding notion in measure theory. Some sources may also use the term ''Dirac delta distribution''.
Generally, when the term ''Dirac delta function'' is used, it is in the sense of distributions rather than measures, the [[Dirac measure]] being among several terms for the corresponding notion in measure theory. Some sources may also use the term ''Dirac delta distribution''.


===Generalizations===
===Generalizations===
Line 161: Line 161:
<math display="block">\int_{\mathbf{R}^n} f(\mathbf{x})\,\delta(d\mathbf{x}) = f(\mathbf{0})</math>
<math display="block">\int_{\mathbf{R}^n} f(\mathbf{x})\,\delta(d\mathbf{x}) = f(\mathbf{0})</math>


for every compactly supported continuous function {{mvar|f}}. As a measure, the {{mvar|n}}-dimensional delta function is the [[product measure]] of the 1-dimensional delta functions in each variable separately. Thus, formally, with {{math|1='''x''' = (''x''<sub>1</sub>, ''x''<sub>2</sub>, ..., ''x''<sub>''n''</sub>)}}, one has{{sfn|Bracewell|1986|loc=Chapter 5}}
for every compactly supported continuous function {{mvar|f}}. As a measure, the {{mvar|n}}-dimensional delta function is the [[product measure]] of the 1-dimensional delta functions in each variable separately. Thus, formally, with {{math|1='''x''' = (''x''<sub>1</sub>, ''x''<sub>2</sub>, ..., ''x''<sub>''n''</sub>)}}, one has{{sfn|Bracewell|1986|loc=Chapter 5}}


{{NumBlk2|:|<math>\delta(\mathbf{x}) = \delta(x_1)\,\delta(x_2)\cdots\delta(x_n).</math>|2}}
{{NumBlk2|:|<math>\delta(\mathbf{x}) = \delta(x_1)\,\delta(x_2)\cdots\delta(x_n).</math>|2}}


The delta function can also be defined in the sense of distributions exactly as above in the one-dimensional case.{{sfn|Hörmander|1983|loc=§3.1}} However, despite widespread use in engineering contexts, ({{EquationNote|2}}) should be manipulated with care, since the product of distributions can only be defined under quite narrow circumstances.{{sfn|Strichartz|1994|loc=§2.3}}{{sfn|Hörmander|1983|loc=§8.2}}
The delta function can also be defined in the sense of distributions exactly as above in the one-dimensional case.{{sfn|Hörmander|1983|loc=§3.1}} However, despite widespread use in engineering contexts, ({{EquationNote|2}}) should be manipulated with care, since the product of distributions can only be defined under quite narrow circumstances.{{sfn|Strichartz|1994|loc=§2.3}}{{sfn|Hörmander|1983|loc=§8.2}}


The notion of a '''[[Dirac measure]]''' makes sense on any set.{{sfn|Rudin |1966 |loc=§1.20}} Thus if {{mvar|X}} is a set, {{math|''x''<sub>0</sub> ∈ ''X''}} is a marked point, and {{math|Σ}} is any [[sigma algebra]] of subsets of {{mvar|X}}, then the measure defined on sets {{math|''A'' ∈ Σ}} by
The notion of a [[Dirac measure]] makes sense on any set.{{sfn|Rudin |1966 |loc=§1.20}} Thus if {{mvar|X}} is a set, {{math|''x''<sub>0</sub> ∈ ''X''}} is a marked point, and {{math|Σ}} is any [[sigma algebra]] of subsets of {{mvar|X}}, then the measure defined on sets {{math|''A'' ∈ Σ}} by


<math display="block">\delta_{x_0}(A)=\begin{cases}
<math display="block">\delta_{x_0}(A)=\begin{cases}
Line 176: Line 176:
is the delta measure or unit mass concentrated at {{math|''x''<sub>0</sub>}}.
is the delta measure or unit mass concentrated at {{math|''x''<sub>0</sub>}}.


Another common generalization of the delta function is to a [[differentiable manifold]] where most of its properties as a distribution can also be exploited because of the [[differentiable structure]]. The delta function on a manifold {{mvar|M}} centered at the point {{math|''x''<sub>0</sub> ∈ ''M''}} is defined as the following distribution:
Another common generalization of the delta function is to a [[differentiable manifold]] where most of its properties as a distribution can also be exploited because of the [[differentiable structure]]. The delta function on a manifold {{mvar|M}} centered at the point {{math|''x''<sub>0</sub> ∈ ''M''}} is defined as the following distribution:


{{NumBlk2|:|<math>\delta_{x_0}[\varphi] = \varphi(x_0)</math>|3}}
{{NumBlk2|:|<math>\delta_{x_0}[\varphi] = \varphi(x_0)</math>|3}}
Line 197: Line 197:


Scaling property proof:
Scaling property proof:
<math display="block">\int\limits_{-\infty}^{\infty} dx\ g(x) \delta (ax) = \frac{1}{a}\int\limits_{-\infty}^{\infty} dx'\ g\left(\frac{x'}{a}\right) \delta (x') = \frac{1}{   a  }g(0).
<math display="block">\int\limits_{-\infty}^{\infty} dx\ g(x) \delta (\alpha x) = \frac{1}{\alpha}\int\limits_{-\infty}^{\infty} dx'\ g\left(\frac{x'}{\alpha}\right) \delta (x') = \frac{1}{\alpha}g(0).
</math>
</math>
where a change of variable {{math|1=''x&prime;'' = ''ax''}} is used. If {{mvar|a}} is negative, i.e., {{math|1=''a'' = &minus;{{!}}''a''{{!}}}}, then
where a change of variable {{math|1=''x&prime;'' = ''&alpha;x''}} is used. If {{mvar|&alpha;}} is negative, i.e., {{math|1=''&alpha;'' = &minus;{{!}}''a''{{!}}}}, then
<math display="block">\int\limits_{-\infty}^{\infty} dx\ g(x) \delta (ax) = \frac{1}{-\left \vert a \right \vert}\int\limits_{\infty}^{-\infty} dx'\ g\left(\frac{x'}{a}\right) \delta (x') = \frac{1}{\left \vert a \right \vert}\int\limits_{-\infty}^{\infty} dx'\ g\left(\frac{x'}{a}\right) \delta (x') = \frac{1}{\left \vert a \right \vert}g(0).
<math display="block">\int\limits_{-\infty}^{\infty} dx\ g(x) \delta (\alpha x) = \frac{1}{-\left \vert \alpha \right \vert}\int\limits_{\infty}^{-\infty} dx'\ g\left(\frac{x'}{\alpha}\right) \delta (x') = \frac{1}{\left \vert \alpha \right \vert}\int\limits_{-\infty}^{\infty} dx'\ g\left(\frac{x'}{\alpha}\right) \delta (x') = \frac{1}{\left \vert \alpha \right \vert}g(0).
</math>
</math>
Thus, {{nowrap|<math>\delta (ax) = \frac{1}{\left \vert a \right \vert} \delta(x)</math>.}}
Thus, {{nowrap|<math>\delta (\alpha x) = \frac{1}{\left \vert \alpha \right \vert} \delta(x)</math>.}}


In particular, the delta function is an [[even function|even]] distribution (symmetry), in the sense that
In particular, the delta function is an [[even function|even]] distribution (symmetry), in the sense that
Line 230: Line 230:
This is sometimes referred to as the ''sifting property''<ref>{{MathWorld|urlname=SiftingProperty|title=Sifting Property}}</ref> or the ''sampling property''.<ref>{{Cite book|last=Karris|first=Steven T.|url={{google books |plainurl=y |id=f0RdM1zv_dkC}}| title=Signals and Systems with MATLAB Applications|date=2003|publisher=Orchard Publications|isbn=978-0-9709511-6-8|language=en| page=[{{google books |plainurl=y |id=f0RdM1zv_dkC&pg=SA1-PA15 }} 15]}}</ref> The delta function is said to "sift out" the value of ''f(t)'' at ''t'' = ''T''.<ref>{{Cite book|last=Roden|first=Martin S.|url={{google books |plainurl=y |id=YEKeBQAAQBAJ}}|title=Introduction to Communication Theory|date=2014-05-17|publisher=Elsevier|isbn=978-1-4831-4556-3|language=en|page=[{{google books |plainurl=y |id=YEKeBQAAQBAJ|page=40}}]}}</ref>
This is sometimes referred to as the ''sifting property''<ref>{{MathWorld|urlname=SiftingProperty|title=Sifting Property}}</ref> or the ''sampling property''.<ref>{{Cite book|last=Karris|first=Steven T.|url={{google books |plainurl=y |id=f0RdM1zv_dkC}}| title=Signals and Systems with MATLAB Applications|date=2003|publisher=Orchard Publications|isbn=978-0-9709511-6-8|language=en| page=[{{google books |plainurl=y |id=f0RdM1zv_dkC&pg=SA1-PA15 }} 15]}}</ref> The delta function is said to "sift out" the value of ''f(t)'' at ''t'' = ''T''.<ref>{{Cite book|last=Roden|first=Martin S.|url={{google books |plainurl=y |id=YEKeBQAAQBAJ}}|title=Introduction to Communication Theory|date=2014-05-17|publisher=Elsevier|isbn=978-1-4831-4556-3|language=en|page=[{{google books |plainurl=y |id=YEKeBQAAQBAJ|page=40}}]}}</ref>


It follows that the effect of [[Convolution|convolving]] a function {{math|''f''(''t'')}} with the time-delayed Dirac delta is to time-delay {{math|''f''(''t'')}} by the same amount:<ref>{{Cite book|last1=Rottwitt|first1=Karsten|url={{google books |plainurl=y |id=G1jSBQAAQBAJ}}|title=Nonlinear Optics: Principles and Applications|last2=Tidemand-Lichtenberg|first2=Peter| date=2014-12-11| publisher=CRC Press|isbn=978-1-4665-6583-8|language=en|page=[{{google books |plainurl=y |id=G1jSBQAAQBAJ|page=276}}] 276}}</ref>
It follows that the effect of [[Convolution|convolving]] a function {{math|''f''(''t'')}} with the time-delayed Dirac delta is to time-delay {{math|''f''(''t'')}} by the same amount:<ref>{{Cite book|last1=Rottwitt|first1=Karsten|url={{google books |plainurl=y |id=G1jSBQAAQBAJ}}|title=Nonlinear Optics: Principles and Applications|last2=Tidemand-Lichtenberg|first2=Peter| date=2014-12-11| publisher=CRC Press|isbn=978-1-4665-6583-8|language=en|page=[{{google books |plainurl=y |id=G1jSBQAAQBAJ|page=276}}] 276}}</ref>


<math display="block">\begin{align}
<math display="block">\begin{align}
Line 247: Line 247:
<math display="block">\int_{\R} \delta\bigl(g(x)\bigr) f\bigl(g(x)\bigr) \left|g'(x)\right| dx = \int_{g(\R)} \delta(u)\,f(u)\,du</math>
<math display="block">\int_{\R} \delta\bigl(g(x)\bigr) f\bigl(g(x)\bigr) \left|g'(x)\right| dx = \int_{g(\R)} \delta(u)\,f(u)\,du</math>


provided that {{mvar|g}} is a [[continuously differentiable]] function with {{math|''g&prime;''}} nowhere zero.{{sfn|Gelfand|Shilov|1966–1968|loc=Vol. 1, §II.2.5}} That is, there is a unique way to assign meaning to the distribution <math>\delta\circ g</math> so that this identity holds for all compactly supported test functions {{mvar|f}}. Therefore, the domain must be broken up to exclude the {{math|1=''g&prime;'' = 0}} point. This distribution satisfies {{math|1=''δ''(''g''(''x'')) = 0}} if {{mvar|g}} is nowhere zero, and otherwise if {{mvar|g}} has a real [[root of a function|root]] at {{math|''x''<sub>0</sub>}}, then
provided that {{mvar|g}} is a [[continuously differentiable]] function with {{math|''g&prime;''}} nowhere zero.{{sfn|Gelfand|Shilov|1966–1968|loc=Vol. 1, §II.2.5}} That is, there is a unique way to assign meaning to the distribution <math>\delta\circ g</math> so that this identity holds for all compactly supported test functions {{mvar|f}}. Therefore, the domain must be broken up to exclude the {{math|1=''g&prime;'' = 0}} point. This distribution satisfies {{math|1=''δ''(''g''(''x'')) = 0}} if {{mvar|g}} is nowhere zero, and otherwise if {{mvar|g}} has a real [[root of a function|root]] at {{math|''x''<sub>0</sub>}}, then


<math display="block">\delta(g(x)) = \frac{\delta(x-x_0)}{|g'(x_0)|}.</math>
<math display="block">\delta(g(x)) = \frac{\delta(x-x_0)}{|g'(x_0)|}.</math>
Line 255: Line 255:
<math display="block">\delta(g(x)) = \sum_i \frac{\delta(x-x_i)}{|g'(x_i)|}</math>
<math display="block">\delta(g(x)) = \sum_i \frac{\delta(x-x_i)}{|g'(x_i)|}</math>


where the sum extends over all roots of {{mvar|''g''(''x'')}}, which are assumed to be [[simple root|simple]]. Thus, for example
where the sum extends over all roots of {{mvar|''g''(''x'')}}, which are assumed to be [[simple root|simple]]. Thus, for example


<math display="block">\delta\left(x^2-\alpha^2\right) = \frac{1}{2|\alpha|} \Big[\delta\left(x+\alpha\right)+\delta\left(x-\alpha\right)\Big].</math>
<math display="block">\delta\left(x^2-\alpha^2\right) = \frac{1}{2|\alpha|} \Big[\delta\left(x+\alpha\right)+\delta\left(x-\alpha\right)\Big].</math>
Line 282: Line 282:
Using the [[coarea formula]] from [[geometric measure theory]], one can also define the composition of the delta function with a [[submersion (mathematics)|submersion]] from one Euclidean space to another one of different dimension; the result is a type of [[current (mathematics)|current]]. In the special case of a continuously differentiable function {{math|''g'' : '''R'''<sup>''n''</sup> → '''R'''}} such that the [[gradient]] of {{mvar|g}} is nowhere zero, the following identity holds{{sfn|Hörmander|1983|loc=§6.1}}
Using the [[coarea formula]] from [[geometric measure theory]], one can also define the composition of the delta function with a [[submersion (mathematics)|submersion]] from one Euclidean space to another one of different dimension; the result is a type of [[current (mathematics)|current]]. In the special case of a continuously differentiable function {{math|''g'' : '''R'''<sup>''n''</sup> → '''R'''}} such that the [[gradient]] of {{mvar|g}} is nowhere zero, the following identity holds{{sfn|Hörmander|1983|loc=§6.1}}
<math display="block">\int_{\R^n} f(\boldsymbol{x}) \, \delta(g(\boldsymbol{x})) \,d\boldsymbol{x} = \int_{g^{-1}(0)}\frac{f(\boldsymbol{x})}{|\boldsymbol{\nabla}g|}\,d\sigma(\boldsymbol{x}) </math>
<math display="block">\int_{\R^n} f(\boldsymbol{x}) \, \delta(g(\boldsymbol{x})) \,d\boldsymbol{x} = \int_{g^{-1}(0)}\frac{f(\boldsymbol{x})}{|\boldsymbol{\nabla}g|}\,d\sigma(\boldsymbol{x}) </math>
where the integral on the right is over {{math|''g''<sup>−1</sup>(0)}}, the {{math|(''n'' − 1)}}-dimensional surface defined by {{math|1=''g''('''x''') = 0}} with respect to the [[Minkowski content]] measure. This is known as a ''simple layer'' integral.
where the integral on the right is over {{math|''g''<sup>−1</sup>(0)}}, the {{math|(''n'' − 1)}}-dimensional surface defined by {{math|1=''g''('''x''') = 0}} with respect to the [[Minkowski content]] measure. This is known as a ''simple layer'' integral.


More generally, if {{mvar|S}} is a smooth hypersurface of {{math|'''R'''<sup>''n''</sup>}}, then we can associate to {{mvar|S}} the distribution that integrates any compactly supported smooth function {{mvar|g}} over {{mvar|S}}:
More generally, if {{mvar|S}} is a smooth hypersurface of {{math|'''R'''<sup>''n''</sup>}}, then we can associate to {{mvar|S}} the distribution that integrates any compactly supported smooth function {{mvar|g}} over {{mvar|S}}:
<math display="block">\delta_S[g] = \int_S g(\boldsymbol{s})\,d\sigma(\boldsymbol{s})</math>
<math display="block">\delta_S[g] = \int_S g(\boldsymbol{s})\,d\sigma(\boldsymbol{s})</math>


where {{mvar|σ}} is the hypersurface measure associated to {{mvar|S}}. This generalization is associated with the [[potential theory]] of [[simple layer potential]]s on {{mvar|S}}. If {{mvar|D}} is a [[domain (mathematical analysis)|domain]] in {{math|'''R'''<sup>''n''</sup>}} with smooth boundary {{mvar|S}}, then {{math|''δ''<sub>''S''</sub>}} is equal to the [[normal derivative]] of the [[indicator function]] of {{mvar|D}} in the distribution sense,
where {{mvar|σ}} is the hypersurface measure associated to {{mvar|S}}. This generalization is associated with the [[potential theory]] of [[simple layer potential]]s on {{mvar|S}}. If {{mvar|D}} is a [[domain (mathematical analysis)|domain]] in {{math|'''R'''<sup>''n''</sup>}} with smooth boundary {{mvar|S}}, then {{math|''δ''<sub>''S''</sub>}} is equal to the [[normal derivative]] of the [[indicator function]] of {{mvar|D}} in the distribution sense,


<math display="block">-\int_{\R^n}g(\boldsymbol{x})\,\frac{\partial 1_D(\boldsymbol{x})}{\partial n}\,d\boldsymbol{x}=\int_S\,g(\boldsymbol{s})\, d\sigma(\boldsymbol{s}),</math>
<math display="block">-\int_{\R^n}g(\boldsymbol{x})\,\frac{\partial 1_D(\boldsymbol{x})}{\partial n}\,d\boldsymbol{x}=\int_S\,g(\boldsymbol{s})\, d\sigma(\boldsymbol{s}),</math>


where {{mvar|n}} is the outward normal.{{sfn|Lange|2012|loc=pp.29–30}}{{sfn|Gelfand|Shilov|1966–1968|p=212}} For a proof, see e.g. the article on the [[surface delta function]].
where {{mvar|n}} is the outward normal.{{sfn|Lange|2012|loc=pp.29–30}}{{sfn|Gelfand|Shilov|1966–1968|p=212}}


In three dimensions, the delta function is represented in spherical coordinates by:
In three dimensions, the delta function is represented in spherical coordinates by:
Line 326: Line 326:
<math display="block">(\tau_h S)[\varphi] = S[\tau_{-h}\varphi].</math>
<math display="block">(\tau_h S)[\varphi] = S[\tau_{-h}\varphi].</math>


In the theory of [[electromagnetism]], the first derivative of the delta function represents a point magnetic [[dipole]] situated at the origin. Accordingly, it is referred to as a dipole or the [[unit doublet|doublet function]].<ref>{{MathWorld|title=Doublet Function|urlname=DoubletFunction}}</ref>
In the theory of [[electromagnetism]], the first derivative of the delta function represents a point magnetic [[dipole]] situated at the origin. Accordingly, it is referred to as a dipole or the [[unit doublet|doublet function]].<ref>{{MathWorld|title=Doublet Function|urlname=DoubletFunction}}</ref>


The derivative of the delta function satisfies a number of basic properties, including:{{sfn|Bracewell|2000|p=86}}
The derivative of the delta function satisfies a number of basic properties, including:{{sfn|Bracewell|2000|p=86}}
Line 337: Line 337:
which can be shown by applying a test function and integrating by parts.
which can be shown by applying a test function and integrating by parts.


The latter of these properties can also be demonstrated by applying distributional derivative definition, Leibniz&nbsp;'s theorem and linearity of inner product:<ref>{{Cite web|url=https://www.matematicamente.it/forum/viewtopic.php?f=36&t=62388&start=10#wrap|title=Gugo82's comment on the distributional derivative of Dirac's delta|date=12 September 2010|website=matematicamente.it}}</ref>
The latter of these properties can also be demonstrated by applying distributional derivative definition, Leibniz&nbsp;'s theorem and linearity of inner product:<ref>{{Cite web|url=https://www.matematicamente.it/forum/viewtopic.php?f=36&t=62388&start=10#wrap|title=Gugo82's comment on the distributional derivative of Dirac's delta|date=12 September 2010|website=matematicamente.it}}</ref>{{better source|date=July 2025}}


<math display="block">
<math display="block">
Line 358: Line 358:
More generally, on an [[open set]] {{mvar|U}} in the {{mvar|n}}-dimensional [[Euclidean space]] <math>\mathbb{R}^n</math>, the Dirac delta distribution centered at a point {{math|''a'' ∈ ''U''}} is defined by{{sfn|Hörmander|1983|p=56}}
More generally, on an [[open set]] {{mvar|U}} in the {{mvar|n}}-dimensional [[Euclidean space]] <math>\mathbb{R}^n</math>, the Dirac delta distribution centered at a point {{math|''a'' ∈ ''U''}} is defined by{{sfn|Hörmander|1983|p=56}}
<math display="block">\delta_a[\varphi]=\varphi(a)</math>
<math display="block">\delta_a[\varphi]=\varphi(a)</math>
for all <math>\varphi \in C_c^\infty(U)</math>, the space of all smooth functions with compact support on {{mvar|U}}. If <math>\alpha = (\alpha_1, \ldots, \alpha_n)</math> is any [[multi-index]] with <math> |\alpha|=\alpha_1+\cdots+\alpha_n</math> and <math>\partial^\alpha</math> denotes the associated mixed [[partial derivative]] operator, then the {{mvar|α}}-th derivative {{mvar|∂<sup>α</sup>δ<sub>a</sub>}} of {{mvar|δ<sub>a</sub>}} is given by{{sfn|Hörmander|1983|p=56}}
for all <math>\varphi \in C_c^\infty(U)</math>, the space of all smooth functions with compact support on {{mvar|U}}. If <math>\alpha = (\alpha_1, \ldots, \alpha_n)</math> is any [[multi-index]] with <math> |\alpha|=\alpha_1+\cdots+\alpha_n</math> and <math>\partial^\alpha</math> denotes the associated mixed [[partial derivative]] operator, then the {{mvar|α}}-th derivative {{mvar|∂<sup>α</sup>δ<sub>a</sub>}} of {{mvar|δ<sub>a</sub>}} is given by{{sfn|Hörmander|1983|p=56}}


<math display="block">\left\langle \partial^\alpha \delta_{a}, \, \varphi \right\rangle = (-1)^{| \alpha |} \left\langle \delta_{a}, \partial^{\alpha} \varphi \right\rangle = (-1)^{| \alpha |} \partial^\alpha \varphi (x) \Big|_{x = a} \quad \text{ for all } \varphi \in C_c^\infty(U).</math>
<math display="block">\left\langle \partial^\alpha \delta_{a}, \, \varphi \right\rangle = (-1)^{| \alpha |} \left\langle \delta_{a}, \partial^{\alpha} \varphi \right\rangle = (-1)^{| \alpha |} \partial^\alpha \varphi (x) \Big|_{x = a} \quad \text{ for all } \varphi \in C_c^\infty(U).</math>
Line 364: Line 364:
That is, the {{mvar|α}}-th derivative of {{mvar|δ<sub>a</sub>}} is the distribution whose value on any test function {{mvar|φ}} is the {{mvar|α}}-th derivative of {{mvar|φ}} at {{mvar|a}} (with the appropriate positive or negative sign).
That is, the {{mvar|α}}-th derivative of {{mvar|δ<sub>a</sub>}} is the distribution whose value on any test function {{mvar|φ}} is the {{mvar|α}}-th derivative of {{mvar|φ}} at {{mvar|a}} (with the appropriate positive or negative sign).


The first partial derivatives of the delta function are thought of as [[double layer potential|double layers]] along the coordinate planes. More generally, the [[normal derivative]] of a simple layer supported on a surface is a double layer supported on that surface and represents a laminar magnetic monopole. Higher derivatives of the delta function are known in physics as [[multipole]]s.{{cn|date=June 2025}}
The first partial derivatives of the delta function are thought of as [[double layer potential|double layers]] along the coordinate planes. More generally, the [[normal derivative]] of a simple layer supported on a surface is a double layer supported on that surface and represents a laminar magnetic monopole. Higher derivatives of the delta function are known in physics as [[multipole]]s.<ref>{{cite journal|title=Application of the Dirac delta function to electric charge and multipole distributions|first=Victor|last=Namias|date=July 1977|journal=[[American Journal of Physics]]|doi=10.1119/1.10779|volume=45|issue=7|pages=624–630}}</ref>


Higher derivatives enter into mathematics naturally as the building blocks for the complete structure of distributions with point support. If {{mvar|S}} is any distribution on {{mvar|U}} supported on the set {{math|{{brace|''a''}}}} consisting of a single point, then there is an integer {{mvar|m}} and coefficients {{mvar|c<sub>α</sub>}} such that{{sfn|Hörmander|1983|p=56}}{{sfn|Rudin|1991|loc=Theorem 6.25}}
Higher derivatives enter into mathematics naturally as the building blocks for the complete structure of distributions with point support. If {{mvar|S}} is any distribution on {{mvar|U}} supported on the set {{math|{{brace|''a''}}}} consisting of a single point, then there is an integer {{mvar|m}} and coefficients {{mvar|c<sub>α</sub>}} such that{{sfn|Hörmander|1983|p=56}}{{sfn|Rudin|1991|loc=Theorem 6.25}}
<math display="block">S = \sum_{|\alpha|\le m} c_\alpha \partial^\alpha\delta_a.</math>
<math display="block">S = \sum_{|\alpha|\le m} c_\alpha \partial^\alpha\delta_a.</math>


Line 379: Line 379:
for all [[continuous function|continuous]] functions {{mvar|f}} having [[compact support]], or that this limit holds for all [[smooth function|smooth]] functions {{mvar|f}} with compact support. The former is convergence in the [[vague topology]] of measures, and the latter is convergence in the sense of [[distribution (mathematics)|distributions]].
for all [[continuous function|continuous]] functions {{mvar|f}} having [[compact support]], or that this limit holds for all [[smooth function|smooth]] functions {{mvar|f}} with compact support. The former is convergence in the [[vague topology]] of measures, and the latter is convergence in the sense of [[distribution (mathematics)|distributions]].


====Approximations to the identity====
=== Approximations to the identity ===
An approximate delta function {{mvar|η<sub>ε</sub>}} can be constructed in the following manner. Let {{mvar|η}} be an absolutely integrable function on {{math|'''R'''}} of total integral {{math|1}}, and define
An approximate delta function {{mvar|η<sub>ε</sub>}} can be constructed in the following manner. Let {{mvar|η}} be an absolutely integrable function on {{math|'''R'''}} of total integral {{math|1}}, and define
<math display="block">\eta_\varepsilon(x) = \varepsilon^{-1} \eta \left (\frac{x}{\varepsilon} \right). </math>
<math display="block">\eta_\varepsilon(x) = \varepsilon^{-1} \eta \left (\frac{x}{\varepsilon} \right). </math>


Line 386: Line 386:
<math display="block">\eta_\varepsilon(x) = \varepsilon^{-n} \eta \left (\frac{x}{\varepsilon} \right). </math>
<math display="block">\eta_\varepsilon(x) = \varepsilon^{-n} \eta \left (\frac{x}{\varepsilon} \right). </math>


Then a simple change of variables shows that {{mvar|η<sub>ε</sub>}} also has integral {{math|1}}. One may show that ({{EquationNote|5}}) holds for all continuous compactly supported functions {{mvar|f}},{{sfn|Stein|Weiss|1971|loc=Theorem 1.18}} and so {{mvar|η<sub>ε</sub>}} converges weakly to {{mvar|δ}} in the sense of measures.
Then a simple change of variables shows that {{mvar|η<sub>ε</sub>}} also has integral {{math|1}}. One may show that ({{EquationNote|5}}) holds for all continuous compactly supported functions {{mvar|f}},{{sfn|Stein|Weiss|1971|loc=Theorem 1.18}} and so {{mvar|η<sub>ε</sub>}} converges weakly to {{mvar|δ}} in the sense of measures.


The {{mvar|η<sub>ε</sub>}} constructed in this way are known as an '''approximation to the identity'''.{{sfn|Rudin|1991|loc=§II.6.31}} This terminology is because the space {{math|''L''<sup>1</sup>('''R''')}} of absolutely integrable functions is closed under the operation of [[convolution]] of functions: {{math|''f'' ∗ ''g'' ∈ ''L''<sup>1</sup>('''R''')}} whenever {{mvar|f}} and {{mvar|g}} are in {{math|''L''<sup>1</sup>('''R''')}}. However, there is no identity in {{math|''L''<sup>1</sup>('''R''')}} for the convolution product: no element {{mvar|h}} such that {{math|1=''f'' ∗ ''h'' = ''f''}} for all {{mvar|f}}. Nevertheless, the sequence {{mvar|η<sub>ε</sub>}} does approximate such an identity in the sense that
The {{mvar|η<sub>ε</sub>}} constructed in this way are known as an '''approximation to the identity'''.{{sfn|Rudin|1991|loc=§II.6.31}} This terminology is because the space {{math|''L''<sup>1</sup>('''R''')}} of absolutely integrable functions is closed under the operation of [[convolution]] of functions: {{math|''f'' ∗ ''g'' ∈ ''L''<sup>1</sup>('''R''')}} whenever {{mvar|f}} and {{mvar|g}} are in {{math|''L''<sup>1</sup>('''R''')}}. However, there is no identity in {{math|''L''<sup>1</sup>('''R''')}} for the convolution product: no element {{mvar|h}} such that {{math|1=''f'' ∗ ''h'' = ''f''}} for all {{mvar|f}}. Nevertheless, the sequence {{mvar|η<sub>ε</sub>}} does approximate such an identity in the sense that


<math display="block">f*\eta_\varepsilon \to f \quad \text{as }\varepsilon\to 0.</math>
<math display="block">f*\eta_\varepsilon \to f \quad \text{as }\varepsilon\to 0.</math>


This limit holds in the sense of [[mean convergence]] (convergence in {{math|''L''<sup>1</sup>}}). Further conditions on the {{mvar|η<sub>ε</sub>}}, for instance that it be a mollifier associated to a compactly supported function,<ref>More generally, one only needs {{math|1=''η'' = ''η''<sub>1</sub>}} to have an integrable radially symmetric decreasing rearrangement.</ref> are needed to ensure pointwise convergence [[almost everywhere]].
This limit holds in the sense of [[mean convergence]] (convergence in {{math|''L''<sup>1</sup>}}). Further conditions on the {{mvar|η<sub>ε</sub>}}, for instance that it be a mollifier associated to a compactly supported function,<ref>More generally, one only needs {{math|1=''η'' = ''η''<sub>1</sub>}} to have an integrable radially symmetric decreasing rearrangement.</ref> are needed to ensure pointwise convergence [[almost everywhere]].


If the initial {{math|1=''η'' = ''η''<sub>1</sub>}} is itself smooth and compactly supported then the sequence is called a [[mollifier]]. The standard mollifier is obtained by choosing {{mvar|η}} to be a suitably normalized [[bump function]], for instance
If the initial {{math|1=''η'' = ''η''<sub>1</sub>}} is itself smooth and compactly supported then the sequence is called a [[mollifier]]. The standard mollifier is obtained by choosing {{mvar|η}} to be a suitably normalized [[bump function]], for instance


<math display="block">\eta(x) = \begin{cases}
<math display="block">\eta(x) = \begin{cases}
Line 402: Line 402:
(<math>I_n</math> ensuring that the total integral is 1).
(<math>I_n</math> ensuring that the total integral is 1).


In some situations such as [[numerical analysis]], a [[piecewise linear function|piecewise linear]] approximation to the identity is desirable. This can be obtained by taking {{math|''η''<sub>1</sub>}} to be a [[hat function]]. With this choice of {{math|''η''<sub>1</sub>}}, one has
In some situations such as [[numerical analysis]], a [[piecewise linear function|piecewise linear]] approximation to the identity is desirable. This can be obtained by taking {{math|''η''<sub>1</sub>}} to be a [[hat function]]. With this choice of {{math|''η''<sub>1</sub>}}, one has


<math display="block"> \eta_\varepsilon(x) = \varepsilon^{-1}\max \left (1-\left|\frac{x}{\varepsilon}\right|,0 \right) </math>
<math display="block"> \eta_\varepsilon(x) = \varepsilon^{-1}\max \left (1-\left|\frac{x}{\varepsilon}\right|,0 \right) </math>
Line 408: Line 408:
which are all continuous and compactly supported, although not smooth and so not a mollifier.
which are all continuous and compactly supported, although not smooth and so not a mollifier.


====Probabilistic considerations====
=== Probabilistic considerations ===
In the context of [[probability theory]], it is natural to impose the additional condition that the initial {{math|''η''<sub>1</sub>}} in an approximation to the identity should be positive, as such a function then represents a [[probability distribution]]. Convolution with a probability distribution is sometimes favorable because it does not result in [[overshoot (signal)|overshoot]] or undershoot, as the output is a [[convex combination]] of the input values, and thus falls between the maximum and minimum of the input function. Taking {{math|''η''<sub>1</sub>}} to be any probability distribution at all, and letting {{math|1=''η<sub>ε</sub>''(''x'') = ''η''<sub>1</sub>(''x''/''ε'')/''ε''}} as above will give rise to an approximation to the identity. In general this converges more rapidly to a delta function if, in addition, {{mvar|η}} has mean {{math|0}} and has small higher moments. For instance, if {{math|''η''<sub>1</sub>}} is the [[uniform distribution (continuous)|uniform distribution]] on {{nowrap|1=<math display="inline">\left[-\frac{1}{2},\frac{1}{2}\right]</math>,}} also known as the [[rectangular function]], then:{{sfn|Saichev|Woyczyński|1997|loc=§1.1 The "delta function" as viewed by a physicist and an engineer, p. 3}}
In the context of [[probability theory]], it is natural to impose the additional condition that the initial {{math|''η''<sub>1</sub>}} in an approximation to the identity should be positive, as such a function then represents a [[probability distribution]]. Convolution with a probability distribution is sometimes favorable because it does not result in [[overshoot (signal)|overshoot]] or undershoot, as the output is a [[convex combination]] of the input values, and thus falls between the maximum and minimum of the input function. Taking {{math|''η''<sub>1</sub>}} to be any probability distribution at all, and letting {{math|1=''η<sub>ε</sub>''(''x'') = ''η''<sub>1</sub>(''x''/''ε'')/''ε''}} as above will give rise to an approximation to the identity. In general this converges more rapidly to a delta function if, in addition, {{mvar|η}} has mean {{math|0}} and has small higher moments. For instance, if {{math|''η''<sub>1</sub>}} is the [[uniform distribution (continuous)|uniform distribution]] on {{nowrap|1=<math display="inline">\left[-\frac{1}{2},\frac{1}{2}\right]</math>,}} also known as the [[rectangular function]], then:{{sfn|Saichev|Woyczyński|1997|loc=§1.1 The "delta function" as viewed by a physicist and an engineer, p. 3}}
<math display="block">
<math display="block">
\eta_\varepsilon(x) = \frac{1}{\varepsilon}\operatorname{rect}\left(\frac{x}{\varepsilon}\right)=
\eta_\varepsilon(x) = \frac{1}{\varepsilon}\operatorname{rect}\left(\frac{x}{\varepsilon}\right)=
Line 425: Line 425:
This is continuous and compactly supported, but not a mollifier because it is not smooth.
This is continuous and compactly supported, but not a mollifier because it is not smooth.


====Semigroups====
=== Semigroups ===
Approximations to the delta functions often arise as convolution [[semigroup]]s.<ref>{{Cite book|last1=Milovanović|first1=Gradimir V.|url={{google books |plainurl=y |id=4U-5BQAAQBAJ}}|title=Analytic Number Theory, Approximation Theory, and Special Functions: In Honor of Hari M. Srivastava|last2=Rassias|first2=Michael Th|date=2014-07-08|publisher=Springer|isbn=978-1-4939-0258-3|language=en|page=[{{google books |plainurl=y |id=4U-5BQAAQBAJ|page=748 }} 748]}}</ref> This amounts to the further constraint that the convolution of {{mvar|η<sub>ε</sub>}} with {{mvar|η<sub>δ</sub>}} must satisfy
Approximations to the delta functions often arise as convolution [[semigroup]]s.<ref>{{Cite book|last1=Milovanović|first1=Gradimir V.|url={{google books |plainurl=y |id=4U-5BQAAQBAJ}}|title=Analytic Number Theory, Approximation Theory, and Special Functions: In Honor of Hari M. Srivastava|last2=Rassias|first2=Michael Th|date=2014-07-08|publisher=Springer|isbn=978-1-4939-0258-3|language=en|page=[{{google books |plainurl=y |id=4U-5BQAAQBAJ|page=748 }} 748]}}</ref> This amounts to the further constraint that the convolution of {{mvar|η<sub>ε</sub>}} with {{mvar|η<sub>δ</sub>}} must satisfy
<math display="block">\eta_\varepsilon * \eta_\delta = \eta_{\varepsilon+\delta}</math>
<math display="block">\eta_\varepsilon * \eta_\delta = \eta_{\varepsilon+\delta}</math>
Line 431: Line 431:
for all {{math|1=''ε'', ''δ'' > 0}}. Convolution semigroups in {{math|''L''<sup>1</sup>}} that approximate the delta function are always an approximation to the identity in the above sense, however the semigroup condition is quite a strong restriction.
for all {{math|1=''ε'', ''δ'' > 0}}. Convolution semigroups in {{math|''L''<sup>1</sup>}} that approximate the delta function are always an approximation to the identity in the above sense, however the semigroup condition is quite a strong restriction.


In practice, semigroups approximating the delta function arise as [[fundamental solution]]s or [[Green's function]]s to physically motivated [[elliptic partial differential equation|elliptic]] or [[parabolic partial differential equation|parabolic]] [[partial differential equations]]. In the context of [[applied mathematics]], semigroups arise as the output of a [[linear time-invariant system]]. Abstractly, if ''A'' is a linear operator acting on functions of ''x'', then a convolution semigroup arises by solving the [[initial value problem]]
In practice, semigroups approximating the delta function arise as [[fundamental solution]]s or [[Green's function]]s to physically motivated [[elliptic partial differential equation|elliptic]] or [[parabolic partial differential equation|parabolic]] [[partial differential equations]]. In the context of [[applied mathematics]], semigroups arise as the output of a [[linear time-invariant system]]. Abstractly, if ''A'' is a linear operator acting on functions of ''x'', then a convolution semigroup arises by solving the [[initial value problem]]


<math display="block">\begin{cases}
<math display="block">\begin{cases}
Line 438: Line 438:
\end{cases}</math>
\end{cases}</math>


in which the limit is as usual understood in the weak sense. Setting {{math|1=''η<sub>ε</sub>''(''x'') = ''η''(''ε'', ''x'')}} gives the associated approximate delta function.
in which the limit is as usual understood in the weak sense. Setting {{math|1=''η<sub>ε</sub>''(''x'') = ''η''(''ε'', ''x'')}} gives the associated approximate delta function.


Some examples of physically important convolution semigroups arising from such a fundamental solution include the following.
Some examples of physically important convolution semigroups arising from such a fundamental solution include the following.


=====The heat kernel=====
==== The heat kernel ====
The [[heat kernel]], defined by
The [[heat kernel]], defined by{{sfn|Stein|Shakarchi|2005|p=111}}


<math display="block">\eta_\varepsilon(x) = \frac{1}{\sqrt{2\pi\varepsilon}} \mathrm{e}^{-\frac{x^2}{2\varepsilon}}</math>
<math display="block">\eta_\varepsilon(x) = \frac{1}{\sqrt{2\pi\varepsilon}} \mathrm{e}^{-\frac{x^2}{2\varepsilon}}</math>
Line 457: Line 457:
and has the same physical interpretation, {{lang|la|[[mutatis mutandis]]}}. It also represents an approximation to the delta function in the sense that {{math|''η<sub>ε</sub>'' → ''δ''}} in the distribution sense as {{math|''ε'' → 0}}.
and has the same physical interpretation, {{lang|la|[[mutatis mutandis]]}}. It also represents an approximation to the delta function in the sense that {{math|''η<sub>ε</sub>'' → ''δ''}} in the distribution sense as {{math|''ε'' → 0}}.


=====The Poisson kernel=====
==== The Poisson kernel ====
The [[Poisson kernel]]
The [[Poisson kernel]]
<math display="block">\eta_\varepsilon(x) = \frac{1}{\pi}\mathrm{Im}\left\{\frac{1}{x-\mathrm{i}\varepsilon}\right\}=\frac{1}{\pi} \frac{\varepsilon}{\varepsilon^2 + x^2}=\frac{1}{2\pi}\int_{-\infty}^{\infty}\mathrm{e}^{\mathrm{i} \xi x-|\varepsilon \xi|}\,d\xi</math>
<math display="block">\eta_\varepsilon(x) = \frac{1}{\pi}\mathrm{Im}\left\{\frac{1}{x-\mathrm{i}\varepsilon}\right\}=\frac{1}{\pi} \frac{\varepsilon}{\varepsilon^2 + x^2}=\frac{1}{2\pi}\int_{-\infty}^{\infty}\mathrm{e}^{\mathrm{i} \xi x-|\varepsilon \xi|}\,d\xi</math>


is the fundamental solution of the [[Laplace equation]] in the upper half-plane.{{sfn|Stein|Weiss|1971|loc=§I.1}} It represents the [[electrostatic potential]] in a semi-infinite plate whose potential along the edge is held at fixed at the delta function. The Poisson kernel is also closely related to the [[Cauchy distribution]] and [[Kernel (statistics)#Kernel functions in common use|Epanechnikov and Gaussian kernel]] functions.<ref>{{Cite book|last=Mader|first=Heidy M.|url={{google books |plainurl=y |id=e5Y_RRPxdyYC}}|title=Statistics in Volcanology|date=2006|publisher=Geological Society of London|isbn=978-1-86239-208-3|language=en|editor-link=Heidy Mader|page=[{{google books |plainurl=y |id=e5Y_RRPxdyYC|page=81}} 81]}}</ref> This semigroup evolves according to the equation
is the fundamental solution of the [[Laplace equation]] in the upper half-plane.{{sfn|Stein|Weiss|1971|loc=§I.1}} It represents the [[electrostatic potential]] in a semi-infinite plate whose potential along the edge is held at fixed at the delta function. The Poisson kernel is also closely related to the [[Cauchy distribution]] and [[Kernel (statistics)#Kernel functions in common use|Epanechnikov and Gaussian kernel]] functions.<ref>{{Cite book|last=Mader|first=Heidy M.|url={{google books |plainurl=y |id=e5Y_RRPxdyYC}}|title=Statistics in Volcanology|date=2006|publisher=Geological Society of London|isbn=978-1-86239-208-3|language=en|editor-link=Heidy Mader|page=[{{google books |plainurl=y |id=e5Y_RRPxdyYC|page=81}} 81]}}</ref> This semigroup evolves according to the equation
<math display="block">\frac{\partial u}{\partial t} = -\left (-\frac{\partial^2}{\partial x^2} \right)^{\frac{1}{2}}u(t,x)</math>
<math display="block">\frac{\partial u}{\partial t} = -\left (-\frac{\partial^2}{\partial x^2} \right)^{\frac{1}{2}}u(t,x)</math>


Line 467: Line 467:
<math display="block">\mathcal{F}\left[\left(-\frac{\partial^2}{\partial x^2} \right)^{\frac{1}{2}}f\right](\xi) = |2\pi\xi|\mathcal{F}f(\xi).</math>
<math display="block">\mathcal{F}\left[\left(-\frac{\partial^2}{\partial x^2} \right)^{\frac{1}{2}}f\right](\xi) = |2\pi\xi|\mathcal{F}f(\xi).</math>


====Oscillatory integrals====
=== Oscillatory integrals ===
In areas of physics such as [[wave propagation]] and [[wave|wave mechanics]], the equations involved are [[hyperbolic partial differential equations|hyperbolic]] and so may have more singular solutions. As a result, the approximate delta functions that arise as fundamental solutions of the associated [[Cauchy problem]]s are generally [[oscillatory integral]]s. An example, which comes from a solution of the [[Euler–Tricomi equation]] of [[transonic]] [[gas dynamics]],{{sfn|Vallée|Soares|2004|loc=§7.2}} is the rescaled [[Airy function]]
In areas of physics such as [[wave propagation]] and [[wave|wave mechanics]], the equations involved are [[hyperbolic partial differential equations|hyperbolic]] and so may have more singular solutions. As a result, the approximate delta functions that arise as fundamental solutions of the associated [[Cauchy problem]]s are generally [[oscillatory integral]]s. An example, which comes from a solution of the [[Euler–Tricomi equation]] of [[transonic]] [[gas dynamics]],{{sfn|Vallée|Soares|2004|loc=§7.2}} is the rescaled [[Airy function]]
<math display="block">\varepsilon^{-1/3}\operatorname{Ai}\left (x\varepsilon^{-1/3} \right). </math>
<math display="block">\varepsilon^{-1/3}\operatorname{Ai}\left (x\varepsilon^{-1/3} \right). </math>


Although using the Fourier transform, it is easy to see that this generates a semigroup in some sense—it is not absolutely integrable and so cannot define a semigroup in the above strong sense. Many approximate delta functions constructed as oscillatory integrals only converge in the sense of distributions (an example is the [[Dirichlet kernel]] below), rather than in the sense of measures.
Although using the Fourier transform, it is easy to see that this generates a semigroup in some sense—it is not absolutely integrable and so cannot define a semigroup in the above strong sense. Many approximate delta functions constructed as oscillatory integrals only converge in the sense of distributions (an example is the [[Dirichlet kernel]] below), rather than in the sense of measures.


Another example is the Cauchy problem for the [[wave equation]] in {{math|'''R'''<sup>1+1</sup>}}:{{sfn|Hörmander|1983|loc=§7.8}}
Another example is the Cauchy problem for the [[wave equation]] in {{math|'''R'''<sup>1+1</sup>}}:{{sfn|Hörmander|1983|loc=§7.8}}
Line 494: Line 494:
<math display="block">L[u]=\delta.</math>
<math display="block">L[u]=\delta.</math>


When {{mvar|L}} is particularly simple, this problem can often be resolved using the Fourier transform directly (as in the case of the Poisson kernel and heat kernel already mentioned). For more complicated operators, it is sometimes easier first to consider an equation of the form
When {{mvar|L}} is particularly simple, this problem can often be resolved using the Fourier transform directly (as in the case of the Poisson kernel and heat kernel already mentioned). For more complicated operators, it is sometimes easier first to consider an equation of the form
<math display="block">L[u]=h</math>
<math display="block">L[u]=h</math>


Line 500: Line 500:
<math display="block">h = h(x\cdot\xi)</math>
<math display="block">h = h(x\cdot\xi)</math>


for some vector {{mvar|ξ}}. Such an equation can be resolved (if the coefficients of {{mvar|L}} are [[analytic function]]s) by the [[Cauchy–Kovalevskaya theorem]] or (if the coefficients of {{mvar|L}} are constant) by quadrature. So, if the delta function can be decomposed into plane waves, then one can in principle solve linear partial differential equations.
for some vector {{mvar|ξ}}. Such an equation can be resolved (if the coefficients of {{mvar|L}} are [[analytic function]]s) by the [[Cauchy–Kovalevskaya theorem]] or (if the coefficients of {{mvar|L}} are constant) by quadrature. So, if the delta function can be decomposed into plane waves, then one can in principle solve linear partial differential equations.


Such a decomposition of the delta function into plane waves was part of a general technique first introduced essentially by [[Johann Radon]], and then developed in this form by [[Fritz John]] ([[#CITEREFJohn1955|1955]]).{{sfn|Courant|Hilbert|1962|loc=§14}} Choose {{mvar|k}} so that {{math|''n'' + ''k''}} is an even integer, and for a real number {{mvar|s}}, put
Such a decomposition of the delta function into plane waves was part of a general technique first introduced essentially by [[Johann Radon]], and then developed in this form by [[Fritz John]] ([[#CITEREFJohn1955|1955]]).{{sfn|Courant|Hilbert|1962|loc=§14}} Choose {{mvar|k}} so that {{math|''n'' + ''k''}} is an even integer, and for a real number {{mvar|s}}, put
Line 515: Line 515:
<math display="block">\varphi(x) = \int_{\mathbf{R}^n}\varphi(y)\,dy\,\Delta_x^{\frac{n+k}{2}} \int_{S^{n-1}} g((x-y)\cdot\xi)\,d\omega_\xi.</math>
<math display="block">\varphi(x) = \int_{\mathbf{R}^n}\varphi(y)\,dy\,\Delta_x^{\frac{n+k}{2}} \int_{S^{n-1}} g((x-y)\cdot\xi)\,d\omega_\xi.</math>


The result follows from the formula for the [[Newtonian potential]] (the fundamental solution of Poisson's equation). This is essentially a form of the inversion formula for the [[Radon transform]] because it recovers the value of {{math|''φ''(''x'')}} from its integrals over hyperplanes. For instance, if {{mvar|n}} is odd and {{math|1=''k'' = 1}}, then the integral on the right hand side is
The result follows from the formula for the [[Newtonian potential]] (the fundamental solution of Poisson's equation). This is essentially a form of the inversion formula for the [[Radon transform]] because it recovers the value of {{math|''φ''(''x'')}} from its integrals over hyperplanes.{{sfn|John|1955}} For instance, if {{mvar|n}} is odd and {{math|1=''k'' = 1}}, then the integral on the right hand side is
<math display="block">
<math display="block">
\begin{align}
\begin{align}
Line 533: Line 533:


===Fourier transform===
===Fourier transform===
The delta function is a [[Distribution (mathematics)#Tempered distributions and Fourier transform|tempered distribution]], and therefore it has a well-defined [[Fourier transform]]. Formally, one finds<ref>The numerical factors depend on the [[Fourier transform#Other conventions|conventions]] for the Fourier transform.</ref>
The delta function is a [[Distribution (mathematics)#Tempered distributions and Fourier transform|tempered distribution]], and therefore it has a well-defined [[Fourier transform]]. Formally, one finds<ref>The numerical factors depend on the [[Fourier transform#Other conventions|conventions]] for the Fourier transform.</ref>


<math display="block">\widehat{\delta}(\xi)=\int_{-\infty}^\infty e^{-2\pi i x \xi} \,\delta(x)dx = 1.</math>
<math display="block">\widehat{\delta}(\xi)=\int_{-\infty}^\infty e^{-2\pi i x \xi} \,\delta(x)dx = 1.</math>


Properly speaking, the Fourier transform of a distribution is defined by imposing [[self-adjoint]]ness of the Fourier transform under the [[Dual_system|duality pairing]] <math>\langle\cdot,\cdot\rangle</math> of tempered distributions with [[Schwartz functions]]. Thus <math>\widehat{\delta}</math> is defined as the unique tempered distribution satisfying
Properly speaking, the Fourier transform of a distribution is defined by imposing [[self-adjoint]]ness of the Fourier transform under the [[Dual_system|duality pairing]] <math>\langle\cdot,\cdot\rangle</math> of tempered distributions with [[Schwartz functions]]. Thus <math>\widehat{\delta}</math> is defined as the unique tempered distribution satisfying


<math display="block">\langle\widehat{\delta},\varphi\rangle = \langle\delta,\widehat{\varphi}\rangle</math>
<math display="block">\langle\widehat{\delta},\varphi\rangle = \langle\delta,\widehat{\varphi}\rangle</math>
Line 547: Line 547:
<math display="block">S*\delta = S.</math>
<math display="block">S*\delta = S.</math>


That is to say that {{mvar|δ}} is an [[identity element]] for the convolution on tempered distributions, and in fact, the space of compactly supported distributions under convolution is an [[associative algebra]] with identity the delta function. This property is fundamental in [[signal processing]], as convolution with a tempered distribution is a [[linear time-invariant system]], and applying the linear time-invariant system measures its [[impulse response]]. The impulse response can be computed to any desired degree of accuracy by choosing a suitable approximation for {{mvar|δ}}, and once it is known, it characterizes the system completely. See {{section link | LTI system theory |Impulse response and convolution}}.
That is to say that {{mvar|δ}} is an [[identity element]] for the convolution on tempered distributions, and in fact, the space of compactly supported distributions under convolution is an [[associative algebra]] with identity the delta function. This property is fundamental in [[signal processing]], as convolution with a tempered distribution is a [[linear time-invariant system]], and applying the linear time-invariant system measures its [[impulse response]]. The impulse response can be computed to any desired degree of accuracy by choosing a suitable approximation for {{mvar|δ}}, and once it is known, it characterizes the system completely. See {{section link | LTI system theory |Impulse response and convolution}}.


The inverse Fourier transform of the tempered distribution {{math|1=''f''(''ξ'') = 1}} is the delta function. Formally, this is expressed as
The inverse Fourier transform of the tempered distribution {{math|1=''f''(''ξ'') = 1}} is the delta function. Formally, this is expressed as
Line 555: Line 555:
for all Schwartz functions {{mvar|''f''}}.
for all Schwartz functions {{mvar|''f''}}.


In these terms, the delta function provides a suggestive statement of the orthogonality property of the Fourier kernel on {{math|'''R'''}}. Formally, one has
In these terms, the delta function provides a suggestive statement of the orthogonality property of the Fourier kernel on {{math|'''R'''}}. Formally, one has
<math display="block">\int_{-\infty}^\infty e^{i 2\pi \xi_1 t} \left[e^{i 2\pi \xi_2 t}\right]^*\,dt = \int_{-\infty}^\infty e^{-i 2\pi (\xi_2 - \xi_1) t} \,dt = \delta(\xi_2 - \xi_1).</math>
<math display="block">\int_{-\infty}^\infty e^{i 2\pi \xi_1 t} \left[e^{i 2\pi \xi_2 t}\right]^*\,dt = \int_{-\infty}^\infty e^{-i 2\pi (\xi_2 - \xi_1) t} \,dt = \delta(\xi_2 - \xi_1).</math>


Line 569: Line 569:
====Fourier kernels====
====Fourier kernels====
{{See also|Convergence of Fourier series}}
{{See also|Convergence of Fourier series}}
In the study of [[Fourier series]], a major question consists of determining whether and in what sense the Fourier series associated with a [[periodic function]] converges to the function. The {{mvar|n}}-th partial sum of the Fourier series of a function {{mvar|f}} of period {{math|2π}} is defined by convolution (on the interval {{closed-closed|−π,π}}) with the [[Dirichlet kernel]]:
In the study of [[Fourier series]], a major question consists of determining whether and in what sense the Fourier series associated with a [[periodic function]] converges to the function. The {{mvar|n}}-th partial sum of the Fourier series of a function {{mvar|f}} of period {{math|2π}} is defined by convolution (on the interval {{closed-closed|−π,π}}) with the [[Dirichlet kernel]]:
<math display="block">D_N(x) = \sum_{n=-N}^N e^{inx} = \frac{\sin\left(\left(N+\frac12\right)x\right)}{\sin(x/2)}.</math>
<math display="block">D_N(x) = \sum_{n=-N}^N e^{inx} = \frac{\sin\left(\left(N+\frac12\right)x\right)}{\sin(x/2)}.</math>
Thus,
Thus,
Line 575: Line 575:
where
where
<math display="block">a_n = \frac{1}{2\pi}\int_{-\pi}^\pi f(y)e^{-iny}\,dy.</math>
<math display="block">a_n = \frac{1}{2\pi}\int_{-\pi}^\pi f(y)e^{-iny}\,dy.</math>
A fundamental result of elementary Fourier series states that the Dirichlet kernel restricted to the interval&nbsp;{{closed-closed|−π,π}} tends to a multiple of the delta function as {{math|''N'' → ∞}}. This is interpreted in the distribution sense, that
A fundamental result of elementary Fourier series states that the Dirichlet kernel restricted to the interval&nbsp;{{closed-closed|−π,π}} tends to a multiple of the delta function as {{math|''N'' → ∞}}. This is interpreted in the distribution sense, that
<math display="block">s_N(f)(0) = \int_{-\pi}^{\pi} D_N(x)f(x)\,dx \to 2\pi f(0)</math>
<math display="block">s_N(f)(0) = \int_{-\pi}^{\pi} D_N(x)f(x)\,dx \to 2\pi f(0)</math>
for every compactly supported {{em|smooth}} function {{mvar|f}}. Thus, formally one has
for every compactly supported {{em|smooth}} function {{mvar|f}}. Thus, formally one has
<math display="block">\delta(x) = \frac1{2\pi} \sum_{n=-\infty}^\infty e^{inx}</math>
<math display="block">\delta(x) = \frac1{2\pi} \sum_{n=-\infty}^\infty e^{inx}</math>
on the interval {{closed-closed|−π,π}}.
on the interval {{closed-closed|−π,π}}.


Despite this, the result does not hold for all compactly supported {{em|continuous}} functions: that is {{math|''D<sub>N</sub>''}} does not converge weakly in the sense of measures. The lack of convergence of the Fourier series has led to the introduction of a variety of [[summability methods]] to produce convergence. The method of [[Cesàro summation]] leads to the [[Fejér kernel]]{{sfn|Lang|1997|p=312}}
Despite this, the result does not hold for all compactly supported {{em|continuous}} functions: that is {{math|''D<sub>N</sub>''}} does not converge weakly in the sense of measures. The lack of convergence of the Fourier series has led to the introduction of a variety of [[summability methods]] to produce convergence. The method of [[Cesàro summation]] leads to the [[Fejér kernel]]{{sfn|Lang|1997|p=312}}


<math display="block">F_N(x) = \frac1N\sum_{n=0}^{N-1} D_n(x) = \frac{1}{N}\left(\frac{\sin \frac{Nx}{2}}{\sin \frac{x}{2}}\right)^2.</math>
<math display="block">F_N(x) = \frac1N\sum_{n=0}^{N-1} D_n(x) = \frac{1}{N}\left(\frac{\sin \frac{Nx}{2}}{\sin \frac{x}{2}}\right)^2.</math>
Line 589: Line 589:
<math display="block">\int_{-\pi}^{\pi} F_N(x)f(x)\,dx \to 2\pi f(0)</math>
<math display="block">\int_{-\pi}^{\pi} F_N(x)f(x)\,dx \to 2\pi f(0)</math>


for every compactly supported {{em|continuous}} function {{mvar|f}}. The implication is that the Fourier series of any continuous function is Cesàro summable to the value of the function at every point.
for every compactly supported {{em|continuous}} function {{mvar|f}}. The implication is that the Fourier series of any continuous function is Cesàro summable to the value of the function at every point.


===Hilbert space theory===
===Hilbert space theory===
The Dirac delta distribution is a [[densely defined]] [[unbounded operator|unbounded]] [[linear functional]] on the [[Hilbert space]] [[Lp space|L<sup>2</sup>]] of [[square-integrable function]]s. Indeed, smooth compactly supported functions are [[dense set|dense]] in {{math|''L''<sup>2</sup>}}, and the action of the delta distribution on such functions is well-defined. In many applications, it is possible to identify subspaces of {{math|''L''<sup>2</sup>}} and to give a stronger [[topology]] on which the delta function defines a [[bounded linear functional]].
The Dirac delta distribution is a [[densely defined]] [[unbounded operator|unbounded]] [[linear functional]] on the [[Hilbert space]] [[Lp space|L<sup>2</sup>]] of [[square-integrable function]]s.{{sfn|Reed|Simon|1980|loc=Ch. II–III, VIII}} Indeed, smooth compactly supported functions are [[dense set|dense]] in {{math|''L''<sup>2</sup>}}, and the action of the delta distribution on such functions is well-defined. In many applications, it is possible to identify subspaces of {{math|''L''<sup>2</sup>}} and to give a stronger [[topology]] on which the delta function defines a [[bounded linear functional]].


====Sobolev spaces====
====Sobolev spaces====
Line 603: Line 603:
<math display="block">\delta[f]=|f(0)| < C \|f\|_{H^1}.</math>
<math display="block">\delta[f]=|f(0)| < C \|f\|_{H^1}.</math>


Thus {{mvar|δ}} is a bounded linear functional on the Sobolev space {{math|''H''<sup>1</sup>}}. Equivalently {{mvar|δ}} is an element of the [[continuous dual space]] {{math|''H''<sup>−1</sup>}} of {{math|''H''<sup>1</sup>}}. More generally, in {{mvar|n}} dimensions, one has {{math|''δ'' ∈ ''H''<sup>−''s''</sup>('''R'''<sup>''n''</sup>)}} provided {{math|''s'' > {{sfrac|''n''|2}}}}.
Thus {{mvar|δ}} is a bounded linear functional on the Sobolev space {{math|''H''<sup>1</sup>}}.{{sfn|Adams|Fournier|2003|p=71}} Equivalently {{mvar|δ}} is an element of the [[continuous dual space]] {{math|''H''<sup>−1</sup>}} of {{math|''H''<sup>1</sup>}}. More generally, in {{mvar|n}} dimensions, one has {{math|''δ'' ∈ ''H''<sup>−''s''</sup>('''R'''<sup>''n''</sup>)}} provided {{math|''s'' > {{sfrac|''n''|2}}}}.


====Spaces of holomorphic functions====
====Spaces of holomorphic functions====
Line 610: Line 610:
<math display="block">f(z) = \frac{1}{2\pi i} \oint_{\partial D} \frac{f(\zeta)\,d\zeta}{\zeta-z},\quad z\in D</math>
<math display="block">f(z) = \frac{1}{2\pi i} \oint_{\partial D} \frac{f(\zeta)\,d\zeta}{\zeta-z},\quad z\in D</math>


for all [[holomorphic function]]s {{mvar|f}} in {{mvar|D}} that are continuous on the closure of {{mvar|D}}. As a result, the delta function {{math|''δ''<sub>''z''</sub>}} is represented in this class of holomorphic functions by the Cauchy integral:
for all [[holomorphic function]]s {{mvar|f}} in {{mvar|D}} that are continuous on the closure of {{mvar|D}}. As a result, the delta function {{math|''δ''<sub>''z''</sub>}} is represented in this class of holomorphic functions by the Cauchy integral:


<math display="block">\delta_z[f] = f(z) = \frac{1}{2\pi i} \oint_{\partial D} \frac{f(\zeta)\,d\zeta}{\zeta-z}.</math>
<math display="block">\delta_z[f] = f(z) = \frac{1}{2\pi i} \oint_{\partial D} \frac{f(\zeta)\,d\zeta}{\zeta-z}.</math>


Moreover, let {{math|''H''<sup>2</sup>(∂''D'')}} be the [[Hardy space]] consisting of the closure in {{math|''L''<sup>2</sup>(∂''D'')}} of all holomorphic functions in {{mvar|D}} continuous up to the boundary of {{mvar|D}}. Then functions in {{math|''H''<sup>2</sup>(∂''D'')}} uniquely extend to holomorphic functions in {{mvar|D}}, and the Cauchy integral formula continues to hold. In particular for {{math|''z'' ∈ ''D''}}, the delta function {{mvar|δ<sub>z</sub>}} is a continuous linear functional on {{math|''H''<sup>2</sup>(∂''D'')}}. This is a special case of the situation in [[several complex variables]] in which, for smooth domains {{mvar|D}}, the [[Szegő kernel]] plays the role of the Cauchy integral.{{sfn|Hazewinkel|1995|p=[{{google books |plainurl=y |id=PE1a-EIG22kC|page=357}} 357]}}
Moreover, let {{math|''H''<sup>2</sup>(∂''D'')}} be the [[Hardy space]] consisting of the closure in {{math|''L''<sup>2</sup>(∂''D'')}} of all holomorphic functions in {{mvar|D}} continuous up to the boundary of {{mvar|D}}. Then functions in {{math|''H''<sup>2</sup>(∂''D'')}} uniquely extend to holomorphic functions in {{mvar|D}}, and the Cauchy integral formula continues to hold. In particular for {{math|''z'' ∈ ''D''}}, the delta function {{mvar|δ<sub>z</sub>}} is a continuous linear functional on {{math|''H''<sup>2</sup>(∂''D'')}}. This is a special case of the situation in [[several complex variables]] in which, for smooth domains {{mvar|D}}, the [[Szegő kernel]] plays the role of the Cauchy integral.{{sfn|Hazewinkel|1995|p=[{{google books |plainurl=y |id=PE1a-EIG22kC|page=357}} 357]}}


Another representation of the delta function in a space of holomorphic functions is on the space <math>H(D)\cap L^2(D)</math> of square-integrable holomorphic functions in an open set <math>D\subset\mathbb C^n</math>. This is a closed subspace of <math>L^2(D)</math>, and therefore is a Hilbert space. On the other hand, the functional that evaluates a holomorphic function in <math>H(D)\cap L^2(D)</math> at a point <math>z</math> of <math>D</math> is a continuous functional, and so by the Riesz representation theorem, is represented by integration against a kernel <math>K_z(\zeta)</math>, the [[Bergman kernel]]. This kernel is the analog of the delta function in this Hilbert space. A Hilbert space having such a kernel is called a [[reproducing kernel Hilbert space]]. In the special case of the unit disc, one has
Another representation of the delta function in a space of holomorphic functions is on the space <math>H(D)\cap L^2(D)</math> of square-integrable holomorphic functions in an open set <math>D\subset\mathbb C^n</math>. This is a closed subspace of <math>L^2(D)</math>, and therefore is a Hilbert space. On the other hand, the functional that evaluates a holomorphic function in <math>H(D)\cap L^2(D)</math> at a point <math>z</math> of <math>D</math> is a continuous functional, and so by the Riesz representation theorem, is represented by integration against a kernel <math>K_z(\zeta)</math>, the [[Bergman kernel]].{{sfn|Zhu|2007|loc=Ch. 4}} This kernel is the analog of the delta function in this Hilbert space. A Hilbert space having such a kernel is called a [[reproducing kernel Hilbert space]]. In the special case of the unit disc, one has
<math display="block">\delta_w[f] = f(w) = \frac1\pi\iint_{|z|<1} \frac{f(z)\,dx\,dy}{(1-\bar zw)^2}.</math>
<math display="block">\delta_w[f] = f(w) = \frac1\pi\iint_{|z|<1} \frac{f(z)\,dx\,dy}{(1-\bar zw)^2}.</math>


Line 626: Line 626:
which may be represented by the notation:
which may be represented by the notation:
<math display="block">\alpha_n =  \varphi_n^\dagger f, </math>
<math display="block">\alpha_n =  \varphi_n^\dagger f, </math>
a form of the [[bra–ket notation]] of Dirac.<ref>
a form of the [[bra–ket notation]] of Dirac.{{sfn|Levin|2002|p=109}} Adopting this notation, the expansion of {{mvar|f}} takes the [[Dyadic tensor|dyadic]] form:{{sfn|Davis|Thomson|2000|p=343}}
 
The development of this section in bra–ket notation is found in {{harv|Levin|2002|loc= Coordinate-space wave functions and completeness, pp.=109''ff''}}</ref> Adopting this notation, the expansion of {{mvar|f}} takes the [[Dyadic tensor|dyadic]] form:{{sfn|Davis|Thomson|2000|loc=Perfect operators, p.344}}


<math display="block">f =  \sum_{n=1}^\infty \varphi_n \left ( \varphi_n^\dagger f \right). </math>
<math display="block">f =  \sum_{n=1}^\infty \varphi_n \left ( \varphi_n^\dagger f \right). </math>
Line 644: Line 642:
<math display="block">f(x) = \sum_{n=1}^\infty \int_D\, \left( \varphi_n (x) \varphi_n^*(\xi)\right) f(\xi) \, d \xi.</math>
<math display="block">f(x) = \sum_{n=1}^\infty \int_D\, \left( \varphi_n (x) \varphi_n^*(\xi)\right) f(\xi) \, d \xi.</math>


The right-hand side converges to {{mvar|f}} in the {{math|''L''<sup>2</sup>}} sense. It need not hold in a pointwise sense, even when {{mvar|f}} is a continuous function. Nevertheless, it is common to abuse notation and write
The right-hand side converges to {{mvar|f}} in the {{math|''L''<sup>2</sup>}} sense. It need not hold in a pointwise sense, even when {{mvar|f}} is a continuous function. Nevertheless, it is common to abuse notation and write


<math display="block">f(x) = \int \, \delta(x-\xi) f (\xi)\, d\xi, </math>
<math display="block">f(x) = \int \, \delta(x-\xi) f (\xi)\, d\xi, </math>


resulting in the representation of the delta function:{{sfn|Davis|Thomson|2000|loc=Equation 8.9.11, p. 344}}
resulting in the representation of the delta function:{{sfn|Davis|Thomson|2000|p=344}}


<math display="block">\delta(x-\xi) = \sum_{n=1}^\infty  \varphi_n (x) \varphi_n^*(\xi). </math>
<math display="block">\delta(x-\xi) = \sum_{n=1}^\infty  \varphi_n (x) \varphi_n^*(\xi). </math>


With a suitable [[rigged Hilbert space]] {{math|(Φ, ''L''<sup>2</sup>(''D''), Φ*)}} where {{math|Φ ⊂ ''L''<sup>2</sup>(''D'')}} contains all compactly supported smooth functions, this summation may converge in {{math|Φ*}}, depending on the properties of the basis {{math|''φ''<sub>''n''</sub>}}. In most cases of practical interest, the orthonormal basis comes from an integral or differential operator (e.g. the [[heat kernel]]), in which case the series converges in the [[Distribution (mathematics)#Distributions|distribution]] sense.{{sfn|de la Madrid|Bohm|Gadella|2002}}
With a suitable [[rigged Hilbert space]] {{math|(Φ, ''L''<sup>2</sup>(''D''), Φ*)}} where {{math|Φ ⊂ ''L''<sup>2</sup>(''D'')}} contains all compactly supported smooth functions, this summation may converge in {{math|Φ*}}, depending on the properties of the basis {{math|''φ''<sub>''n''</sub>}}. In most cases of practical interest, the orthonormal basis comes from an integral or differential operator (e.g. the [[heat kernel]]), in which case the series converges in the [[Distribution (mathematics)#Distributions|distribution]] sense.{{sfn|de la Madrid|Bohm|Gadella|2002}}


===Infinitesimal delta functions===
===Infinitesimal delta functions===
[[Cauchy]] used an infinitesimal {{mvar|α}} to write down a unit impulse, infinitely tall and narrow Dirac-type delta function {{mvar|δ<sub>α</sub>}} satisfying <math display="inline">\int F(x)\delta_\alpha(x) \,dx = F(0)</math> in a number of articles in 1827.{{sfn|Laugwitz|1989}} Cauchy defined an infinitesimal in ''[[Cours d'Analyse]]'' (1827) in terms of a sequence tending to zero. Namely, such a null sequence becomes an infinitesimal in Cauchy's and [[Lazare Carnot]]'s terminology.
[[Cauchy]] used an infinitesimal {{mvar|α}} to write down a unit impulse, infinitely tall and narrow Dirac-type delta function {{mvar|δ<sub>α</sub>}} satisfying <math display="inline">\int F(x)\delta_\alpha(x) \,dx = F(0)</math> in a number of articles in 1827.{{sfn|Laugwitz|1989}} Cauchy defined an infinitesimal in ''[[Cours d'Analyse]]'' (1827) in terms of a sequence tending to zero. Namely, such a null sequence becomes an infinitesimal in Cauchy's and [[Lazare Carnot]]'s terminology.


[[Non-standard analysis]] allows one to rigorously treat infinitesimals. The article by {{harvtxt|Yamashita|2007}} contains a bibliography on modern Dirac delta functions in the context of an infinitesimal-enriched continuum provided by the [[hyperreal number|hyperreals]]. Here the Dirac delta can be given by an actual function, having the property that for every real function {{mvar|F}} one has <math display="inline">\int F(x)\delta_\alpha(x) \, dx = F(0)</math> as anticipated by Fourier and Cauchy.
[[Non-standard analysis]] allows one to rigorously treat infinitesimals. The article by {{harvtxt|Yamashita|2007}} contains a bibliography on modern Dirac delta functions in the context of an infinitesimal-enriched continuum provided by the [[hyperreal number|hyperreals]]. Here the Dirac delta can be given by an actual function, having the property that for every real function {{mvar|F}} one has <math display="inline">\int F(x)\delta_\alpha(x) \, dx = F(0)</math> as anticipated by Fourier and Cauchy.


==Dirac comb==
==Dirac comb==
{{Main|Dirac comb}}
{{Main|Dirac comb}}
[[File:Dirac comb.svg|thumb|A Dirac comb is an infinite series of Dirac delta functions spaced at intervals of {{mvar|T}}]]
[[File:Dirac comb.svg|thumb|A Dirac comb is an infinite series of Dirac delta functions spaced at intervals of {{mvar|T}}]]
A so-called uniform "pulse train" of Dirac delta measures, which is known as a [[Dirac comb]], or as the [[Sha (Cyrillic)|Sha]] distribution, creates a [[sampling (signal processing)|sampling]] function, often used in [[digital signal processing]] (DSP) and discrete time signal analysis. The Dirac comb is given as the [[infinite sum]], whose limit is understood in the distribution sense,
A so-called uniform "pulse train" of Dirac delta measures, which is known as a [[Dirac comb]], or as the [[Sha (Cyrillic)|Sha]] distribution, creates a [[sampling (signal processing)|sampling]] function, often used in [[digital signal processing]] (DSP) and discrete time signal analysis. The Dirac comb is given as the [[infinite sum]], whose limit is understood in the distribution sense,


<math display="block">\operatorname{\text{Ш}}(x) = \sum_{n=-\infty}^\infty \delta(x-n),</math>
<math display="block">\operatorname{\text{Ш}}(x) = \sum_{n=-\infty}^\infty \delta(x-n),</math>
Line 668: Line 666:
which is a sequence of point masses at each of the integers.
which is a sequence of point masses at each of the integers.


Up to an overall normalizing constant, the Dirac comb is equal to its own Fourier transform. This is significant because if {{mvar|f}} is any [[Schwartz space|Schwartz function]], then the [[Wrapped distribution|periodization]] of {{mvar|f}} is given by the convolution
Up to an overall normalizing constant, the Dirac comb is equal to its own Fourier transform. This is significant because if {{mvar|f}} is any [[Schwartz space|Schwartz function]], then the [[Wrapped distribution|periodization]] of {{mvar|f}} is given by the convolution
<math display="block">(f * \operatorname{\text{Ш}})(x) = \sum_{n=-\infty}^\infty f(x-n).</math>
<math display="block">(f * \operatorname{\text{Ш}})(x) = \sum_{n=-\infty}^\infty f(x-n).</math>
In particular,
In particular,
Line 693: Line 691:
<math display="block">\delta_{ij} = \begin{cases} 1 & i=j\\ 0 &i\not=j \end{cases} </math>
<math display="block">\delta_{ij} = \begin{cases} 1 & i=j\\ 0 &i\not=j \end{cases} </math>


for all integers {{mvar|i}}, {{mvar|j}}. This function then satisfies the following analog of the sifting property: if {{mvar|a<sub>i</sub>}} (for {{mvar|i}} in the set of all integers) is any [[Infinite sequence#Doubly-infinite sequences|doubly infinite sequence]], then
for all integers {{mvar|i}}, {{mvar|j}}. This function then satisfies the following analog of the sifting property: if {{mvar|a<sub>i</sub>}} (for {{mvar|i}} in the set of all integers) is any [[Infinite sequence#Doubly-infinite sequences|doubly infinite sequence]], then


<math display="block">\sum_{i=-\infty}^\infty a_i \delta_{ik}=a_k.</math>
<math display="block">\sum_{i=-\infty}^\infty a_i \delta_{ik}=a_k.</math>
Line 706: Line 704:


===Probability theory===
===Probability theory===
In [[probability theory]] and [[statistics]], the Dirac delta function is often used to represent a [[discrete distribution]], or a partially discrete, partially [[continuous distribution]], using a [[probability density function]] (which is normally used to represent absolutely continuous distributions). For example, the probability density function {{math|''f''(''x'')}} of a discrete distribution consisting of points {{math|1='''x''' = {{brace|''x''<sub>1</sub>, ..., ''x<sub>n</sub>''}}}}, with corresponding probabilities {{math|''p''<sub>1</sub>, ..., ''p<sub>n</sub>''}}, can be written as
{{see also|Probability distribution#Dirac delta representation}}
In [[probability theory]] and [[statistics]], the Dirac delta function is often used to represent a [[discrete distribution]], or a partially discrete, partially [[continuous distribution]], using a [[probability density function]] (which is normally used to represent absolutely continuous distributions). For example, the probability density function {{math|''f''(''x'')}} of a discrete distribution consisting of points {{math|1='''x''' = {{brace|''x''<sub>1</sub>, ..., ''x<sub>n</sub>''}}}}, with corresponding probabilities {{math|''p''<sub>1</sub>, ..., ''p<sub>n</sub>''}}, can be written as<ref name=Kanwal>{{cite book | last=Kanwal | first=Ram P. | title=Generalized Functions Theory and Technique | publisher=Birkhäuser Boston | publication-place=Boston, MA | date=1997 | isbn=978-1-4684-0037-3 | doi=10.1007/978-1-4684-0035-9 | doi-access=free | chapter=15.1. Applications to Probability and Random Processes}}</ref>


<math display="block">f(x) = \sum_{i=1}^n p_i \delta(x-x_i).</math>
<math display="block">f(x) = \sum_{i=1}^n p_i \delta(x-x_i).</math>


As another example, consider a distribution in which 6/10 of the time returns a standard [[normal distribution]], and 4/10 of the time returns exactly the value 3.5 (i.e. a partly continuous, partly discrete [[mixture distribution]]). The density function of this distribution can be written as
As another example, consider a distribution in which 6/10 of the time returns a standard [[normal distribution]], and 4/10 of the time returns exactly the value 3.5 (i.e. a partly continuous, partly discrete [[mixture distribution]]). The density function of this distribution can be written as


<math display="block">f(x) = 0.6 \, \frac {1}{\sqrt{2\pi}} e^{-\frac{x^2}{2}} + 0.4 \, \delta(x-3.5).</math>
<math display="block">f(x) = 0.6 \, \frac {1}{\sqrt{2\pi}} e^{-\frac{x^2}{2}} + 0.4 \, \delta(x-3.5).</math>
Line 718: Line 717:
<math display="block">f_Y(y) = \int_{-\infty}^{+\infty} f_X(x) \delta(y-g(x)) \,dx.  </math>
<math display="block">f_Y(y) = \int_{-\infty}^{+\infty} f_X(x) \delta(y-g(x)) \,dx.  </math>


The delta function is also used in a completely different way to represent the [[local time (mathematics)|local time]] of a [[diffusion process]] (like [[Brownian motion]]). The local time of a stochastic process {{math|''B''(''t'')}} is given by
The delta function is also used in a completely different way to represent the [[local time (mathematics)|local time]] of a [[diffusion process]] (like [[Brownian motion]]).{{sfn|Karatzas|Shreve|1998|p=204}} The local time of a stochastic process {{math|''B''(''t'')}} is given by
<math display="block">\ell(x,t) = \int_0^t \delta(x-B(s))\,ds</math>
<math display="block">\ell(x,t) = \int_0^t \delta(x-B(s))\,ds</math>
and represents the amount of time that the process spends at the point {{mvar|x}} in the range of the process. More precisely, in one dimension this integral can be written
and represents the amount of time that the process spends at the point {{mvar|x}} in the range of the process. More precisely, in one dimension this integral can be written
<math display="block">\ell(x,t) = \lim_{\varepsilon\to 0^+}\frac{1}{2\varepsilon}\int_0^t \mathbf{1}_{[x-\varepsilon,x+\varepsilon]}(B(s))\,ds</math>
<math display="block">\ell(x,t) = \lim_{\varepsilon\to 0^+}\frac{1}{2\varepsilon}\int_0^t \mathbf{1}_{[x-\varepsilon,x+\varepsilon]}(B(s))\,ds</math>
where <math>\mathbf{1}_{[x-\varepsilon,x+\varepsilon]}</math> is the [[indicator function]] of the interval <math>[x-\varepsilon,x+\varepsilon].</math>
where <math>\mathbf{1}_{[x-\varepsilon,x+\varepsilon]}</math> is the [[indicator function]] of the interval <math>[x-\varepsilon,x+\varepsilon].</math>


===Quantum mechanics===
===Quantum mechanics===
The delta function is expedient in [[quantum mechanics]]. The [[wave function]] of a particle gives the [[probability amplitude]] of finding a particle within a given region of space. Wave functions are assumed to be elements of the Hilbert space {{math|''L''<sup>2</sup>}} of [[square-integrable function]]s, and the total probability of finding a particle within a given interval is the integral of the magnitude of the wave function squared over the interval. A set {{math|{{brace|{{ket|''φ<sub>n</sub>''}}}}}} of wave functions is orthonormal if
The delta function is expedient in [[quantum mechanics]]. The [[wave function]] of a particle gives the [[probability amplitude]] of finding a particle within a given region of space. Wave functions are assumed to be elements of the Hilbert space {{math|''L''<sup>2</sup>}} of [[square-integrable function]]s, and the total probability of finding a particle within a given interval is the integral of the magnitude of the wave function squared over the interval. A set {{math|{{brace|{{ket|''φ<sub>n</sub>''}}}}}} of wave functions is orthonormal if


<math display="block">\langle\varphi_n \mid \varphi_m\rangle = \delta_{nm},</math>
<math display="block">\langle\varphi_n \mid \varphi_m\rangle = \delta_{nm},</math>


where {{mvar|δ<sub>nm</sub>}} is the Kronecker delta. A set of orthonormal wave functions is complete in the space of square-integrable functions if any wave function {{math|{{ket|ψ}}}} can be expressed as a linear combination of the {{math|{{brace|{{ket|''φ<sub>n</sub>''}}}}}} with complex coefficients:
where {{mvar|δ<sub>nm</sub>}} is the Kronecker delta. A set of orthonormal wave functions is complete in the space of square-integrable functions if any wave function {{math|{{ket|ψ}}}} can be expressed as a linear combination of the {{math|{{brace|{{ket|''φ<sub>n</sub>''}}}}}} with complex coefficients:


<math display="block"> \psi = \sum c_n \varphi_n, </math>
<math display="block"> \psi = \sum c_n \varphi_n, </math>


where {{math|1=''c<sub>n</sub>'' = {{bra-ket|''φ<sub>n</sub>''|''ψ''}}}}. Complete orthonormal systems of wave functions appear naturally as the [[eigenfunction]]s of the [[Hamiltonian (quantum mechanics)|Hamiltonian]] (of a [[bound state|bound system]]) in quantum mechanics that measures the energy levels, which are called the eigenvalues. The set of eigenvalues, in this case, is known as the [[Spectrum (functional analysis)|spectrum]] of the Hamiltonian. In [[bra–ket notation]] this equality implies the [[Borel functional calculus#Resolution of the identity|resolution of the identity]]:
where {{math|1=''c<sub>n</sub>'' = {{bra-ket|''φ<sub>n</sub>''|''ψ''}}}}. Complete orthonormal systems of wave functions appear naturally as the [[eigenfunction]]s of the [[Hamiltonian (quantum mechanics)|Hamiltonian]] (of a [[bound state|bound system]]) in quantum mechanics that measures the energy levels, which are called the eigenvalues. The set of eigenvalues, in this case, is known as the [[Spectrum (functional analysis)|spectrum]] of the Hamiltonian. In [[bra–ket notation]] this equality implies the [[Borel functional calculus#Resolution of the identity|resolution of the identity]]:


<math display="block">I = \sum |\varphi_n\rangle\langle\varphi_n|.</math>
<math display="block">I = \sum |\varphi_n\rangle\langle\varphi_n|.</math>


Here the eigenvalues are assumed to be discrete, but the set of eigenvalues of an [[observable]] can also be continuous. An example is the [[position operator]], {{math|1=''Qψ''(''x'') = ''x''ψ(''x'')}}. The spectrum of the position (in one dimension) is the entire real line and is called a [[Spectrum (physical sciences)#In quantum mechanics|continuous spectrum]]. However, unlike the Hamiltonian, the position operator lacks proper eigenfunctions. The conventional way to overcome this shortcoming is to widen the class of available functions by allowing distributions as well, i.e., to replace the Hilbert space with a [[rigged Hilbert space]].{{sfn|Isham|1995|loc=§6.2}} In this context, the position operator has a complete set of ''generalized eigenfunctions'',{{sfn|Gelfand|Shilov|1966–1968|loc=Vol. 4, §I.4.1}} labeled by the points {{mvar|y}} of the real line, given by
Here the eigenvalues are assumed to be discrete, but the set of eigenvalues of an [[observable]] can also be continuous. An example is the [[position operator]], {{math|1=''Qψ''(''x'') = ''x''ψ(''x'')}}. The spectrum of the position (in one dimension) is the entire real line and is called a [[Spectrum (physical sciences)#In quantum mechanics|continuous spectrum]]. However, unlike the Hamiltonian, the position operator lacks proper eigenfunctions. The conventional way to overcome this shortcoming is to widen the class of available functions by allowing distributions as well, i.e., to replace the Hilbert space with a [[rigged Hilbert space]].{{sfn|Isham|1995|loc=§6.2}} In this context, the position operator has a complete set of ''generalized eigenfunctions'',{{sfn|Gelfand|Shilov|1966–1968|loc=Vol. 4, §I.4.1}} labeled by the points {{mvar|y}} of the real line, given by


<math display="block">\varphi_y(x) = \delta(x-y).</math>
<math display="block">\varphi_y(x) = \delta(x-y).</math>
Line 759: Line 758:
<math display="block">I = \int_\Omega |\varphi_y\rangle\, \langle\varphi_y|\,dy</math>
<math display="block">I = \int_\Omega |\varphi_y\rangle\, \langle\varphi_y|\,dy</math>


where the operator-valued integral is again understood in the weak sense. If the spectrum of {{mvar|P}} has both continuous and discrete parts, then the resolution of the identity involves a summation over the discrete spectrum and an integral over the continuous spectrum.
where the operator-valued integral is again understood in the weak sense. If the spectrum of {{mvar|P}} has both continuous and discrete parts, then the resolution of the identity involves a summation over the discrete spectrum and an integral over the continuous spectrum.


The delta function also has many more specialized applications in quantum mechanics, such as the [[delta potential]] models for a single and double potential well.
The delta function also has many more specialized applications in quantum mechanics, such as the [[delta potential]] models for a single and double potential well.


===Structural mechanics===
===Structural mechanics===
The delta function can be used in [[structural mechanics]] to describe transient loads or point loads acting on structures. The governing equation of a simple [[Harmonic oscillator|mass–spring system]] excited by a sudden force [[impulse (physics)|impulse]] {{mvar|I}} at time {{math|1=''t'' = 0}} can be written
The delta function can be used in [[structural mechanics]] to describe transient loads or point loads acting on structures. The governing equation of a simple [[Harmonic oscillator|mass–spring system]] excited by a sudden force [[impulse (physics)|impulse]] {{mvar|I}} at time {{math|1=''t'' = 0}} can be written{{sfn|Arfken|Weber|2005|pp=975-976}}{{sfn|Boyce|DiPrima|Meade|2017|pp=270-273}}
 
<math display="block">m \frac{d^2 \xi}{dt^2} + k \xi = I \delta(t),</math>
<math display="block">m \frac{d^2 \xi}{dt^2} + k \xi = I \delta(t),</math>
where {{mvar|m}} is the mass, {{mvar|ξ}} is the deflection, and {{mvar|k}} is the [[spring constant]].
where {{mvar|m}} is the mass, {{mvar|ξ}} is the deflection, and {{mvar|k}} is the [[spring constant]].


Line 802: Line 799:


==References==
==References==
{{Refbegin}}
* {{cite book
| last1 = Adams
| first1 = Robert A.
| last2 = Fournier
| first2 = John J. F.
| title = Sobolev Spaces
| edition = 2nd
| publisher = Academic Press
| year = 2003
| isbn = 978-0-12-044143-3
}}
* {{cite book | last=Arfken | first=George Brown | last2=Weber | first2=Hans-Jurgen | title=Mathematical Methods for Physicists | publisher=Academic Press | publication-place=Boston | date=2005 | isbn=0-12-088584-0}}
*{{citation |last1=Aratyn |first1=Henrik |last2=Rasinariu |first2 =Constantin| title=A short course in mathematical methods with Maple|url=https://books.google.com/books?id=JFmUQGd1I3IC&pg=PA314 |isbn=978-981-256-461-0 |publisher=World Scientific |year=2006}}.
*{{citation |last1=Aratyn |first1=Henrik |last2=Rasinariu |first2 =Constantin| title=A short course in mathematical methods with Maple|url=https://books.google.com/books?id=JFmUQGd1I3IC&pg=PA314 |isbn=978-981-256-461-0 |publisher=World Scientific |year=2006}}.
*{{Citation | last1=Arfken | first1=G. B. | author-link1=George B. Arfken | last2=Weber | first2=H. J. | title=Mathematical Methods for Physicists | publisher=[[Academic Press]] | location=Boston, Massachusetts | edition=5th | isbn=978-0-12-059825-0 | year=2000}}.
*{{Citation | last1=Arfken | first1=G. B. | author-link1=George B. Arfken | last2=Weber | first2=H. J. | title=Mathematical Methods for Physicists | publisher=[[Academic Press]] | location=Boston, Massachusetts | edition=5th | isbn=978-0-12-059825-0 | year=2000}}.
* {{citation | author=atis | year = 2013 | title = ATIS Telecom Glossary | url = http://www.atis.org/glossary/definition.aspx?id=743| archive-url = https://web.archive.org/web/20130313211636/http://www.atis.org/glossary/definition.aspx?id=743 | archive-date = 2013-03-13 }}
* {{citation | author=atis | year = 2013 | title = ATIS Telecom Glossary | url = http://www.atis.org/glossary/definition.aspx?id=743| archive-url = https://web.archive.org/web/20130313211636/http://www.atis.org/glossary/definition.aspx?id=743 | archive-date = 2013-03-13 }}
* {{citation | author=Billingsley | title=Probability and measure | year=1986 | edition=2nd}}
* {{cite book | last=Boyce | first=William E. | last2=DiPrima | first2=Richard C. | last3=Meade | first3=Douglas B. | title=Elementary differential equations and boundary value problems | publisher=Wiley | publication-place=Hoboken, NJ | date=2017 | isbn=978-1-119-37792-4}}
* {{citation | last=Bracewell |first= R. N.|author-link=Ronald N. Bracewell| title=The Fourier Transform and Its Applications| edition=2nd | publisher=McGraw-Hill | year=1986|bibcode= 1986ftia.book.....B}}.
* {{citation | last=Bracewell |first= R. N.|author-link=Ronald N. Bracewell| title=The Fourier Transform and Its Applications| edition=2nd | publisher=McGraw-Hill | year=1986|bibcode= 1986ftia.book.....B}}.
* {{citation | last=Bracewell |first= R. N.|author-link=Ronald N. Bracewell| title=The Fourier Transform and Its Applications| edition=3rd | publisher=McGraw-Hill | year=2000}}.
* {{citation | last=Bracewell |first= R. N.|author-link=Ronald N. Bracewell| title=The Fourier Transform and Its Applications| edition=3rd | publisher=McGraw-Hill | year=2000}}.
Line 818: Line 830:
* {{citation|last=Gannon|first=Terry|chapter=Vertex operator algebras|chapter-url=https://books.google.com/books?id=ZOfUsvemJDMC&pg=PA539|title=Princeton Companion to Mathematics|publisher=Princeton University Press|year=2008|isbn=978-1400830398}}.
* {{citation|last=Gannon|first=Terry|chapter=Vertex operator algebras|chapter-url=https://books.google.com/books?id=ZOfUsvemJDMC&pg=PA539|title=Princeton Companion to Mathematics|publisher=Princeton University Press|year=2008|isbn=978-1400830398}}.
*{{citation|first1=I. M.|last1=Gelfand|author-link1=Israel Gelfand|first2=G. E.|last2=Shilov|title=Generalized functions|url=https://books.google.com/books?id=nAnjBQAAQBAJ|volume=1–5|publisher=Academic Press|year=1966–1968|isbn=9781483262246}}.
*{{citation|first1=I. M.|last1=Gelfand|author-link1=Israel Gelfand|first2=G. E.|last2=Shilov|title=Generalized functions|url=https://books.google.com/books?id=nAnjBQAAQBAJ|volume=1–5|publisher=Academic Press|year=1966–1968|isbn=9781483262246}}.
* {{cite book | last1=Halperin | first1=Israel | last2=Schwartz | first2=Laurent | title=Introduction to the Theory of Distributions | publisher=University of Toronto Press | publication-place=Toronto [Ontario] | date=1952 | isbn=978-1-4875-9132-8}}.
*{{citation |last=Hartmann |first=William M. |title=Signals, sound, and sensation |publisher=Springer |year=1997 |isbn=978-1-56396-283-7 |url=https://books.google.com/books?id=3N72rIoTHiEC}}.
*{{citation |last=Hartmann |first=William M. |title=Signals, sound, and sensation |publisher=Springer |year=1997 |isbn=978-1-56396-283-7 |url=https://books.google.com/books?id=3N72rIoTHiEC}}.
* <!-- Both 1995 and 2011 are required unless someone can confirm correct page numbers for the refs. Page 357 of 1995 does not match 357 in 2011 -->{{cite book |last1=Hazewinkel |first1=Michiel |title=Encyclopaedia of Mathematics (set) |date=1995 |publisher=Springer Science & Business Media |isbn=978-1-55608-010-4 |url=https://books.google.com/books?id=PE1a-EIG22kC&pg=PA357 |language=en}}
* <!-- Both 1995 and 2011 are required unless someone can confirm correct page numbers for the refs. Page 357 of 1995 does not match 357 in 2011 -->{{cite book |last1=Hazewinkel |first1=Michiel |title=Encyclopaedia of Mathematics (set) |date=1995 |publisher=Springer Science & Business Media |isbn=978-1-55608-010-4 |url=https://books.google.com/books?id=PE1a-EIG22kC&pg=PA357 |language=en}}
Line 824: Line 837:
*{{citation|mr=0717035|first=L.|last= Hörmander|author-link=Lars Hörmander|title=The analysis of linear partial differential operators I|url=https://books.google.com/books?id=aaLrCAAAQBAJ|series= Grundl. Math. Wissenschaft. |volume= 256 |publisher= Springer  |year=1983|isbn=978-3-540-12104-6 |doi=10.1007/978-3-642-96750-4}}.
*{{citation|mr=0717035|first=L.|last= Hörmander|author-link=Lars Hörmander|title=The analysis of linear partial differential operators I|url=https://books.google.com/books?id=aaLrCAAAQBAJ|series= Grundl. Math. Wissenschaft. |volume= 256 |publisher= Springer  |year=1983|isbn=978-3-540-12104-6 |doi=10.1007/978-3-642-96750-4}}.
* {{citation|first=C. J.|last=Isham|title=Lectures on quantum theory: mathematical and structural foundations|url=https://books.google.com/books?id=dVs8PcZ0Hd8C|year=1995|publisher=Imperial College Press|bibcode=1995lqtm.book.....I |isbn= 978-81-7764-190-5}}.
* {{citation|first=C. J.|last=Isham|title=Lectures on quantum theory: mathematical and structural foundations|url=https://books.google.com/books?id=dVs8PcZ0Hd8C|year=1995|publisher=Imperial College Press|bibcode=1995lqtm.book.....I |isbn= 978-81-7764-190-5}}.
* {{citation
| last = Jeffrey | first = Alan
| year = 1993
| title = Linear Algebra and Ordinary Differential Equations
| page = 639
| publisher = CRC Press
}}
*{{Citation | last1=John | first1=Fritz |author-link=Fritz John| title=Plane waves and spherical means applied to partial differential equations |url=| publisher=Interscience Publishers, New York-London | mr=0075429 | year=1955 }}. [https://books.google.com/books?id=JP6ABuUnhecC Reprinted], Dover Publications, 2004, {{isbn|9780486438047}}.
*{{Citation | last1=John | first1=Fritz |author-link=Fritz John| title=Plane waves and spherical means applied to partial differential equations |url=| publisher=Interscience Publishers, New York-London | mr=0075429 | year=1955 }}. [https://books.google.com/books?id=JP6ABuUnhecC Reprinted], Dover Publications, 2004, {{isbn|9780486438047}}.
* {{citation | last1=Karatzas | first1=Ioannis | last2=Shreve | first2=Steven E. | title=Brownian Motion and Stochastic Calculus | publisher=Springer New York | publication-place=New York, NY | volume=113 | date=1998 | isbn=978-0-387-97655-6 | doi=10.1007/978-1-4612-0949-2 | doi-access=free}}
*{{Citation | last1=Lang | first1=Serge | author1-link=Serge Lang | title=Undergraduate analysis | publisher=Springer-Verlag | location=Berlin, New York | edition=2nd | series=[[Undergraduate Texts in Mathematics]] | isbn=978-0-387-94841-6 | mr=1476913 | year=1997 | doi=10.1007/978-1-4757-2698-5}}.
*{{Citation | last1=Lang | first1=Serge | author1-link=Serge Lang | title=Undergraduate analysis | publisher=Springer-Verlag | location=Berlin, New York | edition=2nd | series=[[Undergraduate Texts in Mathematics]] | isbn=978-0-387-94841-6 | mr=1476913 | year=1997 | doi=10.1007/978-1-4757-2698-5}}.
*{{citation|last=Lange|first=Rutger-Jan|year=2012|title=Potential theory, path integrals and the Laplacian of the indicator|journal=Journal of High Energy Physics|volume=2012|pages=29–30|issue=11|bibcode=2012JHEP...11..032L|doi=10.1007/JHEP11(2012)032|arxiv = 1302.0864 |s2cid=56188533}}.
*{{citation|last=Lange|first=Rutger-Jan|year=2012|title=Potential theory, path integrals and the Laplacian of the indicator|journal=Journal of High Energy Physics|volume=2012|pages=29–30|issue=11|article-number=32 |bibcode=2012JHEP...11..032L|doi=10.1007/JHEP11(2012)032|arxiv = 1302.0864 |s2cid=56188533}}.
*{{citation|doi=10.1007/BF00329867|author-link=Detlef Laugwitz|last=Laugwitz|first=D.|year=1989|title=Definite values of infinite sums: aspects of the foundations of infinitesimal analysis around 1820|journal=Arch. Hist. Exact Sci.|volume=39|issue=3|pages=195–245|s2cid=120890300}}.
*{{citation|doi=10.1007/BF00329867|author-link=Detlef Laugwitz|last=Laugwitz|first=D.|year=1989|title=Definite values of infinite sums: aspects of the foundations of infinitesimal analysis around 1820|journal=Arch. Hist. Exact Sci.|volume=39|issue=3|pages=195–245|s2cid=120890300}}.
*{{Citation |title=An introduction to quantum theory |last=Levin |first=Frank S.  |year=2002 |chapter-url=https://books.google.com/books?id=oc64f4EspFgC&pg=PA109 |pages=109''ff'' |chapter=Coordinate-space wave functions and completeness |isbn=978-0-521-59841-5 |publisher=Cambridge University Press}}
*{{Citation |title=An introduction to quantum theory |last=Levin |first=Frank S.  |year=2002 |chapter-url=https://books.google.com/books?id=oc64f4EspFgC&pg=PA109 |pages=109''ff'' |chapter=Coordinate-space wave functions and completeness |isbn=978-0-521-59841-5 |publisher=Cambridge University Press}}
Line 834: Line 855:
*{{citation|last=McMahon|first=D.|title=Quantum Mechanics Demystified, A Self-Teaching Guide|chapter-url=http://index-of.co.uk/Misc/McGraw-Hill%20-%20Quantum%20Mechanics%20Demystified%20(2006).pdf|access-date=2008-03-17|series=Demystified Series|date=2005-11-22|publisher=McGraw-Hill|location=New York|isbn=978-0-07-145546-6|page=108|chapter=An Introduction to State Space}}.
*{{citation|last=McMahon|first=D.|title=Quantum Mechanics Demystified, A Self-Teaching Guide|chapter-url=http://index-of.co.uk/Misc/McGraw-Hill%20-%20Quantum%20Mechanics%20Demystified%20(2006).pdf|access-date=2008-03-17|series=Demystified Series|date=2005-11-22|publisher=McGraw-Hill|location=New York|isbn=978-0-07-145546-6|page=108|chapter=An Introduction to State Space}}.
*{{Citation | last1=van der Pol | first1=Balth. | last2=Bremmer | first2=H. | title=Operational calculus | publisher=Chelsea Publishing Co. | location=New York | edition=3rd | isbn=978-0-8284-0327-6 | mr=904873 | year=1987}}.
*{{Citation | last1=van der Pol | first1=Balth. | last2=Bremmer | first2=H. | title=Operational calculus | publisher=Chelsea Publishing Co. | location=New York | edition=3rd | isbn=978-0-8284-0327-6 | mr=904873 | year=1987}}.
* {{cite book
| last1 = Reed
| first1 = Michael
| last2 = Simon
| first2 = Barry
| title = Methods of Modern Mathematical Physics, Volume I: Functional Analysis
| publisher = Academic Press
| year = 1980
| isbn = 9780125850506}}
* {{cite book |last1=Rudin |first1=Walter |author1-link=Walter Rudin |title=Real and complex analysis |date=1966 |publisher=McGraw-Hill |location=New York |isbn=0-07-100276-6 |edition=3rd |publication-date=1987 |editor-first=Peter R. |editor-last=Devine}}
* {{cite book |last1=Rudin |first1=Walter |author1-link=Walter Rudin |title=Real and complex analysis |date=1966 |publisher=McGraw-Hill |location=New York |isbn=0-07-100276-6 |edition=3rd |publication-date=1987 |editor-first=Peter R. |editor-last=Devine}}
* {{citation|first=Walter|last=Rudin|author-link=Walter Rudin|title=Functional Analysis|edition=2nd|publisher=McGraw-Hill|year=1991|isbn=978-0-07-054236-5|url-access=registration|url=https://archive.org/details/functionalanalys00rudi}}.
* {{citation|first=Walter|last=Rudin|author-link=Walter Rudin|title=Functional Analysis|edition=2nd|publisher=McGraw-Hill|year=1991|isbn=978-0-07-054236-5|url-access=registration|url=https://archive.org/details/functionalanalys00rudi}}.
Line 852: Line 882:
* {{citation|last=Yamashita |first=H. |year=2006 |title= Pointwise analysis of scalar fields: A nonstandard approach |volume=47 |issue =9 |pages=092301|journal=[[Journal of Mathematical Physics]]|doi=10.1063/1.2339017|bibcode = 2006JMP....47i2301Y }}
* {{citation|last=Yamashita |first=H. |year=2006 |title= Pointwise analysis of scalar fields: A nonstandard approach |volume=47 |issue =9 |pages=092301|journal=[[Journal of Mathematical Physics]]|doi=10.1063/1.2339017|bibcode = 2006JMP....47i2301Y }}
* {{citation|last=Yamashita |first=H. |year=2007 |title= Comment on "Pointwise analysis of scalar fields: A nonstandard approach" [J. Math. Phys. 47, 092301 (2006)] |volume=48 |issue =8 |pages=084101|journal=[[Journal of Mathematical Physics]]|doi=10.1063/1.2771422|bibcode = 2007JMP....48h4101Y |doi-access=free }}
* {{citation|last=Yamashita |first=H. |year=2007 |title= Comment on "Pointwise analysis of scalar fields: A nonstandard approach" [J. Math. Phys. 47, 092301 (2006)] |volume=48 |issue =8 |pages=084101|journal=[[Journal of Mathematical Physics]]|doi=10.1063/1.2771422|bibcode = 2007JMP....48h4101Y |doi-access=free }}
* {{citation
* {{citation |last = Zhao |first= Ji-Cheng|title=Methods for Phase Diagram Determination |year = 2011 |publisher = Elsevier |isbn = 978-0-08-054996-5
| last = Zhao | first= Ji-Cheng
  |language = en |url = https://books.google.com/books?id=blZYGDREpk8C}}
  | url = https://books.google.com/books?id=blZYGDREpk8C}}|title=Methods for Phase Diagram Determination
* {{cite book
  | year = 2011
| last = Zhu
  | publisher = Elsevier
| first = Kehe
  | isbn = 978-0-08-054996-5
| title = Operator Theory in Function Spaces
  | language = en
  | series = Mathematical Surveys and Monographs
  | volume = 138
  | publisher = American Mathematical Society
  | year = 2007
}}
}}
{{Refend}}


==External links==
==External links==

Latest revision as of 17:25, 19 October 2025

Template:Short description Script error: No such module "redirect hatnote". Template:Use American English

File:Dirac distribution PDF.svg
Schematic representation of the Dirac delta function by a line surmounted by an arrow. The height of the arrow is usually meant to specify the value of any multiplicative constant, which will give the area under the function. The other convention is to write the area next to the arrowhead.
File:Dirac function approximation.gif
The Dirac delta as the limit as a0 (in the sense of distributions) of the sequence of zero-centered normal distributions δa(x)=1|a|πe(x/a)2

Template:Differential equations

In mathematical analysis, the Dirac delta function (or Template:Mvar distribution), also known as the unit impulse,Template:Sfn is a generalized function on the real numbers, whose value is zero everywhere except at zero, and whose integral over the entire real line is equal to one.Template:SfnTemplate:SfnTemplate:Sfn Thus it can be represented heuristically as

δ(x)={0,x0,x=0

such that

δ(x)dx=1.

Since there is no function having this property, modelling the delta "function" rigorously involves the use of limits or, as is common in mathematics, measure theory and the theory of distributions.

The delta function was introduced by physicist Paul Dirac, and has since been applied routinely in physics and engineering to model point masses and instantaneous impulses. It is called the delta function because it is a continuous analogue of the Kronecker delta function, which is usually defined on a discrete domain and takes values 0 and 1. The mathematical rigor of the delta function was disputed until Laurent Schwartz developed the theory of distributions, where it is defined as a linear form acting on functions.

Motivation and overview

The graph of the Dirac delta is usually thought of as following the whole x-axis and the positive y-axis.Template:Sfn The Dirac delta is used to model a tall narrow spike function (an impulse), and other similar abstractions such as a point charge or point mass.Template:SfnTemplate:Sfn For example, to calculate the dynamics of a billiard ball being struck, one can approximate the force of the impact by a Dirac delta. In doing so, one can simplify the equations and calculate the motion of the ball by only considering the total impulse of the collision, without a detailed model of all of the elastic energy transfer at subatomic levels (for instance).

To be specific, suppose that a billiard ball is at rest. At time t=0 it is struck by another ball, imparting it with a momentum Template:Mvar, with units kg⋅m⋅s−1. The exchange of momentum is not actually instantaneous, being mediated by elastic processes at the molecular and subatomic level, but for practical purposes it is convenient to consider that energy transfer as effectively instantaneous. The force therefore is Template:Math; the units of Template:Math are s−1.

To model this situation more rigorously, suppose that the force instead is uniformly distributed over a small time interval Δt=[0,T]. That is,

FΔt(t)={P/Δt0<tT,0otherwise.

Then the momentum at any time Template:Mvar is found by integration:

p(t)=0tFΔt(τ)dτ={PtTPt/Δt0tT0otherwise.

Now, the model situation of an instantaneous transfer of momentum requires taking the limit as Template:Math, giving a result everywhere except at Template:Math:

p(t)={Pt>00t<0.

Here the functions FΔt are thought of as useful approximations to the idea of instantaneous transfer of momentum.

The delta function allows us to construct an idealized limit of these approximations. Unfortunately, the actual limit of the functions (in the sense of pointwise convergence) limΔt0+FΔt is zero everywhere but a single point, where it is infinite. To make proper sense of the Dirac delta, we should instead insist that the property

FΔt(t)dt=P,

which holds for all Δt>0, should continue to hold in the limit. So, in the equation F(t)=Pδ(t)=limΔt0FΔt(t), it is understood that the limit is always taken Template:Em.

In applied mathematics, as we have done here, the delta function is often manipulated as a kind of limit (a weak limit) of a sequence of functions, each member of which has a tall spike at the origin: for example, a sequence of Gaussian distributions centered at the origin with variance tending to zero. (However, even in some applications, highly oscillatory functions are used as approximations to the delta function, see below.)

The Dirac delta, given the desired properties outlined above, cannot be a function with domain and range in real numbers.Template:Sfn For example, the objects Template:Math and Template:Math are equal everywhere except at Template:Math yet have integrals that are different. According to Lebesgue integration theory, if Template:Mvar and Template:Mvar are functions such that Template:Math almost everywhere, then Template:Mvar is integrable if and only if Template:Mvar is integrable and the integrals of Template:Mvar and Template:Mvar are identical.Template:Sfn A rigorous approach to regarding the Dirac delta function as a mathematical object in its own right uses measure theory or the theory of distributions.Template:Sfn

History

In physics, the Dirac delta function was popularized by Paul Dirac in this book The Principles of Quantum Mechanics published in 1930.Template:Sfn However, Oliver Heaviside, 35 years before Dirac, described an impulsive function called the Heaviside step function for purposes and with properties analogous to Dirac's work. Even earlier several mathematicians and physicists used limits of sharply peaked functions in derivations.[1] An infinitesimal formula for an infinitely tall, unit impulse delta function (infinitesimal version of Cauchy distribution) explicitly appears in an 1827 text of Augustin-Louis Cauchy.Template:Sfn Siméon Denis Poisson considered the issue in connection with the study of wave propagation as did Gustav Kirchhoff somewhat later. Kirchhoff and Hermann von Helmholtz also introduced the unit impulse as a limit of Gaussians, which also corresponded to Lord Kelvin's notion of a point heat source.[2] The Dirac delta function as such was introduced by Paul Dirac in his 1927 paper The Physical Interpretation of the Quantum Dynamics.[3] He called it the "delta function" since he used it as a continuum analogue of the discrete Kronecker delta.Template:Sfn

Mathematicians refer to the same concept as a distribution rather than a function.[4] Joseph Fourier presented what is now called the Fourier integral theorem in his treatise Théorie analytique de la chaleur in the form:[5]

f(x)=12π  dαf(α) dp cos(pxpα) ,

which is tantamount to the introduction of the Template:Mvar-function in the form:[6]

δ(xα)=12πdp cos(pxpα) .

Later, Augustin Cauchy expressed the theorem using exponentials:[7][8]

f(x)=12π eipx(eipαf(α)dα)dp.

Cauchy pointed out that in some circumstances the order of integration is significant in this result (contrast Fubini's theorem).[9][10]

As justified using the theory of distributions, the Cauchy equation can be rearranged to resemble Fourier's original formulation and expose the δ-function as

f(x)=12πeipx(eipαf(α)dα)dp=12π(eipxeipαdp)f(α)dα=δ(xα)f(α)dα,

where the δ-function is expressed as

δ(xα)=12πeip(xα)dp .

A rigorous interpretation of the exponential form and the various limitations upon the function f necessary for its application extended over several centuries. The problems with a classical interpretation are explained as follows:[11]

The greatest drawback of the classical Fourier transformation is a rather narrow class of functions (originals) for which it can be effectively computed. Namely, it is necessary that these functions decrease sufficiently rapidly to zero (in the neighborhood of infinity) to ensure the existence of the Fourier integral. For example, the Fourier transform of such simple functions as polynomials does not exist in the classical sense. The extension of the classical Fourier transformation to distributions considerably enlarged the class of functions that could be transformed and this removed many obstacles.

Further developments included generalization of the Fourier integral, "beginning with Plancherel's pathbreaking L2-theory (1910), continuing with Wiener's and Bochner's works (around 1930) and culminating with the amalgamation into L. Schwartz's theory of distributions (1945) ...",[12] and leading to the formal development of the Dirac delta function.

Definitions

The Dirac delta function δ(x) can be loosely thought of as a function on the real line which is zero everywhere except at the origin, where it is infinite,

δ(x){+,x=00,x0

and which is also constrained to satisfy the identityTemplate:Sfn

δ(x)dx=1.

This is merely a heuristic characterization. The Dirac delta is not a function in the traditional sense as no extended real number valued function defined on the real numbers has these properties.Template:Sfn

As a measure

One way to rigorously capture the notion of the Dirac delta function is to define a measure, called Dirac measure, which accepts a subset Template:Mvar of the real line Template:Math as an argument, and returns Template:Math if Template:Math, and Template:Math otherwise.[13] If the delta function is conceptualized as modeling an idealized point mass at 0, then Template:Math represents the mass contained in the set Template:Mvar. One may then define the integral against Template:Mvar as the integral of a function against this mass distribution. Formally, the Lebesgue integral provides the necessary analytic device. The Lebesgue integral with respect to the measure Template:Mvar satisfies

f(x)δ(dx)=f(0)

for all continuous compactly supported functions Template:Mvar. The measure Template:Mvar is not absolutely continuous with respect to the Lebesgue measure—in fact, it is a singular measure. Consequently, the delta measure has no Radon–Nikodym derivative (with respect to Lebesgue measure)—no true function for which the property

f(x)δ(x)dx=f(0)

holds.Template:Sfn As a result, the latter notation is a convenient abuse of notation, and not a standard (Riemann or Lebesgue) integral.Template:Sfn

As a probability measure on Template:Math, the delta measure is characterized by its cumulative distribution function, which is the unit step function.[14]

H(x)={1if x00if x<0.

This means that Template:Math is the integral of the cumulative indicator function Template:Math with respect to the measure Template:Mvar; to wit,

H(x)=𝐑𝟏(,x](t)δ(dt)=δ((,x]),

the latter being the measure of this interval. Thus in particular the integration of the delta function against a continuous function can be properly understood as a Riemann–Stieltjes integral:Template:Sfn

f(x)δ(dx)=f(x)dH(x).

All higher moments of Template:Mvar are zero. In particular, characteristic function and moment generating function are both equal to one.Template:Sfn

As a distribution

In the theory of distributions, a generalized function is considered not a function in itself but only through how it affects other functions when "integrated" against them.Template:Sfn In keeping with this philosophy, to define the delta function properly, it is enough to say what the "integral" of the delta function is against a sufficiently "good" test function Template:Mvar.Template:Sfn If the delta function is already understood as a measure, then the Lebesgue integral of a test function against that measure supplies the necessary integral.Template:Sfn

A typical space of test functions consists of all smooth functions on Template:Math with compact support that have as many derivatives as required. As a distribution, the Dirac delta is a linear functional on the space of test functions and is defined byTemplate:Sfn

Template:NumBlk2

for every test function Template:Mvar.

For Template:Mvar to be properly a distribution, it must be continuous in a suitable topology on the space of test functions. In general, for a linear functional Template:Mvar on the space of test functions to define a distribution, it is necessary and sufficient that, for every positive integer Template:Mvar there is an integer Template:Math and a constant Template:Mvar such that for every test function Template:Mvar, one has the inequalityTemplate:Sfn

|S[φ]|CNk=0MNsupx[N,N]|φ(k)(x)|

where Template:Math represents the supremum. With the Template:Mvar distribution, one has such an inequality (with Template:Math with Template:Math for all Template:Mvar. Thus Template:Mvar is a distribution of order zero. It is, furthermore, a distribution with compact support (the support being Template:Math).

The delta distribution can also be defined in several equivalent ways. For instance, it is the distributional derivative of the Heaviside step function. This means that for every test function Template:Mvar, one has

δ[φ]=φ(x)H(x)dx.

Intuitively, if integration by parts were permitted, then the latter integral should simplify to

φ(x)H(x)dx=φ(x)δ(x)dx,

and indeed, a form of integration by parts is permitted for the Stieltjes integral, and in that case, one does have

φ(x)H(x)dx=φ(x)dH(x).

In the context of measure theory, the Dirac measure gives rise to distribution by integration. Conversely, equation (Template:EquationNote) defines a Daniell integral on the space of all compactly supported continuous functions Template:Mvar which, by the Riesz representation theorem, can be represented as the Lebesgue integral of Template:Mvar with respect to some Radon measure.Template:Sfn}

Generally, when the term Dirac delta function is used, it is in the sense of distributions rather than measures, the Dirac measure being among several terms for the corresponding notion in measure theory. Some sources may also use the term Dirac delta distribution.

Generalizations

The delta function can be defined in Template:Mvar-dimensional Euclidean space Template:Math as the measure such that

𝐑nf(𝐱)δ(d𝐱)=f(𝟎)

for every compactly supported continuous function Template:Mvar. As a measure, the Template:Mvar-dimensional delta function is the product measure of the 1-dimensional delta functions in each variable separately. Thus, formally, with Template:Math, one hasTemplate:Sfn

Template:NumBlk2

The delta function can also be defined in the sense of distributions exactly as above in the one-dimensional case.Template:Sfn However, despite widespread use in engineering contexts, (Template:EquationNote) should be manipulated with care, since the product of distributions can only be defined under quite narrow circumstances.Template:SfnTemplate:Sfn

The notion of a Dirac measure makes sense on any set.Template:Sfn Thus if Template:Mvar is a set, Template:Math is a marked point, and Template:Math is any sigma algebra of subsets of Template:Mvar, then the measure defined on sets Template:Math by

δx0(A)={1if x0A0if x0A

is the delta measure or unit mass concentrated at Template:Math.

Another common generalization of the delta function is to a differentiable manifold where most of its properties as a distribution can also be exploited because of the differentiable structure. The delta function on a manifold Template:Mvar centered at the point Template:Math is defined as the following distribution:

Template:NumBlk2

for all compactly supported smooth real-valued functions Template:Mvar on Template:Mvar.Template:Sfn A common special case of this construction is a case in which Template:Mvar is an open set in the Euclidean space Template:Math.

On a locally compact Hausdorff space Template:Mvar, the Dirac delta measure concentrated at a point Template:Mvar is the Radon measure associated with the Daniell integral (Template:EquationNote) on compactly supported continuous functions Template:Mvar.[15] At this level of generality, calculus as such is no longer possible, however a variety of techniques from abstract analysis are available. For instance, the mapping x0δx0 is a continuous embedding of Template:Mvar into the space of finite Radon measures on Template:Mvar, equipped with its vague topology. Moreover, the convex hull of the image of Template:Mvar under this embedding is dense in the space of probability measures on Template:Mvar.Template:Sfn

Properties

Scaling and symmetry

The delta function satisfies the following scaling property for a non-zero scalar Template:Mvar:Template:SfnTemplate:Sfn

δ(αx)dx=δ(u)du|α|=1|α|

and so Template:NumBlk2

Scaling property proof: dx g(x)δ(αx)=1αdx g(xα)δ(x)=1αg(0). where a change of variable Template:Math is used. If Template:Mvar is negative, i.e., Template:Math, then dx g(x)δ(αx)=1|α|dx g(xα)δ(x)=1|α|dx g(xα)δ(x)=1|α|g(0). Thus, δ(αx)=1|α|δ(x).

In particular, the delta function is an even distribution (symmetry), in the sense that

δ(x)=δ(x)

which is homogeneous of degree Template:Math.

Algebraic properties

The distributional product of Template:Mvar with Template:Mvar is equal to zero:

xδ(x)=0.

More generally, (xa)nδ(xa)=0 for all positive integers n.

Conversely, if Template:Math, where Template:Mvar and Template:Mvar are distributions, then

f(x)=g(x)+cδ(x)

for some constant Template:Mvar.Template:Sfn

Translation

The integral of any function multiplied by the time-delayed Dirac delta δT(t)=δ(tT) is

f(t)δ(tT)dt=f(T).

This is sometimes referred to as the sifting property[16] or the sampling property.[17] The delta function is said to "sift out" the value of f(t) at t = T.[18]

It follows that the effect of convolving a function Template:Math with the time-delayed Dirac delta is to time-delay Template:Math by the same amount:[19]

(f*δT)(t) =def f(τ)δ(tTτ)dτ=f(τ)δ(τ(tT))dτsinceδ(x)=δ(x)by (4)=f(tT).

The sifting property holds under the precise condition that Template:Mvar be a tempered distribution (see the discussion of the Fourier transform below). As a special case, for instance, we have the identity (understood in the distribution sense)

δ(ξx)δ(xη)dx=δ(ηξ).

Composition with a function

More generally, the delta distribution may be composed with a smooth function Template:Math in such a way that the familiar change of variables formula holds (where u=g(x)), that

δ(g(x))f(g(x))|g(x)|dx=g()δ(u)f(u)du

provided that Template:Mvar is a continuously differentiable function with Template:Math nowhere zero.Template:Sfn That is, there is a unique way to assign meaning to the distribution δg so that this identity holds for all compactly supported test functions Template:Mvar. Therefore, the domain must be broken up to exclude the Template:Math point. This distribution satisfies Template:Math if Template:Mvar is nowhere zero, and otherwise if Template:Mvar has a real root at Template:Math, then

δ(g(x))=δ(xx0)|g(x0)|.

It is natural therefore to Template:Em the composition Template:Math for continuously differentiable functions Template:Mvar by

δ(g(x))=iδ(xxi)|g(xi)|

where the sum extends over all roots of Template:Mvar, which are assumed to be simple. Thus, for example

δ(x2α2)=12|α|[δ(x+α)+δ(xα)].

In the integral form, the generalized scaling property may be written as

f(x)δ(g(x))dx=if(xi)|g(xi)|.

Indefinite integral

For a constant a and a "well-behaved" arbitrary real-valued function Template:Math, y(x)δ(xa)dx=y(a)H(xa)+c, where Template:Math is the Heaviside step function and Template:Math is an integration constant.

Properties in n dimensions

The delta distribution in an Template:Mvar-dimensional space satisfies the following scaling property instead, δ(αx)=|α|nδ(x), so that Template:Mvar is a homogeneous distribution of degree Template:Math.

Under any reflection or rotation Template:Mvar, the delta function is invariant, δ(ρx)=δ(x).

As in the one-variable case, it is possible to define the composition of Template:Mvar with a bi-Lipschitz function[20] Template:Math uniquely so that the following holds nδ(g(x))f(g(x))|detg(x)|dx=g(n)δ(u)f(u)du for all compactly supported functions Template:Mvar.

Using the coarea formula from geometric measure theory, one can also define the composition of the delta function with a submersion from one Euclidean space to another one of different dimension; the result is a type of current. In the special case of a continuously differentiable function Template:Math such that the gradient of Template:Mvar is nowhere zero, the following identity holdsTemplate:Sfn nf(x)δ(g(x))dx=g1(0)f(x)|g|dσ(x) where the integral on the right is over Template:Math, the Template:Math-dimensional surface defined by Template:Math with respect to the Minkowski content measure. This is known as a simple layer integral.

More generally, if Template:Mvar is a smooth hypersurface of Template:Math, then we can associate to Template:Mvar the distribution that integrates any compactly supported smooth function Template:Mvar over Template:Mvar: δS[g]=Sg(s)dσ(s)

where Template:Mvar is the hypersurface measure associated to Template:Mvar. This generalization is associated with the potential theory of simple layer potentials on Template:Mvar. If Template:Mvar is a domain in Template:Math with smooth boundary Template:Mvar, then Template:Math is equal to the normal derivative of the indicator function of Template:Mvar in the distribution sense,

ng(x)1D(x)ndx=Sg(s)dσ(s),

where Template:Mvar is the outward normal.Template:SfnTemplate:Sfn

In three dimensions, the delta function is represented in spherical coordinates by:

δ(rr0)={1r2sinθδ(rr0)δ(θθ0)δ(ϕϕ0)x0,y0,z0012πr2sinθδ(rr0)δ(θθ0)x0=y0=0, z0014πr2δ(rr0)x0=y0=z0=0

Derivatives

The derivative of the Dirac delta distribution, denoted Template:Math and also called the Dirac delta prime or Dirac delta derivative, is defined on compactly supported smooth test functions Template:Mvar byTemplate:Sfn δ[φ]=δ[φ]=φ(0).

The first equality here is a kind of integration by parts, for if Template:Mvar were a true function then δ(x)φ(x)dx=δ(x)φ(x)|δ(x)φ(x)dx=δ(x)φ(x)dx=φ(0).

By mathematical induction, the Template:Mvar-th derivative of Template:Mvar is defined similarly as the distribution given on test functions by

δ(k)[φ]=(1)kφ(k)(0).

In particular, Template:Mvar is an infinitely differentiable distribution.

The first derivative of the delta function is the distributional limit of the difference quotients:Template:Sfn δ(x)=limh0δ(x+h)δ(x)h.

More properly, one has δ=limh01h(τhδδ) where Template:Mvar is the translation operator, defined on functions by Template:Math, and on a distribution Template:Mvar by (τhS)[φ]=S[τhφ].

In the theory of electromagnetism, the first derivative of the delta function represents a point magnetic dipole situated at the origin. Accordingly, it is referred to as a dipole or the doublet function.[21]

The derivative of the delta function satisfies a number of basic properties, including:Template:Sfn δ(x)=δ(x)xδ(x)=δ(x) which can be shown by applying a test function and integrating by parts.

The latter of these properties can also be demonstrated by applying distributional derivative definition, Leibniz 's theorem and linearity of inner product:[22]Template:Better source

xδ,φ=δ,xφ=δ,(xφ)=δ,xφ+xφ=δ,xφδ,xφ=x(0)φ(0)x(0)φ(0)=x(0)δ,φx(0)δ,φ=x(0)δ,φ+x(0)δ,φ=x(0)δx(0)δ,φx(t)δ(t)=x(0)δ(t)x(0)δ(t)=x(0)δ(t)=δ(t)

Furthermore, the convolution of Template:Mvar with a compactly-supported, smooth function Template:Mvar is

δ*f=δ*f=f,

which follows from the properties of the distributional derivative of a convolution.

Higher dimensions

More generally, on an open set Template:Mvar in the Template:Mvar-dimensional Euclidean space n, the Dirac delta distribution centered at a point Template:Math is defined byTemplate:Sfn δa[φ]=φ(a) for all φCc(U), the space of all smooth functions with compact support on Template:Mvar. If α=(α1,,αn) is any multi-index with |α|=α1++αn and α denotes the associated mixed partial derivative operator, then the Template:Mvar-th derivative Template:Mvar of Template:Mvar is given byTemplate:Sfn

αδa,φ=(1)|α|δa,αφ=(1)|α|αφ(x)|x=a for all φCc(U).

That is, the Template:Mvar-th derivative of Template:Mvar is the distribution whose value on any test function Template:Mvar is the Template:Mvar-th derivative of Template:Mvar at Template:Mvar (with the appropriate positive or negative sign).

The first partial derivatives of the delta function are thought of as double layers along the coordinate planes. More generally, the normal derivative of a simple layer supported on a surface is a double layer supported on that surface and represents a laminar magnetic monopole. Higher derivatives of the delta function are known in physics as multipoles.[23]

Higher derivatives enter into mathematics naturally as the building blocks for the complete structure of distributions with point support. If Template:Mvar is any distribution on Template:Mvar supported on the set Template:Math consisting of a single point, then there is an integer Template:Mvar and coefficients Template:Mvar such thatTemplate:SfnTemplate:Sfn S=|α|mcααδa.

Representations

The delta function can be viewed as the limit of a sequence of functions

δ(x)=limε0+ηε(x). This limit is meant in a weak sense: either that

Template:NumBlk2

for all continuous functions Template:Mvar having compact support, or that this limit holds for all smooth functions Template:Mvar with compact support. The former is convergence in the vague topology of measures, and the latter is convergence in the sense of distributions.

Approximations to the identity

An approximate delta function Template:Mvar can be constructed in the following manner. Let Template:Mvar be an absolutely integrable function on Template:Math of total integral Template:Math, and define ηε(x)=ε1η(xε).

In Template:Mvar dimensions, one uses instead the scaling ηε(x)=εnη(xε).

Then a simple change of variables shows that Template:Mvar also has integral Template:Math. One may show that (Template:EquationNote) holds for all continuous compactly supported functions Template:Mvar,Template:Sfn and so Template:Mvar converges weakly to Template:Mvar in the sense of measures.

The Template:Mvar constructed in this way are known as an approximation to the identity.Template:Sfn This terminology is because the space Template:Math of absolutely integrable functions is closed under the operation of convolution of functions: Template:Math whenever Template:Mvar and Template:Mvar are in Template:Math. However, there is no identity in Template:Math for the convolution product: no element Template:Mvar such that Template:Math for all Template:Mvar. Nevertheless, the sequence Template:Mvar does approximate such an identity in the sense that

f*ηεfas ε0.

This limit holds in the sense of mean convergence (convergence in Template:Math). Further conditions on the Template:Mvar, for instance that it be a mollifier associated to a compactly supported function,[24] are needed to ensure pointwise convergence almost everywhere.

If the initial Template:Math is itself smooth and compactly supported then the sequence is called a mollifier. The standard mollifier is obtained by choosing Template:Mvar to be a suitably normalized bump function, for instance

η(x)={1Inexp(11|x|2)if |x|<10if |x|1. (In ensuring that the total integral is 1).

In some situations such as numerical analysis, a piecewise linear approximation to the identity is desirable. This can be obtained by taking Template:Math to be a hat function. With this choice of Template:Math, one has

ηε(x)=ε1max(1|xε|,0)

which are all continuous and compactly supported, although not smooth and so not a mollifier.

Probabilistic considerations

In the context of probability theory, it is natural to impose the additional condition that the initial Template:Math in an approximation to the identity should be positive, as such a function then represents a probability distribution. Convolution with a probability distribution is sometimes favorable because it does not result in overshoot or undershoot, as the output is a convex combination of the input values, and thus falls between the maximum and minimum of the input function. Taking Template:Math to be any probability distribution at all, and letting Template:Math as above will give rise to an approximation to the identity. In general this converges more rapidly to a delta function if, in addition, Template:Mvar has mean Template:Math and has small higher moments. For instance, if Template:Math is the uniform distribution on [12,12], also known as the rectangular function, then:Template:Sfn ηε(x)=1εrect(xε)={1ε,ε2<x<ε2,0,otherwise.

Another example is with the Wigner semicircle distribution ηε(x)={2πε2ε2x2,ε<x<ε,0,otherwise.

This is continuous and compactly supported, but not a mollifier because it is not smooth.

Semigroups

Approximations to the delta functions often arise as convolution semigroups.[25] This amounts to the further constraint that the convolution of Template:Mvar with Template:Mvar must satisfy ηε*ηδ=ηε+δ

for all Template:Math. Convolution semigroups in Template:Math that approximate the delta function are always an approximation to the identity in the above sense, however the semigroup condition is quite a strong restriction.

In practice, semigroups approximating the delta function arise as fundamental solutions or Green's functions to physically motivated elliptic or parabolic partial differential equations. In the context of applied mathematics, semigroups arise as the output of a linear time-invariant system. Abstractly, if A is a linear operator acting on functions of x, then a convolution semigroup arises by solving the initial value problem

{tη(t,x)=Aη(t,x),t>0limt0+η(t,x)=δ(x)

in which the limit is as usual understood in the weak sense. Setting Template:Math gives the associated approximate delta function.

Some examples of physically important convolution semigroups arising from such a fundamental solution include the following.

The heat kernel

The heat kernel, defined byTemplate:Sfn

ηε(x)=12πεex22ε

represents the temperature in an infinite wire at time Template:Math, if a unit of heat energy is stored at the origin of the wire at time Template:Math. This semigroup evolves according to the one-dimensional heat equation:

ut=122ux2.

In probability theory, Template:Math is a normal distribution of variance Template:Mvar and mean Template:Math. It represents the probability density at time Template:Math of the position of a particle starting at the origin following a standard Brownian motion. In this context, the semigroup condition is then an expression of the Markov property of Brownian motion.

In higher-dimensional Euclidean space Template:Math, the heat kernel is ηε=1(2πε)n/2exx2ε, and has the same physical interpretation, Script error: No such module "Lang".. It also represents an approximation to the delta function in the sense that Template:Math in the distribution sense as Template:Math.

The Poisson kernel

The Poisson kernel ηε(x)=1πIm{1xiε}=1πεε2+x2=12πeiξx|εξ|dξ

is the fundamental solution of the Laplace equation in the upper half-plane.Template:Sfn It represents the electrostatic potential in a semi-infinite plate whose potential along the edge is held at fixed at the delta function. The Poisson kernel is also closely related to the Cauchy distribution and Epanechnikov and Gaussian kernel functions.[26] This semigroup evolves according to the equation ut=(2x2)12u(t,x)

where the operator is rigorously defined as the Fourier multiplier [(2x2)12f](ξ)=|2πξ|f(ξ).

Oscillatory integrals

In areas of physics such as wave propagation and wave mechanics, the equations involved are hyperbolic and so may have more singular solutions. As a result, the approximate delta functions that arise as fundamental solutions of the associated Cauchy problems are generally oscillatory integrals. An example, which comes from a solution of the Euler–Tricomi equation of transonic gas dynamics,Template:Sfn is the rescaled Airy function ε1/3Ai(xε1/3).

Although using the Fourier transform, it is easy to see that this generates a semigroup in some sense—it is not absolutely integrable and so cannot define a semigroup in the above strong sense. Many approximate delta functions constructed as oscillatory integrals only converge in the sense of distributions (an example is the Dirichlet kernel below), rather than in the sense of measures.

Another example is the Cauchy problem for the wave equation in Template:Math:Template:Sfn c22ut2Δu=0u=0,ut=δfor t=0.

The solution Template:Mvar represents the displacement from equilibrium of an infinite elastic string, with an initial disturbance at the origin.

Other approximations to the identity of this kind include the sinc function (used widely in electronics and telecommunications) ηε(x)=1πxsin(xε)=12π1ε1εcos(kx)dk

and the Bessel function ηε(x)=1εJ1ε(x+1ε).

Plane wave decomposition

One approach to the study of a linear partial differential equation L[u]=f,

where Template:Mvar is a differential operator on Template:Math, is to seek first a fundamental solution, which is a solution of the equation L[u]=δ.

When Template:Mvar is particularly simple, this problem can often be resolved using the Fourier transform directly (as in the case of the Poisson kernel and heat kernel already mentioned). For more complicated operators, it is sometimes easier first to consider an equation of the form L[u]=h

where Template:Mvar is a plane wave function, meaning that it has the form h=h(xξ)

for some vector Template:Mvar. Such an equation can be resolved (if the coefficients of Template:Mvar are analytic functions) by the Cauchy–Kovalevskaya theorem or (if the coefficients of Template:Mvar are constant) by quadrature. So, if the delta function can be decomposed into plane waves, then one can in principle solve linear partial differential equations.

Such a decomposition of the delta function into plane waves was part of a general technique first introduced essentially by Johann Radon, and then developed in this form by Fritz John (1955).Template:Sfn Choose Template:Mvar so that Template:Math is an even integer, and for a real number Template:Mvar, put g(s)=Re[sklog(is)k!(2πi)n]={|s|k4k!(2πi)n1n odd|s|klog|s|k!(2πi)nn even.

Then Template:Mvar is obtained by applying a power of the Laplacian to the integral with respect to the unit sphere measure Template:Mvar of Template:Math for Template:Mvar in the unit sphere Template:Math: δ(x)=Δx(n+k)/2Sn1g(xξ)dωξ.

The Laplacian here is interpreted as a weak derivative, so that this equation is taken to mean that, for any test function Template:Mvar, φ(x)=𝐑nφ(y)dyΔxn+k2Sn1g((xy)ξ)dωξ.

The result follows from the formula for the Newtonian potential (the fundamental solution of Poisson's equation). This is essentially a form of the inversion formula for the Radon transform because it recovers the value of Template:Math from its integrals over hyperplanes.Template:Sfn For instance, if Template:Mvar is odd and Template:Math, then the integral on the right hand side is cnΔxn+12Sn1φ(y)|(yx)ξ|dωξdy=cnΔx(n+1)/2Sn1dωξ|p|Rφ(ξ,p+xξ)dp

where Template:Math is the Radon transform of Template:Mvar: Rφ(ξ,p)=xξ=pf(x)dn1x.

An alternative equivalent expression of the plane wave decomposition is:Template:Sfn δ(x)={(n1)!(2πi)nSn1(xξ)ndωξn even12(2πi)n1Sn1δ(n1)(xξ)dωξn odd.

Fourier transform

The delta function is a tempered distribution, and therefore it has a well-defined Fourier transform. Formally, one finds[27]

δ^(ξ)=e2πixξδ(x)dx=1.

Properly speaking, the Fourier transform of a distribution is defined by imposing self-adjointness of the Fourier transform under the duality pairing , of tempered distributions with Schwartz functions. Thus δ^ is defined as the unique tempered distribution satisfying

δ^,φ=δ,φ^

for all Schwartz functions Template:Mvar. And indeed it follows from this that δ^=1.

As a result of this identity, the convolution of the delta function with any other tempered distribution Template:Mvar is simply Template:Mvar:

S*δ=S.

That is to say that Template:Mvar is an identity element for the convolution on tempered distributions, and in fact, the space of compactly supported distributions under convolution is an associative algebra with identity the delta function. This property is fundamental in signal processing, as convolution with a tempered distribution is a linear time-invariant system, and applying the linear time-invariant system measures its impulse response. The impulse response can be computed to any desired degree of accuracy by choosing a suitable approximation for Template:Mvar, and once it is known, it characterizes the system completely. See Template:Section link.

The inverse Fourier transform of the tempered distribution Template:Math is the delta function. Formally, this is expressed as 1e2πixξdξ=δ(x) and more rigorously, it follows since 1,f^=f(0)=δ,f for all Schwartz functions Template:Mvar.

In these terms, the delta function provides a suggestive statement of the orthogonality property of the Fourier kernel on Template:Math. Formally, one has ei2πξ1t[ei2πξ2t]*dt=ei2π(ξ2ξ1)tdt=δ(ξ2ξ1).

This is, of course, shorthand for the assertion that the Fourier transform of the tempered distribution f(t)=ei2πξ1t is f^(ξ2)=δ(ξ1ξ2) which again follows by imposing self-adjointness of the Fourier transform.

By analytic continuation of the Fourier transform, the Laplace transform of the delta function is found to beTemplate:Sfn 0δ(ta)estdt=esa.

Fourier kernels

Script error: No such module "Labelled list hatnote". In the study of Fourier series, a major question consists of determining whether and in what sense the Fourier series associated with a periodic function converges to the function. The Template:Mvar-th partial sum of the Fourier series of a function Template:Mvar of period Template:Math is defined by convolution (on the interval Template:Closed-closed) with the Dirichlet kernel: DN(x)=n=NNeinx=sin((N+12)x)sin(x/2). Thus, sN(f)(x)=DN*f(x)=n=NNaneinx where an=12πππf(y)einydy. A fundamental result of elementary Fourier series states that the Dirichlet kernel restricted to the interval Template:Closed-closed tends to a multiple of the delta function as Template:Math. This is interpreted in the distribution sense, that sN(f)(0)=ππDN(x)f(x)dx2πf(0) for every compactly supported Template:Em function Template:Mvar. Thus, formally one has δ(x)=12πn=einx on the interval Template:Closed-closed.

Despite this, the result does not hold for all compactly supported Template:Em functions: that is Template:Math does not converge weakly in the sense of measures. The lack of convergence of the Fourier series has led to the introduction of a variety of summability methods to produce convergence. The method of Cesàro summation leads to the Fejér kernelTemplate:Sfn

FN(x)=1Nn=0N1Dn(x)=1N(sinNx2sinx2)2.

The Fejér kernels tend to the delta function in a stronger sense that[28]

ππFN(x)f(x)dx2πf(0)

for every compactly supported Template:Em function Template:Mvar. The implication is that the Fourier series of any continuous function is Cesàro summable to the value of the function at every point.

Hilbert space theory

The Dirac delta distribution is a densely defined unbounded linear functional on the Hilbert space L2 of square-integrable functions.Template:Sfn Indeed, smooth compactly supported functions are dense in Template:Math, and the action of the delta distribution on such functions is well-defined. In many applications, it is possible to identify subspaces of Template:Math and to give a stronger topology on which the delta function defines a bounded linear functional.

Sobolev spaces

The Sobolev embedding theorem for Sobolev spaces on the real line Template:Math implies that any square-integrable function Template:Mvar such that

fH12=|f^(ξ)|2(1+|ξ|2)dξ<

is automatically continuous, and satisfies in particular

δ[f]=|f(0)|<CfH1.

Thus Template:Mvar is a bounded linear functional on the Sobolev space Template:Math.Template:Sfn Equivalently Template:Mvar is an element of the continuous dual space Template:Math of Template:Math. More generally, in Template:Mvar dimensions, one has Template:Math provided Template:Math.

Spaces of holomorphic functions

In complex analysis, the delta function enters via Cauchy's integral formula, which asserts that if Template:Mvar is a domain in the complex plane with smooth boundary, then

f(z)=12πiDf(ζ)dζζz,zD

for all holomorphic functions Template:Mvar in Template:Mvar that are continuous on the closure of Template:Mvar. As a result, the delta function Template:Math is represented in this class of holomorphic functions by the Cauchy integral:

δz[f]=f(z)=12πiDf(ζ)dζζz.

Moreover, let Template:Math be the Hardy space consisting of the closure in Template:Math of all holomorphic functions in Template:Mvar continuous up to the boundary of Template:Mvar. Then functions in Template:Math uniquely extend to holomorphic functions in Template:Mvar, and the Cauchy integral formula continues to hold. In particular for Template:Math, the delta function Template:Mvar is a continuous linear functional on Template:Math. This is a special case of the situation in several complex variables in which, for smooth domains Template:Mvar, the Szegő kernel plays the role of the Cauchy integral.Template:Sfn

Another representation of the delta function in a space of holomorphic functions is on the space H(D)L2(D) of square-integrable holomorphic functions in an open set Dn. This is a closed subspace of L2(D), and therefore is a Hilbert space. On the other hand, the functional that evaluates a holomorphic function in H(D)L2(D) at a point z of D is a continuous functional, and so by the Riesz representation theorem, is represented by integration against a kernel Kz(ζ), the Bergman kernel.Template:Sfn This kernel is the analog of the delta function in this Hilbert space. A Hilbert space having such a kernel is called a reproducing kernel Hilbert space. In the special case of the unit disc, one has δw[f]=f(w)=1π|z|<1f(z)dxdy(1z¯w)2.

Resolutions of the identity

Given a complete orthonormal basis set of functions Template:Math in a separable Hilbert space, for example, the normalized eigenvectors of a compact self-adjoint operator, any vector Template:Mvar can be expressed as f=n=1αnφn. The coefficients {αn} are found as αn=φn,f, which may be represented by the notation: αn=φnf, a form of the bra–ket notation of Dirac.Template:Sfn Adopting this notation, the expansion of Template:Mvar takes the dyadic form:Template:Sfn

f=n=1φn(φnf).

Letting Template:Mvar denote the identity operator on the Hilbert space, the expression

I=n=1φnφn,

is called a resolution of the identity. When the Hilbert space is the space Template:Math of square-integrable functions on a domain Template:Mvar, the quantity:

φnφn,

is an integral operator, and the expression for Template:Mvar can be rewritten

f(x)=n=1D(φn(x)φn*(ξ))f(ξ)dξ.

The right-hand side converges to Template:Mvar in the Template:Math sense. It need not hold in a pointwise sense, even when Template:Mvar is a continuous function. Nevertheless, it is common to abuse notation and write

f(x)=δ(xξ)f(ξ)dξ,

resulting in the representation of the delta function:Template:Sfn

δ(xξ)=n=1φn(x)φn*(ξ).

With a suitable rigged Hilbert space Template:Math where Template:Math contains all compactly supported smooth functions, this summation may converge in Template:Math, depending on the properties of the basis Template:Math. In most cases of practical interest, the orthonormal basis comes from an integral or differential operator (e.g. the heat kernel), in which case the series converges in the distribution sense.Template:Sfn

Infinitesimal delta functions

Cauchy used an infinitesimal Template:Mvar to write down a unit impulse, infinitely tall and narrow Dirac-type delta function Template:Mvar satisfying F(x)δα(x)dx=F(0) in a number of articles in 1827.Template:Sfn Cauchy defined an infinitesimal in Cours d'Analyse (1827) in terms of a sequence tending to zero. Namely, such a null sequence becomes an infinitesimal in Cauchy's and Lazare Carnot's terminology.

Non-standard analysis allows one to rigorously treat infinitesimals. The article by Template:Harvtxt contains a bibliography on modern Dirac delta functions in the context of an infinitesimal-enriched continuum provided by the hyperreals. Here the Dirac delta can be given by an actual function, having the property that for every real function Template:Mvar one has F(x)δα(x)dx=F(0) as anticipated by Fourier and Cauchy.

Dirac comb

Script error: No such module "Labelled list hatnote".

File:Dirac comb.svg
A Dirac comb is an infinite series of Dirac delta functions spaced at intervals of Template:Mvar

A so-called uniform "pulse train" of Dirac delta measures, which is known as a Dirac comb, or as the Sha distribution, creates a sampling function, often used in digital signal processing (DSP) and discrete time signal analysis. The Dirac comb is given as the infinite sum, whose limit is understood in the distribution sense,

Ш(x)=n=δ(xn),

which is a sequence of point masses at each of the integers.

Up to an overall normalizing constant, the Dirac comb is equal to its own Fourier transform. This is significant because if Template:Mvar is any Schwartz function, then the periodization of Template:Mvar is given by the convolution (f*Ш)(x)=n=f(xn). In particular, (f*Ш)=f^Ш^=f^Ш is precisely the Poisson summation formula.Template:SfnTemplate:Sfn More generally, this formula remains to be true if Template:Mvar is a tempered distribution of rapid descent or, equivalently, if f^ is a slowly growing, ordinary function within the space of tempered distributions.

Sokhotski–Plemelj theorem

The Sokhotski–Plemelj theorem, important in quantum mechanics, relates the delta function to the distribution Template:Math, the Cauchy principal value of the function Template:Math, defined by

p.v.1x,φ=limε0+|x|>εφ(x)xdx.

Sokhotsky's formula states thatTemplate:Sfn

limε0+1x±iε=p.v.1xiπδ(x),

Here the limit is understood in the distribution sense, that for all compactly supported smooth functions Template:Mvar,

limε0+f(x)x±iεdx=iπf(0)+limε0+|x|>εf(x)xdx.

Relationship to the Kronecker delta

The Kronecker delta Template:Mvar is the quantity defined by

δij={1i=j0i=j

for all integers Template:Mvar, Template:Mvar. This function then satisfies the following analog of the sifting property: if Template:Mvar (for Template:Mvar in the set of all integers) is any doubly infinite sequence, then

i=aiδik=ak.

Similarly, for any real or complex valued continuous function Template:Mvar on Template:Math, the Dirac delta satisfies the sifting property

f(x)δ(xx0)dx=f(x0).

This exhibits the Kronecker delta function as a discrete analog of the Dirac delta function.Template:Sfn

Applications

Probability theory

Script error: No such module "Labelled list hatnote". In probability theory and statistics, the Dirac delta function is often used to represent a discrete distribution, or a partially discrete, partially continuous distribution, using a probability density function (which is normally used to represent absolutely continuous distributions). For example, the probability density function Template:Math of a discrete distribution consisting of points Template:Math, with corresponding probabilities Template:Math, can be written as[29]

f(x)=i=1npiδ(xxi).

As another example, consider a distribution in which 6/10 of the time returns a standard normal distribution, and 4/10 of the time returns exactly the value 3.5 (i.e. a partly continuous, partly discrete mixture distribution). The density function of this distribution can be written as

f(x)=0.612πex22+0.4δ(x3.5).

The delta function is also used to represent the resulting probability density function of a random variable that is transformed by continuously differentiable function. If Template:Math is a continuous differentiable function, then the density of Template:Mvar can be written as

fY(y)=+fX(x)δ(yg(x))dx.

The delta function is also used in a completely different way to represent the local time of a diffusion process (like Brownian motion).Template:Sfn The local time of a stochastic process Template:Math is given by (x,t)=0tδ(xB(s))ds and represents the amount of time that the process spends at the point Template:Mvar in the range of the process. More precisely, in one dimension this integral can be written (x,t)=limε0+12ε0t𝟏[xε,x+ε](B(s))ds where 𝟏[xε,x+ε] is the indicator function of the interval [xε,x+ε].

Quantum mechanics

The delta function is expedient in quantum mechanics. The wave function of a particle gives the probability amplitude of finding a particle within a given region of space. Wave functions are assumed to be elements of the Hilbert space Template:Math of square-integrable functions, and the total probability of finding a particle within a given interval is the integral of the magnitude of the wave function squared over the interval. A set Template:Math of wave functions is orthonormal if

φnφm=δnm,

where Template:Mvar is the Kronecker delta. A set of orthonormal wave functions is complete in the space of square-integrable functions if any wave function Template:Math can be expressed as a linear combination of the Template:Math with complex coefficients:

ψ=cnφn,

where Template:Math. Complete orthonormal systems of wave functions appear naturally as the eigenfunctions of the Hamiltonian (of a bound system) in quantum mechanics that measures the energy levels, which are called the eigenvalues. The set of eigenvalues, in this case, is known as the spectrum of the Hamiltonian. In bra–ket notation this equality implies the resolution of the identity:

I=|φnφn|.

Here the eigenvalues are assumed to be discrete, but the set of eigenvalues of an observable can also be continuous. An example is the position operator, Template:Math. The spectrum of the position (in one dimension) is the entire real line and is called a continuous spectrum. However, unlike the Hamiltonian, the position operator lacks proper eigenfunctions. The conventional way to overcome this shortcoming is to widen the class of available functions by allowing distributions as well, i.e., to replace the Hilbert space with a rigged Hilbert space.Template:Sfn In this context, the position operator has a complete set of generalized eigenfunctions,Template:Sfn labeled by the points Template:Mvar of the real line, given by

φy(x)=δ(xy).

The generalized eigenfunctions of the position operator are called the eigenkets and are denoted by Template:Math.Template:Sfn

Similar considerations apply to any other (unbounded) self-adjoint operator with continuous spectrum and no degenerate eigenvalues, such as the momentum operator Template:Mvar. In that case, there is a set Template:Math of real numbers (the spectrum) and a collection of distributions Template:Mvar with Template:Math such that

Pφy=yφy.

That is, Template:Mvar are the generalized eigenvectors of Template:Mvar. If they form an "orthonormal basis" in the distribution sense, that is:

φy,φy=δ(yy),

then for any test function Template:Mvar,

ψ(x)=Ωc(y)φy(x)dy

where Template:Math. That is, there is a resolution of the identity

I=Ω|φyφy|dy

where the operator-valued integral is again understood in the weak sense. If the spectrum of Template:Mvar has both continuous and discrete parts, then the resolution of the identity involves a summation over the discrete spectrum and an integral over the continuous spectrum.

The delta function also has many more specialized applications in quantum mechanics, such as the delta potential models for a single and double potential well.

Structural mechanics

The delta function can be used in structural mechanics to describe transient loads or point loads acting on structures. The governing equation of a simple mass–spring system excited by a sudden force impulse Template:Mvar at time Template:Math can be writtenTemplate:SfnTemplate:Sfn md2ξdt2+kξ=Iδ(t), where Template:Mvar is the mass, Template:Mvar is the deflection, and Template:Mvar is the spring constant.

As another example, the equation governing the static deflection of a slender beam is, according to Euler–Bernoulli theory,

EId4wdx4=q(x),

where Template:Mvar is the bending stiffness of the beam, Template:Mvar is the deflection, Template:Mvar is the spatial coordinate, and Template:Math is the load distribution. If a beam is loaded by a point force Template:Mvar at Template:Math, the load distribution is written

q(x)=Fδ(xx0).

As the integration of the delta function results in the Heaviside step function, it follows that the static deflection of a slender beam subject to multiple point loads is described by a set of piecewise polynomials.

Also, a point moment acting on a beam can be described by delta functions. Consider two opposing point forces Template:Mvar at a distance Template:Mvar apart. They then produce a moment Template:Math acting on the beam. Now, let the distance Template:Mvar approach the limit zero, while Template:Mvar is kept constant. The load distribution, assuming a clockwise moment acting at Template:Math, is written

q(x)=limd0(Fδ(x)Fδ(xd))=limd0(Mdδ(x)Mdδ(xd))=Mlimd0δ(x)δ(xd)d=Mδ(x).

Point moments can thus be represented by the derivative of the delta function. Integration of the beam equation again results in piecewise polynomial deflection.

See also

Notes

Template:Reflist

References

Template:Refbegin

  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".. Reprinted, Dover Publications, 2004, Template:Isbn.
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Template:Cite thesis
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "citation/CS1"..
  • Script error: No such module "Template wrapper".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".

Template:Refend

External links

Template:ProbDistributions Template:Differential equations topics

Template:Good article

  1. Script error: No such module "Citation/CS1".
  2. A more complete historical account can be found in Script error: No such module "Footnotes"..
  3. Script error: No such module "Citation/CS1".
  4. Script error: No such module "citation/CS1".
  5. Script error: No such module "citation/CS1"., cf. Template:Google books and pp. 546–551. [[[:Template:Google books]] Original French text].
  6. Script error: No such module "citation/CS1".
  7. Script error: No such module "citation/CS1".
  8. Script error: No such module "citation/CS1".
  9. Script error: No such module "citation/CS1".
  10. See, for example, Script error: No such module "citation/CS1".
  11. Script error: No such module "citation/CS1".
  12. Script error: No such module "citation/CS1".
  13. Script error: No such module "Footnotes".
  14. Script error: No such module "Footnotes". See also Script error: No such module "Footnotes". for a different interpretation. Other conventions for the assigning the value of the Heaviside function at zero exist, and some of these are not consistent with what follows.
  15. Script error: No such module "citation/CS1".
  16. Script error: No such module "Template wrapper".
  17. Script error: No such module "citation/CS1".
  18. Script error: No such module "citation/CS1".
  19. Script error: No such module "citation/CS1".
  20. Further refinement is possible, namely to submersions, although these require a more involved change of variables formula.
  21. Script error: No such module "Template wrapper".
  22. Script error: No such module "citation/CS1".
  23. Script error: No such module "Citation/CS1".
  24. More generally, one only needs Template:Math to have an integrable radially symmetric decreasing rearrangement.
  25. Script error: No such module "citation/CS1".
  26. Script error: No such module "citation/CS1".
  27. The numerical factors depend on the conventions for the Fourier transform.
  28. In the terminology of Template:Harvtxt, the Fejér kernel is a Dirac sequence, whereas the Dirichlet kernel is not.
  29. Script error: No such module "citation/CS1".