In physics, engineering and mathematics, the Fourier transform (FT) is an integral transform that takes a function as input and outputs another function that describes the extent to which various frequencies are present in the original function. The output of the transform is a complex-valued function of frequency. The term Fourier transform refers to both this complex-valued function and the mathematical operation. When a distinction needs to be made the Fourier transform is sometimes called the frequency domain representation of the original function. The Fourier transform is analogous to decomposing the sound of a musical chord into the intensities of its constituent pitches.
Functions that are localized in the time domain have Fourier transforms that are spread out across the frequency domain and vice versa, a phenomenon known as the uncertainty principle. The critical case for this principle is the Gaussian function, of substantial importance in probability theory and statistics as well as in the study of physical phenomena exhibiting normal distribution (e.g., diffusion). The Fourier transform of a Gaussian function is another Gaussian function. Joseph Fourier introduced the transform in his study of heat transfer, where Gaussian functions appear as solutions of the heat equation.
The Fourier transform can be formally defined as an improper Riemann integral, making it an integral transform, although this definition is not suitable for many applications requiring a more sophisticated integration theory.^{[note 1]} For example, many relatively simple applications use the Dirac delta function, which can be treated formally as if it were a function, but the justification requires a mathematically more sophisticated viewpoint.^{[note 2]}
The Fourier transform can also be generalized to functions of several variables on Euclidean space, sending a function of 3-dimensional 'position space' to a function of 3-dimensional momentum (or a function of space and time to a function of 4-momentum). This idea makes the spatial Fourier transform very natural in the study of waves, as well as in quantum mechanics, where it is important to be able to represent wave solutions as functions of either position or momentum and sometimes both. In general, functions to which Fourier methods are applicable are complex-valued, and possibly vector-valued.^{[note 3]} Still further generalization is possible to functions on groups, which, besides the original Fourier transform on R or R^{n}, notably includes the discrete-time Fourier transform (DTFT, group = Z), the discrete Fourier transform (DFT, group = Z mod N) and the Fourier series or circular Fourier transform (group = S^{1}, the unit circle ≈ closed finite interval with endpoints identified). The latter is routinely employed to handle periodic functions. The fast Fourier transform (FFT) is an algorithm for computing the DFT.
The Fourier transform is an analysis process, decomposing a complex-valued function into its constituent frequencies and their amplitudes. The inverse process is synthesis, which recreates from its transform.
We can start with an analogy, the Fourier series, which analyzes on a bounded interval for some positive real number The constituent frequencies are a discrete set of harmonics at frequencies whose amplitude and phase are given by the analysis formula:
The analogy for a function can be obtained formally from the analysis formula by taking the limit as , while at the same time taking so that ^{[1]}^{[2]}^{[3]} Formally carrying this out, we obtain, for rapidly decreasing :^{[note 4]}^{[4]}
(Eq.1) |
It is easy to see, assuming the hypothesis of rapid decreasing, that the integral Eq.1 converges for all real , and (using the Riemann–Lebesgue lemma) that the transformed function is also rapidly decreasing. The validity of this definition for classes of functions that are not necessarily rapidly decreasing is discussed later in this section.
Evaluating Eq.1 for all values of produces the frequency-domain function. The complex number , in polar coordinates, conveys both amplitude and phase of frequency The intuitive interpretation of Eq.1 is that the effect of multiplying by is to subtract from every frequency component of function ^{[note 5]} Only the component that was at frequency can produce a non-zero value of the infinite integral, because (at least formally) all the other shifted components are oscillatory and integrate to zero. (see § Example)
The corresponding synthesis formula for such a function is:
(Eq.2) |
Eq.2 is a representation of as a weighted summation of complex exponential functions.
This is also known as the Fourier inversion theorem, and was first introduced in Fourier's Analytical Theory of Heat.^{[5]}^{[6]}^{[7]}^{[8]}
The functions and are referred to as a Fourier transform pair.^{[9]} A common notation for designating transform pairs is:^{[10]}
Until now, we have been dealing with Schwartz functions, which decay rapidly at infinity, with all derivatives. This excludes many functions of practical importance from the definition, such as the rect function. A measurable function is called (Lebesgue) integrable if the Lebesgue integral of its absolute value is finite:
Definition — The Fourier transform of a Lebesgue integrable function is defined by the formula Eq.1.
The integral Eq.1 is well-defined for all because of the assumption . (It can be shown that the function is bounded and uniformly continuous in the frequency domain, and moreover, by the Riemann–Lebesgue lemma, it is zero at infinity.)
However, the class of Lebesgue integrable functions is not ideal from the point of view of the Fourier transform because there is no easy characterization of the image, and thus no easy characterization of the inverse transform.
While Eq.1 defines the Fourier transform for (complex-valued) functions in , it is easy to see that it is not well-defined for other integrability classes, most importantly . For functions in , and with the conventions of Eq.1, the Fourier transform is a unitary operator with respect to the Hilbert inner product on , restricted to the dense subspace of integrable functions. Therefore, it admits a unique continuous extension to a unitary operator on , also called the Fourier transform. This extension is important in part because the Fourier transform preserves the space so that, unlike the case of , the Fourier transform and inverse transform are on the same footing, being transformations of the same space of functions to itself.
Importantly, for functions in , the Fourier transform is no longer given by Eq.1 (interpreted as a Lebesgue integral). For example, the function is in but not , so the integral Eq.1 diverges. In such cases, the Fourier transform can be obtained explicitly by regularizing the integral, and then passing to a limit. In practice, the integral is often regarded as an improper integral instead of a proper Lebesgue integral, but sometimes for convergence one needs to use weak limit or principal value instead of the (pointwise) limits implicit in an improper integral. Titchmarsh (1986) and Dym & McKean (1985) each gives three rigorous ways of extending the Fourier transform to square integrable functions using this procedure.
The conventions chosen in this article are those of harmonic analysis, and are characterized as the unique conventions such that the Fourier transform is both unitary on L^{2} and an algebra homomorphism from L^{1} to L^{∞}, without renormalizing the Lebesgue measure.^{[12]}
When the independent variable ( ) represents time (often denoted by ), the transform variable ( ) represents frequency (often denoted by ). For example, if time is measured in seconds, then frequency is in hertz. The Fourier transform can also be written in terms of angular frequency, whose units are radians per second.
The substitution into Eq.1 produces this convention, where function is relabeled
ordinary frequency ξ (Hz) | unitary | |
---|---|---|
angular frequency ω (rad/s) | unitary | |
non-unitary |
ordinary frequency ξ (Hz) | unitary | |
---|---|---|
angular frequency ω (rad/s) | unitary | |
non-unitary |
For , the Fourier transform can be defined on by Marcinkiewicz interpolation.
The Fourier transform can be defined on domains other than the real line. The Fourier transform on Euclidean space and the Fourier transform on locally abelian groups are discussed later in the article.
The Fourier transform can also be defined for tempered distributions, dual to the space of rapidly decreasing functions (Schwartz functions). A Schwartz function is a smooth function that decays at infinity, along with all of its derivatives. The space of Schwartz functions is denoted by , and its dual is the space of tempered distributions. It is easy to see, by differentiating under the integral and applying the Riemann-Lebesgue lemma, that the Fourier transform of a Schwartz function (defined by the formula Eq.1) is again a Schwartz function. The Fourier transform of a tempered distribution is defined by duality:
Many other characterizations of the Fourier transform exist. For example, one uses the Stone–von Neumann theorem: the Fourier transform is the unique unitary intertwiner for the symplectic and Euclidean Schrödinger representations of the Heisenberg group.
In 1822, Fourier claimed (see Joseph Fourier § The Analytic Theory of Heat) that any function, whether continuous or discontinuous, can be expanded into a series of sines.^{[13]} That important work was corrected and expanded upon by others to provide the foundation for the various forms of the Fourier transform used since.
In general, the coefficients are complex numbers, which have two equivalent forms (see Euler's formula):
The product with (Eq.2) has these forms:
It is noteworthy how easily the product was simplified using the polar form, and how easily the rectangular form was deduced by an application of Euler's formula.
Euler's formula introduces the possibility of negative And Eq.1 is defined Only certain complex-valued have transforms (See Analytic signal. A simple example is ) But negative frequency is necessary to characterize all other complex-valued found in signal processing, partial differential equations, radar, nonlinear optics, quantum mechanics, and others.
For a real-valued Eq.1 has the symmetry property (see § Conjugation below). This redundancy enables Eq.2 to distinguish from But of course it cannot tell us the actual sign of because and are indistinguishable on just the real numbers line.
The Fourier transform of a periodic function cannot be defined using the integral formula directly. In order for integral in Eq.1 to be defined the function must be absolutely integrable. Instead it is common to use Fourier series. It is possible to extend the definition to include periodic functions by viewing them as tempered distributions.
This makes it possible to see a connection between the Fourier series and the Fourier transform for periodic functions that have a convergent Fourier series. If is a periodic function, with period , that has a convergent Fourier series, then:
The Fourier transform of an integrable function can be sampled at regular intervals of arbitrary length These samples can be deduced from one cycle of a periodic function which has Fourier series coefficients proportional to those samples by the Poisson summation formula:
The integrability of ensures the periodic summation converges. Therefore, the samples can be determined by Fourier series analysis:
When has compact support, has a finite number of terms within the interval of integration. When does not have compact support, numerical evaluation of requires an approximation, such as tapering or truncating the number of terms.
The following figures provide a visual illustration of how the Fourier transform measures whether a frequency is present in a particular function. The depicted function oscillates at 3 Hz (if measures seconds) and tends quickly to 0. (The second factor in this equation is an envelope function that shapes the continuous sinusoid into a short pulse.). was specially chosen to have a real Fourier transform that can be easily plotted. The first image is its graph. In order to calculate we must integrate the product The next 2 images are the real and imaginary parts of that product. The real part of the integrand has a non-negative average value, because the alternating signs of and oscillate at the same rate and same phase, whereas and are same rate but orthogonal phase. The result is that when you integrate the real part of the integrand you get a relatively large number (in this case ). Also, when you try to measure a frequency that is not present, as in the case when we look at both real and imaginary component of the product vary rapidly between positive and negative values. Therefore, the integral is very small and the value for the Fourier transform for that frequency is nearly zero. The general situation is usually more complicated than this, but heuristically this is how the Fourier transform measures how much of an individual frequency is present in a function
To re-enforce an earlier point, the reason for the response at Hz is because and are indistinguishable. The transform of would have just one response, whose amplitude is the integral of the smooth envelope: whereas (second graph above) is
Let and represent integrable functions Lebesgue-measurable on the real line satisfying:
The Fourier transform has the following basic properties:^{[14]}
When the real and imaginary parts of a complex function are decomposed into their even and odd parts, there are four components, denoted below by the subscripts RE, RO, IE, and IO. And there is a one-to-one mapping between the four components of a complex time function and the four components of its complex frequency transform:
From this, various relationships are apparent, for example:
In particular, if is real, then is even symmetric (aka Hermitian function):
And if is purely imaginary, then is odd symmetric:
Substituting in the definition, we obtain:
The integral of over its domain is known as the average value or DC bias of the function.
Under suitable conditions on the function , it can be recovered from its Fourier transform . Indeed, denoting the Fourier transform operator by , so , then for suitable functions, applying the Fourier transform twice simply flips the function: , which can be interpreted as "reversing time". Since reversing time is two-periodic, applying this twice yields , so the Fourier transform operator is four-periodic, and similarly the inverse Fourier transform can be obtained by applying the Fourier transform three times: . In particular the Fourier transform is invertible (under suitable conditions).
More precisely, defining the parity operator such that , we have:
This fourfold periodicity of the Fourier transform is similar to a rotation of the plane by 90°, particularly as the two-fold iteration yields a reversal, and in fact this analogy can be made precise. While the Fourier transform can simply be interpreted as switching the time domain and the frequency domain, with the inverse Fourier transform switching them back, more geometrically it can be interpreted as a rotation by 90° in the time–frequency domain (considering time as the x-axis and frequency as the y-axis), and the Fourier transform can be generalized to the fractional Fourier transform, which involves rotations by other angles. This can be further generalized to linear canonical transformations, which can be visualized as the action of the special linear group SL_{2}(R) on the time–frequency plane, with the preserved symplectic form corresponding to the uncertainty principle, below. This approach is particularly studied in signal processing, under time–frequency analysis.
The frequency variable must have inverse units to the units of the original function's domain (typically named t or x). For example, if t is measured in seconds, ξ should be in cycles per second or hertz. If the scale of time is in units of 2π seconds, then another Greek letter ω typically is used instead to represent angular frequency (where ω = 2πξ) in units of radians per second. If using x for units of length, then ξ must be in inverse length, e.g., wavenumbers. That is to say, there are two versions of the real line: one which is the range of t and measured in units of t, and the other which is the range of ξ and measured in inverse units to the units of t. These two distinct versions of the real line cannot be equated with each other. Therefore, the Fourier transform goes from one space of functions to a different space of functions: functions which have a different domain of definition.
In general, ξ must always be taken to be a linear form on the space of its domain, which is to say that the second real line is the dual space of the first real line. See the article on linear algebra for a more formal explanation and for more details. This point of view becomes essential in generalizations of the Fourier transform to general symmetry groups, including the case of Fourier series.
That there is no one preferred way (often, one says "no canonical way") to compare the two versions of the real line which are involved in the Fourier transform—fixing the units on one line does not force the scale of the units on the other line—is the reason for the plethora of rival conventions on the definition of the Fourier transform. The various definitions resulting from different choices of units differ by various constants.
In other conventions, the Fourier transform has i in the exponent instead of −i, and vice versa for the inversion formula. This convention is common in modern physics^{[15]} and is the default for Wolfram Alpha, and does not mean that the frequency has become negative, since there is no canonical definition of positivity for frequency of a complex wave. It simply means that is the amplitude of the wave instead of the wave (the former, with its minus sign, is often seen in the time dependence for Sinusoidal plane-wave solutions of the electromagnetic wave equation, or in the time dependence for quantum wave functions). Many of the identities involving the Fourier transform remain valid in those conventions, provided all terms that explicitly involve i have it replaced by −i. In Electrical engineering the letter j is typically used for the imaginary unit instead of i because i is used for current.
When using dimensionless units, the constant factors might not even be written in the transform definition. For instance, in probability theory, the characteristic function Φ of the probability density function f of a random variable X of continuous type is defined without a negative sign in the exponential, and since the units of x are ignored, there is no 2π either:
(In probability theory, and in mathematical statistics, the use of the Fourier—Stieltjes transform is preferred, because so many random variables are not of continuous type, and do not possess a density function, and one must treat not functions but distributions, i.e., measures which possess "atoms".)
From the higher point of view of group characters, which is much more abstract, all these arbitrary choices disappear, as will be explained in the later section of this article, which treats the notion of the Fourier transform of a function on a locally compact Abelian group.
The Fourier transform may be defined in some cases for non-integrable functions, but the Fourier transforms of integrable functions have several strong properties.
The Fourier transform f̂ of any integrable function f is uniformly continuous and^{[16]}
By the Riemann–Lebesgue lemma,^{[11]}
However, need not be integrable. For example, the Fourier transform of the rectangular function, which is integrable, is the sinc function, which is not Lebesgue integrable, because its improper integrals behave analogously to the alternating harmonic series, in converging to a sum without being absolutely convergent.
It is not generally possible to write the inverse transform as a Lebesgue integral. However, when both f and are integrable, the inverse equality
Let f(x) and g(x) be integrable, and let f̂(ξ) and ĝ(ξ) be their Fourier transforms. If f(x) and g(x) are also square-integrable, then the Parseval formula follows:^{[17]}
The Plancherel theorem, which follows from the above, states that^{[18]}
Plancherel's theorem makes it possible to extend the Fourier transform, by a continuity argument, to a unitary operator on L^{2}(R). On L^{1}(R) ∩ L^{2}(R), this extension agrees with original Fourier transform defined on L^{1}(R), thus enlarging the domain of the Fourier transform to L^{1}(R) + L^{2}(R) (and consequently to L^{p}(R) for 1 ≤ p ≤ 2). Plancherel's theorem has the interpretation in the sciences that the Fourier transform preserves the energy of the original quantity. The terminology of these formulas is not quite standardised. Parseval's theorem was proved only for Fourier series, and was first proved by Lyapunov. But Parseval's formula makes sense for the Fourier transform as well, and so even though in the context of the Fourier transform it was proved by Plancherel, it is still often referred to as Parseval's formula, or Parseval's relation, or even Parseval's theorem.
See Pontryagin duality for a general formulation of this concept in the context of locally compact abelian groups.
The Poisson summation formula (PSF) is an equation that relates the Fourier series coefficients of the periodic summation of a function to values of the function's continuous Fourier transform. The Poisson summation formula says that for sufficiently regular functions f,
It has a variety of useful forms that are derived from the basic one by application of the Fourier transform's scaling and time-shifting properties. The formula has applications in engineering, physics, and number theory. The frequency-domain dual of the standard Poisson summation formula is also called the discrete-time Fourier transform.
Poisson summation is generally associated with the physics of periodic media, such as heat conduction on a circle. The fundamental solution of the heat equation on a circle is called a theta function. It is used in number theory to prove the transformation properties of theta functions, which turn out to be a type of modular form, and it is connected more generally to the theory of automorphic forms where it appears on one side of the Selberg trace formula.
Suppose f(x) is an absolutely continuous differentiable function, and both f and its derivative f′ are integrable. Then the Fourier transform of the derivative is given by
Analogically, , so
By applying the Fourier transform and using these formulas, some ordinary differential equations can be transformed into algebraic equations, which are much easier to solve. These formulas also give rise to the rule of thumb "f(x) is smooth if and only if f̂(ξ) quickly falls to 0 for |ξ| → ∞." By using the analogous rules for the inverse Fourier transform, one can also say "f(x) quickly falls to 0 for |x| → ∞ if and only if f̂(ξ) is smooth."
The Fourier transform translates between convolution and multiplication of functions. If f(x) and g(x) are integrable functions with Fourier transforms f̂(ξ) and ĝ(ξ) respectively, then the Fourier transform of the convolution is given by the product of the Fourier transforms f̂(ξ) and ĝ(ξ) (under other conventions for the definition of the Fourier transform a constant factor may appear).
This means that if:
In linear time invariant (LTI) system theory, it is common to interpret g(x) as the impulse response of an LTI system with input f(x) and output h(x), since substituting the unit impulse for f(x) yields h(x) = g(x). In this case, ĝ(ξ) represents the frequency response of the system.
Conversely, if f(x) can be decomposed as the product of two square integrable functions p(x) and q(x), then the Fourier transform of f(x) is given by the convolution of the respective Fourier transforms p̂(ξ) and q̂(ξ).
In an analogous manner, it can be shown that if h(x) is the cross-correlation of f(x) and g(x):
As a special case, the autocorrelation of function f(x) is:
The Fourier transform is a linear transform which has eigenfunctions obeying with
A set of eigenfunctions is found by noting that the homogeneous differential equation
More generally, a set of eigenfunctions is also found by noting that the differentiation rules imply that the ordinary differential equation
Under this convention for the Fourier transform, we have that
In other words, the Hermite functions form a complete orthonormal system of eigenfunctions for the Fourier transform on L^{2}(R).^{[14]}^{[21]} However, this choice of eigenfunctions is not unique. Because of there are only four different eigenvalues of the Fourier transform (the fourth roots of unity ±1 and ±i) and any linear combination of eigenfunctions with the same eigenvalue gives another eigenfunction.^{[22]} As a consequence of this, it is possible to decompose L^{2}(R) as a direct sum of four spaces H_{0}, H_{1}, H_{2}, and H_{3} where the Fourier transform acts on He_{k} simply by multiplication by i^{k}.
Since the complete set of Hermite functions ψ_{n} provides a resolution of the identity they diagonalize the Fourier operator, i.e. the Fourier transform can be represented by such a sum of terms weighted by the above eigenvalues, and these sums can be explicitly summed:
This approach to define the Fourier transform was first proposed by Norbert Wiener.^{[23]} Among other properties, Hermite functions decrease exponentially fast in both frequency and time domains, and they are thus used to define a generalization of the Fourier transform, namely the fractional Fourier transform used in time–frequency analysis.^{[24]} In physics, this transform was introduced by Edward Condon.^{[25]} This change of basis functions becomes possible because the Fourier transform is a unitary transform when using the right conventions. Consequently, under the proper conditions it may be expected to result from a self-adjoint generator via^{[26]}
The operator is the number operator of the quantum harmonic oscillator written as^{[27]}^{[28]}
It can be interpreted as the generator of fractional Fourier transforms for arbitrary values of t, and of the conventional continuous Fourier transform for the particular value with the Mehler kernel implementing the corresponding active transform. The eigenfunctions of are the Hermite functions which are therefore also eigenfunctions of
Upon extending the Fourier transform to distributions the Dirac comb is also an eigenfunction of the Fourier transform.
The Heisenberg group is a certain group of unitary operators on the Hilbert space L^{2}(R) of square integrable complex valued functions f on the real line, generated by the translations (T_{y} f)(x) = f (x + y) and multiplication by e^{i2πξx}, (M_{ξ} f)(x) = e^{i2πξx} f (x). These operators do not commute, as their (group) commutator is
Denote the Heisenberg group by H_{1}. The above procedure describes not only the group structure, but also a standard unitary representation of H_{1} on a Hilbert space, which we denote by ρ : H_{1} → B(L^{2}(R)). Define the linear automorphism of R^{2} by
According to the Stone–von Neumann theorem, the unitary representations ρ and ρ ∘ j are unitarily equivalent, so there is a unique intertwiner W ∈ U(L^{2}(R)) such that
Many of the standard properties of the Fourier transform are immediate consequences of this more general framework.^{[29]} For example, the square of the Fourier transform, W^{2}, is an intertwiner associated with J^{2} = −I, and so we have (W^{2}f)(x) = f (−x) is the reflection of the original function f.
The integral for the Fourier transform
The Paley–Wiener theorem says that f is smooth (i.e., n-times differentiable for all positive integers n) and compactly supported if and only if f̂ (σ + iτ) is a holomorphic function for which there exists a constant a > 0 such that for any integer n ≥ 0,
(If f is not smooth, but only L^{2}, the statement still holds provided n = 0.^{[32]}) The space of such functions of a complex variable is called the Paley—Wiener space. This theorem has been generalised to semisimple Lie groups.^{[33]}
If f is supported on the half-line t ≥ 0, then f is said to be "causal" because the impulse response function of a physically realisable filter must have this property, as no effect can precede its cause. Paley and Wiener showed that then f̂ extends to a holomorphic function on the complex lower half-plane τ < 0 which tends to zero as τ goes to infinity.^{[34]} The converse is false and it is not known how to characterise the Fourier transform of a causal function.^{[35]}
The Fourier transform f̂(ξ) is related to the Laplace transform F(s), which is also used for the solution of differential equations and the analysis of filters.
It may happen that a function f for which the Fourier integral does not converge on the real axis at all, nevertheless has a complex Fourier transform defined in some region of the complex plane.
For example, if f(t) is of exponential growth, i.e.,
The more usual version ("one-sided") of the Laplace transform is
If f is also causal, and analytical, then: Thus, extending the Fourier transform to the complex domain means it includes the Laplace transform as a special case in the case of causal functions—but with the change of variable s = i2πξ.
From another, perhaps more classical viewpoint, the Laplace transform by its form involves an additional exponential regulating term which lets it converge outside of the imaginary line where the Fourier transform is defined. As such it can converge for at most exponentially divergent series and integrals, whereas the original Fourier decomposition cannot, enabling analysis of systems with divergent or critical elements. Two particular examples from linear signal processing are the construction of allpass filter networks from critical comb and mitigating filters via exact pole-zero cancellation on the unit circle. Such designs are common in audio processing, where highly nonlinear phase response is sought for, as in reverb.
Furthermore, when extended pulselike impulse responses are sought for signal processing work, the easiest way to produce them is to have one circuit which produces a divergent time response, and then to cancel its divergence through a delayed opposite and compensatory response. There, only the delay circuit in-between admits a classical Fourier description, which is critical. Both the circuits to the side are unstable, and do not admit a convergent Fourier decomposition. However, they do admit a Laplace domain description, with identical half-planes of convergence in the complex plane (or in the discrete case, the Z-plane), wherein their effects cancel.
In modern mathematics the Laplace transform is conventionally subsumed under the aegis Fourier methods. Both of them are subsumed by the far more general, and more abstract, idea of harmonic analysis.
Still with , if is complex analytic for a ≤ τ ≤ b, then
Theorem: If f(t) = 0 for t < 0, and |f(t)| < Ce^{a|t|} for some constants C, a > 0, then
This theorem implies the Mellin inversion formula for the Laplace transformation,^{[36]}
The hypotheses can be weakened, as in the results of Carleson and Hunt, to f(t) e^{−at} being L^{1}, provided that f is of bounded variation in a closed neighborhood of t (cf. Dirichlet–Dini theorem), the value of f at t is taken to be the arithmetic mean of the left and right limits, and provided that the integrals are taken in the sense of Cauchy principal values.^{[38]}
L^{2} versions of these inversion formulas are also available.^{[39]}
The Fourier transform can be defined in any arbitrary number of dimensions n. As with the one-dimensional case, there are many conventions. For an integrable function f(x), this article takes the definition:
All of the basic properties listed above hold for the n-dimensional Fourier transform, as do Plancherel's and Parseval's theorem. When the function is integrable, the Fourier transform is still uniformly continuous and the Riemann–Lebesgue lemma holds.^{[11]}
Generally speaking, the more concentrated f(x) is, the more spread out its Fourier transform f̂(ξ) must be. In particular, the scaling property of the Fourier transform may be seen as saying: if we squeeze a function in x, its Fourier transform stretches out in ξ. It is not possible to arbitrarily concentrate both a function and its Fourier transform.
The trade-off between the compaction of a function and its Fourier transform can be formalized in the form of an uncertainty principle by viewing a function and its Fourier transform as conjugate variables with respect to the symplectic form on the time–frequency domain: from the point of view of the linear canonical transformation, the Fourier transform is rotation by 90° in the time–frequency domain, and preserves the symplectic form.
Suppose f(x) is an integrable and square-integrable function. Without loss of generality, assume that f(x) is normalized:
It follows from the Plancherel theorem that f̂(ξ) is also normalized.
The spread around x = 0 may be measured by the dispersion about zero^{[40]} defined by
In probability terms, this is the second moment of |f(x)|^{2} about zero.
The uncertainty principle states that, if f(x) is absolutely continuous and the functions x·f(x) and f′(x) are square integrable, then^{[14]}
The equality is attained only in the case
In fact, this inequality implies that:
In quantum mechanics, the momentum and position wave functions are Fourier transform pairs, up to a factor of the Planck constant. With this constant properly taken into account, the inequality above becomes the statement of the Heisenberg uncertainty principle.^{[42]}
A stronger uncertainty principle is the Hirschman uncertainty principle, which is expressed as:
Fourier's original formulation of the transform did not use complex numbers, but rather sines and cosines. Statisticians and others still use this form. An absolutely integrable function f for which Fourier inversion holds can be expanded in terms of genuine frequencies (avoiding negative frequencies, which are sometimes considered hard to interpret physically^{[43]}) λ by
This is called an expansion as a trigonometric integral, or a Fourier integral expansion. The coefficient functions a and b can be found by using variants of the Fourier cosine transform and the Fourier sine transform (the normalisations are, again, not standardised):
Older literature refers to the two transform functions, the Fourier cosine transform, a, and the Fourier sine transform, b.
The function f can be recovered from the sine and cosine transform using
Let the set of homogeneous harmonic polynomials of degree k on R^{n} be denoted by A_{k}. The set A_{k} consists of the solid spherical harmonics of degree k. The solid spherical harmonics play a similar role in higher dimensions to the Hermite polynomials in dimension one. Specifically, if f(x) = e^{−π|x|2}P(x) for some P(x) in A_{k}, then f̂(ξ) = i^{−k} f(ξ). Let the set H_{k} be the closure in L^{2}(R^{n}) of linear combinations of functions of the form f(|x|)P(x) where P(x) is in A_{k}. The space L^{2}(R^{n}) is then a direct sum of the spaces H_{k} and the Fourier transform maps each space H_{k} to itself and is possible to characterize the action of the Fourier transform on each space H_{k}.^{[11]}
Let f(x) = f_{0}(|x|)P(x) (with P(x) in A_{k}), then
Here J_{(n + 2k − 2)/2} denotes the Bessel function of the first kind with order n + 2k − 2/2. When k = 0 this gives a useful formula for the Fourier transform of a radial function.^{[47]} This is essentially the Hankel transform. Moreover, there is a simple recursion relating the cases n + 2 and n^{[48]} allowing to compute, e.g., the three-dimensional Fourier transform of a radial function from the one-dimensional one.
In higher dimensions it becomes interesting to study restriction problems for the Fourier transform. The Fourier transform of an integrable function is continuous and the restriction of this function to any set is defined. But for a square-integrable function the Fourier transform could be a general class of square integrable functions. As such, the restriction of the Fourier transform of an L^{2}(R^{n}) function cannot be defined on sets of measure 0. It is still an active area of study to understand restriction problems in L^{p} for 1 < p < 2. It is possible in some cases to define the restriction of a Fourier transform to a set S, provided S has non-zero curvature. The case when S is the unit sphere in R^{n} is of particular interest. In this case the Tomas–Stein restriction theorem states that the restriction of the Fourier transform to the unit sphere in R^{n} is a bounded operator on L^{p} provided 1 ≤ p ≤ 2n + 2/n + 3.
One notable difference between the Fourier transform in 1 dimension versus higher dimensions concerns the partial sum operator. Consider an increasing collection of measurable sets E_{R} indexed by R ∈ (0,∞): such as balls of radius R centered at the origin, or cubes of side 2R. For a given integrable function f, consider the function f_{R} defined by:
Suppose in addition that f ∈ L^{p}(R^{n}). For n = 1 and 1 < p < ∞, if one takes E_{R} = (−R, R), then f_{R} converges to f in L^{p} as R tends to infinity, by the boundedness of the Hilbert transform. Naively one may hope the same holds true for n > 1. In the case that E_{R} is taken to be a cube with side length R, then convergence still holds. Another natural candidate is the Euclidean ball E_{R} = {ξ : |ξ| < R}. In order for this partial sum operator to converge, it is necessary that the multiplier for the unit ball be bounded in L^{p}(R^{n}). For n ≥ 2 it is a celebrated theorem of Charles Fefferman that the multiplier for the unit ball is never bounded unless p = 2.^{[23]} In fact, when p ≠ 2, this shows that not only may f_{R} fail to converge to f in L^{p}, but for some functions f ∈ L^{p}(R^{n}), f_{R} is not even an element of L^{p}.
The definition of the Fourier transform by the integral formula
The Fourier transform F : L^{1}(R^{n}) → L^{∞}(R^{n}) is a bounded operator. This follows from the observation that
Since compactly supported smooth functions are integrable and dense in L^{2}(R^{n}), the Plancherel theorem allows us to extend the definition of the Fourier transform to general functions in L^{2}(R^{n}) by continuity arguments. The Fourier transform in L^{2}(R^{n}) is no longer given by an ordinary Lebesgue integral, although it can be computed by an improper integral, here meaning that for an L^{2} function f,
Many of the properties of the Fourier transform in L^{1} carry over to L^{2}, by a suitable limiting argument.
Furthermore, F : L^{2}(R^{n}) → L^{2}(R^{n}) is a unitary operator.^{[50]} For an operator to be unitary it is sufficient to show that it is bijective and preserves the inner product, so in this case these follow from the Fourier inversion theorem combined with the fact that for any f, g ∈ L^{2}(R^{n}) we have
In particular, the image of L^{2}(R^{n}) is itself under the Fourier transform.
The definition of the Fourier transform can be extended to functions in L^{p}(R^{n}) for 1 ≤ p ≤ 2 by decomposing such functions into a fat tail part in L^{2} plus a fat body part in L^{1}. In each of these spaces, the Fourier transform of a function in L^{p}(R^{n}) is in L^{q}(R^{n}), where q = p/p − 1 is the Hölder conjugate of p (by the Hausdorff–Young inequality). However, except for p = 2, the image is not easily characterized. Further extensions become more technical. The Fourier transform of functions in L^{p} for the range 2 < p < ∞ requires the study of distributions.^{[16]} In fact, it can be shown that there are functions in L^{p} with p > 2 so that the Fourier transform is not defined as a function.^{[11]}
One might consider enlarging the domain of the Fourier transform from L^{1} + L^{2} by considering generalized functions, or distributions. A distribution on R^{n} is a continuous linear functional on the space C_{c}(R^{n}) of compactly supported smooth functions, equipped with a suitable topology. The strategy is then to consider the action of the Fourier transform on C_{c}(R^{n}) and pass to distributions by duality. The obstruction to doing this is that the Fourier transform does not map C_{c}(R^{n}) to C_{c}(R^{n}). In fact the Fourier transform of an element in C_{c}(R^{n}) can not vanish on an open set; see the above discussion on the uncertainty principle. The right space here is the slightly larger space of Schwartz functions. The Fourier transform is an automorphism on the Schwartz space, as a topological vector space, and thus induces an automorphism on its dual, the space of tempered distributions.^{[11]} The tempered distributions include all the integrable functions mentioned above, as well as well-behaved functions of polynomial growth and distributions of compact support.
For the definition of the Fourier transform of a tempered distribution, let f and g be integrable functions, and let f̂ and ĝ be their Fourier transforms respectively. Then the Fourier transform obeys the following multiplication formula,^{[11]}
Every integrable function f defines (induces) a distribution T_{f} by the relation
Distributions can be differentiated and the above-mentioned compatibility of the Fourier transform with differentiation and convolution remains true for tempered distributions.
The Fourier transform of a finite Borel measure μ on R^{n} is given by:^{[51]}
This transform continues to enjoy many of the properties of the Fourier transform of integrable functions. One notable difference is that the Riemann–Lebesgue lemma fails for measures.^{[16]} In the case that dμ = f(x) dx, then the formula above reduces to the usual definition for the Fourier transform of f. In the case that μ is the probability distribution associated to a random variable X, the Fourier–Stieltjes transform is closely related to the characteristic function, but the typical conventions in probability theory take e^{iξx} instead of e^{−i2πξx}.^{[14]} In the case when the distribution has a probability density function this definition reduces to the Fourier transform applied to the probability density function, again with a different choice of constants.
The Fourier transform may be used to give a characterization of measures. Bochner's theorem characterizes which functions may arise as the Fourier–Stieltjes transform of a positive measure on the circle.^{[16]}
Furthermore, the Dirac delta function, although not a function, is a finite Borel measure. Its Fourier transform is a constant function (whose specific value depends upon the form of the Fourier transform used).
The Fourier transform may be generalized to any locally compact abelian group. A locally compact abelian group is an abelian group that is at the same time a locally compact Hausdorff topological space so that the group operation is continuous. If G is a locally compact abelian group, it has a translation invariant measure μ, called Haar measure. For a locally compact abelian group G, the set of irreducible, i.e. one-dimensional, unitary representations are called its characters. With its natural group structure and the topology of uniform convergence on compact sets (that is, the topology induced by the compact-open topology on the space of all continuous functions from to the circle group), the set of characters Ĝ is itself a locally compact abelian group, called the Pontryagin dual of G. For a function f in L^{1}(G), its Fourier transform is defined by^{[16]}
The Riemann–Lebesgue lemma holds in this case; f̂(ξ) is a function vanishing at infinity on Ĝ.
The Fourier transform on T = R/Z is an example; here T is a locally compact abelian group, and the Haar measure μ on T can be thought of as the Lebesgue measure on [0,1). Consider the representation of T on the complex plane C that is a 1-dimensional complex vector space. There are a group of representations (which are irreducible since C is 1-dim) where for .
The character of such representation, that is the trace of for each and , is itself. In the case of representation of finite group, the character table of the group G are rows of vectors such that each row is the character of one irreducible representation of G, and these vectors form an orthonormal basis of the space of class functions that map from G to C by Schur's lemma. Now the group T is no longer finite but still compact, and it preserves the orthonormality of character table. Each row of the table is the function of and the inner product between two class functions (all functions being class functions since T is abelian) is defined as with the normalizing factor . The sequence is an orthonormal basis of the space of class functions .
For any representation V of a finite group G, can be expressed as the span ( are the irreps of G), such that . Similarly for and , . The Pontriagin dual is and for , is its Fourier transform for .
The Fourier transform is also a special case of Gelfand transform. In this particular context, it is closely related to the Pontryagin duality map defined above.
Given an abelian locally compact Hausdorff topological group G, as before we consider space L^{1}(G), defined using a Haar measure. With convolution as multiplication, L^{1}(G) is an abelian Banach algebra. It also has an involution * given by
Taking the completion with respect to the largest possibly C*-norm gives its enveloping C*-algebra, called the group C*-algebra C*(G) of G. (Any C*-norm on L^{1}(G) is bounded by the L^{1} norm, therefore their supremum exists.)
Given any abelian C*-algebra A, the Gelfand transform gives an isomorphism between A and C_{0}(A^), where A^ is the multiplicative linear functionals, i.e. one-dimensional representations, on A with the weak-* topology. The map is simply given by
The Fourier transform can also be defined for functions on a non-abelian group, provided that the group is compact. Removing the assumption that the underlying group is abelian, irreducible unitary representations need not always be one-dimensional. This means the Fourier transform on a non-abelian group takes values as Hilbert space operators.^{[52]} The Fourier transform on compact groups is a major tool in representation theory^{[53]} and non-commutative harmonic analysis.
Let G be a compact Hausdorff topological group. Let Σ denote the collection of all isomorphism classes of finite-dimensional irreducible unitary representations, along with a definite choice of representation U^{(σ)} on the Hilbert space H_{σ} of finite dimension d_{σ} for each σ ∈ Σ. If μ is a finite Borel measure on G, then the Fourier–Stieltjes transform of μ is the operator on H_{σ} defined by
The mapping
The Peter–Weyl theorem holds, and a version of the Fourier inversion formula (Plancherel's theorem) follows: if f ∈ L^{2}(G), then
The generalization of the Fourier transform to the noncommutative situation has also in part contributed to the development of noncommutative geometry.^{[citation needed]} In this context, a categorical generalization of the Fourier transform to noncommutative groups is Tannaka–Krein duality, which replaces the group of characters with the category of representations. However, this loses the connection with harmonic functions.
In signal processing terms, a function (of time) is a representation of a signal with perfect time resolution, but no frequency information, while the Fourier transform has perfect frequency resolution, but no time information: the magnitude of the Fourier transform at a point is how much frequency content there is, but location is only given by phase (argument of the Fourier transform at a point), and standing waves are not localized in time – a sine wave continues out to infinity, without decaying. This limits the usefulness of the Fourier transform for analyzing signals that are localized in time, notably transients, or any signal of finite extent.
As alternatives to the Fourier transform, in time–frequency analysis, one uses time–frequency transforms or time–frequency distributions to represent signals in a form that has some time information and some frequency information – by the uncertainty principle, there is a trade-off between these. These can be generalizations of the Fourier transform, such as the short-time Fourier transform, fractional Fourier transform, Synchrosqueezing Fourier transform,^{[54]} or other functions to represent signals, as in wavelet transforms and chirplet transforms, with the wavelet analog of the (continuous) Fourier transform being the continuous wavelet transform.^{[24]}
Linear operations performed in one domain (time or frequency) have corresponding operations in the other domain, which are sometimes easier to perform. The operation of differentiation in the time domain corresponds to multiplication by the frequency,^{[note 7]} so some differential equations are easier to analyze in the frequency domain. Also, convolution in the time domain corresponds to ordinary multiplication in the frequency domain (see Convolution theorem). After performing the desired operations, transformation of the result can be made back to the time domain. Harmonic analysis is the systematic study of the relationship between the frequency and time domains, including the kinds of functions or operations that are "simpler" in one or the other, and has deep connections to many areas of modern mathematics.
Perhaps the most important use of the Fourier transformation is to solve partial differential equations. Many of the equations of the mathematical physics of the nineteenth century can be treated this way. Fourier studied the heat equation, which in one dimension and in dimensionless units is
As usual, the problem is not to find a solution: there are infinitely many. The problem is that of the so-called "boundary problem": find a solution which satisfies the "boundary conditions"
Here, f and g are given functions. For the heat equation, only one boundary condition can be required (usually the first one). But for the wave equation, there are still infinitely many solutions y which satisfy the first boundary condition. But when one imposes both conditions, there is only one possible solution.
It is easier to find the Fourier transform ŷ of the solution than to find the solution directly. This is because the Fourier transformation takes differentiation into multiplication by the Fourier-dual variable, and so a partial differential equation applied to the original function is transformed into multiplication by polynomial functions of the dual variables applied to the transformed function. After ŷ is determined, we can apply the inverse Fourier transformation to find y.
Fourier's method is as follows. First, note that any function of the forms
Second, note that therefore any integral
Now this resembles the formula for the Fourier synthesis of a function. In fact, this is the real inverse Fourier transform of a_{±} and b_{±} in the variable x.
The third step is to examine how to find the specific unknown coefficient functions a_{±} and b_{±} that will lead to y satisfying the boundary conditions. We are interested in the values of these solutions at t = 0. So we will set t = 0. Assuming that the conditions needed for Fourier inversion are satisfied, we can then find the Fourier sine and cosine transforms (in the variable x) of both sides and obtain
Similarly, taking the derivative of y with respect to t and then applying the Fourier sine and cosine transformations yields
These are four linear equations for the four unknowns a_{±} and b_{±}, in terms of the Fourier sine and cosine transforms of the boundary conditions, which are easily solved by elementary algebra, provided that these transforms can be found.
In summary, we chose a set of elementary solutions, parametrized by ξ, of which the general solution would be a (continuous) linear combination in the form of an integral over the parameter ξ. But this integral was in the form of a Fourier integral. The next step was to express the boundary conditions in terms of these integrals, and set them equal to the given functions f and g. But these expressions also took the form of a Fourier integral because of the properties of the Fourier transform of a derivative. The last step was to exploit Fourier inversion by applying the Fourier transformation to both sides, thus obtaining expressions for the coefficient functions a_{±} and b_{±} in terms of the given boundary conditions f and g.
From a higher point of view, Fourier's procedure can be reformulated more conceptually. Since there are two variables, we will use the Fourier transformation in both x and t rather than operate as Fourier did, who only transformed in the spatial variables. Note that ŷ must be considered in the sense of a distribution since y(x, t) is not going to be L^{1}: as a wave, it will persist through time and thus is not a transient phenomenon. But it will be bounded and so its Fourier transform can be defined as a distribution. The operational properties of the Fourier transformation that are relevant to this equation are that it takes differentiation in x to multiplication by i2πξ and differentiation with respect to t to multiplication by i2πf where f is the frequency. Then the wave equation becomes an algebraic equation in ŷ:
We may as well consider the distributions supported on the conic that are given by distributions of one variable on the line ξ = f plus distributions on the line ξ = −f as follows: if Φ is any test function,
Then Fourier inversion gives, for the boundary conditions, something very similar to what we had more concretely above (put Φ(ξ, f) = e^{i2π(xξ+tf)}, which is clearly of polynomial growth):
Now, as before, applying the one-variable Fourier transformation in the variable x to these functions of x yields two equations in the two unknown distributions s_{±} (which can be taken to be ordinary functions if the boundary conditions are L^{1} or L^{2}).
From a calculational point of view, the drawback of course is that one must first calculate the Fourier transforms of the boundary conditions, then assemble the solution from these, and then calculate an inverse Fourier transform. Closed form formulas are rare, except when there is some geometric symmetry that can be exploited, and the numerical calculations are difficult because of the oscillatory nature of the integrals, which makes convergence slow and hard to estimate. For practical calculations, other methods are often used.
The twentieth century has seen the extension of these methods to all linear partial differential equations with polynomial coefficients, and by extending the notion of Fourier transformation to include Fourier integral operators, some non-linear equations as well.
The Fourier transform is also used in nuclear magnetic resonance (NMR) and in other kinds of spectroscopy, e.g. infrared (FTIR). In NMR an exponentially shaped free induction decay (FID) signal is acquired in the time domain and Fourier-transformed to a Lorentzian line-shape in the frequency domain. The Fourier transform is also used in magnetic resonance imaging (MRI) and mass spectrometry.
The Fourier transform is useful in quantum mechanics in at least two different ways. To begin with, the basic conceptual structure of quantum mechanics postulates the existence of pairs of complementary variables, connected by the Heisenberg uncertainty principle. For example, in one dimension, the spatial variable q of, say, a particle, can only be measured by the quantum mechanical "position operator" at the cost of losing information about the momentum p of the particle. Therefore, the physical state of the particle can either be described by a function, called "the wave function", of q or by a function of p but not by a function of both variables. The variable p is called the conjugate variable to q. In classical mechanics, the physical state of a particle (existing in one dimension, for simplicity of exposition) would be given by assigning definite values to both p and q simultaneously. Thus, the set of all possible physical states is the two-dimensional real vector space with a p-axis and a q-axis called the phase space.
In contrast, quantum mechanics chooses a polarisation of this space in the sense that it picks a subspace of one-half the dimension, for example, the q-axis alone, but instead of considering only points, takes the set of all complex-valued "wave functions" on this axis. Nevertheless, choosing the p-axis is an equally valid polarisation, yielding a different representation of the set of possible physical states of the particle. Both representations of the wavefunction are related by a Fourier transform, such that
Physically realisable states are L^{2}, and so by the Plancherel theorem, their Fourier transforms are also L^{2}. (Note that since q is in units of distance and p is in units of momentum, the presence of the Planck constant in the exponent makes the exponent dimensionless, as it should be.)
Therefore, the Fourier transform can be used to pass from one way of representing the state of the particle, by a wave function of position, to another way of representing the state of the particle: by a wave function of momentum. Infinitely many different polarisations are possible, and all are equally valid. Being able to transform states from one representation to another by the Fourier transform is not only convenient but also the underlying reason of the Heisenberg uncertainty principle.
The other use of the Fourier transform in both quantum mechanics and quantum field theory is to solve the applicable wave equation. In non-relativistic quantum mechanics, Schrödinger's equation for a time-varying wave function in one-dimension, not subject to external forces, is
This is the same as the heat equation except for the presence of the imaginary unit i. Fourier methods can be used to solve this equation.
In the presence of a potential, given by the potential energy function V(x), the equation becomes
The "elementary solutions", as we referred to them above, are the so-called "stationary states" of the particle, and Fourier's algorithm, as described above, can still be used to solve the boundary value problem of the future evolution of ψ given its values for t = 0. Neither of these approaches is of much practical use in quantum mechanics. Boundary value problems and the time-evolution of the wave function is not of much practical interest: it is the stationary states that are most important.
In relativistic quantum mechanics, Schrödinger's equation becomes a wave equation as was usual in classical physics, except that complex-valued waves are considered. A simple example, in the absence of interactions with other particles or fields, is the free one-dimensional Klein–Gordon–Schrödinger–Fock equation, this time in dimensionless units,
This is, from the mathematical point of view, the same as the wave equation of classical physics solved above (but with a complex-valued wave, which makes no difference in the methods). This is of great use in quantum field theory: each separate Fourier component of a wave can be treated as a separate harmonic oscillator and then quantized, a procedure known as "second quantization". Fourier methods have been adapted to also deal with non-trivial interactions.
Finally, the number operator of the quantum harmonic oscillator can be interpreted, for example via the Mehler kernel, as the generator of the Fourier transform .^{[27]}
The Fourier transform is used for the spectral analysis of time-series. The subject of statistical signal processing does not, however, usually apply the Fourier transformation to the signal itself. Even if a real signal is indeed transient, it has been found in practice advisable to model a signal by a function (or, alternatively, a stochastic process) which is stationary in the sense that its characteristic properties are constant over all time. The Fourier transform of such a function does not exist in the usual sense, and it has been found more useful for the analysis of signals to instead take the Fourier transform of its autocorrelation function.
The autocorrelation function R of a function f is defined by
This function is a function of the time-lag τ elapsing between the values of f to be correlated.
For most functions f that occur in practice, R is a bounded even function of the time-lag τ and for typical noisy signals it turns out to be uniformly continuous with a maximum at τ = 0.
The autocorrelation function, more properly called the autocovariance function unless it is normalized in some appropriate fashion, measures the strength of the correlation between the values of f separated by a time lag. This is a way of searching for the correlation of f with its own past. It is useful even for other statistical tasks besides the analysis of signals. For example, if f(t) represents the temperature at time t, one expects a strong correlation with the temperature at a time lag of 24 hours.
It possesses a Fourier transform,
This Fourier transform is called the power spectral density function of f. (Unless all periodic components are first filtered out from f, this integral will diverge, but it is easy to filter out such periodicities.)
The power spectrum, as indicated by this density function P, measures the amount of variance contributed to the data by the frequency ξ. In electrical signals, the variance is proportional to the average power (energy per unit time), and so the power spectrum describes how much the different frequencies contribute to the average power of the signal. This process is called the spectral analysis of time-series and is analogous to the usual analysis of variance of data that is not a time-series (ANOVA).
Knowledge of which frequencies are "important" in this sense is crucial for the proper design of filters and for the proper evaluation of measuring apparatuses. It can also be useful for the scientific analysis of the phenomena responsible for producing the data.
The power spectrum of a signal can also be approximately measured directly by measuring the average power that remains in a signal after all the frequencies outside a narrow band have been filtered out.
Spectral analysis is carried out for visual signals as well. The power spectrum ignores all phase relations, which is good enough for many purposes, but for video signals other types of spectral analysis must also be employed, still using the Fourier transform as a tool.
Other common notations for include:
In the sciences and engineering it is also common to make substitutions like these:
So the transform pair can become
A disadvantage of the capital letter notation is when expressing a transform such as or which become the more awkward and
In some contexts such as particle physics, the same symbol may be used for both for a function as well as it Fourier transform, with the two only distinguished by their argument I.e. would refer to the Fourier transform because of the momentum argument, while would refer to the original function because of the positional argument. Although tildes may be used as in to indicate Fourier transforms, tildes may also be used to indicate a modification of a quantity with a more Lorentz invariant form, such as , so care must be taken. Similarly, often denotes the Hilbert transform of .
The interpretation of the complex function f̂(ξ) may be aided by expressing it in polar coordinate form
Then the inverse transform can be written:
The Fourier transform may be thought of as a mapping on function spaces. This mapping is here denoted F and F(f) is used to denote the Fourier transform of the function f. This mapping is linear, which means that F can also be seen as a linear transformation on the function space and implies that the standard notation in linear algebra of applying a linear transformation to a vector (here the function f) can be used to write F f instead of F(f). Since the result of applying the Fourier transform is again a function, we can be interested in the value of this function evaluated at the value ξ for its variable, and this is denoted either as F f(ξ) or as (F f)(ξ). Notice that in the former case, it is implicitly understood that F is applied first to f and then the resulting function is evaluated at ξ, not the other way around.
In mathematics and various applied sciences, it is often necessary to distinguish between a function f and the value of f when its variable equals x, denoted f(x). This means that a notation like F(f(x)) formally can be interpreted as the Fourier transform of the values of f at x. Despite this flaw, the previous notation appears frequently, often when a particular function or a function of a particular variable is to be transformed. For example,
Notice, that the last example is only correct under the assumption that the transformed function is a function of x, not of x_{0}.
As discussed above, the characteristic function of a random variable is the same as the Fourier–Stieltjes transform of its distribution measure, but in this context it is typical to take a different convention for the constants. Typically characteristic function is defined
As in the case of the "non-unitary angular frequency" convention above, the factor of 2π appears in neither the normalizing constant nor the exponent. Unlike any of the conventions appearing above, this convention takes the opposite sign in the exponent.
The appropriate computation method largely depends how the original mathematical function is represented and the desired form of the output function. In this section we consider both functions of a continuous variable, and functions of a discrete variable (i.e. ordered pairs of and values). For discrete-valued the transform integral becomes a summation of sinusoids, which is still a continuous function of frequency ( or ). When the sinusoids are harmonically-related (i.e. when the -values are spaced at integer multiples of an interval), the transform is called discrete-time Fourier transform (DTFT).
Sampling the DTFT at equally-spaced values of frequency is the most common modern method of computation. Efficient procedures, depending on the frequency resolution needed, are described at Discrete-time Fourier transform § Sampling the DTFT. The discrete Fourier transform (DFT), used there, is usually computed by a fast Fourier transform (FFT) algorithm.
Tables of closed-form Fourier transforms, such as § Square-integrable functions, one-dimensional and § Table of discrete-time Fourier transforms, are created by mathematically evaluating the Fourier analysis integral (or summation) into another closed-form function of frequency ( or ).^{[55]} When mathematically possible, this provides a transform for a continuum of frequency values.
Many computer algebra systems such as Matlab and Mathematica that are capable of symbolic integration are capable of computing Fourier transforms analytically. For example, to compute the Fourier transform of cos(6πt) e^{−πt2} one might enter the command integrate cos(6*pi*t) exp(−pi*t^2) exp(-i*2*pi*f*t) from -inf to inf
into Wolfram Alpha.^{[note 8]}
Discrete sampling of the Fourier transform can also be done by numerical integration of the definition at each value of frequency for which transform is desired.^{[56]}^{[57]}^{[58]} The numerical integration approach works on a much broader class of functions than the analytic approach.
If the input function is a series of ordered pairs, numerical integration reduces to just a summation over the set of data pairs.^{[59]} The DTFT is a common subcase of this more general situation.
The following tables record some closed-form Fourier transforms. For functions f(x) and g(x) denote their Fourier transforms by f̂ and ĝ. Only the three most common conventions are included. It may be useful to notice that entry 105 gives a relationship between the Fourier transform of a function and the original function, which can be seen as relating the Fourier transform and its inverse.
The Fourier transforms in this table may be found in Erdélyi (1954) or Kammler (2000, appendix).
Function | Fourier transform unitary, ordinary frequency |
Fourier transform unitary, angular frequency |
Fourier transform non-unitary, angular frequency |
Remarks | |
---|---|---|---|---|---|
Definitions | |||||
101 | Linearity | ||||
102 | Shift in time domain | ||||
103 | Shift in frequency domain, dual of 102 | ||||
104 | Scaling in the time domain. If |a| is large, then f(ax) is concentrated around 0 and spreads out and flattens. | ||||
105 | The same transform is applied twice, but x replaces the frequency variable (ξ or ω) after the first transform. | ||||
106 | n^{th}-order derivative.
As f is a Schwartz function | ||||
106.5 |