9
$\begingroup$

A few weeks ago, I looked into the theory of locally convex spaces. Now I want to continue with the theory of distributions. The following description is somewhat confusing to me:

(*)

The basic idea in the theory of distributions is, to comprehend a function $f$ as a functional $T_f$ via $T_f(\phi)=\int f(x) \phi(x) dx$ on a suitable function space.


Here is what I know: $(**)$

For $\Omega \subseteq \mathbb{R}^n$, we have the vector space of smooth functions of $\Omega$, $\mathcal{E}(\Omega):=C^{\infty}(\Omega):=\{\phi: \Omega \rightarrow \mathbb{R}| \phi \text{ is a } C^{\infty} \text{ function}\}$.

Further,for $C \subset \mathbb{R}^n$ compact denote $\mathcal{D}_C(\Omega):= \{\phi \in C^{\infty}| \text{supp}(\phi) \subseteq C\}$.

Then one can define the space of text functions as $\mathcal{D}(\Omega):=\underset{C \subset \Omega, C \text{ compact} }{\bigcup} \mathcal{D}_C(\Omega)$.

On $\mathcal{D}_C(\Omega)$ one can define a locally convex space via the seminorms $p_m(\phi):=\underset{|\alpha| \leq m}{\max} \underset{x \in \Omega}{sup|\partial^{\alpha \phi(x)}|}$, where $\phi \in \mathcal{D}_C(\Omega)$.

On $\mathcal{D}(\Omega)$ the set of seminorms $p$ on $\mathcal{D}(\Omega)$ such that $p_{|\mathcal{D}_C(\Omega)}$ is continuous w.r.t $\tau_C $for every compact $C \subset \Omega$ generates a locally convex topology on $\mathcal{D}_C(\Omega)$.


My question is the following:

In detail, what is the basic idea of distributions?

Especially, since the description of the idea in $(*)$ sounds a lot like some (Riesz-) Representation theorem.

For context, I do know the basics about locally convex spaces. I haven't encounter distributions thus far.

$\endgroup$
8
  • $\begingroup$ What is your background in Laplace and Fourier transforms? How about differential equations? $\endgroup$
    – J W
    Commented 18 hours ago
  • 2
    $\begingroup$ My own experience was that I began to understand distributions when I began to understand the Dirac delta function. This might or might not work for you. $\endgroup$ Commented 18 hours ago
  • $\begingroup$ If $f\in L^\infty$, then for all $g$ with compact support ("test functions"), you can look at $\int_{\mathbb R} fg$. The map $g\mapsto \inte_\RR fg$ is a distribution. It faithfully represents $f$. So every function like $f$ is a distribution. So distributions are generalizations of functions. They are useful for understanding PDE. $\endgroup$
    – Wulfhartus
    Commented 17 hours ago
  • 1
    $\begingroup$ Have you ever heard of the Dirac delta function? That is the most basic distribution - it is a misnomer to call it a function (although it is commonly described and used as one.) $\endgroup$ Commented 17 hours ago
  • 1
    $\begingroup$ My two comments above come together in realizing that the Dirac $\delta$ "function" is the distribution you get taking the Fourier transform of the constant function $f(x)=1.$ $\endgroup$ Commented 17 hours ago

1 Answer 1

16
$\begingroup$

The idea of distributions did not come immediately to L. Schwartz. The final definition came to him after several attempts, as he describes in his wonderful autobiography. Here I will give a non-historical account.

Weakly Differentiable Functions

There are many functions in mathematics, calculus, and engineering that are not differentiable at every point.
A simple example is
$$ F(x) = |x|. $$ Its derivative is $-1$ for $x<0$ and $1$ for $x>0$, but it is not defined at $x=0$.
However, if we consider
$$ f(x) = \operatorname{sgn}(x) := \begin{cases} -1, & x<0,\\ 0, & x=0,\\ 1, & x>0, \end{cases} $$ we see that
$$ F(x)-F(y) = \int_x^y f(t)\, dt, $$ for every $x,y \in \mathbb{R}$.

Thus, $F$ behaves like a primitive of $f$ on $\mathbb{R}$, except that $f$ is not regular enough to apply the classical rules of calculus, where primitives are defined for continuous functions. In particular, we cannot conclude that $F'(0) = f(0)$, since $f$ is discontinuous at $0$.

Nevertheless, in some weak sense, $f$ is the derivative of $F$. The idea of weak derivatives was introduced by Sobolev, prior to Schwartz.

This might look trivial, but there are natural situations where weak derivatives are convenient. A classical example arises in the earliest partial differential equations in the literature: the wave equation and the heat equation.

Consider someone playing a guitar. Under some (idealized) assumptions, the motion of the string is described by the wave equation
$$ u_{tt}(x,t) = c^2\, u_{xx}(x,t), $$ where $u(x,t)$ is the height of the string at position $x$ and time $t$. Here, $u_{tt}$ and $u_{xx}$ are second-order partial derivatives in $t$ and $x$.

A simple difficulty occurs if the musician plucks the string: the initial shape of the string at $t=0$ may resemble the graph of $|x|$. Then $u(x,0)$ is not differentiable with respect to $x$ at some point. How can we use a differential equation if the initial shape is not differentiable?

The idea is to differentiate without ever differentiating directly, using the principle of integration by parts.

Integration by Parts

The classical Integration by Parts formula states that if $f$ and $\phi$ are smooth (say, $C^1$) then
$$ \int_a^b f'(x)\phi(x)\, dx = - \int_a^b f(x)\phi'(x)\, dx + f(b)\phi(b) - f(a)\phi(a). $$

If $\phi$ vanishes outside $(a,b)$, this simplifies to
$$ \int_\mathbb{R} f'(x)\phi(x)\, dx = - \int_\mathbb{R} f(x)\phi'(x)\, dx. $$

The surprising fact is that the right-hand side still makes sense even if $f$ is not classically differentiable, provided $\phi$ is smooth enough. Thus we can compute
$$ \int f'(x)\phi(x)\, dx $$ without ever differentiating $f$ directly.

Very Regular Functions as Probes

In practice, you never observe a function $f(x)$ directly. For example, if $f(x)$ is the temperature of a metal bar, you only measure temperature averages near a point. You are probing $f$ using a smoother function $\phi$. Formally, define

$$ T_f(\phi) := \int_\mathbb{R} f(x)\,\phi(x)\, dx. $$

An interesting example is when $\phi$ is a $C^\infty$ function with compact support so that $\phi\ge0$ and $\int \phi \, dx = 1$. Then $T_f(\phi)$ is a kind of weighted average of $f$ over the support of $\phi$.

Importantly, if $f$ and $g$ are continuous and
$$ T_f(\phi) = T_g(\phi) $$ for every $\phi\in C_c^\infty(\mathbb{R})$, then $f=g$. Here $C_c^\infty(\mathbb{R})$ denotes the vector space of all $C^\infty$ functions with compact support. Thus, the functional $T_f$ contains all the information about $f$, so we can identify $f$ with this linear functional on the space of test functions.

Moreover, if $f\in C^\infty$, integration by parts yields
$$ T_{f'}(\phi) = -\,T_f(\phi'), \tag{1} $$ so we can recover the derivative of $f$ in this weak sense.

We define $g$ to be the weak derivative of $f$ if
$$ T_g(\phi) = -\, T_f(\phi') $$ for every test function $\phi$.

The Breakthrough of Schwartz

Sobolev used weak derivatives very successfully. But in Sobolev’s approach, weak derivatives were still functions.

Schwartz went beyond this: he observed that identifying functions with their functionals $T_f$ on $C^\infty_c$ allowed him to keep differentiating arbitrary linear functionals, even if they do not come from any function.

He defined a distribution as any linear functional
$$ T : C^\infty_c(\mathbb{R}) \to \mathbb{R}, $$ and its derivative by
$$ T'(\phi) := -\,T(\phi'). $$

This allows us to differentiate objects that are not functions and to obtain derivatives that are also not functions.

For example, differentiating $F(x)=|x|$ in the sense of distributions gives
$$ T'_F = T_{\operatorname{sgn}}, $$ where $\operatorname{sgn}$ is the sign function defined above.

Differentiating again,
$$ T''_F = T'_{\operatorname{sgn}} = 2\,\delta_0, $$ where $\delta_0$ is Dirac’s delta at $0$, defined by
$$ \delta_0(\phi) := \phi(0). $$

This distribution does not come from any classical function. It comes from a measure assigning weight $1$ to the point $0$. And we can keep differentiating ad infinitum.

The space $C_c^\infty$ is a classical space of test functions. We can also considered other test function spaces (such as $C^\infty$ functions with rapid decay), which we omit here.

That is a major advantage of distributions: they allow us to place many disparate objects in the same space: $L^p$ functions, wildly discontinuous functions, continuous functions, $C^k$ functions, regular measures, and even more exotic objects, are all distributions, functionals on a suitable space of test functions.

Within this framework, we can with great freedom take linear combinations, form series, and differentiate these otherwise unrelated objects. Distributions allow us to treat all of them on equal footing.

Why Distributions Are Useful

Distributions are extremely useful because they allow you to break down the proof of the existence of solutions to problems involving functions and its derivatives (such as partial differential equations) into two steps:

  1. Existence as a distribution: First, prove that a solution exists in the sense of distributions. At this stage, one can freely manipulate derivatives of arbitrary order without worrying too much about its existence in a strong sense and its regularity.

  2. Regularity improvement: Then, step by step, establish that this distribution is more regular, eventually showing that it corresponds to an actual function with the desired regularity.

A final example

I have also encountered distributions in my own work (and with my students), not in the classical PDE setting, but in the study of functions similar to the Weierstrass function $$ f(x)=\sum_{n=0}^\infty \frac{1}{2^n} \cos(2^n x). $$

This belongs to the famous family of examples by Weierstrass of nowhere differentiable functions. However, one is tempted to formally differentiate term by term:
$$ f'(x)=\sum_{n=0}^\infty -\,\sin(2^n x), \tag{2} $$ which does not make sense in the classical sense. Nevertheless, the right-hand side does define a distribution. Indeed, for a $2\pi$-periodic $C^\infty$ function $\phi$,
$$ T'_f(\phi) = - \int_0^{2\pi} f(x)\,\phi'(x)\, dx = \sum_{n=0}^\infty \int_0^{2\pi} -\,\sin(2^n x)\,\phi(x)\, dx. $$

Thus, the formal derivative in (2) makes sense in the sense of distributions, and this approach can be used to study the regularity of the Weierstrass function and similar examples.

$\endgroup$
7
  • 1
    $\begingroup$ This explanation is on par with ones seen in advanced PDE classes (and in some sense more intuitive). Would you mind demonstrating how $T'_F=T_{\text{sgn}}$ for $F(x)=|x|$? :) $\endgroup$
    – Man-I-Fold
    Commented 8 hours ago
  • 2
    $\begingroup$ @Man-I-Fold Write $\int_{-\infty}^\infty |x|\phi'(x)\,dx = -\int_{-\infty}^0 x\phi'(x)\,dx + \int_0^\infty x\phi'(x)\,dx$ and integrate by parts to get $-\int_{-\infty}^\infty \text{sgn}(x)\phi(x)\,dx$. $\endgroup$ Commented 8 hours ago
  • 1
    $\begingroup$ Here's another example that's easy on the eyes. Define $u,v:(0,2)\to\mathbb{R}$ by $u(x):=\begin{cases}x&\text{ if }0<x\leq 1\\1&\text{ if }1\leq x<2\end{cases}$, $v(x):=\begin{cases}1&\text{ if }0<x\leq 1\\0&\text{ if }1<x<2\end{cases}$. We claim that $v$ is a derivative of $u$ on $(0,2)$ in the weak sense from Evans' 2nd ed of PDEs. To see this, fix any $\phi\in C^\infty_c((0,2))$. We calculate \begin{align*}\int_0^2u\phi'dx&=\int_0^1x\phi'dx+\int_1^2\phi'dx\\&=-\int_0^1\phi dx+\phi(1)-\phi(1)\\&=-\int_0^11\phi dx-\int_1^20\phi dx=-\int_0^2v\phi dx.\end{align*} $\endgroup$
    – Man-I-Fold
    Commented 7 hours ago
  • 1
    $\begingroup$ In the IBP section you say “The surprising fact is that the right-hand side still makes sense even if…”, but that’s not really surprising is it. It is very trivial that under very mild assumptions on $f$, and one regularity assumption on $\phi$, the quantity $\int f\phi’$ makes sense. Instead I think the ingenuity comes in the fact that we can use the RHS pretty much as a (generalized) definition for the LHS. And this entire idea is predicated on the principle that integrals are “better” (easier to handle/manipulate) than derivatives. $\endgroup$
    – peek-a-boo
    Commented 7 hours ago
  • 1
    $\begingroup$ Also, it should be emphasized that generalizing definitions purely for the sake of generality is a silly idea. But the utility of distributions comes, as you’ve explained, because of how effectively they solve several problems (especially when dealing with PDEs, even ‘simple’ ones like Laplace’s equation and the flat linear wave equation). And thinking ahead, the introduction of the Schwartz space is a perfect intermediate between $C^{\infty}_c$ and $C^{\infty}$ :) $\endgroup$
    – peek-a-boo
    Commented 7 hours ago

You must log in to answer this question.

Start asking to get answers

Find the answer to your question by asking.

Ask question

Explore related questions

See similar questions with these tags.