Variational derivative

See @olver86 page 244.

Introduction

For functionals, the variational derivative plays the role of the gradient of functions of several variables.
Given a smooth function f, the gradient is a 1-form df such that for a vector V

dfx(V)=limϵ0f(x+ϵV)f(x)ϵ=ddϵf(x+ϵV)|ϵ=0,

that is, it tells us how much the function varies along the direction specified by V.

If we think of x not like a finite dimensional vector (x1,,xp) but like a function x:{1,,p}R we can generalize this to the case x:RR, tx(t), being now f not a function but a functional.

The question is: for a functional f, is there any mathematical object δfx such that for a function V(t) gives us the new function

δfx(V)=limϵ0f(x+ϵV)f(x)ϵ=ddϵf(x+ϵV)|ϵ=0

?
Observe that in the usual case of x being a vector and f being a function we can interpret df as the gradient vector f (assuming the standard inner product, see here for more information) which satisfies

dfx(V)=(f)x,V=ddϵf(x+ϵV)|ϵ=0.

We can replace , with the usual inner product dt for the Hilbert space L2, so we can alternatively require to our new object δfx, being f a functional again, to satisfy

δfxVdt=ddϵf(x+ϵV)|ϵ=0

There are lots of technical details we are missing here, but this is the idea.

Definition

Definition (@olver86 page 245)
Let J[u] be a variational problem (for functions u:RpRq). The variational derivative of J is the unique q-tuple

δJ[u]=(δ1J[u],,δqJ[u])

such that given functions f,η:ΩRpRq, η with compact support, satisfies:

(1)ddϵJ[f+ϵη]|ϵ=0=ΩδJ[f(x)]η(x)dx.

Some facts

Remarks
a) To my knowledge, it doesn't have to exist...

b) I think that the quantity (1) should also be denoted by

δJf(η)

to agree with dfx(V). In this sense δJ would be something similar to differential of a function. The expression δJf=δJ[f(x)] can be thought as another function of x or, alternatively, as something eating functions η and turning back numbers, i.e., something analogous to the 1-form dfx in the usual case of functions of Rn. Something like the duality between the gradient fx and dfx (vectors correspond to functions).

c) Since the functional J is coming from a variational problem we can compute an explicit formula for its variational derivative. Let's focus in the case of functions u:RR:

ΩδJ[f(x)]η(x)dx=ddϵJ[f+ϵη]|ϵ=0==ddϵ|ϵ=0ΩL(x,prn(f+ϵη))dx==ΩLuη+Lu1η++Lunηn)dx

being ηj) the jth derivative of η.
We can apply integration by parts, leaving apart the first term, and obtain that the expression above equals

ΩLuηdx+ΩDxLu1ηdx++ΩDxLunηn1)dx

and after n steps:

=ΩLuηdx+ΩDxLu1ηdx++Ω(Dx)nLunηdx==Ωj=0n(Dx)jLujηdx

So it should be

δJ[f(x)]=j=0n(Dx)jLuj

This last expression is called the Euler operator.

Proposition (@olver86 page 246)
If f(x) is an extrema of J[u] then δJ[f(x)]0, the null function on Ω.

The Euler-Lagrange equations then appear from this proposition together with Remark c) above.

Old stuff (integrate with above)

Functional derivative. Gradient flow

Given a functional F:HR, being H a Hilbert space of functions, the associated gradient flow is given by the equation

(1)ρt=Fρ.

for ρ(t)H.

In other words, ρ decreases along the gradient of F. The terminology stems from the 'finite dimensional' case, where a function f(x,y,z) produces a vector field V=f, which is called its 'gradient vector field'. Then, as with any vector field, one can study the flow induced by that vector field, i.e. the flow of the dynamical system given by x˙=V(x).

In (1), the notation Fρ denotes the so-called functional derivative of F to ρ, which generalises the 'gradient' notion for functions. There exist multiple versions of the functional derivative, mainly because its definition depends on the function space on which F acts. Anyway, the idea is to perturb ρ a bit, i.e. to substitute ρρ+ϵϕ with 0<ϵ1, and work out the resulting expression.

Example

Let H=L2(Rn), and F such that for uL2(Rn):

F(u)=12|u(x)|2dx

That is, F is the Dirichlet energy. Then the Heat Equation

tu=2u

is the associated gradient flow problem.