Three line summary
-
The Malliavin derivative is an operator defined by manipulating the chaos expansion of a square-integrable random variable.
-
The Malliavin derivative transforms square-integrable variables into square integrable processes.
-
The Malliavin derivative shares some properties with the classic derivative such as the product rule and chain rule, but the fundamental theorem of calculus only holds in some special cases.
Why should I care?
The Malliavin derivative is (somewhat unsurprisingly) a fundamental object of Malliavin calculus and has many applications in finance, numerical methods, and optimal control.
Notation
The same as in previous posts. Furthermore we will shorten the notation $L^2(\Omega,\mathcal{F}\U T)$ and $L^2(I\times\Omega,\mathcal{B}(I)\otimes \mathcal{F}\U T)$ to $L^2(\Omega)$ and $L^2(I\times\Omega)$ respectively.
Introduction
The Malliavin derivative was originally introduced as an operator associated with the Fréchet differential of random variables $X: C(I)\to {\mathbb R}$. The aforementioned construction provides some motivation behind the Malliavin derivative and will be developed in the next post. However, a more general construction can be obtained via the chaos expansion.
Definition 1. Given $X=\sum\U {n=0}^{\infty} I\U n(f\U n)\in L^2(\Omega)$ we say that $X$ is Malliavin differentiable if
and denote the space of Malliavin differentiable functions by
Furthermore, we define the Malliavin derivative of $X$ as
These kinds of manipulations of the terms in the chaos expansion should be familiar from our definition of the Skorohod integral. Note for example that $f\U {n}(\cdot ,t)$ is a square integrable symmetric function of $n-1$ variables and thus the terms $I\U {n-1}(f\U {n}(\cdot ,t))$ make sese. This checked, the first questions to be asked is: “what kind of object is the Malliavin derivative of a random variable? What is the link between $\mathbb{D}\U {1,2}$ and $D\U t$? Is $|\cdot |\U {\mathbb{D}\U {1,2}}$ even a norm?” We answer this in the next proposition and corollary.
Proposition 1. The Malliavin derivative is well defined on $\mathbb{D}\U {1,2}$ and establishes a linear isometry
Proof. The proof of the first part of the proposition is a straightforward application of Itô’s $n$-th isometry and the monotone convergence theorem as we have that
Finally, the linearity of $D$ follows from the linearity of the iterated Itô integrals (which itself is a consequence of the linearity of the Itô integral). ◻
In summary, the Malliavin derivative turns a square-integrable random variable into a possibly non-adapted, stochastic process. You may recall from our previous posts that we had an operator that went in the opposite direction. The Skorohod integral $\delta$. In fact, the Malliavin derivative and the Skorohod integral are adjoint operators in a sense that will be made precise in the next post. For now, we show that, as occurs with the ordinary derivative, a random variable has Malliavin derivative $0$ if and only if it is constant.
Corollary 1. $(\mathbb{D}\U {1,2},\norm{\cdot }\U {\mathbb{D}\U {1,2}})$ is a seminormed space. Furthermore
Proof. The triangle inequality and the absolute homogeneity are direct consequences of the isometry of the previous proposition. This shows that $\norm{\cdot }\U {\mathbb{D}\U {1,2}}$ is a seminorm. The second part follows from the fact that
So $\norm{X}\U {\mathbb{D}\U {1,2}}=0$ if and only if $f\U n=0$ for all $n\geq 1$, which in turn is equivalent to $X=I\U 0(f\U 0):=f\U 0$. Where we recall that by convention $L^2 (S_0):={\mathbb R}$ and $I\U 0$ was defined as the identity on ${\mathbb R}$. This concludes the proof. ◻
Before moving on we show a motivating example. Let us consider some deterministic function $f\in L^2(I)$ and set
Then, by construction, we have $X=I\U 1(f)$ so
This is a nice result and it
might suggest something akin to the fundamental theorem of calculus such
as $D\U t(\delta Y)=Y(t)$ for any Skorohod integrable process $Y$. However,
this will not hold in general and as, will be seen in the next post,
occurs if and only if $Y$ is a deterministic function in $L^2(\Omega)$.
This said, we now show that $\mathbb{D}\U {1,2}$ is closed in the sense
that: given a convergent sequence $X\U m\to X$, if the derivatives
$DX\U m$ converge then also $D X\U m\to D X$.
Proposition 2. Let $X\U m \in \mathbb{D}\U {1,2}$ such that $X\U m$ is a Cauchy sequence in both $L^2(\Omega)$ and in $\mathbb{D}\U {1,2}$. Then, there exists $X \in \mathbb{D}\U {1,2}$ such that
Proof. First of all, we note that since $L^2(\Omega)$ is complete $X\U m$ must converge to some $X\in L^2(\Omega)$. Let us write the respective chaos expansions as
By Itô’s isometry and the convergence $X\U m\to X\in L^2(\Omega)$ we deduce that also $f^{(m)}\U n\to f\U n\in L^2(I^n)$ for each $n$. By now applying Fatou’s lemma and the fact that $X\U m$ is by hypothesis Cauchy in $\mathbb{D}\U {1,2}$ we obtain that
As desired. ◻
We now conclude this post by stating two properties of the Malliavin derivative that are analogous to those verified by the derivative of ordinary functions. Firstly, an analog to the chain rule for the ordinary derivative. The proof can be found on page $29$ of Nunno and Øksendal’s book 1 but is rather technical and relies on Hermite polynomials which were not discussed previously, so we omit it.
Definition 2. We write $\mathbb{D}\U {1,2}^0\subset L^2(\Omega)$ for the space of square integrable random variables $X=\sum\U {n=0}^{\infty} I\U n(f\U n)$ such that $f\U n=0$ for all but finitely many $n$.
Proposition 3 (Product rule). Given $X\U 1,X\U 2\in \mathbb{D}^0\U {1,2}$ it holds that
Finally, though we shall not use it, we mention that if $\Omega=\mathcal{S}^\star({\mathbb R})$ is the dual of the Schwartz space and we construct a probability measure $\mathbb{P}$ called the white noise probability measure then the following version of the chain rule also holds (see 2 page $89$).
Proposition 4 (Chain rule). Consider $\varphi\in C\U 1({\mathbb R}^d)$ with $\nabla \varphi\in {L^\infty({\mathbb R}^d\to{\mathbb R}^d)}$ and $X=(X\U 1,\ldots,X\U d)$ such that $X\U i\in \mathbb{D}\U {1,2}$ for each $i=1,\ldots,d$. Then it holds that $\varphi(X)\in \mathbb{D}\U {1,2}$ with
Extending past p=2
The Malliavin derivative lets us define a derivative on a subset of $L^2(\Omega)$. However, it may also be useful to have a concept of derivative on random variables in $L^p(\Omega)$. We now explain how to do this via an alternative construction of the Malliavin derivative. First of all, consider the set of cylindrical variables
That is, $\mathbb{W}$ is the set of all smooth functions with bounded derivatives of Wiener integrals of deterministic functions. Let us use the abbreviation $W(h):=\int\U {I}h(t) dW(t)$. Then, the results in the previous section show that the Malliavin differential of a cylindrical variable is
One can also start directly with the above equation as the defintion of Malliavin differential. In this case it is not clear that $D\U t$ is well defined (that is, independent of the representation of $X=\varphi(W(h))$). However it is, see 3 page 10. Analogously to how one defines the norm on Sobolev spaces, we now take any $1\leq p< \infty$ and define a norm on $\mathbb{W}$ by
Then, $D$ is a continuous linear operator on $(\mathbb{W},\norm{\cdot }\U {\mathbb{D}^{1,p}})$ to $L^2(\Omega\to L^2(I))$. As a result, $D$ may be extended to the closure of $(\mathbb{W},\norm{\cdot }\U {\mathbb{D}^{1,p}})$. We denote this closure by $\mathbb{D}^{1,p}$ and by abuse of notation also write $D$ for the continuous extension of $D$ to $\mathbb{D}^{1,p}$. Note that by definition of the norm $\norm{\cdot }\U {\mathbb{D}^{1,p}}$, necessarily $\mathbb{D}^{1,p}$ is a subset of $L^p(\Omega)$. In this way, we have been able to extend the Malliavin differential to $\mathbb{D}^{1,p}\subset L^p(\Omega)$. Explicitly, we have that
Where $X\U n \in \mathbb{W}$ is a sequence converging to $X$ in $\mathbb{D}^{1,p}$. Furthermore, we note that by the previous discussion $D$ coincides with our previous definition of the Malliavin differential when $p=2$. For the case $p=\infty$ we define
We now conclude with an extension of the chain rule which can be used even when $\varphi$ does not have bounded derivative.
Proposition 5 (Chain rule for $\mathbb{D}^{1,p}$). Let $X\in \mathbb{D}^{1,p}$ and consider $\varphi\in C^1({\mathbb R}^n)$ such that $\norm{\nabla\varphi(x)}\leq C(1+\norm{x}^\alpha)$ for some $0\leq \alpha\leq p-1$. Then $\varphi(X)\in \mathbb{D}^{1,q}$, where $q=p/(\alpha+1)$. Furthermore,
Proof. By the mean value inequality we have that
As a result we have that
Furthermore, by Hölder’s inequality applied to $r=(\alpha+1) / \alpha, s=\alpha+1$ we have that
We now take a sequence of cylindrical random variables $X\U n$ converging to $X$ in $\mathbb{D}^{1,p}$ and an approximation to the identity $\delta\U n$. Let us set $\varphi\U n:=\varphi * \delta\U n$
Where the final equality is due to the same method that gave the previous two inclusions and the way $X\U n,\varphi\U n$ converge to $X,\varphi$ respectively. ◻
Essentially the previous proposition says that, if $X$ is differentiable and the derivative of $\varphi$ doesn’t grow to fast (depending on the integrability of $DX$), then $\varphi(X)$ is also differentiable and we can apply the chain rule. The integrability of $D\varphi(X)$ depending on the integrability of $DX$ and the growth of $\nabla \varphi$.
Example application
For example, in the case $p=2$ we could take $\varphi(x)=x^2$ to deduce that
If $X=1\U A$ is an indicator function for some $A\in \mathcal{F}$ we obtain that
From here we deduce that $D1\U A=0$. As we have seen, this occurs if and only if $1\U A$ is constant, so necessarily $1\U A=0$ or $1\U A=1$. Identifying sets with functions, we have just proved the following
Corollary 2. Given $A\in \mathcal{F}$ we have that $1\U A\in \mathbb{D}^{1,2}$ if and only if (almost everywhere) $A=\emptyset$ or $A=\Omega$.
Multiple derivatives
Finally we comment on how it is possible to iterate the Malliavin derivative. Given a cylindrical process $X=\varphi(W(h))\in \mathbb{W}$ we should have that, with Einstein notation
As a result, we define given $X\in \mathbb{W}$
In the same fashion as before, we can now define the $k$-th differential norm as
Where we use the convention $D^0 X:=X, L^2(I^0):= {\mathbb R}$. Then we simply define $\mathbb{D}^{p,k}$ to be the completion of $\mathbb{W}$ with this norm. Finally, we define