1. Brownian Motion
1.1 Overview
This entry describes the fundamental properties and key characteristics of Brownian motion. I present its defining axioms, derive its moment structure, and establish the covariance formula that governs its dependence across time.
Brownian motion is the central driving noise in stochastic calculus. Rather than being defined by a formula, it is characterized by a small set of structural properties that completely determine its behavior. These properties encode randomness, independence, and scaling in time.
1.2 Definition of Brownian motion
A stochastic process \((W_t)_{t \ge 0}\) is called a standard Brownian motion if it satisfies:
- \(W_0 = 0\) almost surely
- The paths \(t \mapsto W_t(\omega)\) are almost surely continuous
- The process has independent increments
- For all \(0 \le s < t\), the increment \(W_t - W_s\) is normally distributed with mean \(0\) and variance \(t - s\)
The last property can be written as
1.3 Mean of Brownian motion
Setting \(s = 0\) in \(\eqref{eq:brownian_increment_distribution}\), we obtain
Thus, the law of Brownian motion at a fixed time is completely determined by the elapsed time \(t\). The expected value of \(W_t\) follows directly from its Gaussian distribution. Using \(\eqref{eq:brownian_marginal_distribution}\),
This reflects the absence of drift: Brownian motion fluctuates symmetrically around zero.
1.3.1 Variance expressed via second moments
The variance of \(W_t\) is defined as
Since Brownian motion has zero mean \(\mathbb{E}[W_t] = 0\), the variance simplifies to
1.3.2 Explicit computation using the Gaussian density
Let \(f_{W_t}\) denote the density of \(W_t\). From \(\eqref{eq:brownian_marginal_distribution}\) (mean zero, and variance \(t\)).
The second moment is therefore 1
Substituting \(\eqref{eq:brownian_density}\),
1.3.3 Reduction to the standard Gaussian moment
Perform the change of variables
Then \(\eqref{eq:second_moment_expanded}\) becomes
The remaining integral is the second moment of a standard normal random variable and equals \(1\). Hence,
Combining with \(\eqref{eq:variance_second_moment}\), we conclude
1.4 Covariance structure of Brownian motion
We now derive the covariance between two times \(s\) and \(t\).
Assume without loss of generality that \(0 \le s \le t\).
1.4.1 Definition of covariance
The covariance is defined by
Since Brownian motion has zero mean, this simplifies to
1.4.2 Decomposition using independent increments
Write \(W_t\) as
Substituting into \(\eqref{eq:covariance_simplified}\),
1.4.3 Vanishing of the mixed term
The increment \(W_t - W_s\) is independent of \(\mathcal{F}_s\) and has zero mean.
Since \(W_s\) is \(\mathcal{F}_s\)-measurable,
1.4.4 Final covariance computation
Using \(\eqref{eq:mixed_term_zero}\),
From \(\eqref{eq:variance_final}\), we know that
Therefore,
Since covariance is symmetric, this can be written compactly as
1.5 Expected Values of Brownian Motion Differences
A common task in stochastic calculus is computing expectations involving differences of Brownian motion at multiple time points. This section systematically derives these expected values, building on the independent increments property.
1.5.1 Single Increment: \(\mathbb{E}[W_t - W_s]\)
For any \(0 \le s < t\), the increment \(W_t - W_s\) is normally distributed:
Therefore, the expected value is immediately:
Why This Matters
This zero-mean property is fundamental: Brownian motion has no drift. Given the process value at time \(s\), the expected change over any future interval is zero. This makes Brownian motion a martingale.
1.5.2 Product of Two Increments: \(\mathbb{E}[(W_t - W_s)(W_u - W_v)]\)
Consider two time intervals \([s,t]\) and \([v,u]\). The expected value of the product of increments depends on whether the intervals overlap.
1.5.2.1 Case 1: Disjoint Intervals (\(t \le v\) or \(u \le s\))
When the intervals are disjoint, the increments are independent. For example, if \(t \le v\):
Independent Increments Property
When intervals \([s,t]\) and \([v,u]\) don't overlap, the corresponding increments are independent random variables. For independent zero-mean variables, the expectation of their product equals the product of their expectations (both zero).
1.5.2.2 Case 2: Overlapping Intervals
When intervals overlap, we must decompose the increments carefully. Assume \(s < v < t < u\). Then:
Expanding the product:
Using independence of non-overlapping increments: - \(\mathbb{E}[(W_v - W_s)(W_t - W_v)] = 0\) - \(\mathbb{E}[(W_v - W_s)(W_u - W_t)] = 0\) - \(\mathbb{E}[(W_t - W_v)(W_u - W_t)] = 0\)
Only the variance term survives:
General Pattern
For overlapping intervals, the expected product equals the length of the overlap:
This generalizes both the disjoint case (overlap = 0) and nested cases.
1.5.3 Special Case: \(\mathbb{E}[W_s W_t]\) Revisited
Setting \(s=0\) and \(v=0\) in the overlapping case gives us the covariance formula we already derived:
This confirms our earlier covariance result \(\eqref{eq:covariance_final}\).
1.5.4 Three-Point Differences: \(\mathbb{E}[(W_t - W_s)(W_s - W_r)]\)
For ordered times \(0 \le r < s < t\), consider the product of consecutive increments:
These increments correspond to disjoint intervals \([r,s]\) and \([s,t]\), so they are independent:
Common Mistake
Don't confuse this with \(\mathbb{E}[W_t W_s]\). While \(W_t\) and \(W_s\) are correlated (both contain \(W_s\)), the increments \(W_t - W_s\) and \(W_s - W_r\) are independent because they occur over disjoint time intervals.
1.5.5 Squared Difference: \(\mathbb{E}[(W_t - W_s)^2]\)
For \(0 \le s < t\):
This is simply the variance of the increment, since the mean is zero.
Connection to Quadratic Variation
The fact that \(\mathbb{E}[(W_t - W_s)^2] = t - s\) is the foundation for the quadratic variation of Brownian motion. Summing over partitions leads to the result that the quadratic variation of Brownian motion over \([0,t]\) equals \(t\).
1.6 Quadratic Variation of Brownian Motion
Quadratic variation measures how "rough" or "wiggly" a path is. For smooth functions like polynomials, quadratic variation is zero. But Brownian motion is nowhere differentiable, as its quadratic variation is non-zero and fundamental to stochastic calculus.
1.6.1 Definition via Partition
For a partition of \([0,t]\):
the quadratic variation is defined as the limit of squared increments:
where the limit is taken as the mesh size \(\max_i (t_{i+1} - t_i) \to 0\), and convergence is in probability.
Why Quadratic Variation?
Why focus on squared increments? Because:
- First variation (sum of absolute increments) is infinite for Brownian motion
- Quadratic variation (sum of squared increments) converges to a finite, non-random limit
- Quadratic variation is what matters in Itô calculus — it replaces the role of classical derivatives
1.6.2 Deriving the Result
1.6.2.1 Compute Expected Value of Squared Increments
For a single partition interval \([t_i, t_{i+1}]\), we know from earlier that:
1.6.2.2 Sum Over All Increments
The expected value of the sum of squared increments is:
1.6.2.3 Evaluate the Sum
The sum telescopes to give the total elapsed time:
Therefore:
1.6.2.4 Show Convergence in Probability
Now we need to show that the actual sum (not just its expectation) converges. The variance of the sum is:
Computing Variance of Squared Gaussian
For a zero-mean normal variable \(X \sim \mathcal{N}(0, \sigma^2)\):
For our increment \(W(t_{i+1}) - W(t_i) \sim \mathcal{N}(0, \Delta t_i)\):
Each term contributes \(2(\Delta t_i)^2\):
As \(n \to \infty\) and mesh size \(\to 0\), the sum \(\sum_{i=0}^{n-1} (\Delta t_i)^2 \to 0\). Therefore:
1.6.2.5 Conclude
By Chebyshev's inequality, if the variance goes to zero, the sum converges in probability to its expected value:
Chebyshev's Inequality and Convergence in Probability
Chebyshev's Inequality states that for any random variable \(X\) and any \(\epsilon > 0\):
Applying to our quadratic variation sum:
Let \(X_n = \sum_{i=0}^{n-1} \left(W(t_{i+1}) - W(t_i)\right)^2\). We have:
- \(\mathbb{E}[X_n] = t\) (from \(\ref{eq:qv_expectation}\))
- \(\mathrm{Var}[X_n] \to 0\) as \(n \to \infty\) (from \(\ref{eq:qv_variance_limit}\))
By Chebyshev's inequality:
Taking the limit: As \(n \to \infty\), the variance \(\mathrm{Var}(X_n) \to 0\), so:
This holds for any fixed \(\epsilon > 0\). This is exactly the definition of convergence in probability:
which is precisely equation \(\eqref{eq:qv_convergence}\).
Therefore, the quadratic variation of Brownian motion over \([0,t]\) is:
1.6.3 Key Insight: Randomness Cancels
This result is counterintuitive: even though Brownian motion is highly random, the sum of squared increments produces a deterministic result \([W]_t = t\).
The randomness in individual increments cancels out when squared and summed: - Some increments are large, some small - Some positive, some negative - But on average, they square to exactly \(t\)
Why This Matters for Stochastic Calculus
The quadratic variation \([W]_t = t\) is the foundation of Itô's lemma. In ordinary calculus:
But in stochastic calculus, the quadratic variation contributes an extra term:
The term \(dt\) comes directly from the fact that \((dW)^2 \approx dt\) (more precisely, has quadratic variation \(dt\)).
-
See the nice, an straigthforward, post in Open Source Quant about moments. ↩