Stochastic Processes

Probability & StatisticsDifficulty: ████Depth: 6Unlocks: 0

Poisson processes, Brownian motion, Wiener process. Continuous-time stochastic models. Ito calculus foundations.

▶ Advanced Learning Details

Graph Position

115
Depth Cost
0
Fan-Out (ROI)
0
Bottleneck Score
6
Chain Length

Random events in time and noisy continuous signals are everywhere: from phone-call arrivals to stock prices and particle diffusion — stochastic processes give the precise language and tools to model and analyse them.

TL;DR:

Stochastic processes study time-indexed random phenomena; Poisson processes model random discrete events, Brownian/Wiener processes model continuous Gaussian noise, and Ito calculus provides the integration and chain rule needed to manipulate continuous-time stochastic differential equations (SDEs).

What Is a Stochastic Process?

A stochastic process is a collection of random variables indexed by time: {Xt}tT\{X_t\}_{t\in T}, where TT is typically {0,1,2,}\{0,1,2,\dots\} (discrete time) or an interval [0,)[0,\infty) (continuous time). Intuitively, a stochastic process describes the evolution of a random system observed at different times. Two canonical continuous-time families are Poisson processes (jump/counting processes) and Brownian/Wiener processes (continuous-path Gaussian noise).

Why study these two? Poisson processes capture "events" that occur at random times (phone calls, earthquakes), while Brownian motion captures the accumulation of tiny, independent random disturbances (particle diffusion, financial returns at fine time scales). They also form building blocks for more complicated continuous-time models (jump-diffusions, renewal processes, SDEs) used across queueing, finance, physics and engineering.

Connection to prerequisites

  • In Markov Chains, we learned about memoryless dynamics (the Markov property) and transition matrices. Many continuous-time processes inherit a memoryless property (e.g., exponential interarrival times in Poisson processes) and are continuous-time Markov processes.
  • In Common Distributions, we learned the Poisson, exponential and normal laws; these are the marginal/increment distributions for Poisson and Wiener processes.
  • In Integrals, we learned Riemann sums and limits; in continuous-time stochastic calculus we use mean-square limits of Riemann-type sums to define stochastic integrals.

Poisson process — definition and intuition

A counting process N(t)N(t) is a Poisson process with rate λ>0\lambda>0 if:

1) N(0)=0N(0)=0.

2) It has independent increments: for 0s<t0\le s<t, N(t)N(s)N(t)-N(s) is independent of the past up to time ss (this is a continuous-time analog of the Markov property learned in Markov Chains).

3) It has stationary increments: distribution depends only on tst-s.

4) P(N(h)=1)=λh+o(h)P(N(h)=1)=\lambda h+o(h), P(N(h)2)=o(h)P(N(h)\ge2)=o(h) as h0h\downarrow0 (no multiple jumps in infinitesimal time).

The exact distribution: for t>0t>0,

P(N(t)=k)=eλt(λt)kk!.P(N(t)=k)=e^{-\lambda t}\frac{(\lambda t)^k}{k!}.

Concrete numeric example: with λ=3\lambda=3 events/hour and t=2t=2 hours, N(2)Poisson(6)N(2)\sim\mathrm{Poisson}(6) so

P(N(2)=5)=e6655!0.1606.P(N(2)=5)=e^{-6}\frac{6^5}{5!}\approx 0.1606.

Interarrival times T1,T2,T_1,T_2,\dots are iid exponential(λ)(\lambda); e.g., with λ=3\lambda=3, P(T1>1)=e310.0498P(T_1>1) = e^{-3\cdot1}\approx 0.0498 (probability the first event takes more than 1 hour).

Brownian motion / Wiener process — definition and intuition

A standard Brownian motion (or Wiener process) {Wt}t0\{W_t\}_{t\ge0} satisfies:

1) W0=0W_0=0.

2) Independent increments.

3) Stationary Gaussian increments: WtWsN(0,ts)W_t-W_s\sim N(0, t-s) for t>st>s.

4) Almost surely continuous paths.

This process is the central continuous-time Gaussian model and arises as a scaling limit of random walks (Donsker's invariance principle). Numeric example: W2W1N(0,1)W_2-W_1\sim N(0,1), so P(W2W1>1.96)0.05P(|W_2-W_1|>1.96)\approx0.05.

Key qualitative facts:

  • Scalability: aWt/a2aW_{t/a^2} is also Brownian motion (diffusive scaling).
  • Paths are a.s. nowhere differentiable, but of finite quadratic variation: the sum of squared increments over a partition of [0,t][0,t] tends to tt.

Continuous-time stochastic models

Two canonical classes: pure jump (Poisson) and continuous diffusion (Brownian). Real-world models often combine both (jump-diffusions). The mathematical machinery to manipulate SDEs driven by Brownian motion is Ito calculus, which modifies the ordinary chain rule to account for the quadratic variation of Brownian paths.

This section sets the stage: the next sections derive core formulas for Poisson processes and Brownian/Ito calculus and show worked examples.

Core Mechanic 1: Poisson Processes — distributions, interarrivals, thinning and superposition

Distribution and derivation (binomial limit)

A standard construction shows the Poisson law as the limit of Binomial(n,pn)(n,p_n) with pn=λ/np_n=\lambda/n and nn\to\infty: for fixed kk,

(nk)pnk(1pn)nkeλλkk!.\binom{n}{k}p_n^k(1-p_n)^{n-k}\to e^{-\lambda}\frac{\lambda^k}{k!}.

Concrete numeric check: take λ=2\lambda=2, n=1000n=1000, pn=0.002p_n=0.002. The probability of k=0k=0 is approximately (10.002)1000e20.1353(1-0.002)^{1000}\approx e^{-2}\approx0.1353.

Interarrival times and memoryless property

From the Poisson process with rate λ\lambda the waiting time until the first event T1T_1 satisfies

P(T1>t)=P(N(t)=0)=eλt,P(T_1>t)=P(N(t)=0)=e^{-\lambda t},

so T1Exp(λ)T_1\sim\mathrm{Exp}(\lambda). Exponential distributions are memoryless: P(T1>t+sT1>t)=P(T1>s)P(T_1>t+s\mid T_1>t)=P(T_1>s). In Markov Chains, we saw discrete memoryless geometric waiting times; exponential is the continuous analogue.

Order statistics representation

Given N(t)=nN(t)=n, the nn arrival times conditional on N(t)=nN(t)=n are distributed as the order statistics of nn iid Uniform(0,t)(0,t) variables. Example: with λ=3\lambda=3, t=2t=2 and conditioning on N(2)=2N(2)=2, the two arrival times have joint density equal to 2!/222!/2^2 on 0<u1<u2<20< u_1<u_2<2; marginally each arrival is likely near the center.

Superposition and thinning

  • Superposition: if N1,N2N_1,N_2 are independent Poisson processes with rates λ1,λ2\lambda_1,\lambda_2, then N=N1+N2N=N_1+N_2 is Poisson(λ1+λ2)(\lambda_1+\lambda_2).

Numeric example: merging two independent streams at rates 2 and 5 per hour yields a Poisson rate 7 per hour.

  • Thinning: each arrival of a Poisson(λ)(\lambda) process is kept independently with probability pp to produce a sub-process; the kept events form Poisson(pλ)(p\lambda) and the discarded ones Poisson((1p)λ)((1-p)\lambda), independent.

Numeric example: thinning with p=0.3p=0.3 a Poisson process with λ=10\lambda=10 gives a kept process of rate $3$.

Moment generating and PGF

The probability generating function (PGF) for N(t)N(t) is

GN(s)=E[sN(t)]=exp(λt(s1)).G_N(s)=E[s^{N(t)}]=\exp\big(\lambda t(s-1)\big).

Numeric example: with λ=4\lambda=4, t=0.5t=0.5, GN(0.5)=exp(40.5(0.51))=exp(2(0.5))=e10.3679G_N(0.5)=\exp(4\cdot0.5(0.5-1))=\exp(2(-0.5))=e^{-1}\approx0.3679.

A simple applied calculation — probability of at least k events

Question: rate λ=2\lambda=2 per hour, time t=3t=3 hours. What is P(N(3)3)P(N(3)\ge3)?

Solution: N(3)Poisson(6)N(3)\sim\mathrm{Poisson}(6), so

P(N(3)3)=1k=02e66kk!=1e6(1+6+362)1e6(1+6+18)1e62510.002478250.9380.P(N(3)\ge3)=1-\sum_{k=0}^2 e^{-6}\frac{6^k}{k!}=1- e^{-6}\left(1+6+\frac{36}{2}\right)\approx1- e^{-6}(1+6+18)\approx1- e^{-6}\cdot25\approx 1-0.002478\cdot25\approx 0.9380.

Generator viewpoint (continuous-time Markov chains)

For a pure birth Poisson process (counting upward by ones), its forward generator acting on bounded functions f:Z0Rf:\mathbb{Z}_{\ge0}\to\mathbb{R} is

(Lf)(n)=λ(f(n+1)f(n)).(\mathcal{L}f)(n)=\lambda\big(f(n+1)-f(n)\big).

This mirrors the discrete Markov Chains generator learned earlier, now with rate λ\lambda for jumps. For example, choose f(n)=nf(n)=n. Then (Lf)(n)=λ(\mathcal{L}f)(n)=\lambda and solves the ODE dE[Nt]/dt=E[(Lf)(Nt)]=λdE[N_t]/dt=E[(\mathcal{L}f)(N_t)]=\lambda, consistent with E[Nt]=λtE[N_t]=\lambda t.

Takeaway from this section: Poisson processes give a clean, tractable model for random discrete events; many useful transformations (conditioning, thinning, superposition) are exact and have simple probabilistic proofs that rely on independent and stationary increments and the exponential memoryless property. All formulas above had concrete numeric instantiations to make computation immediate.

Core Mechanic 2: Brownian Motion, Quadratic Variation, and Ito Calculus

Brownian motion (Wiener process) recap and basic computations

Recall WtW_t is standard Brownian motion with independent stationary Gaussian increments: WtWsN(0,ts)W_t-W_s\sim N(0,t-s). Key moment: E[Wt]=0E[W_t]=0, Var(Wt)=t\mathrm{Var}(W_t)=t. Concrete numeric example: for t=4t=4, W4N(0,4)W_4\sim N(0,4) so P(W4>2)=P(N(0,1)>1)0.3173P(|W_4|>2) = P(|N(0,1)|>1)\approx0.3173 because 2/4=12/\sqrt{4}=1.

Quadratic variation — the source of Ito's extra term

Take a partition Πn={0=t0<t1<<tn=t}\Pi_n=\{0=t_0<t_1<\dots<t_n=t\} with mesh max(ti+1ti)0\max(t_{i+1}-t_i)\to0. Define the quadratic variation along the partition:

Q(Πn)=i=0n1(Wti+1Wti)2.Q(\Pi_n)=\sum_{i=0}^{n-1}\big(W_{t_{i+1}}-W_{t_i}\big)^2.

Because increments are independent with variance ti+1tit_{i+1}-t_i, we have

E[Q(Πn)]=i(ti+1ti)=t.E[Q(\Pi_n)]=\sum_{i}(t_{i+1}-t_i)=t.

Also Var(Q(Πn))0\mathrm{Var}(Q(\Pi_n))\to0 as mesh shrinks, so Q(Πn)tQ(\Pi_n)\to t in probability and almost surely along appropriate subsequences. Concrete numeric check: take uniform partition into 100 intervals on [0,1][0,1]; each increment has variance $0.01$, expected sum of squares is 1.

This nonzero quadratic variation (unlike smooth paths where it is 0) causes Ito calculus to acquire an extra term relative to ordinary calculus.

Ito integral — definition sketch

Let {ϕ(t)}\{\phi(t)\} be a predictable process (non-anticipating, i.e., depends only on the past). For simple processes that are piecewise constant on partitions, define

In=i=0n1ϕ(ti)(Wti+1Wti).I_n=\sum_{i=0}^{n-1} \phi(t_i)\big(W_{t_{i+1}}-W_{t_i}\big).

The Ito integral is the mean-square limit as the mesh goes to zero:

0tϕ(s)dWs:=limmesh0In\int_0^t \phi(s)\,dW_s := \lim_{\text{mesh}\to0} I_n

with convergence in L^2. Example: if ϕ(s)=1\phi(s)=1 constant, then the integral is WtW_t itself: 0t1dWs=Wt\int_0^t 1\,dW_s=W_t.

Isometry and computations

The Ito isometry gives

E[(0tϕ(s)dWs)2]=E[0tϕ(s)2ds].E\left[\left(\int_0^t \phi(s)\,dW_s\right)^2\right]=E\left[\int_0^t \phi(s)^2\,ds\right].

Numeric example: if ϕ(s)=2\phi(s)=2 constant on [0,1][0,1], then E[(012dWs)2]=E[014ds]=4E[(\int_0^1 2\,dW_s)^2]=E[\int_0^1 4\,ds]=4. Indeed 012dWsN(0,4)\int_0^1 2\,dW_s\sim N(0,4).

Ito's formula (stochastic chain rule)

If XtX_t solves an SDE

dXt=a(t,Xt)dt+b(t,Xt)dWtdX_t = a(t,X_t)\,dt + b(t,X_t)\,dW_t

and f(t,x)f(t,x) is C1,2C^{1,2} (once differentiable in tt, twice in xx), then

df(t,Xt)=(tf+axf+12b2xxf)(t,Xt)dt+(bxf)(t,Xt)dWt.df(t,X_t) = \left(\partial_t f + a\partial_x f + \tfrac12 b^2 \partial_{xx}f\right)(t,X_t)\,dt + (b\partial_x f)(t,X_t)\,dW_t.

Note the 12b2xxf\tfrac12 b^2 \partial_{xx}f term coming from quadratic variation. Concrete numeric application: let f(x)=x2f(x)=x^2 and Xt=WtX_t=W_t (so a=0,b=1a=0,b=1). Then Ito's formula yields

d(Wt2)=2WtdWt+1dt.d(W_t^2) = 2W_t\,dW_t + 1\,dt.

Take expectation to get dE[Wt2]=dtdE[W_t^2]=dt, so E[Wt2]=tE[W_t^2]=t, matching the variance property. Numeric check: at t=3t=3, E[W32]=3E[W_3^2]=3.

Proof sketch of Ito's formula for f(x)f(x) (time-homogeneous case)

Use Taylor expansion on increments:

f(Xt+Δt)f(Xt)f(Xt)ΔXt+12f(Xt)(ΔXt)2+o((ΔXt)2).f(X_{t+\Delta t})-f(X_t) \approx f'(X_t)\Delta X_t + \tfrac12 f''(X_t)(\Delta X_t)^2 + o((\Delta X_t)^2).

For ΔXt=aΔt+bΔWt\Delta X_t = a\Delta t + b\Delta W_t, the linear term gives f(Xt)(aΔt+bΔWt)f'(X_t)(a\Delta t + b\Delta W_t); the quadratic term yields 12f(Xt)b2(ΔWt)2\tfrac12 f''(X_t)b^2(\Delta W_t)^2. But (ΔWt)2Δt(\Delta W_t)^2\approx \Delta t (quadratic variation), so the second-order term contributes 12b2f(Xt)Δt\tfrac12 b^2 f''(X_t)\Delta t. Higher-order terms vanish in the limit because ΔWt=O(Δt)\Delta W_t = O(\sqrt{\Delta t}).

Martingales and exponential martingales

A useful family: for constant θ\theta, the process

Mt=exp(θWt12θ2t)M_t = \exp\left(\theta W_t - \tfrac12\theta^2 t\right)

is a martingale. Numeric example: with θ=1\theta=1 and t=2t=2, E[M2]=1E[M_2]=1 and M2=exp(W21)M_2=\exp(W_2 - 1).

SDE example and solution technique

Consider the linear SDE (Ornstein-Uhlenbeck variant) for constants θ,σ\theta,\sigma:

dXt=θXtdt+σdWt,X0=x0.dX_t = -\theta X_t\,dt + \sigma\,dW_t,\qquad X_0=x_0.

The integrating factor solution (variation of constants) yields

Xt=x0eθt+σ0teθ(ts)dWs.X_t = x_0 e^{-\theta t} + \sigma\int_0^t e^{-\theta (t-s)}\,dW_s.

Numeric example: with θ=1,σ=2,x0=1,t=1\theta=1,\sigma=2,x_0=1,t=1, the expectation is E[X1]=e10.3679E[X_1]= e^{-1}\approx0.3679 and variance

Var(X1)=σ201e2(1s)ds=401e2(1s)ds=401e2udu=4(1e2)/2=2(1e2)2(10.1353)1.7294.\mathrm{Var}(X_1)=\sigma^2\int_0^1 e^{-2(1-s)}\,ds=4\int_0^1 e^{-2(1-s)}\,ds=4\int_0^1 e^{-2u}du=4(1-e^{-2})/2 =2(1-e^{-2})\approx2(1-0.1353)\approx1.7294.

Takeaway: Ito calculus alters the ordinary calculus chain rule by a quadratic-variation term. The Ito integral is a mean-square limit defined for non-anticipating integrands, and Ito's formula is the workhorse for manipulating functions of SDE solutions.

Applications and Connections: where these tools go and why they matter

Black–Scholes and quantitative finance

One of the clearest applications is option pricing. Model a stock price by the geometric SDE

dSt=μStdt+σStdWt,S0=s0.dS_t = \mu S_t\,dt + \sigma S_t\,dW_t,\qquad S_0=s_0.

Ito's formula applied to logSt\log S_t gives

dlogSt=(μ12σ2)dt+σdWt,d\log S_t = \left(\mu - \tfrac12\sigma^2\right)dt + \sigma\,dW_t,

so the explicit solution is

St=s0exp((μ12σ2)t+σWt).S_t = s_0\exp\left(\left(\mu - \tfrac12\sigma^2\right)t + \sigma W_t\right).

Concrete numeric example: take s0=100s_0=100, μ=0.05\mu=0.05, σ=0.2\sigma=0.2, t=1t=1 year. Then

E[S1]=s0eμt=100e0.051001.05127105.127.E[S_1] = s_0 e^{\mu t} = 100e^{0.05}\approx100\cdot1.05127\approx 105.127.

Black–Scholes uses risk-neutral pricing (μ\mu replaced by risk-free rate rr) and properties of lognormal distributions to price European options analytically.

Queueing, telecommunications and reliability

Poisson processes are the standard model for arrival processes in queues (e.g., M/M/1 queue). Key performance measures — waiting times and queue lengths — are derived from Poisson/exponential properties. Example numerical calculation: with arrival rate λ=5\lambda=5/hr and service rate μ=6\mu=6/hr, utilization ρ=λ/μ0.833\rho=\lambda/\mu\approx0.833; the stationary average number in system for M/M/1 is ρ/(1ρ)5\rho/(1-\rho)\approx5 customers.

Physics and diffusion

Brownian motion models particle diffusion: the heat equation is the forward equation (Fokker–Planck) for the probability density of Brownian motion. The diffusion constant ties Var(Wt)\mathrm{Var}(W_t) to physical diffusivity.

Stochastic control, filtering and estimation

Ito calculus enables stochastic optimal control (Hamilton–Jacobi–Bellman PDEs) and stochastic filtering (Kalman–Bucy filter for linear Gaussian SDEs). For example, the linear SDE + Gaussian noise assumptions produce closed-form filters because all conditional distributions remain Gaussian.

Statistics for stochastic processes

Parameter estimation for rates λ\lambda in Poisson models or drift/diffusion coefficients in SDEs uses likelihoods based on increments and Girsanov transformations. For example, by observing a Poisson process on [0,T][0,T] with N(T)=nN(T)=n, the MLE for λ\lambda is λ^=n/T\hat{\lambda}=n/T.

Machine learning and stochastic optimisation

Stochastic gradient methods can be viewed as discrete-time stochastic processes; diffusion limits lead to SDE approximations describing algorithm behaviour and escape probabilities from basins of attraction.

Hybrid models and jump-diffusions

Real applications often combine Poisson jumps and Brownian diffusion: e.g., financial returns may have continuous Gaussian noise plus occasional large jumps modeled by a compound Poisson process. SDEs with jumps require an extended Ito formula incorporating jump terms.

Practical modeling checklist

  • Decide whether events are discrete (Poisson) or continuous/noisy (Brownian) or both.
  • Check stationarity/independent increments assumptions; these yield exact tractability.
  • Use Ito's formula to convert between SDEs and deterministic PDEs (Fokker–Planck, backward Kolmogorov).

Downstream methods enabled

  • SDE solution techniques and explicit formulas (Black–Scholes, Ornstein–Uhlenbeck).
  • Stochastic stability, large deviations, and exit-time problems.
  • Nonlinear filtering and stochastic control.

Concrete final illustration: pricing expectation under geometric Brownian motion. Using the StS_t formula above with s0=100,μ=0.05,σ=0.2,t=1s_0=100,\mu=0.05,\sigma=0.2,t=1, the distribution of S1S_1 is lognormal, and the probability P(S1>110)=P(σW1>log(1.1)(μ12σ2))P(S_1>110)=P\left(\sigma W_1 > \log(1.1) - (\mu-\tfrac12\sigma^2)\right). Numeric compute: log(1.1)0.09531\log(1.1)\approx0.09531, (μ0.5σ2)=0.050.02=0.03(\mu-0.5\sigma^2)=0.05-0.02=0.03, so threshold for W1W_1 is (0.095310.03)/0.20.32755(0.09531-0.03)/0.2\approx0.32755. Thus P(S1>110)=P(W1>0.32755)0.3716P(S_1>110)=P(W_1>0.32755)\approx0.3716.

This section shows how Poisson processes, Brownian motion and Ito calculus are not abstract curiosities but precise tools that produce explicit models, closed-form calculations, and pathwise constructions for a wide range of applications.

Worked Examples (3)

Poisson count probability

Rate λ=2\lambda=2 events/hour; find P(N(3)3)P(N(3)\ge3) for t=3t=3 hours.

  1. Recognize N(3)Poisson(λt)=Poisson(23)=Poisson(6)N(3)\sim\mathrm{Poisson}(\lambda t)=\mathrm{Poisson}(2\cdot3)=\mathrm{Poisson}(6).

  2. Compute probabilities for k=0,1,2k=0,1,2 and subtract from 1: P(N(3)3)=1k=02e66kk!P(N(3)\ge3)=1-\sum_{k=0}^2 e^{-6}\frac{6^k}{k!}.

  3. Calculate term-by-term: e6600!=e60.00247875e^{-6}\frac{6^0}{0!}=e^{-6}\approx0.00247875.

  4. Next: e6611!=6e60.0148725e^{-6}\frac{6^1}{1!}=6e^{-6}\approx0.0148725; then e6622!=18e60.0446175e^{-6}\frac{6^2}{2!}=18e^{-6}\approx0.0446175.

  5. Sum the three: 0.00247875+0.0148725+0.04461750.061968750.00247875+0.0148725+0.0446175\approx0.06196875. Subtract from 1 to get 0.93803125\approx0.93803125.

Insight: This example uses the defining Poisson distribution formula and shows how to compute tail probabilities via finite sums. It reinforces intuition that rare low counts are unlikely when the mean is large (mean 6).

Ito formula on $f(W_t)=W_t^2$

Let WtW_t be standard Brownian motion. Use Ito's formula to compute d(Wt2)d(W_t^2) and then find E[Wt2]E[W_t^2] for t=3t=3.

  1. Set f(x)=x2f(x)=x^2. Then f(x)=2xf'(x)=2x, f(x)=2f''(x)=2.

  2. Apply Ito's formula (time-homogeneous case): df(Wt)=f(Wt)dWt+12f(Wt)dtdf(W_t)=f'(W_t)dW_t + \tfrac12 f''(W_t) dt.

  3. Substitute derivatives: d(Wt2)=2WtdWt+122dt=2WtdWt+dtd(W_t^2)=2W_t\,dW_t + \tfrac12\cdot2\,dt = 2W_t\,dW_t + dt.

  4. Take expectations: E[d(Wt2)]=E[2WtdWt]+E[dt]E[d(W_t^2)] = E[2W_t\,dW_t] + E[dt]. The stochastic integral has zero expectation, so dE[Wt2]=dtdE[W_t^2]=dt.

  5. Integrate from 0 to 3: E[W32]=03ds=3E[W_3^2]=\int_0^3 ds = 3.

Insight: Ito's formula produces an extra deterministic dtdt term absent in classical chain rule; that term exactly accounts for the quadratic variation and yields the known variance of Brownian motion.

Ornstein–Uhlenbeck moments

Consider dXt=Xtdt+2dWtdX_t = -X_t\,dt + 2\,dW_t, X0=1X_0=1. Compute E[X1]E[X_1] and Var(X1)\mathrm{Var}(X_1).

  1. Solve via integrating factor: multiply by ete^{t} to get d(etXt)=et2dWtd(e^{t}X_t)= e^{t}\cdot 2\,dW_t.

  2. Integrate: etXt=X0+20tesdWse^{t}X_t = X_0 + 2\int_0^t e^{s}\,dW_s, so Xt=X0et+20te(ts)dWsX_t = X_0 e^{-t} + 2\int_0^t e^{-(t-s)}\,dW_s.

  3. Take expectation: E[Xt]=X0et=etE[X_t]=X_0 e^{-t}=e^{-t}. For t=1t=1, E[X1]=e10.3679E[X_1]=e^{-1}\approx0.3679.

  4. Compute variance using Ito isometry: Var(Xt)=40te2(ts)ds=40te2udu\mathrm{Var}(X_t)=4\int_0^t e^{-2(t-s)}\,ds=4\int_0^t e^{-2u}\,du with u=tsu=t-s.

  5. Evaluate for t=1t=1: Var(X1)=4(1e2)/2=2(1e2)2(10.1353)1.7294\mathrm{Var}(X_1)=4(1-e^{-2})/2 =2(1-e^{-2})\approx2(1-0.1353)\approx1.7294.

Insight: Linear SDEs can be solved explicitly; integrals against Brownian motion yield Gaussian random variables whose variance follows from the Ito isometry. The result shows mean reversion (exponential decay) and stationary variance as tt\to\infty.

Key Takeaways

  • A stochastic process is a time-indexed family of random variables; Poisson processes model discrete random events, while Brownian/Wiener processes model continuous Gaussian noise.

  • Poisson processes have independent, stationary increments with Poisson marginals; interarrival times are iid exponential (memoryless).

  • Brownian motion has Gaussian independent increments and nonzero quadratic variation: sums of squared increments over a partition converge to elapsed time.

  • The Ito integral is defined for non-anticipating integrands as an L^2 limit; the Ito isometry relates second moments of the integral to the integral of the squared integrand.

  • Ito's formula extends the chain rule by adding a half the second derivative times the diffusion coefficient squared (the quadratic variation term).

  • Many applied models (Black–Scholes, queuing, diffusion, filtering) follow directly from these building blocks; linear SDEs often admit explicit solutions via integrating factors.

  • Always check assumptions: independent increments, stationarity, continuity of paths (or presence of jumps) determine which tools apply.

Common Mistakes

  • Treating Brownian paths as differentiable: Brownian motion is almost surely nowhere differentiable; attempts to apply ordinary calculus to sample paths produce wrong terms (you need Ito calculus).

  • Forgetting the Ito correction: applying the classical chain rule to SDEs and omitting the 12b2f\tfrac12 b^2 f'' term leads to incorrect drift terms (a common error in derivations).

  • Confusing independent increments with independent values at time points: increments over disjoint intervals are independent, but values like WtW_t and WsW_s are not independent unless tst-s covers the interval from 0 (i.e., unless one is difference from the other).

  • Misusing memoryless property: exponential interarrival times are memoryless, but conditional distributions such as arrival times given counts are order statistics, not independent exponentials.

Practice

easy

Easy: A Poisson process has rate λ=4\lambda=4 per hour. What is the probability of exactly 3 events in a 30-minute interval?

Hint: Compute λt\lambda t for t=0.5t=0.5 hours and use the Poisson pmf.

Show solution

Here λt=40.5=2\lambda t = 4\cdot0.5=2. So P(N(0.5)=3)=e2233!=e286=43e21.33330.135340.18045P(N(0.5)=3)=e^{-2}\frac{2^3}{3!}=e^{-2}\frac{8}{6}=\frac{4}{3}e^{-2}\approx1.3333\cdot0.13534\approx0.18045.

medium

Medium: Let WtW_t be standard Brownian motion. Use Ito's formula to compute dYtdY_t when Yt=exp(at+bWt)Y_t=\exp(a t + b W_t) for constants a,ba,b. Then compute E[Yt]E[Y_t] for given a=0.1,b=0.5,t=2a=0.1,b=0.5,t=2.

Hint: Apply Ito to f(t,x)=exp(at+bx)f(t,x)=\exp(a t + b x). Remember tf=af\partial_t f = a f, xf=bf\partial_x f = b f, xxf=b2f\partial_{xx}f = b^2 f.

Show solution

Ito gives

dYt=(af+12b2f)dt+bfdWt=f(a+12b2)dt+bfdWt.dY_t = (a f + \tfrac12 b^2 f)\,dt + b f\,dW_t = f\Big(a + \tfrac12 b^2\Big)dt + b f\,dW_t.

Taking expectations kills the dWtdW_t term: dE[Yt]=E[f](a+12b2)dtdE[Y_t]=E[f]\Big(a+\tfrac12 b^2\Big)dt, so E[Yt]=e(a+12b2)tE[Y0]E[Y_t]=e^{(a+\tfrac12 b^2)t}E[Y_0]. With Y0=1Y_0=1, a=0.1,b=0.5,t=2a=0.1,b=0.5,t=2, we get exponent (0.1+0.52/2)2=(0.1+0.125)2=0.2252=0.45(0.1+0.5^2/2)\cdot2=(0.1+0.125)\cdot2=0.225\cdot2=0.45, so E[Y2]=e0.451.571E[Y_2]=e^{0.45}\approx1.571.

hard

Hard: Consider the SDE dXt=μXtdt+σXtdWtdX_t = \mu X_t\,dt + \sigma X_t\,dW_t with X0=x0>0X_0=x_0>0. (This is geometric Brownian motion.) Derive the explicit solution and compute Var(Xt)\mathrm{Var}(X_t) in terms of x0,μ,σ,tx_0,\mu,\sigma,t.

Hint: Apply Ito to logXt\log X_t to linearize; then use the known moments of the lognormal distribution.

Show solution

Using Ito on f(x)=logxf(x)=\log x: dlogXt=(μ12σ2)dt+σdWtd\log X_t = (\mu - \tfrac12\sigma^2)dt + \sigma dW_t. Integrate to get

logXt=logx0+(μ12σ2)t+σWt.\log X_t = \log x_0 + (\mu - \tfrac12\sigma^2)t + \sigma W_t.

Exponentiate:

Xt=x0exp((μ12σ2)t+σWt).X_t = x_0\exp\left((\mu - \tfrac12\sigma^2)t + \sigma W_t\right).

Since σWtN(0,σ2t)\sigma W_t\sim N(0,\sigma^2 t), XtX_t is lognormal. Its mean is E[Xt]=x0eμtE[X_t]=x_0 e^{\mu t}. Its second moment is

E[Xt2]=x02e2μt+σ2tE[X_t^2] = x_0^2 e^{2\mu t + \sigma^2 t}

(since E[e2σWt]=e22σ2t/2=e2σ2tE[e^{2\sigma W_t}] = e^{2^2\sigma^2 t/2} = e^{2\sigma^2 t}, combine with exponent). Therefore

Var(Xt)=E[Xt2](E[Xt])2=x02e2μt(eσ2t1).\mathrm{Var}(X_t) = E[X_t^2] - (E[X_t])^2 = x_0^2 e^{2\mu t}\left(e^{\sigma^2 t}-1\right).

This completes the derivation.

Connections

Looking back: In Markov Chains we learned memoryless transitions and generators; Poisson processes are continuous-time Markov chains with exponential holding times, and their generator Lf(n)=λ(f(n+1)f(n))\mathcal{L}f(n)=\lambda(f(n+1)-f(n)) mirrors discrete generators. From Common Distributions we directly use Poisson, exponential and normal laws as the marginals/increments of Poisson and Brownian processes. From Integrals, the idea of Riemann sums and limits underlies the construction of the Ito integral (mean-square limits of adapted Riemann sums).

Looking forward: Mastery of Poisson processes and Ito calculus is essential for studying stochastic differential equations (SDEs), which underpin Black–Scholes option pricing, stochastic control and filtering (e.g., Kalman–Bucy, nonlinear filters), and for linking probabilistic models to PDEs (Fokker–Planck and backward Kolmogorov equations). Advanced topics that rely on these foundations include large deviations for stochastic processes, Malliavin calculus (stochastic calculus of variations), jump-diffusion models, and modern stochastic numerical methods (Euler–Maruyama, Milstein schemes). Specific prerequisite-to-downstream map: Poisson/exponential results -> queueing theory and point-process statistics; Brownian/Ito -> SDE theory, PDE connections, financial mathematics, stochastic filtering and control.

Quality: pending (0.0/5)