I'm trying to study the Merton Model for portfolio optimization and the document doesn't explain a quite important step : if V(t,x)=sup{E[U(XT(ϕ)) | Xt=x] | ϕ an admissible trading strategy}
What are those regularity assumptions ? How can we prove them ?
Answer
This is an optimal control problem.
Consider a self-financing strategy π:=(πs)s∈[t,T] over the horizon [t,T] consisting in, over each infinitesimal period of time [t,t+dt[, investing a fraction πt of the current wealth in a risky asset St and placing the remaining part in the risk free asset Bt. Given the following dynamics dSt=St(μtdt+σtdWt)
Consider the value function V(t,x;(πs)s∈[t,T])=Et[U(Xπ,xT)]
The optimal control π∗t is the stochastic process such that (π∗s)s∈[t,T]=argsup(πs)s∈[t,T]V(t,x;(πs)s∈[t,T])
while the optimal cost is V(t,x)=Et[U(Xπ∗,xT)]
The optimal cost function solves the Hamilton-Jacobi-Bellman equations.
The proof can be obtained by viewing the control problem as a Dynamic Programming Problem and relying on Bellman's principle of optimality (see (1) below).
As @noob2 mentions, at some point the Itô differential of the optimal cost V(t,x) appears. Therefore regularity conditions are the usual conditions for Itô integration, both for Xt and V(t,x).
Some intuition V(t,x)=Et[U(Xπ∗,xT)]=Et[Et+dt[U(Xπ∗,x+dXt(π∗t)T)]]=Et[V(t+dt,x+dXt(π∗t))]=supπtEt[V(t+dt,x+dXt(πt))]=supπtEt[V(t,x)+dV(t,x)dtdt+dV(t,x)dxdXt+d2V(t,x)dx2d⟨X⟩t]=V(t,x)+dV(t,x)dtdt+supπt(dV(t,x)dxx(rt+πt(μt−rt))+12d2V(t,x)dx2x2π2tσ2t)dt
[edit]
The DPP point of view consists in viewing the optimal control (π∗s)s∈[t,T] as the "union" of what you choose to do over [t,t+dt[ and what you do over [t+dt,T[. Informally: (π∗s)s∈[t,T]=π∗t∪(π∗s)s∈[t+dt,T]
At this point, Bellman's optimality principle tells you that the restriction of the optimal control (π∗s)s∈[t+dt,T] is itself the optimal policy over the horizon [t+dt,T[. This is why in (1) you can write that Et+dt[U(Xπ∗,x+dXt(π∗t)T)]=V(t+dt,x+dXt(π∗t))
No comments:
Post a Comment