0
32kviews
Bellman's principle of optimality.
1 Answer
0
2.3kviews

Bellman’s principle of optimality: An optimal policy (set of decisions) has the property that whatever the initial state and decisions are, the remaining decisions must constitute and optimal policy with regard to the state resulting from the first decision.

Mathematically, this can be written as:

fN(x)=max.[r(dn)+fN1T(x,dn)]dnx

where fN(x)=the optimal return from an N-stage process when initial state is xr(dn)=immediate return due to decisiondnT(x,dn)=the transfer function which gives the resulting state{x}=set of admissible decisions

This equation is also known as a dynamic programming equation. It represents a necessary condition for optimality associated with the mathematical optimization method known as dynamic programming. It writes the value of a decision problem at a certain point in time in terms of the payoff from some initial choices and the value of the remaining decision problem that results from those initial choices. This breaks a dynamic optimization problem into simpler subproblems.

Please log in to add an answer.