|
| PolicyDeterministic () |
| All actions will be optimized.
|
|
| PolicyDeterministic (indvec policy) |
| A partial policy that can be used to fix some actions policy[s] = -1 means that the action should be optimized in the state policy of length 0 means that all actions will be optimized.
|
|
Solution | new_solution (size_t statecount, numvec valuefunction) const |
|
template<class SType > |
prec_t | update_solution (Solution &solution, const SType &state, long stateid, const numvec &valuefunction, prec_t discount) const |
| Computed the Bellman update and updates the solution to the best response It does not update the value function. More...
|
|
template<class SType > |
prec_t | update_value (const Solution &solution, const SType &state, long stateid, const numvec &valuefunction, prec_t discount) const |
| Computes a fixed Bellman update using the current solution policy. More...
|
|
|
indvec | policy |
| Partial policy specification (action -1 is ignored and optimized)
|
|
|
void | process_valuefunction (size_t statecount, numvec &valuefunction) const |
|
indvec | process_policy (size_t statecount) const |
|
◆ update_solution()
template<class SType >
prec_t craam::algorithms::PolicyDeterministic::update_solution |
( |
Solution & |
solution, |
|
|
const SType & |
state, |
|
|
long |
stateid, |
|
|
const numvec & |
valuefunction, |
|
|
prec_t |
discount |
|
) |
| const |
|
inline |
Computed the Bellman update and updates the solution to the best response It does not update the value function.
- Returns
- New value for the state
◆ update_value()
template<class SType >
prec_t craam::algorithms::PolicyDeterministic::update_value |
( |
const Solution & |
solution, |
|
|
const SType & |
state, |
|
|
long |
stateid, |
|
|
const numvec & |
valuefunction, |
|
|
prec_t |
discount |
|
) |
| const |
|
inline |
Computes a fixed Bellman update using the current solution policy.
- Returns
- New value for the state
The documentation for this class was generated from the following file: