CRAAM
2.0.0
Robust and Approximate Markov Decision Processes
|
Represents a single transition between two states after taking an action:
\[ (s, a, s', r, w) \]
where: More...
#include <Samples.hpp>
Public Member Functions | |
Sample (State state_from, Action action, State state_to, prec_t reward, prec_t weight, long step, long run) | |
State | state_from () const |
Original state. | |
Action | action () const |
Action taken. | |
State | state_to () const |
Destination state. | |
prec_t | reward () const |
Reward associated with the sample. | |
prec_t | weight () const |
Sample weight. | |
long | step () const |
Number of the step in an one execution of the simulation. | |
long | run () const |
Number of the actual execution. | |
Protected Attributes | |
State | _state_from |
Original state. | |
Action | _action |
Action taken. | |
State | _state_to |
Destination state. | |
prec_t | _reward |
Reward associated with the sample. | |
prec_t | _weight |
Sample weight. | |
long | _step |
Number of the step in an one execution of the simulation. | |
long | _run |
Number of the actual execution. | |
Represents a single transition between two states after taking an action:
\[ (s, a, s', r, w) \]
where:
In addition, the sample also includes step and the run. These are used for example to compute the return from samples.
State | MDP state: \( s, s'\) |
Action | MDP action: \( a \) |