|
CRAAM
2.0.0
Robust and Approximate Markov Decision Processes
|
Represents a single transition between two states after taking an action:
\[ (s, a, s', r, w) \]
where: More...
#include <Samples.hpp>
Public Member Functions | |
| Sample (State state_from, Action action, State state_to, prec_t reward, prec_t weight, long step, long run) | |
| State | state_from () const |
| Original state. | |
| Action | action () const |
| Action taken. | |
| State | state_to () const |
| Destination state. | |
| prec_t | reward () const |
| Reward associated with the sample. | |
| prec_t | weight () const |
| Sample weight. | |
| long | step () const |
| Number of the step in an one execution of the simulation. | |
| long | run () const |
| Number of the actual execution. | |
Protected Attributes | |
| State | _state_from |
| Original state. | |
| Action | _action |
| Action taken. | |
| State | _state_to |
| Destination state. | |
| prec_t | _reward |
| Reward associated with the sample. | |
| prec_t | _weight |
| Sample weight. | |
| long | _step |
| Number of the step in an one execution of the simulation. | |
| long | _run |
| Number of the actual execution. | |
Represents a single transition between two states after taking an action:
\[ (s, a, s', r, w) \]
where:
In addition, the sample also includes step and the run. These are used for example to compute the return from samples.
| State | MDP state: \( s, s'\) |
| Action | MDP action: \( a \) |
1.8.13