1. How the Values of Each Cell Are Calculated
- : Value of the current state.
- : Action taken in state .
- : Probability of transitioning to state after action .
- : Immediate reward for transitioning to .
- : Discount factor (0 ≤ ≤ 1).
- If moving "right" from state leads to state with reward (step cost) and , the contribution to for this action is .
- The value is the maximum of such contributions across all actions.
- If :
- Future rewards are valued equally with immediate rewards.
- The agent prioritizes minimizing step costs (e.g., shorter paths) to maximize cumulative rewards.
- Example: A path with 5 steps (total cost = -0.05) is better than a path with 10 steps (total cost = -0.10), even if both reach the goal.
- If :
- Future rewards are discounted.
- The agent prefers shorter paths to "lock in" the goal reward sooner (e.g., > ).
- Example: With , a 5-step path yields , while a 10-step path yields .
- has no impact on path length, as the total reward is always 10 (goal reward).
- The agent can take any path (long or short) without penalty.
- Values are computed via the Bellman Equation, incorporating rewards and discounted future values.
- Gamma affects shortest-path preference only if step costs exist. Higher (closer to 1) still prioritizes shorter paths due to cumulative cost minimization, while lower amplifies this effect by devaluing delayed rewards.
Comments
Post a Comment
Please leave you comments