In the context of reinforcement learning and decision making, a **value function** is a function that estimates the expected return (or future rewards) that an agent can achieve from a given state or state-action pair. It plays a fundamental role in evaluating the optimality of policies, guiding the agent's decisions as it seeks to maximize its cumulative rewards over time.
The function being maximized in a optimization problem.
New to topics? Read the docs here!