DvalueF
DvalueF is a term used in decision theory and machine learning to describe a family of value functions that incorporate a time- or state-dependent devaluation of future rewards. It generalizes the standard discounted value function by allowing an additional weighting, D(t) or D(s, a, t), to modulate the contribution of rewards over time or across states. This flexible framing enables modeling scenarios where distant outcomes are valued differently due to risk, uncertainty, or horizon considerations.
Formally, for a policy π in a Markov decision process, the DvalueF of a starting state s can
Computation and learning with DvalueF adapt standard temporal-difference or Monte Carlo methods by replacing the plain
Applications of DvalueF appear in long-horizon planning under uncertainty, risk-sensitive control, robotics, and finance, where near-term