We consider discounted Markov decision processes (MDPs) with countably-infinite state spaces, finite action spaces, and unbounded rewards. Typical examples of such MDPs are inventory management and ...
Results that may be inaccessible to you are currently showing.
Hide inaccessible results