4.7 Article

Reinforcement Learning for Joint Optimization of Multiple Rewards

Journal

JOURNAL OF MACHINE LEARNING RESEARCH
Volume 24, Issue -, Pages -

Publisher

MICROTOME PUBL

Keywords

-

Ask authors/readers for more resources

The paper discusses the problem of optimizing a non-linear function of the long term average rewards and proposes model-based and model-free algorithms to solve it. The proposed algorithms are shown to outperform conventional RL approaches in fairness scheduling and queueing system scheduling.
Finding optimal policies which maximize long term rewards of Markov Decision Processes requires the use of dynamic programming and backward induction to solve the Bellman optimality equation. However, many real-world problems require optimization of an objective that is non-linear in cumulative rewards for which dynamic programming cannot be applied directly. For example, in a resource allocation problem, one of the objectives is to maximize long-term fairness among the users. We notice that when an agent aim to optimize some function of the sum of rewards is considered, the problem loses its Markov nature. This paper addresses and formalizes the problem of optimizing a non-linear function of the long term average of rewards. We propose model-based and model-free algorithms to learn the policy, where the model-based policy is shown to achieve a regret of O similar to LKDS for K objectives combined with a concave L-Lipschitz function. Further, using the fairness in cellular base-station scheduling, and queueing system scheduling as examples, the proposed algorithm is shown to significantly outperform the conventional RL approaches.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available