Reoptimization Nearly Solves Weakly Coupled Markov Decision Processes - Systèmes Répartis, Calcul Parallèle et Réseaux Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2024

Reoptimization Nearly Solves Weakly Coupled Markov Decision Processes

Résumé

We propose a new policy, called the LP-update policy, to solve finite horizon weakly-coupled Markov decision processes. The latter can be seen as multi-constraint multi-action bandits, and generalize the classical restless bandit problems. Our solution is based on re-solving periodically a relaxed version of the original problem, that can be cast as a linear program (LP). When the problem is made of $N$ statistically identical sub-components, we show that the LP-update policy becomes asymptotically optimal at rate $O(T^2/\sqrt{N})$. This rate can be improved to $O(T/\sqrt{N})$ if the problem satisfies some ergodicity property and to $O(1/N)$ if the problem is non-degenerate. The definition of non-degeneracy extends the same notion for restless bandits. By using this property, we also improve the computational efficiency of the LP-update policy. We illustrate the performance of our policy on randomly generated examples, as well as a generalized applicant screening problem, and show that it outperforms existing heuristics.
Fichier principal
Vignette du fichier
LP_update_for_weakly_coupled_MDP.pdf (946.02 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-04570177 , version 1 (06-05-2024)

Licence

Paternité

Identifiants

  • HAL Id : hal-04570177 , version 1

Citer

Nicolas Gast, Bruno Gaujal, Chen Yan. Reoptimization Nearly Solves Weakly Coupled Markov Decision Processes. 2024. ⟨hal-04570177⟩
0 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More