1 paper accepted to L4DC

Our paper on Reinforcement Learning with almost sure constraints [1] has been accepted to the Learning for Dynamics and Control Conference!

[1] A. Castellano, H. Min, J. Bazerque, and E. Mallada, “Reinforcement Learning with Almost Sure Constraints,” in Proceedings of The 4th Annual Learning for Dynamics and Control Conference, 2022, pp. 559-570.
[Bibtex] [Abstract] [Download PDF]

In this work we address the problem of finding feasible policies for Constrained Markov Decision Processes under probability one constraints. We argue that stationary policies are not sufficient for solving this problem, and that a rich class of policies can be found by endowing the controller with a scalar quantity, so called budget, that tracks how close the agent is to violating the constraint. We show that the minimal budget required to act safely can be obtained as the smallest fixed point of a Bellman-like operator, for which we analyze its convergence properties. We also show how to learn this quantity when the true kernel of the Markov decision process is not known, while providing sample-complexity bounds. The utility of knowing this minimal budget relies in that it can aid in the search of optimal or near-optimal policies by shrinking down the region of the state space the agent must navigate. Simulations illustrate the different nature of probability one constraints against the typically used constraints in expectation.

@inproceedings{cmbm2022l4dc,
  abstract = {In this work we address the problem of finding feasible policies for Constrained Markov Decision Processes under probability one constraints. We argue that stationary policies are not sufficient for solving this problem, and that a rich class of policies can be found by endowing the controller with a scalar quantity, so called budget, that tracks how close the agent is to violating the constraint. We show that the minimal budget required to act safely can be obtained as the smallest fixed point of a Bellman-like operator, for which we analyze its convergence properties. We also show how to learn this quantity when the true kernel of the Markov decision process is not known, while providing sample-complexity bounds. The utility of knowing this minimal budget relies in that it can aid in the search of optimal or near-optimal policies by shrinking down the region of the state space the agent must navigate. Simulations illustrate the different nature of probability one constraints against the typically used constraints in expectation.},
  author = {Castellano, Agustin and Min, Hancheng and Bazerque, Juan and Mallada, Enrique},
  booktitle = {Proceedings of The 4th Annual Learning for Dynamics and Control Conference},
  editor = {Firoozi, Roya and Mehr, Negar and Yel, Esen and Antonova, Rika and Bohg, Jeannette and Schwager, Mac and Kochenderfer, Mykel},
  grants = {CAREER-1752362, TRIPODS-1934979, CPS-2136324},
  month = {6},
  pages = {559--570},
  publisher = {PMLR},
  pubstate = {presented, Feb 2022 accepted, submitted Dec 2021},
  series = {Proceedings of Machine Learning Research},
  title = {Reinforcement Learning with Almost Sure Constraints},
  url = {https://mallada.ece.jhu.edu/pubs/2022-L4DC-CMBM.pdf},
  volume = {168},
  year = {2022}
}