Robert Farrell, Rajarshi Das, et al.
AAAI-SS 2010
We devise a control-theoretic reinforcement learning approach to support direct learning of the optimal policy. We establish various theoretical properties of our approach, such as convergence and optimality of our analog of the Bellman operator and Q-learning, a new control-policy-variable gradient theorem, and a specific gradient ascent algorithm based on this theorem within the context of a specific control-theoretic framework. We empirically evaluate the performance of our control-theoretic approach on several classical reinforcement learning tasks, demonstrating significant improvements in solution quality, sample complexity, and running time of our approach over state-of-the-art methods.
Robert Farrell, Rajarshi Das, et al.
AAAI-SS 2010
Fernando Martinez, Juntao Chen, et al.
AAAI 2025
Lisa Hamada, Indra Priyadarsini S, et al.
AAAI 2025
Chen-chia Chang, Wan-hsuan Lin, et al.
ICML 2025