Summary
Model-based reinforcement learning approaches rely on discrete-time state transition models, but many control tasks operate in continuous-time. This paper proposes a continuous-time MBRL framework based on a novel actor-critic method. The approach uses Bayesian neural ODEs to infer unknown state evolution differentials and explicitly solves continuous-time control systems. Experimental results show the robustness of the model against noisy data and its effectiveness in solving control problems. The paper discusses challenges and solutions related to continuous-time reinforcement learning and presents a comprehensive approach to address them.