• DocumentCode
    46518
  • Title

    Reinforcement Learning for Port-Hamiltonian Systems

  • Author

    Sprangers, Olivier ; Babuska, Robert ; Nageshrao, Subramanya P. ; Lopes, Gabriel A. D.

  • Author_Institution
    Delft Center for Syst. & Control, Delft Univ. of Technol., Delft, Netherlands
  • Volume
    45
  • Issue
    5
  • fYear
    2015
  • fDate
    May-15
  • Firstpage
    1003
  • Lastpage
    1013
  • Abstract
    Passivity-based control (PBC) for port-Hamiltonian systems provides an intuitive way of achieving stabilization by rendering a system passive with respect to a desired storage function. However, in most instances the control law is obtained without any performance considerations and it has to be calculated by solving a complex partial differential equation (PDE). In order to address these issues we introduce a reinforcement learning (RL) approach into the energy-balancing passivity-based control (EB-PBC) method, which is a form of PBC in which the closed-loop energy is equal to the difference between the stored and supplied energies. We propose a technique to parameterize EB-PBC that preserves the systems´s PDE matching conditions, does not require the specification of a global desired Hamiltonian, includes performance criteria, and is robust. The parameters of the control law are found by using actor-critic (AC) RL, enabling the search for near-optimal control policies satisfying a desired closed-loop energy landscape. The advantage is that the solutions learned can be interpreted in terms of energy shaping and damping injection, which makes it possible to numerically assess stability using passivity theory. From the RL perspective, our proposal allows for the class of port-Hamiltonian systems to be incorporated in the AC framework, speeding up the learning thanks to the resulting parameterization of the policy. The method has been successfully applied to the pendulum swing-up problem in simulations and real-life experiments.
  • Keywords
    approximation theory; closed loop systems; damping; learning (artificial intelligence); nonlinear dynamical systems; optimal control; stability; AC RL; RL approach; actor-critic RL; closed-loop energy; control law; control law parameters; damping injection; energy shaping; energy-balancing passivity-based control method; near-optimal control policy search; numerically assessed stability; parameterized EB-PBC method; passivity theory; pendulum swing-up problem; performance criteria; port-Hamiltonian systems; reinforcement learning; stabilization; storage function; stored energies; supplied energies; system PDE matching conditions; system rendering; Damping; Equations; Learning (artificial intelligence); Mathematical model; Optimal control; Stability analysis; Vectors; Actor-critic (AC); energy-balancing (EB); passivity-based control (PBC); port-Hamiltonian (PH) systems; reinforcement learning (RL);
  • fLanguage
    English
  • Journal_Title
    Cybernetics, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    2168-2267
  • Type

    jour

  • DOI
    10.1109/TCYB.2014.2343194
  • Filename
    6883207