Home > Research > Publications & Outputs > Deep Reinforcement Learning-Based Control Frame...

Electronic data

  • DDPG_Multilateral_TMRB22

    Rights statement: ©2022 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.

    Accepted author manuscript, 1.26 MB, PDF document

    Available under license: CC BY: Creative Commons Attribution 4.0 International License

Links

Text available via DOI:

View graph of relations

Deep Reinforcement Learning-Based Control Framework for Multilateral Telesurgery

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published
  • Sarah Bacha
  • Weibang Bai
  • Ziwei Wang
  • Bo Xiao
  • Eric Yeatman
Close
<mark>Journal publication date</mark>31/05/2022
<mark>Journal</mark>IEEE Transactions on Medical Robotics and Bionics
Issue number2
Volume4
Number of pages4
Pages (from-to)352-355
Publication StatusPublished
Early online date27/04/22
<mark>Original language</mark>English

Abstract

The upper boundary of time delay is often required in traditional telesurgery control design, which would result in infeasibility of telesurgery across regions. To overcome this issue, this paper introduces a new control framework based on deep deterministic policy gradient (DDPG) reinforcement learning (RL) algorithm. The developed framework effectively overcomes the phase difference and data loss caused by time delays, which facilitates the restoration of surgeon’s intention and interactive force. Kalman filter (KF) is employed to blend multiple surgeons’ commands and predict the final local commands, respectively. The control framework ensures synchronization tracking performance and transparency. Prior knowledge of time delay is therefore not required. Simulation and experiment results have demonstrated the merits of the proposed framework.

Bibliographic note

©2022 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.