TY - GEN
T1 - OOCL-DDQN
T2 - 29th IEEE International Conference on Parallel and Distributed Systems, ICPADS 2023
AU - Li, Huijie
AU - Lin, Wei
AU - Huo, Jian
AU - Luo, Wuman
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Anesthesia control is critical in surgical procedures. Nowadays, various methods have been proposed for automated anesthesia control. However, traditional control model-based methods and PK/PD-based methods cannot fully capture the individual characteristics of patients. Machine learning-based methods are not suitable for dealing with large-scale clinical anaesthesia data. The existing deep reinforcement learning-based (DRL) methods cannot provide both stable and reliable anesthesia strategies while ensuring the agent's adaptability to environmental changes. To address these issues, in this paper, we propose a deep reinforcement learning model named OOCL-DDQN for automated anesthesia control. Specifically, we propose an online evaluation and offline training mechanism to well avoid excessive reliance on the environment models. To enhance the model stability, we design a clip method to optimize the Bellman equation in Double DQN. Besides, we propose a fixed-order random sampling method to enable the agent to effectively learn the relationship between anesthesia state and anesthetic dosage in different patients. In data preprocessing, we design a same-frequency resampling method to ensures that the resampled data obey the Markov property. To evaluate the performance of our proposed method, we conduct extensive experiments on a real-world dataset. The experiment results show that our proposed model gets better performances than the other state-of-the-art methods.
AB - Anesthesia control is critical in surgical procedures. Nowadays, various methods have been proposed for automated anesthesia control. However, traditional control model-based methods and PK/PD-based methods cannot fully capture the individual characteristics of patients. Machine learning-based methods are not suitable for dealing with large-scale clinical anaesthesia data. The existing deep reinforcement learning-based (DRL) methods cannot provide both stable and reliable anesthesia strategies while ensuring the agent's adaptability to environmental changes. To address these issues, in this paper, we propose a deep reinforcement learning model named OOCL-DDQN for automated anesthesia control. Specifically, we propose an online evaluation and offline training mechanism to well avoid excessive reliance on the environment models. To enhance the model stability, we design a clip method to optimize the Bellman equation in Double DQN. Besides, we propose a fixed-order random sampling method to enable the agent to effectively learn the relationship between anesthesia state and anesthetic dosage in different patients. In data preprocessing, we design a same-frequency resampling method to ensures that the resampled data obey the Markov property. To evaluate the performance of our proposed method, we conduct extensive experiments on a real-world dataset. The experiment results show that our proposed model gets better performances than the other state-of-the-art methods.
KW - Automated Anesthesia Control
KW - Deep Reinforcement Learning
KW - Double Deep Q-Network (DDQN)
UR - http://www.scopus.com/inward/record.url?scp=85190283802&partnerID=8YFLogxK
U2 - 10.1109/ICPADS60453.2023.00234
DO - 10.1109/ICPADS60453.2023.00234
M3 - Conference contribution
AN - SCOPUS:85190283802
T3 - Proceedings of the International Conference on Parallel and Distributed Systems - ICPADS
SP - 1676
EP - 1683
BT - Proceedings - 2023 IEEE 29th International Conference on Parallel and Distributed Systems, ICPADS 2023
PB - IEEE Computer Society
Y2 - 17 December 2023 through 21 December 2023
ER -