变压器
计算机科学
运动学
手势
人工智能
循环神经网络
机器人
人工神经网络
水准点(测量)
语音识别
机器学习
工程类
物理
电气工程
电压
经典力学
地理
大地测量学
作者
Simge Nur Kabataş,Duygu Sarıkaya
标识
DOI:10.1109/siu53274.2021.9477969
摘要
Automatic classification and recognition of surgical activities is an important step towards providing feedback in surgical training, preventing adversarial events and medical errors in surgeries. Kinematic data recorded from surgical robots contains information about the surgeon's movements. Using this data, we can model and recognize surgical gestures automatically. In this study, the Transformer model, which has shown better performance than Recurrent Neural Networks (RNNs) with time series data, has been used to recognize surgical gestures with kinematic data. The model learned in this study is compared with the Long Short-Term Memory (LSTM) model, which is widely used in the literature. The average accuracy of for JHU-ISI Gesture And Skill Asessment Working Set (JIGSAWS) the Transformer model is 0.77. According to the results, Transformer model is comparable to the state of the art LSTM models, and has outperformed the LSTM model we have developed in this study as part of the benchmark, and the standard variation is lower. To our knowledge, our study is the first to use Transformer model for surgical activity recognition with kinematic data. Our experiments show the promise of Transformer Network in this domain.
科研通智能强力驱动
Strongly Powered by AbleSci AI