Skip to main content
Erschienen in: World Journal of Urology 8/2021

Open Access 06.11.2020 | Invited Review

Evolving robotic surgery training and improving patient safety, with the integration of novel technologies

verfasst von: I-Hsuan Alan Chen, Ahmed Ghazi, Ashwin Sridhar, Danail Stoyanov, Mark Slack, John D. Kelly, Justin W. Collins

Erschienen in: World Journal of Urology | Ausgabe 8/2021

Abstract

Introduction

Robot-assisted surgery is becoming increasingly adopted by multiple surgical specialties. There is evidence of inherent risks of utilising new technologies that are unfamiliar early in the learning curve. The development of standardised and validated training programmes is crucial to deliver safe introduction. In this review, we aim to evaluate the current evidence and opportunities to integrate novel technologies into modern digitalised robotic training curricula.

Methods

A systematic literature review of the current evidence for novel technologies in surgical training was conducted online and relevant publications and information were identified. Evaluation was made on how these technologies could further enable digitalisation of training.

Results

Overall, the quality of available studies was found to be low with current available evidence consisting largely of expert opinion, consensus statements and small qualitative studies. The review identified that there are several novel technologies already being utilised in robotic surgery training. There is also a trend towards standardised validated robotic training curricula. Currently, the majority of the validated curricula do not incorporate novel technologies and training is delivered with more traditional methods that includes centralisation of training services with wet laboratories that have access to cadavers and dedicated training robots.

Conclusions

Improvements to training standards and understanding performance data have good potential to significantly lower complications in patients. Digitalisation automates data collection and brings data together for analysis. Machine learning has potential to develop automated performance feedback for trainees. Digitalised training aims to build on the current gold standards and to further improve the ‘continuum of training’ by integrating PBP training, 3D-printed models, telementoring, telemetry and machine learning.
Hinweise

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Introduction

It is recognised that errors are more common early in the surgeons learning curve [1] and the combination of simultaneously learning about both technology and technique, on patients, has inherent patient safety risks if training is not optimised [2, 3].
The first validated robotic training curriculum was published in 2015 [4]. This validated curriculum is the current gold standard and has been replicated by several societies in multiple specialties [5, 6]. The standardised structure describes staged training commencing with a baseline evaluation, e-learning and operating-room (OR) observation. With modules of simulation training, including wet-laboratory training in cadavers, pigs and other animal models. However, centralised wet-laboratory training centres are expensive and limit access. Another key issue is the level of competence that the trainee has when they commence operations on patients. Expertise may not be available locally, requiring travelling proctors. Weaknesses in individual’s training and subsequent performance can be missed if training is not objectively assessed, benchmarked and quality assured. In the aviation industry, there are international training standards that are benchmarked and quality assured [7]. Proficiency in performance must be shown before the pilot is allowed to fly a plane with passengers onboard. The same rigorous approach to surgical training has not yet been applied [7].
To improve surgical training, we need awareness of weaknesses, quality assured standards and access to affordable training that are integrated with job planning. The combination of systems thinking with a proficiency-based progression (PBP) approach to training has been shown to be highly successful in reducing errors in aviation training [7], whereas surgical training has historically been an apprentice model, with variabilities in the trainer’s skills as both a surgeon and educator [1]. Ultimately, all stages of training will benefit from digitalisation and automated data collection related to surgeon performance.

Materials and methods

A systematic narrative review was performed with a comprehensive computerised search completed using PubMed and Medline databases. We systematically searched using medical subject headings including ‘robot-assisted surgery training’, ‘robotic surgery training’, ‘curriculum development’ and ‘proficiency-based training’, ‘surgical education’, ‘3D printed models’, ‘telementoring’, ‘eye tracking’, ‘machine learning’ and ‘AI’. Articles of interest included reports of novel technologies used in health-care and surgical training, prospective studies on the impact of robotic simulation training, robotic training curriculum development with validation and systematic reviews on robotic training published between July 2000, when the first robotic systems received FDA approval in the USA [7], and March 2020. Other significant studies cited in the reference list of selected papers were evaluated, as well as studies of interest published before the systematic search. Sections of the review were allocated to six researchers with expertise in that area and the reviewers independently selected papers for detailed review evaluating the abstract and, if necessary, the full-text manuscript. Potential discrepancies for inclusion were resolved by open group discussion.

Findings

Overall, the quality of available studies was found to be low with current available evidence consisting largely of expert opinion, consensus statements and small qualitative studies. The review identified that there are several novel technologies already being utilised in robotic surgery training and also a trend towards standardised validated robotic training curricula. Currently, the majority of validated curricula follow traditional methods that do not incorporate novel technologies.

Establishing surgical performance metrics as a starting point

Surgical training has historically been delivered via a master-apprentice model, where the trainee observes and learns from the experienced trainer, eventually being ‘signed off’ as competent. However, subjective assessments of surgical performance have been shown to be highly variable with poor inter-rater reliability [8]. Skills learning is more efficient when sustained deliberate practice (SDP) is enabled [9]. This requires the skills to be defined with objective metrics of performance that are agreed by both the trainer and student [9]. SDP is an important element of PBP training, which has been shown to reduce error rates, early in the learning curve, by approximately 50% [8]. SDP states that repetition of skills with deliberate practice is key to success and that the defined metrics should be able to be replicated in various settings. Objective metrics once established should be utilised in multiple training settings to enable a continuum of training, from e-learning that describes the metrics in key index procedures, to the development of the simulation models that reflect the metrics [10], to telementoring protocols, to recording and auditing outcome data. There is good evidence that when surgery is standardised, it is easier to identify the subtleties of the technique to improve patient outcomes [11], and to identify postoperative complications related to sub-optimised technique [12].
Various approaches to defining metrics have been defined. The process requires task deconstruction and identification of key elements. Essential elements within each defined phase of the procedure include the tasks to be completed and the errors to avoid. To enable SDP, it is important to have agreement between the trainer and trainee. Consensus on the phases, tasks and errors can be reached using the Delphi process [13]. Strategies to drive standardised training, with a top-down approach, include ‘train-the-trainer’ courses, where trainers learn about the curriculum structure and content, the metrics to assess and how to deliver training safely [1]. An evolving surgical technique is equally likely to be affected by technological advancements in surgery and both optimisation of technique and technology benefit from standardisation of the fundamentals. This will be increasingly utilised with the development of robotic networks and integration of artificial intelligence (AI) and machine learning [14], to scale deep learning development into a more predictable and efficient practice benefits from standardisation. It is a common language or a coordination mechanism for industry, academics and clinicians to accelerate the development of both the technology and technique. Thus, a standardised approach to a key index procedure could be described as a minimal viable product (MVP), rather than a defined end product of established optimised performance (See Fig. 1).

E-learning

Online education has potential to deliver both synchronous and asynchronous learning. Video content is ‘richer’ than text and previous research concludes 1 min of video is worth 1.8 million words. The reasoning is: “if ‘a picture is worth a thousand words’, then a minute of video with 30 frames per second is worth at least 1.8 million words.” [15]. Cost models estimate the costs of developing video on demand for education according to Levin’s Ingredient, model related to personnel, equipment and consumables [16]. Video content development can be planned according to Fig. 2.
Future developments in innovative educational content can be broadly considered as platform functionality related and approaches to enhance engagement, although there is significant overlap. Virtual reality (VR) simulation is already widely available and has been shown to improve patient outcomes, but is not yet fully integrated into training [17]. Although the research concerning the implementation of augmented reality (AR) in surgical education is relatively limited, there are promising results regarding the teaching potential [18]. There is also evidence that gamification of knowledge acquisition is beneficial to learning curves and this is currently being explored in surgical training [19]. Further research is needed to prove if AR and VR can effectively replace or supplement traditional surgical pedagogy methods and whether gamification promotes a desire to learn.

Telepresence

In the current robotic training, without telementoring, an experienced surgeon usually proctors the trainee in their own hospital for their first five to ten operations [4]. If the hospital commencing robotic surgery lacks expertise, then this requires surgeons from centres of excellence to travel to that hospital. Proctorship is therefore expensive and supervision limited to an agreed number of operations. With the likely possibility of the ‘inexperienced’ surgeon being confronted with unusual anatomy or a difficult case after this period of proctorship, this is the time that the surgeon and his/her patient are most vulnerable to the early learning curve. The introduction of telementoring, where training and expert support/guidance can be continued remotely from centres of excellence, is more cost-efficient and accessible [14]. Telementorship consists of audio-visual communication where the remote surgeon can see the operating field, and additional functionalities to aid communication include telestration and image overlay [20].
Several studies have concluded telementoring is an effective training tool [21]. One study showed that residents in the telementoring group performed significantly better compared to non-mentoring group (p < 0.001) [22]. The safety of telementoring has also been established. In a systematic review of 11 studies, 9 concluded that telementoring did not prolong surgery time compared to on-site mentoring; none of them reported an elevated complication rate; and only 3% of the total number of cases reported technical issues [23]. Of note, a study of laparoscopic cholecystectomies conducted by Byrne et al. [24] concluded that telementoring could be used as bridge between on-site supervision to totally unsupervised performance. A telementoring study by Pahlsson et al. [25] focused on endoscopic retrograde cholangiopancreatography (ERCP) confirmed that telementoring might be delivered from a high-volume endoscopist at a tertiary hospital to a low-volume rural hospital, to provide a higher success rate which could be maintained without telementoring support.
One of the original purposes of telementoring was for the battlefield and potentially the biggest advantages may be in the management of emergency scenarios [26]. In 1999, Cubano and colleagues successfully connected USS Abraham Lincoln aircraft with a land-based surgical mentor and completed five laparoscopic hernia repairs under telementoring guidance [27]. Telementoring can currently be delivered with different network infrastructure, including wide access networks (WANs) and 5G [28]. Project 6 was proposed by the Society of American Gastrointestinal and Endoscopic Surgeons (SAGES) in 2015, aiming to promote development of surgical telementoring [29]. A current limitation is surgical telementoring’s requirement for bandwidth and rural areas usually lack it. With the development of 5G, this barrier would likely be overcome [28].

Eye tracking

In general, the mental workload associated with an easy task is low, whereas difficult tasks produce higher mental workload. Mental workload can also be described by the difference between task demands and available attention resources. Thus, a high workload task that is mentally demanding leaves little or no spare attention capacity to deal with new or unexpected events, and the less likelihood of learning. Increased workload during surgery has been associated with inferior task performance, a higher likelihood of errors and the possibility of an incomplete skill transfer to the clinical environment [30, 31]. Until recently, cognitive (mental) load measures were limited to subjective ratings administrated after the task (NASA-TLX), or performance on a secondary-task as opposed to instantaneous load in a given moment [32]. While such measures are well suited to evaluating the relative differences in cognitive load between practice trials, they cannot provide detailed information about whether increased or decreased load is experienced during the performance of a learning task. Task-evoked pupillary responses (TEPRs) that include changes in pupil diameter and patterns in eye movement fixation have been found to occur shortly after the onset of a task and subside quickly after processing is terminated. Studying changes in TEPRs traditionally requires complex and limiting laboratory infrastructure with non-mobile cameras and onerous manual data collection and analysis.
Newly developed portable devices have facilitated this process, as they allow digital recording of pupil changes and a more convenient means for the quantification of TEPRs in dynamic environments. These devices have demonstrated that physicians with more training and experience exhibit less cognitive load than novices when answering questions in their field of expertise [33]. Recently, a publication evaluated the relationship between eye-tracking measures and perceived workload in robotic surgical tasks. Eight surgical trainees performed up to 12 simulated exercises. Pupil diameter and gaze entropy were found to distinguish differences in workload between task difficulty levels, and both metrics increased as task level difficulty increased. It was found that eye-tracking features achieved an accuracy of 84.7% in predicting workload levels [34]. Causer et al. [35] also maintained that gaze training could help trainees lessen the negative effects of anxiety by concentrating on continually relevant information. In future research, we can utilise a lightweight, non-obtrusive eye tracker (Pupil Labs) worn by the participants to evaluate the impact of various robotic training curricula on the trainee’s workload. Owing to the fundamental differences of gaze behaviour between experts and novices, eye-tracking technology also has potential for proficiency assessment [34].
Eye tracking has been shown to be a valuable training tool that can impact the learning curve. Chestwood et al. demonstrated that the projection of an expert’s gaze pattern on a trainee’s laparoscopic screen during a simulation task was found to aid trainee performance. The research concluded that, by simultaneously reflecting a supervisor’s gaze to a trainee, the completion time of laparoscopic tasks and number of errors could be significantly reduced [36]. Additionally, eye tracking can measure the gaze focusing on an area of interest (AOI). One study indicated that as medical students became more familiar with anatomical landmarks, cognitively salient gaze pattern changes within AOI could be observed and so progression of gaze behaviour may be expected during active learning and familiarisation [37]. In future work assessing telementoring, we could assess the use of the projected eye tracking of a trainer to aid in remote proctoring of trainees. Eye tracking can be utilised with both an open console and immersive console (Fig. 3).

Fabricating the ideal surgical simulation platform utilising 3D printing

One area of simulation that has proven to be difficult is the creation of a high-fidelity process that accurately and reproducibly simulates an entire procedure including relevant anatomy and pathology (see Table 1). With increased awareness of the benefits of standardised training, a novel full-immersion simulation using 3D printed models builds on the principles of SDP, with integration of defined performance metrics [10].
Table 1
Comparison of the different surgical training models
Model
Strengths
Weaknesses
Task deconstruction models
Address metrics and are cost effective, e.g. chicken gizzard model for vesico-urethral anastmosis
Limited development to comprehensively address metrics, benchmarks and error management
Porcine model
Flexible training model for tissue handling
Expensive
Not human anatomy
No human pathology
Limited accessibility
Canine cadaver model
Flexible training model for tissue handling
Not human anatomy
No human pathology
Limited accessibility
Human cadaver model
Flexible training model
Expensive
Lacks pathology and does not bleed
Limited accessibility
3D printed models [10]
Flexible training model
Currently, high development costs (lowered if printed casts rather than printed models)
Can incorporate pathology and vascularisation
Increasingly realistic tissue handling
Models that address specific defined metrics need to be developed
Can incorporate metrics and benchmarks [10]
VR simulation
Advanced procedural training models available (e.g. robotic prostatectomy, hysterectomy)
Current scope/range/image quality limited
AR simulation
Potential to develop
Limited development
The Simulation Innovation Laboratory at the University of Rochester Medical Centre has successfully merged polymer casting and 3D printing technologies to fabricate anatomically accurate training platforms permitting realistic dissection, haemostasis and suturing for complex procedures as percutaneous nephrolithotomy [38], robot-assisted radical prostatectomy (RARP) [10] and partial nephrectomy [39]. Taking the RARP model [10] as an example, the MRI of a patient with T1c, Gleason 7 cancer was imported into Mimics 20.0 (Mimics; Materialise, Leuven, Belgium)., for segmentation of each of the patient’s pelvic organs to form a computer-aided design (CAD) anatomical model (Fig. 4a). Individual injection casts were then designed from the CAD and 3D printed using a Fusion3 F400-S 3D printer (Fusion3 Design, Greensboro, NC, USA) (Fig. 4b). PVA hydrogel was injected into these casts based on the desired mechanical properties for each organ and cast in series to replicate their anatomical relationships between various organs (Fig. 4c). The model was perfused through hollow, watertight vessels incorporated into the NVB, dorsal venous complex and prostatic pedicle during the casting process. To replicate the entire procedure, the prostatectomy organ complex was layered in its anatomical configuration within a 3D-printed male pelvis that was fitted with pelvic floor muscles, pelvic fat, and relevant structures made of PVA (Fig. 4d).
Training and assessment go hand in hand, training without formative feedback is reduced to sheer repetition [9]. Naturally, the best method of assessment would be one of clinical relevance that could be tracked to evaluate progress during training. Our casting technique allowed incorporation of Clinically-Relevant Metrics Performance Metrics of Simulation (CRPMS) pertinent to an RARP procedure including key phases of the procedure: nerve tension during NVB dissection (measured through calibrated analogue stretch sensors, aligned within the NVB during the casting process); surgical margins (measured by addition of a chemical that exhibits chemiluminescence in the prostate cast); and vesico-urethral anastomosis (VUA) integrity (that could be tested for any leaks after injecting 180 cc of saline) (Fig. 5).
Simulations by five experts and nine novices were completed to validate incorporated CRPMS and correlate them to standardised objective assessments, e.g. global evaluative assessment of robotic skills (GEARS) and robotic anastomosis competency evaluation (RACE) [10]. Nerve forces applied during the simulation were significantly lower for experts. Higher force sensitivity (Subcategory of GEARS Score) and Total GEARS Score correlated with lower nerve forces applied with total energy (J) – 0.66 (0.019) and − 0.87 (0.000), respectively, which was significantly different between novices and experts (p = 0.003). The VUA leak rate highly correlated with total RACE score − 0.86 (0.000), which was also significantly different between novices and experts (p = 0.003). This study presents a novel method for real-time assessment and feedback during robotic surgery training utilising incorporated CRPMS. These efforts provided a platform for reproducing realistic procedural models with the added capacity to provide objective procedural metrics that permit real-time feedback and assessment during robotic surgery training.

Automated performance metrics

Robotic surgery training curricula aim to objectively assess surgical trainee basic robotic surgical skills or specific procedures using tools such as global skills assessment, robotic-objective structured assessment of technical skills (R-OSATS), global operative assessment of laparoscopic skills (GOALS) and GEARS. While the majority of these tools were utilised for evaluation of basic robotic surgery performance, the GEARS assessed selective steps of specific procedures and associated them with patient outcomes [40]. Nevertheless, these tools counted on human rating, inevitably mingled with subjective bias. In addition, the time and resources of educators or assessors is required. As such, automated performance metrics (APMs) have potential to objectively and analytically evaluate surgical techniques, providing surgical trainees with more equitable and quantifiable assessment across disciplines. Heterogeneity exists in automated methods of surgical skill assessment. Levin et al. [41] identified a pathway of automated methods, including three sessions: data extraction (kinetic and computer vision methods), automated methods (motion tracking of tools, hands and eyes, and muscle contraction analysis), and analysis utilising machine learning, deep learning or performance classification. The procedural approaches most frequently studied were robotic and laparoscopic minimally invasive surgery. The most common utilised method for automation was tool motion tracking. Most previous studies were accomplished in a simulated environment, instead of in the OR.
In the laboratory setting or live surgery, APMs can be categorised into three sectors, encompassing kinematic data (e.g. instrument moving velocity, travelling distance, deceleration and acceleration), systems event data (e.g. master clutch use, camera movement, energy application and third arm swap), and instrument grip force [42]. There are a variety of recording devices, such as ProMISTM, trakSTAR, AcceleGlove, and dVLogger (Intuitive). The most frequently investigated and validated APMs are kinematic data. Both kinematic and systems events data have shown good discrimination between surgeon’s expertise. Although instrument grip force was less frequently studied, Gomez et al. [43] identified that novice robotic surgeons applied higher grip forces than experts when performing laboratory exercises. If these surgical activity recognition models could assess each phase of a surgical procedure, they could be utilised to compute efficiency metrics. Nevertheless, most models reported in the literature ranged from around 50–80% accuracy [44]. Recognition of anatomical landmarks is a current limitation of APMs evaluating surgical skills. Nosrati et al. [45] proposed a new technique to localise both visible and occluded anatomical structures on an endoscopic view of partial nephrectomy. They leveraged both preoperative 3D computed tomography scans and intraoperative endoscopic visual clues as well as vasculature pulsation in order for accurately segmenting the highly fluctuated environment in minimally invasive surgeries.

AI and machine learning (ML)

Quantifying and evaluating surgical ability are intensely researched topics in the medical community and particularly in minimally invasive surgery. AI driven systems are emerging mostly based on vision (endoscopic video), which embeds space–time information about both the instrument motion and the surgical site that can provide richer discriminative power. Because video is inherently available in many modern surgical procedures and captures a rich log of events, it is a promising digital record that can be analysed to infer ability, but of course it can also be complemented with additional information from integrated ORs or from robot kinematics [46]. Crucial to exploiting video or other OR data for assessing competence is the automatic analysis of the signal, because it is impractical to scale systems relying on manual observations of vast volumes of procedures. For automation, various computer vision (CV) or ML algorithms have been introduced to evaluate surgical performance and also to extract higher level information from the surgical video stream [47]. Both model-based [48, 49] and data-driven [39] algorithms have been explored as a means to lift skill metrics from video akin to other AI application areas, and data-driven and especially deep learning-based methods have rapidly emerged as the most effective and robustly performing algorithms for understanding surgical video.
To extract APMs from video, a number of building block algorithmic capabilities are important to develop, for example: detection of surgical instrument presence [50], delineation of surgical tools’ position and motion [51], segmentation of surgical site into objects [52] or the video into key surgical steps [53, 54], activity or significant event detection [55] as well as others like the detection of critical structures [56]. Marked progress in each of these building blocks of surgical process understanding has taken place in recent years, but a significant challenge is still the availability of large, well annotated datasets that can be used to evaluate systems in a fair and comparable manner. This is particularly lacking for data that combines OR information despite recent efforts or on robotic information and video to support fused analysis [57] with the only currently available data being on phantom environments [58]. The transferability of systems across different operating techniques, instrument toolsets or even different procedures is also largely unmet at present.
With maturity of AI systems a number of other capabilities have also been explored like the estimation of the remaining procedural time form the real-time video feed [59] or automatic image to video retrieval [60] or potentially risk estimation [61]. An important area of work that has received limited attention at present is how such AI powered technologies for understanding surgical process and performance can be utilised in situ with effective user interfaces that appropriately support clinical workflow. They reported that the RF-50 algorithm provided the best performance, delivering 87.2% accuracy in predicting length of stay following RARP. However, there was lack of task-based, efficiency metrics. They demonstrated a better performance for surgical activity recognition in RARP, proving the feasibility of automated postoperative efficiency reports, especially for critical tasks in a clinical procedure. A deep-learning network, an artificial neural network with significant layers, has assisted in discovering hidden and abstract implications present in data. Despite great performance of deep-learning methods in multiple fields, the lack of solid theory explaining how they work would be an important concern. Baghdadi et al. [62] also described ML-based analysis of textural and colour visual features on a robotic endoscopic view to localise anatomical landmarks during RARP. Nevertheless, it is still laborious and time-consuming to label all anatomical landmarks in a surgical procedure to feed data into ML algorithms. How to enable precise automatic annotation by AI rather than repetitive manual work remains a fundamental issue.

Discussion

It has previously been estimated that 10–15% of surgical patients in the UK have adverse events whilst in hospital. 50% are in the operating room and 50% are preventable [63]. A recent study from the USA estimated that a third of all medical injuries are due to error [64]. The risk of errors has potential to increase with new technologies. In the US between 2000 and 2013 10,624 adverse events related to robotic procedures were reported [2]. During this period, 144 deaths (1.4% of the 10,624 reports), 1391 patient injuries (13.1%), and 8061 device malfunctions (75.9%) were reported. The USA ‘Agency for healthcare research and quality’, estimated that the annual cost to US healthcare from medical errors is €17.1 billion. Six of the top ten identified medical errors are related to surgical procedures [6]. Rigorous analysis of surgical errors and training to reduce their occurrence will reduce the financial burden to both patients and hospitals. However, the majority of surgical training remains opportunistic, unstructured and delivered in an apprenticeship style.
In 2013, a group of experts expressed concern that robotic surgery training is not standardised and insufficient to ensure patient safety [3]. Two years later, the ECRI institute published their annual independent review on health technology hazards, in which a lack of robotic surgical training was identified as one of the top ten risks to US patients [65]. The ECRI report stated: “Insufficient training of surgeons on robotic technologies can result in surgical errors that lead to prolonged surgery, substandard operation outcomes early in the surgeons’ learning curve, complications that require additional treatment and even serious patient injury or death. Errors can result if training is insufficient or ineffective e.g. if it does not provide an assurance of competency”.
Simulation training, which aims to avoid patients being exposed to the trainees’ early learning curve, could be significantly improved by utilising 3D printed models that bleed and incorporate metrics of surgical performance to aid SDP, real-time assessment and feedback [10]. These models also do not require wet-laboratory facilities, thereby reducing costs and making training more accessible. Novel technologies have potential to support SDP in various settings. An evolved digitalised curriculum could be monitored from a centralised hub and incorporate: standardised asynchronous and synchronous e-learning modules with PBP training [8] that has benchmarks, 3D-printed models [10], telemetry [66], eye-tracking metrics [34] and video performance analysis completed with telementoring in real time [29]. All training aspects aligned with defined metrics and delivering a continuum of training. Incorporating supervised ML algorithms to identify and prioritise key elements of training and performance will enable personalised learning and eventually automated performance feedback [14, 66].

Conclusions

Improvements to training standards and understanding performance data have huge potential to significantly lower complications in patients. Digitalisation automates data collection and brings data together for analysis. Digitalised training aims to build on the current gold standards and to further improve the ‘continuum of training’ by integrating PBP training, 3D-printed models, telementoring, telemetry and machine learning. Study and evaluation of performance metrics and patient outcomes with machine learning have potential to develop automated performance feedback for trainees.
Objective performance metrics will help deliver society-approved and validated robotic surgery curriculums for multiple surgical specialities. This will aid credentialing of surgeons in new medical technologies and their applications.

Compliance with ethical standards

Conflict of interest

Not applicable.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Unsere Produktempfehlungen

e.Med Interdisziplinär

Kombi-Abonnement

Für Ihren Erfolg in Klinik und Praxis - Die beste Hilfe in Ihrem Arbeitsalltag

Mit e.Med Interdisziplinär erhalten Sie Zugang zu allen CME-Fortbildungen und Fachzeitschriften auf SpringerMedizin.de.

e.Med Gynäkologie

Kombi-Abonnement

Mit e.Med Gynäkologie erhalten Sie Zugang zu CME-Fortbildungen der beiden Fachgebiete, den Premium-Inhalten der Fachzeitschriften, inklusive einer gedruckten gynäkologischen oder urologischen Zeitschrift Ihrer Wahl.

e.Med Urologie

Kombi-Abonnement

Mit e.Med Urologie erhalten Sie Zugang zu den urologischen CME-Fortbildungen und Premium-Inhalten der urologischen Fachzeitschriften.

Literatur
1.
Zurück zum Zitat Collins JW, Levy J, Stefanidis D, Gallagher A, Coleman M, Cecil T et al (2019) Utilising the Delphi process to develop a proficiency-based progression train-the-trainer course for robotic surgery training. Eur Urol 75(5):775–785PubMedCrossRef Collins JW, Levy J, Stefanidis D, Gallagher A, Coleman M, Cecil T et al (2019) Utilising the Delphi process to develop a proficiency-based progression train-the-trainer course for robotic surgery training. Eur Urol 75(5):775–785PubMedCrossRef
2.
Zurück zum Zitat Alemzadeh H, Raman J, Leveson N, Kalbarczyk Z, Iyer RK (2016) Adverse events in robotic surgery: a retrospective study of 14 years of FDA data. PLoS ONE 11(4):e0151470PubMedPubMedCentralCrossRef Alemzadeh H, Raman J, Leveson N, Kalbarczyk Z, Iyer RK (2016) Adverse events in robotic surgery: a retrospective study of 14 years of FDA data. PLoS ONE 11(4):e0151470PubMedPubMedCentralCrossRef
3.
Zurück zum Zitat Ahmed K, Khan R, Mottrie A, Lovegrove C, Abaza R, Ahlawat R et al (2015) Development of a standardised training curriculum for robotic surgery: a consensus statement from an international multidisciplinary group of experts. BJU Int 116(1):93–101PubMedCrossRef Ahmed K, Khan R, Mottrie A, Lovegrove C, Abaza R, Ahlawat R et al (2015) Development of a standardised training curriculum for robotic surgery: a consensus statement from an international multidisciplinary group of experts. BJU Int 116(1):93–101PubMedCrossRef
4.
Zurück zum Zitat Volpe A, Ahmed K, Dasgupta P, Ficarra V, Novara G, van der Poel H et al (2015) Pilot validation study of the European association of urology robotic training curriculum. Eur Urol 68(2):292–299PubMedCrossRef Volpe A, Ahmed K, Dasgupta P, Ficarra V, Novara G, van der Poel H et al (2015) Pilot validation study of the European association of urology robotic training curriculum. Eur Urol 68(2):292–299PubMedCrossRef
5.
Zurück zum Zitat Veronesi G, Dorn P, Dunning J, Cardillo G, Schmid RA, Collins J et al (2018) Outcomes from the Delphi process of the thoracic robotic curriculum development committee. Eur J Cardiothorac Surg 53(6):1173–1179PubMedCrossRef Veronesi G, Dorn P, Dunning J, Cardillo G, Schmid RA, Collins J et al (2018) Outcomes from the Delphi process of the thoracic robotic curriculum development committee. Eur J Cardiothorac Surg 53(6):1173–1179PubMedCrossRef
6.
Zurück zum Zitat Rusch P, Ind T, Kimmig R, Maggioni A, Ponce J, Zanagnolo V et al (2019) Recommendations for a standardised educational program in robot assisted gynaecological surgery: consensus from the Society of European Robotic Gynaecological Surgery (SERGS). Facts Views Vis Obgyn 11(1):29–41PubMedPubMedCentral Rusch P, Ind T, Kimmig R, Maggioni A, Ponce J, Zanagnolo V et al (2019) Recommendations for a standardised educational program in robot assisted gynaecological surgery: consensus from the Society of European Robotic Gynaecological Surgery (SERGS). Facts Views Vis Obgyn 11(1):29–41PubMedPubMedCentral
8.
Zurück zum Zitat Angelo RL, Ryu RK, Pedowitz RA, Beach W, Burns J, Dodds J et al (2015) A proficiency-based progression training curriculum coupled with a model simulator results in the acquisition of a superior arthroscopic bankart skill set. Arthroscopy 31(10):1854–1871PubMedCrossRef Angelo RL, Ryu RK, Pedowitz RA, Beach W, Burns J, Dodds J et al (2015) A proficiency-based progression training curriculum coupled with a model simulator results in the acquisition of a superior arthroscopic bankart skill set. Arthroscopy 31(10):1854–1871PubMedCrossRef
9.
Zurück zum Zitat Ericsson KA, Harwell KW (2019) Deliberate practice and proposed limits on the effects of practice on the acquisition of expert performance: why the original definition matters and recommendations for future research. Front Psychol 10:2396PubMedPubMedCentralCrossRef Ericsson KA, Harwell KW (2019) Deliberate practice and proposed limits on the effects of practice on the acquisition of expert performance: why the original definition matters and recommendations for future research. Front Psychol 10:2396PubMedPubMedCentralCrossRef
10.
Zurück zum Zitat Witthaus MW, Farooq S, Melnyk R, Campbell T, Saba P, Mathews E et al (2020) Incorporation and validation of clinically relevant performance metrics of simulation (CRPMS) into a novel full-immersion simulation platform for nerve-sparing robot-assisted radical prostatectomy (NS-RARP) utilizing three-dimensional printing and hydrogel casting technology. BJU Int 125(2):322–332PubMedCrossRef Witthaus MW, Farooq S, Melnyk R, Campbell T, Saba P, Mathews E et al (2020) Incorporation and validation of clinically relevant performance metrics of simulation (CRPMS) into a novel full-immersion simulation platform for nerve-sparing robot-assisted radical prostatectomy (NS-RARP) utilizing three-dimensional printing and hydrogel casting technology. BJU Int 125(2):322–332PubMedCrossRef
11.
Zurück zum Zitat Schlomm T, Heinzer H, Steuber T, Salomon G, Engel O, Michl U et al (2011) Full functional-length urethral sphincter preservation during radical prostatectomy. Eur Urol 60(2):320–329PubMedCrossRef Schlomm T, Heinzer H, Steuber T, Salomon G, Engel O, Michl U et al (2011) Full functional-length urethral sphincter preservation during radical prostatectomy. Eur Urol 60(2):320–329PubMedCrossRef
12.
Zurück zum Zitat Collins JW, Tyritzis S, Nyberg T, Schumacher M, Laurin O, Khazaeli D et al (2013) Robot-assisted radical cystectomy: description of an evolved approach to radical cystectomy. Eur Urol 64(4):654–663PubMedCrossRef Collins JW, Tyritzis S, Nyberg T, Schumacher M, Laurin O, Khazaeli D et al (2013) Robot-assisted radical cystectomy: description of an evolved approach to radical cystectomy. Eur Urol 64(4):654–663PubMedCrossRef
14.
Zurück zum Zitat Collins J, Akre O, Challacombe B, Karim O, Wiklund P (2015) Robotic networks: delivering empowerment through integration. BJU Int 116(2):167–168PubMedCrossRef Collins J, Akre O, Challacombe B, Karim O, Wiklund P (2015) Robotic networks: delivering empowerment through integration. BJU Int 116(2):167–168PubMedCrossRef
15.
Zurück zum Zitat Mcquivey J, De Lussanet M, Wilkos D (2008) How video will take over the world. Forrester Research, Cambridge Mcquivey J, De Lussanet M, Wilkos D (2008) How video will take over the world. Forrester Research, Cambridge
17.
Zurück zum Zitat MacCraith E, Forde JC, Davis NF (2019) Robotic simulation training for urological trainees: a comprehensive review on cost, merits and challenges. J Robot Surg 13(3):371–377PubMedCrossRef MacCraith E, Forde JC, Davis NF (2019) Robotic simulation training for urological trainees: a comprehensive review on cost, merits and challenges. J Robot Surg 13(3):371–377PubMedCrossRef
18.
Zurück zum Zitat Chytas D, Johnson EO, Piagkou M, Mazarakis A, Babis GC, Chronopoulos E et al (2020) The role of augmented reality in Anatomical education: an overview. Ann Anat 229:151463PubMedCrossRef Chytas D, Johnson EO, Piagkou M, Mazarakis A, Babis GC, Chronopoulos E et al (2020) The role of augmented reality in Anatomical education: an overview. Ann Anat 229:151463PubMedCrossRef
19.
Zurück zum Zitat Scaffidi MA, Khan R, Walsh CM, Pearl M, Winger K, Kalaichandran R et al (2019) Protocol for a randomised trial evaluating the effect of applying gamification to simulation-based endoscopy training. BMJ Open 9(2):e024134PubMedPubMedCentralCrossRef Scaffidi MA, Khan R, Walsh CM, Pearl M, Winger K, Kalaichandran R et al (2019) Protocol for a randomised trial evaluating the effect of applying gamification to simulation-based endoscopy training. BMJ Open 9(2):e024134PubMedPubMedCentralCrossRef
20.
Zurück zum Zitat Shin DH, Dalag L, Azhar RA, Santomauro M, Satkunasivam R, Metcalfe C et al (2015) A novel interface for the telementoring of robotic surgery. BJU Int 116(2):302–308PubMedCrossRefPubMedCentral Shin DH, Dalag L, Azhar RA, Santomauro M, Satkunasivam R, Metcalfe C et al (2015) A novel interface for the telementoring of robotic surgery. BJU Int 116(2):302–308PubMedCrossRefPubMedCentral
21.
Zurück zum Zitat Anvari M, McKinley C, Stein H (2005) Establishment of the world's first telerobotic remote surgical service: for provision of advanced laparoscopic surgery in a rural community. Ann Surg 241(3):460–464PubMedPubMedCentralCrossRef Anvari M, McKinley C, Stein H (2005) Establishment of the world's first telerobotic remote surgical service: for provision of advanced laparoscopic surgery in a rural community. Ann Surg 241(3):460–464PubMedPubMedCentralCrossRef
22.
Zurück zum Zitat Panait L, Rafiq A, Tomulescu V, Boanca C, Popescu I, Carbonell A et al (2006) Telementoring versus on-site mentoring in virtual reality-based surgical training. Surg Endosc 20(1):113–118PubMedCrossRef Panait L, Rafiq A, Tomulescu V, Boanca C, Popescu I, Carbonell A et al (2006) Telementoring versus on-site mentoring in virtual reality-based surgical training. Surg Endosc 20(1):113–118PubMedCrossRef
23.
Zurück zum Zitat Bilgic E, Turkdogan S, Watanabe Y, Madani A, Landry T, Lavigne D et al (2017) Effectiveness of telementoring in surgery compared with on-site mentoring: a systematic review. Surg Innov 24(4):379–385PubMedCrossRef Bilgic E, Turkdogan S, Watanabe Y, Madani A, Landry T, Lavigne D et al (2017) Effectiveness of telementoring in surgery compared with on-site mentoring: a systematic review. Surg Innov 24(4):379–385PubMedCrossRef
24.
Zurück zum Zitat Byrne JP, Mughal MM (2000) Telementoring as an adjunct to training and competence-based assessment in laparoscopic cholecystectomy. Surg Endosc 14(12):1159–1161PubMedCrossRef Byrne JP, Mughal MM (2000) Telementoring as an adjunct to training and competence-based assessment in laparoscopic cholecystectomy. Surg Endosc 14(12):1159–1161PubMedCrossRef
25.
Zurück zum Zitat Pahlsson HI, Groth K, Permert J, Swahn F, Lohr M, Enochsson L et al (2013) Telemedicine: an important aid to perform high-quality endoscopic retrograde cholangiopancreatography in low-volume centers. Endoscopy 45(5):357–361PubMedCrossRef Pahlsson HI, Groth K, Permert J, Swahn F, Lohr M, Enochsson L et al (2013) Telemedicine: an important aid to perform high-quality endoscopic retrograde cholangiopancreatography in low-volume centers. Endoscopy 45(5):357–361PubMedCrossRef
26.
Zurück zum Zitat Satava RM (2003) Robotic surgery: from past to future–a personal journey. Surg Clin N Am 83(6):1491–1500PubMedCrossRef Satava RM (2003) Robotic surgery: from past to future–a personal journey. Surg Clin N Am 83(6):1491–1500PubMedCrossRef
27.
Zurück zum Zitat Cubano M, Poulose BK, Talamini MA, Stewart R, Antosek LE, Lentz R et al (1999) Long distance telementoring. A novel tool for laparoscopy aboard the USS Abraham Lincoln. Surg Endosc 13(7):673–678PubMedCrossRef Cubano M, Poulose BK, Talamini MA, Stewart R, Antosek LE, Lentz R et al (1999) Long distance telementoring. A novel tool for laparoscopy aboard the USS Abraham Lincoln. Surg Endosc 13(7):673–678PubMedCrossRef
28.
Zurück zum Zitat Lacy AM, Bravo R, Otero-Pineiro AM, Pena R, De Lacy FB, Menchaca R et al (2019) 5G-assisted telementored surgery. Br J Surg 106(12):1576–1579PubMedCrossRef Lacy AM, Bravo R, Otero-Pineiro AM, Pena R, De Lacy FB, Menchaca R et al (2019) 5G-assisted telementored surgery. Br J Surg 106(12):1576–1579PubMedCrossRef
29.
Zurück zum Zitat Schlachta CM, Nguyen NT, Ponsky T, Dunkin B (2016) Project 6 summit: SAGES telementoring initiative. Surg Endosc 30(9):3665–3672PubMedCrossRef Schlachta CM, Nguyen NT, Ponsky T, Dunkin B (2016) Project 6 summit: SAGES telementoring initiative. Surg Endosc 30(9):3665–3672PubMedCrossRef
30.
Zurück zum Zitat McGaghie WC, Issenberg SB, Cohen ER, Barsuk JH, Wayne DB (2011) Does simulation-based medical education with deliberate practice yield better results than traditional clinical education? A meta-analytic comparative review of the evidence. Acad Med 86(6):706–711PubMedPubMedCentralCrossRef McGaghie WC, Issenberg SB, Cohen ER, Barsuk JH, Wayne DB (2011) Does simulation-based medical education with deliberate practice yield better results than traditional clinical education? A meta-analytic comparative review of the evidence. Acad Med 86(6):706–711PubMedPubMedCentralCrossRef
31.
Zurück zum Zitat van Merrienboer JJ, Sweller J (2010) Cognitive load theory in health professional education: design principles and strategies. Med Educ 44(1):85–93PubMedCrossRef van Merrienboer JJ, Sweller J (2010) Cognitive load theory in health professional education: design principles and strategies. Med Educ 44(1):85–93PubMedCrossRef
32.
Zurück zum Zitat Paas F, Tuovinen JE, Tabbers H, Van Gerven PW (2003) Cognitive load measurement as a means to advance cognitive load theory. Educ Psychol 38(1):63–71CrossRef Paas F, Tuovinen JE, Tabbers H, Van Gerven PW (2003) Cognitive load measurement as a means to advance cognitive load theory. Educ Psychol 38(1):63–71CrossRef
33.
Zurück zum Zitat Szulewski A, Roth N, Howes D (2015) The use of task-evoked pupillary response as an objective measure of cognitive load in novices and trained physicians: a new tool for the assessment of expertise. Acad Med 90(7):981–987PubMedCrossRef Szulewski A, Roth N, Howes D (2015) The use of task-evoked pupillary response as an objective measure of cognitive load in novices and trained physicians: a new tool for the assessment of expertise. Acad Med 90(7):981–987PubMedCrossRef
34.
Zurück zum Zitat Wu C, Cha J, Sulek J, Zhou T, Sundaram CP, Wachs J et al (2019) Eye-tracking metrics predict perceived workload in robotic surgical skills training. Hum Factors 2019:0018720819874544 Wu C, Cha J, Sulek J, Zhou T, Sundaram CP, Wachs J et al (2019) Eye-tracking metrics predict perceived workload in robotic surgical skills training. Hum Factors 2019:0018720819874544
35.
Zurück zum Zitat Causer J, Vickers JN, Snelgrove R, Arsenault G, Harvey A (2014) Performing under pressure: quiet eye training improves surgical knot-tying performance. Surgery 156(5):1089–1096PubMedCrossRef Causer J, Vickers JN, Snelgrove R, Arsenault G, Harvey A (2014) Performing under pressure: quiet eye training improves surgical knot-tying performance. Surgery 156(5):1089–1096PubMedCrossRef
36.
Zurück zum Zitat Chetwood AS, Kwok K-W, Sun L-W, Mylonas GP, Clark J, Darzi A et al (2012) Collaborative eye tracking: a potential training tool in laparoscopic surgery. Surg Endosc 26(7):2003–2009PubMedCrossRef Chetwood AS, Kwok K-W, Sun L-W, Mylonas GP, Clark J, Darzi A et al (2012) Collaborative eye tracking: a potential training tool in laparoscopic surgery. Surg Endosc 26(7):2003–2009PubMedCrossRef
37.
Zurück zum Zitat Zumwalt AC, Iyer A, Ghebremichael A, Frustace BS, Flannery S (2015) Gaze patterns of gross anatomy students change with classroom learning. Anat Sci Educ 8(3):230–241PubMedCrossRef Zumwalt AC, Iyer A, Ghebremichael A, Frustace BS, Flannery S (2015) Gaze patterns of gross anatomy students change with classroom learning. Anat Sci Educ 8(3):230–241PubMedCrossRef
38.
Zurück zum Zitat Ghazi A, Campbell T, Melnyk R, Feng C, Andrusco A, Stone J et al (2017) Validation of a full-immersion simulation platform for percutaneous nephrolithotomy using three-dimensional printing technology. J Endourol 31(12):1314–1320PubMedCrossRef Ghazi A, Campbell T, Melnyk R, Feng C, Andrusco A, Stone J et al (2017) Validation of a full-immersion simulation platform for percutaneous nephrolithotomy using three-dimensional printing technology. J Endourol 31(12):1314–1320PubMedCrossRef
39.
Zurück zum Zitat Melnyk R, Ezzat B, Saba P, Farooq S, Campbell T, McAleavey S et al (2019) Mechanical and functional validation of a perfused, robot-assisted partial nephrectomy simulation platform using a combination of 3D printing and hydrogel casting. World J Urol 2019:1–11 Melnyk R, Ezzat B, Saba P, Farooq S, Campbell T, McAleavey S et al (2019) Mechanical and functional validation of a perfused, robot-assisted partial nephrectomy simulation platform using a combination of 3D printing and hydrogel casting. World J Urol 2019:1–11
40.
Zurück zum Zitat Ghani KR, Comstock B, Miller DC, Dunn RL, Kim T, Linsell S et al (2017) PNFBA-02 technical skill assessment of surgeons performing robot-assisted radical prostatectomy: relationship between crowdsourced review and patient outcomes. J Urol. 197(4S):e609 Ghani KR, Comstock B, Miller DC, Dunn RL, Kim T, Linsell S et al (2017) PNFBA-02 technical skill assessment of surgeons performing robot-assisted radical prostatectomy: relationship between crowdsourced review and patient outcomes. J Urol. 197(4S):e609
42.
Zurück zum Zitat Chen J, Cheng N, Cacciamani G, Oh P, Lin-Brande M, Remulla D et al (2019) Objective assessment of robotic surgical technical skill: a systematic review. J Urol 2013:461–469CrossRef Chen J, Cheng N, Cacciamani G, Oh P, Lin-Brande M, Remulla D et al (2019) Objective assessment of robotic surgical technical skill: a systematic review. J Urol 2013:461–469CrossRef
43.
Zurück zum Zitat Gomez ED, Aggarwal R, McMahan W, Bark K, Kuchenbecker KJ (2016) Objective assessment of robotic surgical skill using instrument contact vibrations. Surg Endosc 30(4):1419–1431PubMedCrossRef Gomez ED, Aggarwal R, McMahan W, Bark K, Kuchenbecker KJ (2016) Objective assessment of robotic surgical skill using instrument contact vibrations. Surg Endosc 30(4):1419–1431PubMedCrossRef
44.
Zurück zum Zitat Zia A, Guo L, Zhou L, Essa I, Jarc A (2019) Novel evaluation of surgical activity recognition models using task-based efficiency metrics. Int J Comput Assist Radiol Surg 14(12):2155–2163PubMedCrossRef Zia A, Guo L, Zhou L, Essa I, Jarc A (2019) Novel evaluation of surgical activity recognition models using task-based efficiency metrics. Int J Comput Assist Radiol Surg 14(12):2155–2163PubMedCrossRef
45.
Zurück zum Zitat Nosrati MS, Amir-Khalili A, Peyrat J-M, Abinahed J, Al-Alao O, Al-Ansari A et al (2016) Endoscopic scene labelling and augmentation using intraoperative pulsatile motion and colour appearance cues with preoperative anatomical priors. Int J Comput Assist Radiol Surg 11(8):1409–1418PubMedCrossRef Nosrati MS, Amir-Khalili A, Peyrat J-M, Abinahed J, Al-Alao O, Al-Ansari A et al (2016) Endoscopic scene labelling and augmentation using intraoperative pulsatile motion and colour appearance cues with preoperative anatomical priors. Int J Comput Assist Radiol Surg 11(8):1409–1418PubMedCrossRef
46.
Zurück zum Zitat Maier-Hein L, Vedula SS, Speidel S, Navab N, Kikinis R, Park A et al (2017) Surgical data science for next-generation interventions. Nature Biomed Eng 1(9):691–696CrossRef Maier-Hein L, Vedula SS, Speidel S, Navab N, Kikinis R, Park A et al (2017) Surgical data science for next-generation interventions. Nature Biomed Eng 1(9):691–696CrossRef
48.
Zurück zum Zitat Nguyen XA, Ljuhar D, Pacilli M, Nataraja RM, Chauhan S (2019) Surgical skill levels: classification and analysis using deep neural network model and motion signals. Comput Methods Programs Biomed 177:1–8PubMedCrossRef Nguyen XA, Ljuhar D, Pacilli M, Nataraja RM, Chauhan S (2019) Surgical skill levels: classification and analysis using deep neural network model and motion signals. Comput Methods Programs Biomed 177:1–8PubMedCrossRef
49.
Zurück zum Zitat Hung AJ, Chen J, Che Z, Nilanon T, Jarc A, Titus M et al (2018) Utilizing machine learning and automated performance metrics to evaluate robot-assisted radical prostatectomy performance and predict outcomes. J Endourol 32(5):438–444PubMedCrossRef Hung AJ, Chen J, Che Z, Nilanon T, Jarc A, Titus M et al (2018) Utilizing machine learning and automated performance metrics to evaluate robot-assisted radical prostatectomy performance and predict outcomes. J Endourol 32(5):438–444PubMedCrossRef
50.
Zurück zum Zitat Ross T, Reinke A, Full PM, Wagner M, Kenngott H, Apitz M, et al (2020) Robust medical instrument segmentation challenge 2019. arXiv:200310299 Ross T, Reinke A, Full PM, Wagner M, Kenngott H, Apitz M, et al (2020) Robust medical instrument segmentation challenge 2019. arXiv:200310299
51.
Zurück zum Zitat Colleoni E, Moccia S, Du X, De Momi E, Stoyanov D (2019) Deep learning based robotic tool detection and articulation estimation with spatio-temporal layers. IEEE Robot Autom Lett 4(3):2714–2721CrossRef Colleoni E, Moccia S, Du X, De Momi E, Stoyanov D (2019) Deep learning based robotic tool detection and articulation estimation with spatio-temporal layers. IEEE Robot Autom Lett 4(3):2714–2721CrossRef
52.
Zurück zum Zitat Allan M, Kondo S, Bodenstedt S, Leger S, Kadkhodamohammadi R, Luengo I et al (2020) 2018 Robotic Scene Segmentation Challenge. arXiv:200111190 Allan M, Kondo S, Bodenstedt S, Leger S, Kadkhodamohammadi R, Luengo I et al (2020) 2018 Robotic Scene Segmentation Challenge. arXiv:200111190
53.
Zurück zum Zitat Twinanda AP, Shehata S, Mutter D, Marescaux J, De Mathelin M, Padoy N (2016) Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans Med Imaging 36(1):86–97PubMedCrossRef Twinanda AP, Shehata S, Mutter D, Marescaux J, De Mathelin M, Padoy N (2016) Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans Med Imaging 36(1):86–97PubMedCrossRef
54.
Zurück zum Zitat Zisimopulos O, Flouty E, Luengo I et al (2018) DeepPhase: surgical phase recognition in CAT ARA CTS videos. In: Frangi AF, Schnabel JA, Davatzikos C, Alberola-Lopez C, Fichtinger G (eds) Medical Image Computing and Computer Assisted Intervention: MICCAI 2018. Springer, Cham, Switzerland, pp 265–272 Zisimopulos O, Flouty E, Luengo I et al (2018) DeepPhase: surgical phase recognition in CAT ARA CTS videos. In: Frangi AF, Schnabel JA, Davatzikos C, Alberola-Lopez C, Fichtinger G (eds) Medical Image Computing and Computer Assisted Intervention: MICCAI 2018. Springer, Cham, Switzerland, pp 265–272
55.
Zurück zum Zitat Czempiel T, Paschali M, Keicher M, Simson W, Feussner H, Kim ST et al (2020) TeCNO: Surgical phase recognition with multi-stage temporal convolutional networks. arXiv preprint arXiv:200310751 Czempiel T, Paschali M, Keicher M, Simson W, Feussner H, Kim ST et al (2020) TeCNO: Surgical phase recognition with multi-stage temporal convolutional networks. arXiv preprint arXiv:200310751
56.
Zurück zum Zitat Oh JS, Minneman JA, Ehlers AP, Deal SB, Alseidi AA, Wright AS (2019) 2019 Scientific Session of the Society of American Gastrointestinal and Endoscopic Surgeons (SAGES), Baltimore, Maryland, USA: Resident Fellow Session Abstracts Oh JS, Minneman JA, Ehlers AP, Deal SB, Alseidi AA, Wright AS (2019) 2019 Scientific Session of the Society of American Gastrointestinal and Endoscopic Surgeons (SAGES), Baltimore, Maryland, USA: Resident Fellow Session Abstracts
57.
Zurück zum Zitat van Amsterdam B, Clarkson MJ, Stoyanov D (2020) Multi-task recurrent neural network for surgical gesture recognition and progress prediction. arXiv preprint arXiv:200304772 van Amsterdam B, Clarkson MJ, Stoyanov D (2020) Multi-task recurrent neural network for surgical gesture recognition and progress prediction. arXiv preprint arXiv:200304772
58.
Zurück zum Zitat Ahmidi N, Tao L, Sefati S, Gao Y, Lea C, Haro BB et al (2017) A dataset and benchmarks for segmentation and recognition of gestures in robotic surgery. IEEE Trans Biomed Eng 64(9):2025–2041PubMedPubMedCentralCrossRef Ahmidi N, Tao L, Sefati S, Gao Y, Lea C, Haro BB et al (2017) A dataset and benchmarks for segmentation and recognition of gestures in robotic surgery. IEEE Trans Biomed Eng 64(9):2025–2041PubMedPubMedCentralCrossRef
59.
Zurück zum Zitat Twinanda AP, Yengera G, Mutter D, Marescaux J, Padoy N (2018) RSDNet: Learning to predict remaining surgery duration from laparoscopic videos without manual annotations. IEEE Trans Med Imaging 38(4):1069–1078PubMedCrossRef Twinanda AP, Yengera G, Mutter D, Marescaux J, Padoy N (2018) RSDNet: Learning to predict remaining surgery duration from laparoscopic videos without manual annotations. IEEE Trans Med Imaging 38(4):1069–1078PubMedCrossRef
60.
Zurück zum Zitat Petscharnig S, Schöffmann K (2018) Binary convolutional neural network features off-the-shelf for image to video linking in endoscopic multimedia databases. Multimedia Tools Appl 77(21):28817–28842CrossRef Petscharnig S, Schöffmann K (2018) Binary convolutional neural network features off-the-shelf for image to video linking in endoscopic multimedia databases. Multimedia Tools Appl 77(21):28817–28842CrossRef
61.
Zurück zum Zitat Yu F, Croso GS, Kim TS, Song Z, Parker F, Hager GD et al (2019) Assessment of automated identification of phases in videos of cataract surgery using machine learning and deep learning techniques. JAMA Net Open 2(4):e191860CrossRef Yu F, Croso GS, Kim TS, Song Z, Parker F, Hager GD et al (2019) Assessment of automated identification of phases in videos of cataract surgery using machine learning and deep learning techniques. JAMA Net Open 2(4):e191860CrossRef
63.
64.
Zurück zum Zitat David G, Gunnarsson CL, Waters HC, Horblyuk R, Kaplan HS (2013) Economic measurement of medical errors using a hospital claims database. Value Health 16(2):305–310PubMedCrossRef David G, Gunnarsson CL, Waters HC, Horblyuk R, Kaplan HS (2013) Economic measurement of medical errors using a hospital claims database. Value Health 16(2):305–310PubMedCrossRef
65.
Zurück zum Zitat Institue E (2014) Top 10 health technology hazards for 2015. Health Dev 1:3–6 Institue E (2014) Top 10 health technology hazards for 2015. Health Dev 1:3–6
66.
Zurück zum Zitat Chen A, Ghodoussipour S, Titus MB, Nguyen JH, Chen J, Ma R et al (2019) Comparison of clinical outcomes and automated performance metrics in robot-assisted radical prostatectomy with and without trainee involvement. World J Urol 2019:1–7 Chen A, Ghodoussipour S, Titus MB, Nguyen JH, Chen J, Ma R et al (2019) Comparison of clinical outcomes and automated performance metrics in robot-assisted radical prostatectomy with and without trainee involvement. World J Urol 2019:1–7
Metadaten
Titel
Evolving robotic surgery training and improving patient safety, with the integration of novel technologies
verfasst von
I-Hsuan Alan Chen
Ahmed Ghazi
Ashwin Sridhar
Danail Stoyanov
Mark Slack
John D. Kelly
Justin W. Collins
Publikationsdatum
06.11.2020
Verlag
Springer Berlin Heidelberg
Erschienen in
World Journal of Urology / Ausgabe 8/2021
Print ISSN: 0724-4983
Elektronische ISSN: 1433-8726
DOI
https://doi.org/10.1007/s00345-020-03467-7

Weitere Artikel der Ausgabe 8/2021

World Journal of Urology 8/2021 Zur Ausgabe

Ambulantisierung: Erste Erfahrungen mit dem Hybrid-DRG

02.05.2024 DCK 2024 Kongressbericht

Die Hybrid-DRG-Verordnung soll dazu führen, dass mehr chirurgische Eingriffe ambulant durchgeführt werden, wie es in anderen Ländern schon länger üblich ist. Die gleiche Vergütung im ambulanten und stationären Sektor hatten Niedergelassene schon lange gefordert. Aber die Umsetzung bereitet ihnen doch Kopfzerbrechen.

Sind Frauen die fähigeren Ärzte?

30.04.2024 Gendermedizin Nachrichten

Patienten, die von Ärztinnen behandelt werden, dürfen offenbar auf bessere Therapieergebnisse hoffen als Patienten von Ärzten. Besonders gilt das offenbar für weibliche Kranke, wie eine Studie zeigt.

Harninkontinenz: Netz-Op. erfordert über lange Zeit intensive Nachsorge

30.04.2024 Harninkontinenz Nachrichten

Frauen mit Belastungsinkontinenz oder Organprolaps sind nach einer Netz-Operation keineswegs beschwerdefrei. Vielmehr scheint die Krankheitslast weiterhin hoch zu sein, sogar höher als von harninkontinenten Frauen, die sich nicht haben operieren lassen.

Adjuvante Immuntherapie verlängert Leben bei RCC

25.04.2024 Nierenkarzinom Nachrichten

Nun gibt es auch Resultate zum Gesamtüberleben: Eine adjuvante Pembrolizumab-Therapie konnte in einer Phase-3-Studie das Leben von Menschen mit Nierenzellkarzinom deutlich verlängern. Die Sterberate war im Vergleich zu Placebo um 38% geringer.

Update Urologie

Bestellen Sie unseren Fach-Newsletter und bleiben Sie gut informiert.