Skip to main content
Erschienen in: International Journal of Computer Assisted Radiology and Surgery 7/2019

11.04.2019 | Original Article

A novel gaze-supported multimodal human–computer interaction for ultrasound machines

verfasst von: Hongzhi Zhu, Septimiu E. Salcudean, Robert N. Rohling

Erschienen in: International Journal of Computer Assisted Radiology and Surgery | Ausgabe 7/2019

Einloggen, um Zugang zu erhalten

Abstract

Purpose

Conventional ultrasound (US) machines employ a physical control panel (PCP) as the primary user interface for machine control. This panel is adjacent to the main machine display that requires the operator’s constant attention. The switch of attention to the control panel can lead to interruptions in the flow of the medical examination. Some ultraportable machines also lack many physical controls. Furthermore, the need to both control the US machine and observe the US image may lead the practitioners to adopt unergonomic postures and repetitive motions that can lead to work-related injuries. Therefore, there is a need for a more efficient human–computer interaction method on US machines.

Methods

To tackle some of the limitations with the PCP, we propose to merge the PCP into the main screen of the US machines. We propose to use gaze tracking and a handheld controller so that machine control can be achieved via a multimodal human–computer interaction (HCI) method that does not require one to touch the screen or look away from the US image. As a first step, a pop-up menu and measurement tool were designed on top of the US image based on gaze position for efficient machine control.

Results

A comparative study was performed on the BK Medical SonixTOUCH US machine. Participants were asked to complete the task of measuring the area of an ellipse-shaped tumor in a phantom using our gaze-supported HCI method as well as the traditional method. The user study indicates that the task completion time can be reduced by \(20.6\%\) when using our gaze-supported HCI, while no extra workload is imposed on the operators.

Conclusions

Our preliminary study suggests that, when combined with a simple handheld controller, eye gaze tracking can be integrated into the US machine HCI for more efficient machine control.
Anhänge
Nur mit Berechtigung zugänglich
Literatur
1.
Zurück zum Zitat Andreoni G, Mazzola M, Matteoli S, D’Onofrio S, Forzoni L (2015) Ultrasound system typologies, user interfaces and probes design: a review. Procedia Manuf 3:112–119CrossRef Andreoni G, Mazzola M, Matteoli S, D’Onofrio S, Forzoni L (2015) Ultrasound system typologies, user interfaces and probes design: a review. Procedia Manuf 3:112–119CrossRef
2.
Zurück zum Zitat Bartol K, Graziano S, Kelton W (2007) Remote wireless control device for an ultrasound machine and method. US Patent App. 10/550,046 Bartol K, Graziano S, Kelton W (2007) Remote wireless control device for an ultrasound machine and method. US Patent App. 10/550,046
3.
Zurück zum Zitat Cobbold RS (2006) Foundations of biomedical ultrasound. Oxford University Press, Oxford Cobbold RS (2006) Foundations of biomedical ultrasound. Oxford University Press, Oxford
4.
Zurück zum Zitat Feit AM, Williams S, Toledo A, Paradiso, A, Kulkarni H, Kane S, Morris MR (2017) Toward everyday gaze input: accuracy and precision of eye tracking and implications for design. In: Proceedings of the 2017 CHI conference on human factors in computing systems. ACM, pp 1118–1130 Feit AM, Williams S, Toledo A, Paradiso, A, Kulkarni H, Kane S, Morris MR (2017) Toward everyday gaze input: accuracy and precision of eye tracking and implications for design. In: Proceedings of the 2017 CHI conference on human factors in computing systems. ACM, pp 1118–1130
5.
Zurück zum Zitat Halwani Y (2017) An investigation of multi-modal gaze-supported zoom and pan interactions in ultrasound machines. Ph.D. thesis, University of British Columbia Halwani Y (2017) An investigation of multi-modal gaze-supported zoom and pan interactions in ultrasound machines. Ph.D. thesis, University of British Columbia
6.
Zurück zum Zitat Halwani Y, Salcudean SE, Lessoway VA, Fels SS (2017) Enhancing zoom and pan in ultrasound machines with a multimodal gaze-based interface. In: Proceedings of the 2017 CHI conference extended abstracts on human factors in computing systems. ACM, pp 1648–1654 Halwani Y, Salcudean SE, Lessoway VA, Fels SS (2017) Enhancing zoom and pan in ultrasound machines with a multimodal gaze-based interface. In: Proceedings of the 2017 CHI conference extended abstracts on human factors in computing systems. ACM, pp 1648–1654
7.
Zurück zum Zitat Han JH, Yang Eh (2015) Method of moving the displays of an ultrasound diagnostic device and ultrasound diagnostic device. US Patent App. 14/314,313 Han JH, Yang Eh (2015) Method of moving the displays of an ultrasound diagnostic device and ultrasound diagnostic device. US Patent App. 14/314,313
8.
Zurück zum Zitat Hart SG (2006) Nasa-task load index (nasa-tlx); 20 years later. In: Proceedings of the human factors and ergonomics society annual meeting, vol 50. Sage Publications, Los Angeles, CA, pp 904–908 Hart SG (2006) Nasa-task load index (nasa-tlx); 20 years later. In: Proceedings of the human factors and ergonomics society annual meeting, vol 50. Sage Publications, Los Angeles, CA, pp 904–908
9.
Zurück zum Zitat Hart SG, Staveland LE (1988) Development of nasa-tlx (task load index): results of empirical and theoretical research. Adv Psychol, vol 52. Elsevier, Amsterdam, pp 139–183 Hart SG, Staveland LE (1988) Development of nasa-tlx (task load index): results of empirical and theoretical research. Adv Psychol, vol 52. Elsevier, Amsterdam, pp 139–183
10.
Zurück zum Zitat Istance H, Bates R, Hyrskykari A, Vickers S (2008) Snap clutch, a moded approach to solving the midas touch problem. In: Proceedings of the 2008 symposium on eye tracking research & applications. ACM, pp 221–228 Istance H, Bates R, Hyrskykari A, Vickers S (2008) Snap clutch, a moded approach to solving the midas touch problem. In: Proceedings of the 2008 symposium on eye tracking research & applications. ACM, pp 221–228
11.
Zurück zum Zitat Jarc AM (2017) Medical devices, systems, and methods using eye gaze tracking for stereo viewer. US Patent App. 15/126,151 Jarc AM (2017) Medical devices, systems, and methods using eye gaze tracking for stereo viewer. US Patent App. 15/126,151
12.
Zurück zum Zitat Murphy S, Need DE (1996) Voice control of a medical ultrasound scanning machine. US Patent 5,544,654 Murphy S, Need DE (1996) Voice control of a medical ultrasound scanning machine. US Patent 5,544,654
13.
Zurück zum Zitat Pelissier L, Zhang B, Bobovsky T (2016) Highly configurable medical ultrasound machine and related methods. US Patent 9,408,587 Pelissier L, Zhang B, Bobovsky T (2016) Highly configurable medical ultrasound machine and related methods. US Patent 9,408,587
14.
Zurück zum Zitat Reiner B (2017) Visually directed human-computer interaction for medical applications. US Patent 9,841,811 Reiner B (2017) Visually directed human-computer interaction for medical applications. US Patent 9,841,811
15.
Zurück zum Zitat Ruhland K, Peters CE, Andrist S, Badler JB, Badler NI, Gleicher M, Mutlu B, McDonnell R (2015) A review of eye gaze in virtual agents, social robotics and HCI: behaviour generation, user interaction and perception. Comput Graph Forum 34(6):299–326CrossRef Ruhland K, Peters CE, Andrist S, Badler JB, Badler NI, Gleicher M, Mutlu B, McDonnell R (2015) A review of eye gaze in virtual agents, social robotics and HCI: behaviour generation, user interaction and perception. Comput Graph Forum 34(6):299–326CrossRef
16.
Zurück zum Zitat Washburn MJ, Hawley BM, Prichard SD (2010) Voice control of a generic input device for an ultrasound system. US Patent 7,698,142 Washburn MJ, Hawley BM, Prichard SD (2010) Voice control of a generic input device for an ultrasound system. US Patent 7,698,142
17.
Zurück zum Zitat Weinger MB, Gardner-Bonneau DJ, Wiklund ME (2010) Handbook of human factors in medical device design. CRC Press, Boca RatonCrossRef Weinger MB, Gardner-Bonneau DJ, Wiklund ME (2010) Handbook of human factors in medical device design. CRC Press, Boca RatonCrossRef
18.
Zurück zum Zitat Wichrowski M (2015) Usability engineering in the prototyping process of software user interfaces for mobile medical ultrasound devices. Comput Sci 16:219–236CrossRef Wichrowski M (2015) Usability engineering in the prototyping process of software user interfaces for mobile medical ultrasound devices. Comput Sci 16:219–236CrossRef
19.
Zurück zum Zitat Yudkovitch LM, Farrokhnia F, Chiao R (2007) Method and apparatus for natural voice control of an ultrasound machine. US Patent 7,247,139 Yudkovitch LM, Farrokhnia F, Chiao R (2007) Method and apparatus for natural voice control of an ultrasound machine. US Patent 7,247,139
20.
Zurück zum Zitat Zhai S, Morimoto C, Ihde S (1999) Manual and gaze input cascaded (magic) pointing. In: Proceedings of the SIGCHI Conference on human factors in computing systems. ACM, pp 246–253 Zhai S, Morimoto C, Ihde S (1999) Manual and gaze input cascaded (magic) pointing. In: Proceedings of the SIGCHI Conference on human factors in computing systems. ACM, pp 246–253
Metadaten
Titel
A novel gaze-supported multimodal human–computer interaction for ultrasound machines
verfasst von
Hongzhi Zhu
Septimiu E. Salcudean
Robert N. Rohling
Publikationsdatum
11.04.2019
Verlag
Springer International Publishing
Erschienen in
International Journal of Computer Assisted Radiology and Surgery / Ausgabe 7/2019
Print ISSN: 1861-6410
Elektronische ISSN: 1861-6429
DOI
https://doi.org/10.1007/s11548-019-01964-8

Weitere Artikel der Ausgabe 7/2019

International Journal of Computer Assisted Radiology and Surgery 7/2019 Zur Ausgabe

Update Radiologie

Bestellen Sie unseren Fach-Newsletter und bleiben Sie gut informiert.