More Than Meets the Eye: Analyzing Anesthesiologists' Visual Attention in the Operating Room Using Deep Learning Models

by   Sapir Gershov, et al.

Patient's vital signs, which are displayed on monitors, make the anesthesiologist's visual attention (VA) a key component in the safe management of patients under general anesthesia; moreover, the distribution of said VA and the ability to acquire specific cues throughout the anesthetic, may have a direct impact on patient's outcome. Currently, most studies employ wearable eye-tracking technologies to analyze anesthesiologists' visual patterns. Albeit being able to produce meticulous data, wearable devices are not a sustainable solution for large-scale or long-term use for data collection in the operating room (OR). Thus, by utilizing a novel eye-tracking method in the form of deep learning models that process monitor-mounted webcams, we collected continuous behavioral data and gained insight into the anesthesiologist's VA distribution with minimal disturbance to their natural workflow. In this study, we collected OR video recordings using the proposed framework and compared different visual behavioral patterns. We distinguished between baseline VA distribution during uneventful periods to patterns associated with active phases or during critical, unanticipated incidents. In the future, such a platform may serve as a crucial component of context-aware assistive technologies in the OR.


page 6

page 7

page 8


Decoding Attention from Gaze: A Benchmark Dataset and End-to-End Models

Eye-tracking has potential to provide rich behavioral data about human c...

Predicting video saliency using crowdsourced mouse-tracking data

This paper presents a new way of getting high-quality saliency maps for ...

Deep Learning-based Eye-Tracking Analysis for Diagnosis of Alzheimer's Disease Using 3D Comprehensive Visual Stimuli

Alzheimer's Disease (AD) causes a continuous decline in memory, thinking...

MLGaze: Machine Learning-Based Analysis of Gaze Error Patterns in Consumer Eye Tracking Systems

Analyzing the gaze accuracy characteristics of an eye tracker is a criti...

TobiiGlassesPySuite: An open-source suite for using the Tobii Pro Glasses 2 in eye-tracking studies

In this paper we present the TobiiGlassesPySuite, an open-source suite w...

Learning Behavioral Representations of Routines From Large-scale Unlabeled Wearable Time-series Data Streams using Hawkes Point Process

Continuously-worn wearable sensors enable researchers to collect copious...

Fostering Bilateral Patient-Clinician Engagement in Active Self-Tracking of Subjective Experience

In this position paper we describe select aspects of our experience with...

Please sign up or login with your details

Forgot password? Click here to reset