Please use this identifier to cite or link to this item:
https://www.um.edu.mt/library/oar/handle/123456789/144351| Title: | A distance robust EOG-based feature for gaze trajectory inference |
| Authors: | Mifsud, Matthew Camilleri, Tracey A. Camilleri, Kenneth P. |
| Keywords: | Electrooculography Eye -- Examination Eye tracking Eye -- Movements Human-computer interaction Biomedical engineering Visual evoked response |
| Issue Date: | 2024 |
| Publisher: | Elsevier Ltd. |
| Citation: | Mifsud, M., Camilleri, T. A., & Camilleri, K. P. (2024). A distance robust EOG-based feature for gaze trajectory inference. Biomedical Signal Processing and Control, 90, 105852. |
| Abstract: | The vast majority of eye controlled applications rely on a mapping function which relates the estimated gaze angle of the user to a pixel coordinate on the screen. In turn, such systems heavily depend on the user maintaining an unnatural, fixed and known distance from the computer screen in order to ensure proper operation of the designed application. This work has investigated the use of trajectory bearing angles as a distance-robust electrooculography (EOG)-based feature which can be used for gaze trajectory inference. The trajectory bearing angles are extracted directly from EOG data and are shown to be robust to the distance of the user from the screen. Three different EOG-based bearing angle estimation methods are investigated and results of a proof-of-concept Hidden Markov Model (HMM) swipe typing application carried out by ten subjects, show that an average top-five rate of over 80% can be achieved across three different user distances from the screen. |
| URI: | https://www.um.edu.mt/library/oar/handle/123456789/144351 |
| Appears in Collections: | Scholarly Works - FacEngSCE |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| A_distance_robust_EOG_based_feature_for_gaze_trajectory_inference_2024.pdf Restricted Access | 1.29 MB | Adobe PDF | View/Open Request a copy |
Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.
