ABSTRACT
Web accessibility and digital inclusion are still one of the great research challenges in computer science. For people with severe motor disabilities, eye-gaze interaction enables hands-free operation of graphical interfaces. Nonetheless, gaze tracking devices typically require frequent user calibration that weakens the overall user experience and performance. In this paper we propose EyePursuitLinks, an eye-pursuit based technique that allows people with severe motor disabilities to browse the web by following multiple moving targets corresponding to the links they want to follow. The main contribution of this paper is the use of Smart Targets (ST) to select potentially large number of hyperlinks within a relatively small area. We have performed a user experiment with 10 volunteers to evaluate the performance of the Smart Targets selection against conventional pursuit selection mechanism using 4, 8, and 16 simultaneous targets. Our results show that the use of ST is significantly more robust than the conventional method for larger number of targets.
- M. Cecilia Baranauskas, Clarisse de Souza, and Roberto Pereira. 2014. I GranDIHC-BR — Grandes Desafios de Pesquisa em Interação Humano-Computador no Brasil. https://doi.org/10.13140/2.1.3651.9201Google ScholarCross Ref
- Alex Torquato S. Carneiro, Carlos Eduardo L. Elmadjian, Candy Gonzales, Flavio L. Coutinho, and Carlos H. Morimoto. 2019. PursuitPass: A Visual Pursuit-Based User Authentication System. In 2019 32nd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI). IEEE, Brazil, 226–233. https://doi.org/10.1109/SIBGRAPI.2019.00038Google ScholarCross Ref
- Marcus Carter, Eduardo Velloso, John Downs, Abigail Sellen, Kenton O’Hara, and Frank Vetere. 2016. PathSync: Multi-User Gestural Interaction with Touchless Rhythmic Path Mimicry. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems (San Jose, California, USA) (CHI ’16). ACM, New York, NY, USA, 3415–3427. https://doi.org/10.1145/2858036.2858284Google ScholarDigital Library
- Christopher Clarke, Alessio Bellino, Augusto Esteves, Eduardo Velloso, and Hans Gellersen. 2016. TraceMatch: a computer vision technique for user input by tracing of animated controls. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing. 298–303.Google ScholarDigital Library
- Robert Dawson. 2011. How significant is a boxplot outlier?Journal of Statistics Education 19, 2 (2011).Google Scholar
- Heiko Drewes, Mohamed Khamis, and Florian Alt. 2019. DialPlates: enabling pursuits-based user interfaces with large target numbers. In Proceedings of the 18th International Conference on Mobile and Ubiquitous Multimedia. 1–10.Google ScholarDigital Library
- Andrew T Duchowski, Nathan Cournia, and Hunter Murphy. 2004. Gaze-contingent displays: A review. CyberPsychology & Behavior 7, 6 (2004), 621–634.Google ScholarCross Ref
- Augusto Esteves, Eduardo Velloso, Andreas Bulling, and Hans Gellersen. 2015. Orbits: Gaze interaction for smart watches using smooth pursuit eye movements. In Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology. ACM, 457–466.Google ScholarDigital Library
- Pedro Figueiredo and Manuel J. Fonseca. 2018. EyeLinks: A Gaze-Only Click Alternative for Heterogeneous Clickables. In Proceedings of the 20th ACM International Conference on Multimodal Interaction (Boulder, CO, USA) (ICMI ’18). Association for Computing Machinery, New York, NY, USA, 307–314. https://doi.org/10.1145/3242969.3243021Google ScholarDigital Library
- David Halliday, Robert Resnick, and Jearl Walker. 2016. Fundamentos de física. Vol. I.Grupo Gen-LTC.Google Scholar
- Christof Lutteroth, Moiz Penkar, and Gerald Weber. 2015. Gaze vs. Mouse: A Fast and Accurate Gaze-Only Click Alternative. In Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology (Charlotte, NC, USA) (UIST ’15). Association for Computing Machinery, New York, NY, USA, 385–394. https://doi.org/10.1145/2807442.2807461Google ScholarDigital Library
- Otto Hans-Martin Lutz, Antje Christine Venjakob, and Stefan Ruff. 2015. SMOOVS: Towards calibration-free text entry by gaze using smooth pursuit movements. Journal of Eye Movement Research 8, 1 (2015), 1–11.Google Scholar
- P. Majaranta, U.K. Ahola, and O. Špakov. 2009. Fast gaze typing with an adjustable dwell time. In Proceedings of the 27th international conference on Human factors in computing systems (CHI ’09). ACM, New York, NY, 357–360. https://doi.org/10.1145/1518701.1518758Google ScholarDigital Library
- Katharina Reiter, Ken Pfeuffer, Augusto Esteves, Tim Mittermeier, and Florian Alt. 2022. Look & Turn: One-handed and Expressive Menu Interaction by Gaze and Arm Turns in VR. In 2022 Symposium on Eye Tracking Research and Applications. 1–7.Google ScholarDigital Library
- John W Tukey 1977. Exploratory data analysis. Vol. 2. Reading, Mass.Google Scholar
- JJ Van Middendorp, F Watkins, C Park, and H Landymore. 2015. Eye-tracking computer systems for inpatients with tetraplegia: findings from a feasibility study. Spinal cord 53, 3 (2015), 221–225.Google Scholar
- Eduardo Velloso, Flavio Luiz Coutinho, Andrew Kurauchi, and Carlos H Morimoto. 2018. Circular orbits detection for gaze interaction using 2D correlation and profile matching algorithms. In Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications. ACM, 25.Google ScholarDigital Library
- Eduardo Velloso and Carlos H Morimoto. 2021. A probabilistic interpretation of motion correlation selection techniques. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–13.Google ScholarDigital Library
- David Verweij, Augusto Esteves, Vassilis-Javed Khan, and Saskia Bakker. 2017. WaveTrace: motion matching input using wrist-worn motion sensors. In Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems. 2180–2186.Google ScholarDigital Library
- Mélodie Vidal, Andreas Bulling, and Hans Gellersen. 2013. Pursuits: Spontaneous Interaction with Displays Based on Smooth Pursuit Eye Movement and Moving Targets. In Proceedings of the 2013 ACM International Joint Conference on Pervasive and Ubiquitous Computing (Zurich, Switzerland) (UbiComp ’13). ACM, New York, NY, USA, 439–448. https://doi.org/10.1145/2493432.2493477Google ScholarDigital Library
- Mélodie Vidal, Andreas Bulling, and Hans Gellersen. 2015. Pursuits: Spontaneous eye-based interaction for dynamic interfaces. GetMobile: Mobile Computing and Communications 18, 4 (2015), 8–10.Google ScholarDigital Library
- Mélodie Vidal, Ken Pfeuffer, Andreas Bulling, and Hans W Gellersen. 2013. Pursuits: eye-based interaction with moving targets. In CHI’13 Extended Abstracts on Human Factors in Computing Systems. 3147–3150.Google Scholar
- John Williamson. 2006. Continuous uncertain interaction. Ph. D. Dissertation. University of Glasgow (United Kingdom).Google Scholar
- John Williamson and Roderick Murray-Smith. 2004. Pointing without a pointer. In CHI’04 Extended Abstracts on Human Factors in Computing Systems. ACM, 1407–1410.Google Scholar
Index Terms
- EyePursuitLinks - an Eye-pursuit Based Interface for Web Browsing Using Smart Targets
Recommendations
Web browsing interface for people with severe speech and motor impairment in India
ASSETS '14: Proceedings of the 16th international ACM SIGACCESS conference on Computers & accessibilityWe present design and development of a web browser that allow easy dissemination of information through World Wide Web for people with cerebral palsy in India. Our focus user group comprises people with severe form of spastic cerebral palsy and highly ...
Guidelines for an accessible web automation interface
ASSETS '11: The proceedings of the 13th international ACM SIGACCESS conference on Computers and accessibilityIn recent years, the Web has become an ever more sophisticated and irreplaceable tool in our daily lives. While the visual Web has been advancing at a rapid pace, assistive technology has not been able to keep up, increasingly putting visually impaired ...
Assistive web browsing with touch interfaces
ASSETS '10: Proceedings of the 12th international ACM SIGACCESS conference on Computers and accessibilityThis demonstration will propose a touch-based directional navigation technique, on touch interface (e.g., iPhone, Macbook) for people with visual disabilities especially blind individuals. Such interfaces coupled with TTS (text-to-speech) systems open ...
Comments