Abstract
Eye gaze trackers are devices designed to identify an individual’s gaze fixation in relation to a screen or another reference point. These tools are widely applied in usability testing as they provide various metrics for studying how people interact with applications. In the past, these tools were expensive and required a controlled environment, as well as trained personnel for proper operation. Although nowadays, new implementations do not require physical hardware to perform these tests, they often rely on license-based models instead of being open source. The objective of this work is to create a standalone system that enables any user to implement a low-cost eye gaze tracker using web technologies. The goal is to facilitate its use in remote and face-to-face studies in a simple way, requiring only a computer and a webcam. We evaluated the impact of three different calibration techniques on the performance of a regression-based prediction algorithm in eye-tracking. In our experiments, the best result of linear regression was obtained with a circular calibration system that uses 160 points. Furthermore, we integrated the system with a web interface and an API, enabling users to record their usability sessions and analyze fixation points through heatmaps.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Remote User Experience LAB (RUXLAB). Available at: https://github.com/uramakilab/remote-usability-lab.
- 2.
ISO/CD 9241-11: Ergonomics of human-system interaction - Part 11: Guidance on usability (1998). Available at: https://www.iso.org/standard/63500.html. Accessed date: September 15, 2023.
- 3.
Global leader in eye tracking for over 20 years - Tobii. Available at: https://www.tobii.com/. Accessed date: September 15, 2023.
- 4.
Pupil Invisible - Eye tracking glasses for the real world - Pupil Labs. Available at: https://pupil-labs.com/. Accessed date: September 15, 2023.
- 5.
Remote User Experience LAB. Available at: https://retlab.web.app/. Accessed date: September 15, 2023.
- 6.
OpenAPI Specification - Version 3.0.3. Available at: https://swagger.io/specification/. Accessed date: September 15, 2023.
- 7.
TensorFlow.js: Face Landmarks Detection. Available at: https://github.com/tensorflow/tfjs-models/tree/master/face-landmarks-detection. Accessed date: September 15, 2023.
- 8.
Tensorflow Javascript. Available at: https://www.tensorflow.org/js. Accessed date: September 15, 2023.
- 9.
MediaPipe. Available at: https://developers.google.com/mediapipe. Accessed date: September 15, 2023.
- 10.
Vuejs. Available at: https://vuejs.org/ Accessed date: September 15, 2023.
- 11.
Vue Component Framework. Available at: https://vuetifyjs.com/. Accessed date: September 15, 2023.
- 12.
Material Design. Available at: https://m2.material.io/design/guidelines-overview Accessed date: September 15, 2023.
- 13.
MediaDevices - Web APIs | MDN. Available at: https://developer.mozilla.org/es/docs/Web/API/MediaDevices. Accessed date: September 15, 2023.
- 14.
MediaRecorder - Web APIs | MDN. Available at: https://developer.mozilla.org/en-US/docs/Web/API/MediaRecorder. Accessed date: September 15, 2023.
- 15.
Firebase. Available at: https://firebase.google.com/. Accessed date: September 15, 2023.
- 16.
Dynamic heatmap for the web. Available at: https://www.patrick-wied.at/static/heatmapjs/. Accessed date: September 15, 2023.
- 17.
Available at: https://github.com/uramakilab/web-eye-tracker. Accessed date: September 15, 2023.
References
Agustin, J.S., et al.: Evaluation of a low-cost open-source gaze tracker. In: Proceedings of the 2010 Symposium on eye-tracking research & applications, pp. 77–80 (2010)
Ahn, H., Jeon, J., Ko, D., Gwak, J., Jeon, M.: Contactless real-time eye gaze-mapping system based on simple Siamese networks. Appl. Sci. 13(9) (2023). https://doi.org/10.3390/app13095374
Bastien, J.C.: Usability testing: a review of some methodological and technical aspects of the method. Int. J. Med. Inform. 79(4), e18–e23 (2010)
Bevan, N.: What is the difference between the purpose of usability and user experience evaluation methods. In: Proceedings of the Workshop UXEM. vol. 9, pp. 1–4. Citeseer (2009)
Biedert, R., Buscher, G., Dengel, A.: The eye book. Informatik-Spektrum 33(3), 272–281 (2009)
Braunschweig, K., Eberius, J., Thiele, M., Lehner, W.: The state of open data. Limits Curr. Open Data Platforms 1, 72–72 (2012)
Brooke, J., et al.: Sus-a quick and dirty usability scale. Usability Eval. Indust. 189(194), 4–7 (1996)
Capdevila, M.G., Saltiveri, T.G.: Heurísticas de usabilidad utilizando una plataforma abierta y colaborativa. V Congreso Internacional de Ciencias de la Computación y Sistemas de Información 2021 (2022)
Carter, B.T., Luke, S.G.: Best practices in eye tracking research. Int. J. Psychophysiol. 155, 49–62 (2020)
Castillo, J.C., Hartson, H.R., Hix, D.: Remote usability evaluation: can users report their own critical incidents? In: CHI 98 Conference Summary on Human Factors In Computing Systems, pp. 253–254 (1998)
Chennamma, H., Yuan, X.: A survey on eye-gaze tracking techniques. arXiv preprint arXiv:1312.6410 (2013)
Dalmaijer, E.S.: Pygaze: an open-source, cross-platform toolbox for minimal-effort programming of eyetracking experiments. Behav. Res. Methods 46(4), 913–931 (2014)
Ferhat, Onur e Vilariño, F.: Rastreamento ocular de baixo custo: o panorama atual. Inteligência computacional e neurociência 2016 (2016)
Ferhat, O., Vilarino, F., Sánchez, F.J.: A cheap portable eye-tracker solution for common setups. J. Eye Movement Res. 7(3) (2014)
Frøkjær, E., Hertzum, M., Hornbæk, K.: Measuring usability: are effectiveness, efficiency, and satisfaction really correlated? In: Proceedings of the SIGCHI conference on Human Factors in Computing Systems, pp. 345–352 (2000)
Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580–587 (2014)
Hart, S.G.: Nasa-task load index (nasa-tlx); 20 years later. In: Proceedings of the human factors and ergonomics society annual meeting. vol. 50, pp. 904–908. Sage publications Sage CA: Los Angeles, CA (2006)
Hassan, H.M., Galal-Edeen, G.H.: From usability to user experience. In: 2017 International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), pp. 216–222. IEEE (2017)
Hertzum, M., Borlund, P., Kristoffersen, K.B.: What do thinking-aloud participants say? a comparison of moderated and unmoderated usability sessions. Int. J. Human-Comput. Interact. 31(9), 557–570 (2015)
Holanda, Corey e Komogortsev, O.: Eye tracking em tablets comuns não modificados: desafios e soluções. In: Proceedings of the Symposium on Eye Tracking Research and Applications, pp. 277–280 (2012)
Holmqvist, K., Nyström, M., Mulvey, F.: Eye tracker data quality: What it is and how to measure it. In: Proceedings of the symposium on eye tracking research and applications, pp. 45–52 (2012)
Hwang, B.J., Chen, H.H., Hsieh, C.H., Huang, D.Y.: Gaze tracking based on concatenating spatial-temporal features. Sensors 22(2), 545 (2022). https://doi.org/10.3390/s22020545
Jain, A.K., Li, S.Z.: Handbook of face recognition, vol. 1. Springer (2011)
Jordan, P.W.: An introduction to usability. CRC Press (1998)
Kanade, P., David, F., Kanade, S.: Convolutional neural networks (CNN) based eye-gaze tracking system using machine learning algorithm. Europ. J. Electr. Eng. Comput. Sci. 5(2), 36–40 (2021)
Karlsson, H., Berglund, E., Larsson, J.: Method and apparatus for eye tracking (2014). https://patents.google.com/patent/US8723875B2/, the Eye Tribe Aps
Kassner, M., Patera, W., Bulling, A.: Pupil: An open source platform for pervasive eye tracking and mobile gaze-based interaction. In: Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunctpublication, pp. 1151–1160 (2014)
Kim, J.H., Jeong, J.W.: Gaze estimation in the dark with generative adversarial networks. In: ACM Symposium on Eye Tracking Research and Applications. ETRA ’20 Adjunct, Association for Computing Machinery, New York, NY, USA (2020). https://doi.org/10.1145/3379157.3391654
Krafka, K., et al.: Eye tracking for everyone. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2176–2184 (2016)
Kuling, E., et al.: Myeye: an open-source wearable gaze tracker. In: Proceedings of the Symposium on Eye Tracking Research and Applications, pp. 1–10. ACM (2019)
Lee, S.: Understanding face detection with the viola-jones object detection framework. Towards data science (2020)
Li, D., Babcock, J., Parkhurst, D.J.: Openeyes: a low-cost head-mounted eye-tracking solution. In: Proceedings of the 2006 symposium on Eye tracking research & applications, pp. 95–100 (2006)
Lim, J.Z., Mountstephens, J., Teo, J.: Emotion recognition using eye-tracking: taxonomy, review and current challenges. Sensors 20(8), 2384 (2020)
Madathil, K.C., Greenstein, J.S.: An investigation of the efficacy of collaborative virtual reality systems for moderated remote usability testing. Appl. Ergon. 65, 501–514 (2017)
Manhartsberger, M., Zellhofer, N.: Eye tracking in usability research: What users really see. In: Usability Symposium. vol. 198, pp. 141–152 (2005)
Martins, A.I., Queirós, A., Silva, A.G., Rocha, N.P.: Usability evaluation methods: a systematic review. Human Factors Softw. Develop. Design 250–273 (2015)
Nielsen, J.: Usability laboratories. Behav. Inform. Technol. 13(1–2), 3–8 (1994)
Nielsen, J., Pernice, K.: Eyetracking web usability. New Riders Publishing (2003)
Ooms, K., Dupont, L., Lapon, L., Popelka, S.: Accuracy and precision of fixation locations recorded with the low-cost eye tribe tracker in different experimental setups. J. Eye Movement Res. 8(1) (2015)
Palmero, C., Selva, J., Bagheri, M.A., Escalera, S.: Recurrent CNN for 3D gaze estimation using appearance and shape cues (2018)
Papoutsaki, A., Laskey, J., Huang, J.: Searchgazer: Webcam eye tracking for remote studies of web search. In: Proceedings of the 2017 Conference On Conference Human Information Interaction and Retrieval, pp. 17–26 (2017)
Pedregosa, F., et al.: Scikit-learn: Machine learning in python. J. Mach. Learn. Res. 12, 2825–2830 (2011)
Pernice, K., Nielsen, J.: How to conduct eyetracking studies. Nielsen Norman Group 945397498 (2009)
Poole, A., Ball, L.J.: Eye tracking in HCI and usability research. In: Encyclopedia of human computer interaction, pp. 211–219. IGI global (2006)
Sauro, J.: things to know about the single ease question (seq). Measuring U 2012 (2012)
Scholtz, J.: Adaptation of traditional usability testing methods for remote testing. In: Proceedings of the 34th Annual Hawaii International Conference on System Sciences, pp. 8-pp. IEEE (2001)
Shehu, I.S., Wang, Y., Athuman, A.M., Fu, X.: Remote eye gaze tracking research: a comparative evaluation on past and recent progress. Electronics 10(24), 3165 (2021)
Sjöberg, A., Rominger, M.: Beyond hand-eye coordination: An exploration of eye-tracking and speech recognition as a navigation tool for interactive systems (2015)
Skodras, E., Kanas, V.G., Fakotakis, N.: On visual gaze tracking based on a single low cost camera. Signal Process. Image Commun. 36, 29–42 (2015)
Sogo, H.: Gazeparser: an open-source and multiplatform library for low-cost eye tracking and analysis. Behav. Res. Methods 45, 684–695 (2013)
Thompson, K.E., Rozanski, E.P., Haake, A.R.: Here, there, anywhere: remote usability testing that works. In: Proceedings of the 5th Conference on Information Technology Education, pp. 132–137 (2004)
Venugopal, D., Amudha, J., Jyotsna, C.: Developing an application using eye tracker. In: 2016 IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), pp. 1518–1522. IEEE (2016)
Voßkühler, A., Nordmeier, V., Kuchinke, L., Jacobs, A.M.: Ogama (open gaze and mouse analyzer): open-source software designed to analyze eye and mouse movements in slideshow study designs. Behav. Res. Methods 40, 1150–1162 (2008)
Wisiecka, K., et al.: Comparison of webcam and remote eye tracking. In: 2022 Symposium on Eye Tracking Research and Applications, pp. 1–7 (2022)
Wood, E., Bulling, A.: Eyetab: Model-based gaze estimation on unmodified tablet computers. In: Proceedings of ETRA (2014). http://www.cl.cam.ac.uk/research/rainbow/projects/eyetab/
Zhang, M., Bulling, A.: Xlabs: A platform for rapid design, prototyping and evaluation of ubiquitous gaze interfaces. In: Proceedings of the Symposium on Eye Tracking Research and Applications, pp. 69–76. ACM (2018)
Zhao, W., Chellappa, R., Phillips, P.J., Rosenfeld, A.: Face recognition: a literature survey. ACM Comput. Surv. (CSUR) 35(4), 399–458 (2003)
Zielinski, P.: Opengazer: open-source gaze tracker for ordinary webcams. Samsung and The Gatsby Charitable Foundation. http://www.inference.phy.cam.ac.uk/opengazer (2007)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Capdevila, M.G., Rodrigues, K.A.P., Jardim, C.F., Silva, R.M. (2023). An Open Source Eye Gaze Tracker System to Perform Remote User Testing Evaluations. In: Naldi, M.C., Bianchi, R.A.C. (eds) Intelligent Systems. BRACIS 2023. Lecture Notes in Computer Science(), vol 14197. Springer, Cham. https://doi.org/10.1007/978-3-031-45392-2_13
Download citation
DOI: https://doi.org/10.1007/978-3-031-45392-2_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-45391-5
Online ISBN: 978-3-031-45392-2
eBook Packages: Computer ScienceComputer Science (R0)