Detalhes bibliográficos
Ano de defesa: |
2023 |
Autor(a) principal: |
Elmadjian, Carlos Eduardo Leão |
Orientador(a): |
Não Informado pela instituição |
Banca de defesa: |
Não Informado pela instituição |
Tipo de documento: |
Tese
|
Tipo de acesso: |
Acesso aberto |
Idioma: |
eng |
Instituição de defesa: |
Biblioteca Digitais de Teses e Dissertações da USP
|
Programa de Pós-Graduação: |
Não Informado pela instituição
|
Departamento: |
Não Informado pela instituição
|
País: |
Não Informado pela instituição
|
Palavras-chave em Português: |
|
Link de acesso: |
https://www.teses.usp.br/teses/disponiveis/45/45134/tde-28062023-204100/
|
Resumo: |
Human-computer interaction has been facing a significant shift from the traditional paradigm of WIMP interfaces using desktop PCs and notebooks to a more distributed and heterogeneous one, where interaction occurs with multiple devices, including mobile phones, smart watches, smart glasses, and IoT-enabled devices. As a result, traditional means of interaction may not provide a satisfactory user experience in such a scenario. In this thesis, we look into gaze-based interaction as an alternative to comply with this new paradigm and some of its fundamental challenges. In particular, we investigate new techniques to recognize eye movement patterns, 3D gaze estimation, and gaze-based interactive methods for micro-interaction scenarios. Though our ultimate goal is to enhance gaze interaction in wearable computing, we broke down this objective into several fronts, namely improving gaze estimation for 3D environments using wearable eye-tracking devices, enhancing real-time eye movement pattern recognition to support context-aware systems using deep temporal convolutional models and scale-invariant methods, and creating novel gaze interaction methods with a just-look-to-select mechanism, in spite of the Midas touch problem. Our results demonstrate that a 3D wearable gaze estimation procedure is feasible, though still very challenging; that our techniques for eye pattern recognition are able to achieve the best of both worlds: state-of-the-art accuracy while being computationally lightweight; and that our interactive techniques, such as GazeBar and V-Switch, have the potential to improve micro-interactions without the common safety mechanisms deployed to gaze-based methods. Though there are still significant obstacles to be addressed in future work, including thorough user studies and complex context detection strategies, this thesis shines new light towards gaze-based interaction and related challenges in this paradigm-shifting environment. |