Motor imagery (MI) is a fundamental brain-machine interface (BMI) paradigm in which users learn how to modulate their brain signals to voluntarily and selectively activate specific areas of the sensorimotor cortex. The self-paced nature of this kinesthetic imagination makes MI well-suited for several human-robot interaction (HRI) scenarios. However, the performance of MI decoding is highly dependent on both the user’s expertise and the quality of the decoding algorithm. To address these challenges, we propose a method that integrates environmental data from robotic sensors to enhance the MI decoding process. Preliminary experiments indicate that this approach improves decoding accuracy and the overall performance of the brain-driven system, opening new opportunities for research on how machines can enhance the usability of BMIs systems.

Enhancing Motor Imagery Decoding with Environmental Context During Robot Control

Stefano Tortora;Emanuele Menegatti;Luca Tonin
2025

Abstract

Motor imagery (MI) is a fundamental brain-machine interface (BMI) paradigm in which users learn how to modulate their brain signals to voluntarily and selectively activate specific areas of the sensorimotor cortex. The self-paced nature of this kinesthetic imagination makes MI well-suited for several human-robot interaction (HRI) scenarios. However, the performance of MI decoding is highly dependent on both the user’s expertise and the quality of the decoding algorithm. To address these challenges, we propose a method that integrates environmental data from robotic sensors to enhance the MI decoding process. Preliminary experiments indicate that this approach improves decoding accuracy and the overall performance of the brain-driven system, opening new opportunities for research on how machines can enhance the usability of BMIs systems.
2025
2025 IEEE International Conference on Systems, Man, and Cybernetics (SMC)
File in questo prodotto:
Non ci sono file associati a questo prodotto.
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11577/3590743
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact