Visual Input Affects the Decoding of Imagined Movements of the Same Limb

Research output: Contribution to conferencePosterResearchpeer-review

Abstract

A better understanding how movements are encoded in electroencephalography (EEG) signals is required to develop a more natural control for motor neuroprostheses. We decoded imagined hand close and supination movements from seven healthy subjects and investigated the influence of the visual input. We found that motor imagination of these movements can be decoded from low-frequency time-domain EEG signals with a maximum average classification accuracy of 57.3 +/- 5.0%. The simultaneous observation of congruent hand movements increased the classification accuracy to 64.1 +/- 8.3%. Furthermore, the sole observation of hand movements yielded discriminable brain patterns (61.9 +/- 5.5%). These findings show that for low-frequency time-domain EEG signals, the type of visual input during classifier training affects the performance and has to be considered in future studies.
Original languageEnglish
Publication statusPublished - 18 Sep 2017
Event7th Graz BCI Conference 2017: From Vision to Reality - Graz, Austria
Duration: 18 Sep 201722 Sep 2017

Conference

Conference7th Graz BCI Conference 2017
CountryAustria
CityGraz
Period18/09/1722/09/17

Fields of Expertise

  • Human- & Biotechnology

Treatment code (Nähere Zuordnung)

  • Basic - Fundamental (Grundlagenforschung)

Fingerprint Dive into the research topics of 'Visual Input Affects the Decoding of Imagined Movements of the Same Limb'. Together they form a unique fingerprint.

  • Cite this

    Ofner, P., Kersch, P., & Müller-Putz, G. (2017). Visual Input Affects the Decoding of Imagined Movements of the Same Limb. Poster session presented at 7th Graz BCI Conference 2017, Graz, Austria.