Multi-Armed Bandits in Brain-Computer Interfaces

Frida Heskebeck, Carolina Bergeling, Bo Bernhardsson

Research output: Contribution to journalReview articlepeer-review

Abstract

The multi-armed bandit (MAB) problem models a decision-maker that optimizes its actions based on current and acquired new knowledge to maximize its reward. This type of online decision is prominent in many procedures of Brain-Computer Interfaces (BCIs) and MAB has previously been used to investigate, e.g., what mental commands to use to optimize BCI performance. However, MAB optimization in the context of BCI is still relatively unexplored, even though it has the potential to improve BCI performance during both calibration and real-time implementation. Therefore, this review aims to further describe the fruitful area of MABs to the BCI community. The review includes a background on MAB problems and standard solution methods, and interpretations related to BCI systems. Moreover, it includes state-of-the-art concepts of MAB in BCI and suggestions for future research.

Original languageEnglish
Article number931085
JournalFrontiers in Human Neuroscience
Volume16
DOIs
Publication statusPublished - 2022 Jul 5

Bibliographical note

Funding Information:
This work was partially supported by the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation. All authors are also members of the ELLIIT Strategic Research Area.

Publisher Copyright:
Copyright © 2022 Heskebeck, Bergeling and Bernhardsson.

Subject classification (UKÄ)

  • Control Engineering

Free keywords

  • Brain-Computer Interface (BCI)
  • calibration
  • multi-armed bandit (MAB)
  • real-time optimization
  • reinforcement learning

Fingerprint

Dive into the research topics of 'Multi-Armed Bandits in Brain-Computer Interfaces'. Together they form a unique fingerprint.

Cite this