Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/144555
Title: Imagined speech decoding by learning consensus graph from RKHS-based multi-view EEG features
Authors: Zhao, Zhenye
Peng, Yong
Camilleri, Kenneth P.
Kong, Wanzeng
Cichocki, Andrzej
Keywords: Electroencephalography
Brain -- Diseases -- Diagnosis
Brain-computer interfaces
Speech disorders -- Patients -- Means of communication
Speech processing systems
Issue Date: 2025
Publisher: Institute of Electrical and Electronics Engineers
Citation: Zhao, Z., Peng, Y., Camilleri, K. P., Kong, W., & Cichocki, A. (2025). Imagined Speech Decoding by Learning Consensus Graph From RKHS-Based Multi-View EEG Features. IEEE Signal Processing Letters, 32, 3944-3948.
Abstract: Electroencephalogram (EEG)-based speech imagery has emerged as a novel brain–computer interface (BCI) paradigm, which holds promise for aiding individuals with speech disorders to achieve more intuitive communication; however, the imagined speech decoding performance is still far from the level required for real applications. Among the possible reasons, one is that there is still no agreement on which domain features are more closely related to imagined speech and the other is that EEG features in the original space may be not be sufficiently discriminative for decoding. To both ends, we propose to learn a Reproducible Kernel Hilbert Space (RKHS)-based consensus Graph from Multi-View (KGMV) EEG features for imagined speech decoding. Specifically, the complementary information from multiple-domain EEG features are sufficiently explored to avoid the information loss caused by single-domain features; besides, the nonlinear modeling capability of KGMV in terms of EEG data is enhanced by implicitly mapping multi-view EEG features into RKHS. Accordingly, a consensus graph is adaptively learned in the kernel space to capture the semantic connections of EEG samples. Experiments are conducted on a speech imagery EEG data set and the results depict that the proposed KGMV model demonstrates superior decoding performance compared to single-view models and several other state-of-the-art approaches.
URI: https://www.um.edu.mt/library/oar/handle/123456789/144555
Appears in Collections:Scholarly Works - FacEngSCE

Files in This Item:
File Description SizeFormat 
Imagined_speech_decoding_by_learning_consensus_graph_from_RKHS_based_multi-view_EEG_features_2025.pdf
  Restricted Access
750.75 kBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.