Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/109431
Title: Using machine learning to predict epileptic seizures from EEG data
Authors: Cauchi, Jonathan
Garg, Lalit
Garg, Gaurav
Keywords: Electroencephalography -- Data processing
Machine learning -- Case studies
Epileptics -- Care
Convulsions -- Forecasting
Artificial intelligence
Issue Date: 2022
Publisher: SSRN
Citation: Cauchi, J., Garg, L., & Garg, G. (2022). Using Machine Learning to Predict Epileptic Seizures from EEG Data. The Lancet, available at http://dx.doi.org/10.2139/ssrn.4157506
Abstract: With the recent Artificial Intelligence (AI) progress, especially Machine Learning (ML), researchers aim to apply techniques for improving and automating certain clinical practice facets. This paper proposes and investigates the efficiency of several machine learning approaches to predict epileptic seizure onsets accurately to prepare patients for recurrent convulsion episodes, enhancing their quality of life. The study was performed in collaboration with industry and analyses the non-invasive scalp Electroencephalography (EEG) signals. The feature space is extracted using statistical, and wavelet transforms. The results from K-Nearest Neighbour (KNN), Support Vector Machines (SVM), and an Ensemble Classifier are compared. The proposed techniques are evaluated using one of the most extensive seizure EEG datasets, the CHB-MIT dataset, which includes 192 seizure readings from 22 patients suffering from intractable seizures. It is among the first seizure prediction studies and shows that the three methods perform similar, although the Ensemble Classifier achieves a higher specificity, sensitivity and accuracy.
URI: https://www.um.edu.mt/library/oar/handle/123456789/109431
Appears in Collections:Scholarly Works - FacICTCIS

Files in This Item:
File Description SizeFormat 
Using machine learning to predict epileptic seizures from EEG data 2022.pdf650.37 kBAdobe PDFView/Open


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.