Menu
Location: Home page > Research > Research
PUMCH Neurology Team Developed AI-based Multimodal Interictal Epileptiform Discharge Detection Method
CopyFrom: PUMCH UpdateTime: 2024-05-21 Font Size: SmallBig

Recently, the EEG and neurophysiology team under the Department of Neurology, PUMCH, in collaboration with the Technology Department, NetEase Media Group, published their original findings on an automatic detection model of interictal epileptiform discharge (IED) in the journal Neural Networks (a tier 1 journal, or among the top 5%, as ranked by the Chinese Academy of Sciences). The study proposed an AI-powered multimodal detection method, vEpiNet, that innovatively incorporates video data into IED detection. Compared with traditional models, vEpiNet demonstrated higher accuracy and efficiency. The research was supported by the National High-Level Hospital Clinical Research Funding and the Key Research and Development Program under the Ministry of Science and Technology.



Epilepsy is a common neurological disorder, and electroencephalogram (EEG) is a crucial tool for its diagnosis. IED, as a characteristic indicator, is essential for the diagnosis, classification, and medication management of epilepsy. Currently, the detection of IED relies heavily on the interpretation and manual annotation by epilepsy specialists, which is time-consuming and subjective. AI can improve EEG interpretation efficiency and reduce human error. In recent years, AI-based capabilities for IED detection and analysis have rapidly evolved. Most existing models at home and abroad, though, rely solely on EEG data and have low specificity.

The employed datasets comprise 24,931 IED video-EEG segments from patients at the PUMCH Epilepsy Center. EEG data were processed through short-time Fourier transform and the EfficientNetV2-S for electrophysiological feature extraction. Video data were first processed by a patient detection model YOLOv5-patient for target localization to minimize environmental and bystander interference. Then the frame difference algorithm was used to discern the body movements and address issues of body covered, and key point detection for subtle facial movements, transforming full-body and facial movements into video features. This approach significantly reduces video data volume, and ensures accurate feature extraction and fast processing speed. Finally, the video and EEG features were fused via a multilayer perceptron, and the model was named vEpiNet.



▲Illustration of the multimodal IED detection model

Results show that this model achieves high levels of sensitivity, specificity, and accuracy, with a lower per-minute false positive rate than previous studies. Video features significantly enhance accuracy and reduce judgement errors, particularly in real clinical settings.



▲Testing results of vEpiNet and nEpiNet which relies on EEG only

The model can be embedded in a compatible EEG machine operating system, where analyzed video EEG data is directly displayed with an annotated detection line on the EEG machine’s playback software. Currently, processing one hour’s video EEG data takes only 5.7 minutes.


▲Detection example: The vertical line is the software detection result, indicating the occurrence of IED within a two-second window before and after (as shown by the red dashed lines)

The integration of AI plays a crucial role in standardizing and normalizing EEG reports to make them more consistent. This study validates the effectiveness and feasibility of using a multimodal approach in the automatic detection of IEDs, providing a new direction for the application of AI in EEG analysis.


Translated by Liu Haiyan

Reviewed by Yang Xunzhe and Wang Yao