###
计算机系统应用英文版:2022,31(2):298-304
本文二维码信息
码上扫一扫!
基于SVM-KNN算法的情绪脑电识别
(曲阜师范大学 工学院, 日照 276826)
Emotion Classification Using EEG Signals Based on SVM-KNN Algorithm
(College of Engineering, Qufu Normal University, Rizhao 276826, China)
摘要
图/表
参考文献
相似文献
本文已被:浏览 838次   下载 1556
Received:April 26, 2021    Revised:May 19, 2021
中文摘要: 情绪识别与日常生活的诸多领域都有很大联系. 然而, 通过单一算法难以获得较高的情绪识别准确率, 为此, 提出一种基于支持向量机(support vector machine, SVM)和K近邻(K-nearest neighbors, KNN)融合算法(SVM-KNN)的情绪脑电识别模型. 在情绪分类时, 首先计算待识别样本与最优分类超平面的空间距离, 若两者距离大于提前设定的阈值, 选用SVM分类器对情绪样本分类, 否则选用KNN分类器. 最后在SEED情感数据集上进行实验测试, 通过对比实验, 得出SVM-KNN算法提高了情绪三分类的准确率. 运用该模型可有效地对情绪类型进行识别, 对于医疗护理方面获取表达障碍患者的情绪状态有积极意义.
Abstract:Emotion recognition is closely related to many facets of our daily lives. However, it is difficult to achieve a satisfying emotion recognition rate by using one single algorithm. Therefore, this study puts forward an emotion recognition model based on electroencephalogram (EEG) with a fusion algorithm that combines the support vector machine (SVM) algorithm with the K-nearest neighbors algorithm (SVM-KNN). In the emotion classification process, the spatial distance between the sample to be identified and the optimal classification hyperplane is calculated. If it is longer than the preset threshold, the SVM classifier is chosen to classify the emotion records. Otherwise, the KNN classifier is chosen. Finally, experiments are carried out on the SJTU emotion EEG dataset (SEED). The comparative experiments show that the SVM-KNN algorithm improves the accuracy of the three-emotion classification. This model can effectively identify the types of emotions and thus has positive significance in obtaining the emotions of patients with expression disorders in medical care.
文章编号:     中图分类号:    文献标志码:
基金项目:山东省科技厅重大创新工程(2019JZZY011111);全国大学生创新训练项目(S202010446028)
引用文本:
滕凯迪,赵倩,谭浩然,郑金和,董宜先,单洪芳.基于SVM-KNN算法的情绪脑电识别.计算机系统应用,2022,31(2):298-304
TENG Kai-Di,ZHAO Qian,TAN Hao-Ran,ZHENG Jin-He,DONG Yi-Xian,SHAN Hong-Fang.Emotion Classification Using EEG Signals Based on SVM-KNN Algorithm.COMPUTER SYSTEMS APPLICATIONS,2022,31(2):298-304