Signal classification is a key of brain-computer interface (BCI). In this paper, we present a new method for classifying the electroencephalogram (EEG) signals of which the features are heterogeneous. This method is called wrapped elastic net feature selection and classification. Firstly, we used the joint application of time-domain statistic, power spectral density (PSD), common spatial pattern (CSP) and autoregressive (AR) model to extract high-dimensional fused features of the preprocessed EEG signals. Then we used the wrapped method for feature selection. We fitted the logistic regression model penalized with elastic net on the training data, and obtained the parameter estimation by coordinate descent method. Then we selected best feature subset by using 10-fold cross-validation. Finally, we classified the test sample using the trained model. Data used in the experiment were the EEG data from international BCI Competition Ⅳ. The results showed that the method proposed was suitable for fused feature selection with high-dimension. For identifying EEG signals, it is more effective and faster, and can single out a more relevant subset to obtain a relatively simple model. The average test accuracy reached 81.78%.
The task of automatic generation of medical image reports faces various challenges, such as diverse types of diseases and a lack of professionalism and fluency in report descriptions. To address these issues, this paper proposes a multimodal medical imaging report based on memory drive method (mMIRmd). Firstly, a hierarchical vision transformer using shifted windows (Swin-Transformer) is utilized to extract multi-perspective visual features of patient medical images, and semantic features of textual medical history information are extracted using bidirectional encoder representations from transformers (BERT). Subsequently, the visual and semantic features are integrated to enhance the model's ability to recognize different disease types. Furthermore, a medical text pre-trained word vector dictionary is employed to encode labels of visual features, thereby enhancing the professionalism of the generated reports. Finally, a memory driven module is introduced in the decoder, addressing long-distance dependencies in medical image data. This study is validated on the chest X-ray dataset collected at Indiana University (IU X-Ray) and the medical information mart for intensive care chest x-ray (MIMIC-CXR) released by the Massachusetts Institute of Technology and Massachusetts General Hospital. Experimental results indicate that the proposed method can better focus on the affected areas, improve the accuracy and fluency of report generation, and assist radiologists in quickly completing medical image report writing.
The result of the emotional state induced by music may provide theoretical support and help for assisted music therapy. The key to assessing the state of emotion is feature extraction of the emotional electroencephalogram (EEG). In this paper, we study the performance optimization of the feature extraction algorithm. A public multimodal database for emotion analysis using physiological signals (DEAP) proposed by Koelstra et al. was applied. Eight kinds of positive and negative emotions were extracted from the dataset, representing the data of fourteen channels from the different regions of brain. Based on wavelet transform, δ, θ, α and β rhythms were extracted. This paper analyzed and compared the performances of three kinds of EEG features for emotion classification, namely wavelet features (wavelet coefficients energy and wavelet entropy), approximate entropy and Hurst exponent. On this basis, an EEG feature fusion algorithm based on principal component analysis (PCA) was proposed. The principal component with a cumulative contribution rate more than 85% was retained, and the parameters which greatly varied in characteristic root were selected. The support vector machine was used to assess the state of emotion. The results showed that the average accuracy rates of emotional classification with wavelet features, approximate entropy and Hurst exponent were respectively 73.15%, 50.00% and 45.54%. By combining these three methods, the features fused with PCA possessed an accuracy of about 85%. The obtained classification accuracy by using the proposed fusion algorithm based on PCA was improved at least 12% than that by using single feature, providing assistance for emotional EEG feature extraction and music therapy.
As the most common active brain-computer interaction paradigm, motor imagery brain-computer interface (MI-BCI) suffers from the bottleneck problems of small instruction set and low accuracy, and its information transmission rate (ITR) and practical application are severely limited. In this study, we designed 6-class imagination actions, collected electroencephalogram (EEG) signals from 19 subjects, and studied the effect of collaborative brain-computer interface (cBCI) collaboration strategy on MI-BCI classification performance, the effects of changes in different group sizes and fusion strategies on group multi-classification performance are compared. The results showed that the most suitable group size was 4 people, and the best fusion strategy was decision fusion. In this condition, the classification accuracy of the group reached 77%, which was higher than that of the feature fusion strategy under the same group size (77.31% vs. 56.34%), and was significantly higher than that of the average single user (77.31% vs. 44.90%). The research in this paper proves that the cBCI collaboration strategy can effectively improve the MI-BCI classification performance, which lays the foundation for MI-cBCI research and its future application.
Diabetic retinopathy (DR) and its complication, diabetic macular edema (DME), are major causes of visual impairment and even blindness. The occurrence of DR and DME is pathologically interconnected, and their clinical diagnoses are closely related. Joint learning can help improve the accuracy of diagnosis. This paper proposed a novel adaptive lesion-aware fusion network (ALFNet) to facilitate the joint grading of DR and DME. ALFNet employed DenseNet-121 as the backbone and incorporated an adaptive lesion attention module (ALAM) to capture the distinct lesion characteristics of DR and DME. A deep feature fusion module (DFFM) with a shared-parameter local attention mechanism was designed to learn the correlation between the two diseases. Furthermore, a four-branch composite loss function was introduced to enhance the network’s multi-task learning capability. Experimental results demonstrated that ALFNet achieved superior joint grading performance on the Messidor dataset, with joint accuracy rates of 0.868 (DR 2 & DME 3), outperforming state-of-the-art methods. These results highlight the unique advantages of the proposed approach in the joint grading of DR and DME, thereby improving the efficiency and accuracy of clinical decision-making.
Remote photoplethysmography is susceptible to motion artifacts and individual physiological variations in complex environments. This paper proposes a remote heart rate estimation method based on frequency regulation and multi-scale spatio-temporal modeling. To address artifact noise issues, a frequency-regulated normalization module is designed to emphasize the dominant heart rate frequency while suppressing noise. To address the issue of individual physiological variations, the proposed method introduces a multi-level spatio-temporal feature fusion module to comprehensively capture physiological information through multi-scale convolutions and cross-layer integration. Subsequently, a dynamic weighting spatio-temporal feature module is introduced during spatio-temporal modeling to enhance long-term dependency modeling. Experimental results demonstrate that the proposed method achieves superior performance in cross-dataset evaluation. When trained on the PURE dataset and tested on the UBFC-rPPG dataset, the mean absolute error decreases from 1.31 to 1.28. Conversely, when trained on the UBFC-rPPG dataset and tested on the PURE dataset, the mean absolute error further decreases from 0.97 to 0.82. These results significantly outperform existing state-of-the-art methods, demonstrating the strong generalization capability and outstanding performance of our model across datasets. From the perspectives of frequency-regulated and multi-scale spatio-temporal modeling, this work enriches the modeling methodology for remote photoplethysmography pulse wave-based heart rate estimation, enhancing the stability and usability of remote heart rate estimation under complex interference and cross-scenario conditions.
In order to solve the current problems in medical equipment maintenance, this study proposed an intelligent fault diagnosis method for medical equipment based on long short term memory network(LSTM). Firstly, in the case of no circuit drawings and unknown circuit board signal direction, the symptom phenomenon and port electrical signal of 7 different fault categories were collected, and the feature coding, normalization, fusion and screening were preprocessed. Then, the intelligent fault diagnosis model was built based on LSTM, and the fused and screened multi-modal features were used to carry out the fault diagnosis classification and identification experiment. The results were compared with those using port electrical signal, symptom phenomenon and the fusion of the two types. In addition, the fault diagnosis algorithm was compared with BP neural network (BPNN), recurrent neural network (RNN) and convolution neural network (CNN). The results show that based on the fused and screened multi-modal features, the average classification accuracy of LSTM algorithm model reaches 0.970 9, which is higher than that of using port electrical signal alone, symptom phenomenon alone or the fusion of the two types. It also has higher accuracy than BPNN, RNN and CNN, which provides a relatively feasible new idea for intelligent fault diagnosis of similar equipment.
The recurrent neural network architecture improves the processing ability of time-series data. However, issues such as exploding gradients and poor feature extraction limit its application in the automatic diagnosis of mild cognitive impairment (MCI). This paper proposed a research approach for building an MCI diagnostic model using a Bayesian-optimized bidirectional long short-term memory network (BO-BiLSTM) to address this problem. The diagnostic model was based on a Bayesian algorithm and combined prior distribution and posterior probability results to optimize the BO-BiLSTM network hyperparameters. It also used multiple feature quantities that fully reflected the cognitive state of the MCI brain, such as power spectral density, fuzzy entropy, and multifractal spectrum, as the input of the diagnostic model to achieve automatic MCI diagnosis. The results showed that the feature-fused Bayesian-optimized BiLSTM network model achieved an MCI diagnostic accuracy of 98.64% and effectively completed the diagnostic assessment of MCI. In conclusion, based on this optimization, the long short-term neural network model has achieved automatic diagnostic assessment of MCI, providing a new diagnostic model for intelligent diagnosis of MCI.
Existing emotion recognition research is typically limited to static laboratory settings and has not fully handle the changes in emotional states in dynamic scenarios. To address this problem, this paper proposes a method for dynamic continuous emotion recognition based on electroencephalography (EEG) and eye movement signals. Firstly, an experimental paradigm was designed to cover six dynamic emotion transition scenarios including happy to calm, calm to happy, sad to calm, calm to sad, nervous to calm, and calm to nervous. EEG and eye movement data were collected simultaneously from 20 subjects to fill the gap in current multimodal dynamic continuous emotion datasets. In the valence-arousal two-dimensional space, emotion ratings for stimulus videos were performed every five seconds on a scale of 1 to 9, and dynamic continuous emotion labels were normalized. Subsequently, frequency band features were extracted from the preprocessed EEG and eye movement data. A cascade feature fusion approach was used to effectively combine EEG and eye movement features, generating an information-rich multimodal feature vector. This feature vector was input into four regression models including support vector regression with radial basis function kernel, decision tree, random forest, and K-nearest neighbors, to develop the dynamic continuous emotion recognition model. The results showed that the proposed method achieved the lowest mean square error for valence and arousal across the six dynamic continuous emotions. This approach can accurately recognize various emotion transitions in dynamic situations, offering higher accuracy and robustness compared to using either EEG or eye movement signals alone, making it well-suited for practical applications.
In the process of lower limb rehabilitation training, fatigue estimation is of great significance to improve the accuracy of intention recognition and avoid secondary injury. However, most of the existing methods only consider surface electromyography (sEMG) features but ignore electrocardiogram (ECG) features when performing in fatigue estimation, which leads to the low and unstable recognition efficiency. Aiming at this problem, a method that uses the fusion features of ECG and sEMG signal to estimate the fatigue during lower limb rehabilitation was proposed, and an improved particle swarm optimization-support vector machine classifier (improved PSO-SVM) was proposed and used to identify the fusion feature vector. Finally, the accurate recognition of the three states of relax, transition and fatigue was achieved, and the recognition rates were 98.5%, 93.5%, and 95.5%, respectively. Comparative experiments showed that the average recognition rate of this method was 4.50% higher than that of sEMG features alone, and 13.66% higher than that of the combined features of ECG and sEMG without feature fusion. It is proved that the feature fusion of ECG and sEMG signals in the process of lower limb rehabilitation training can be used for recognizing fatigue more accurately.