west china medical publishers
Keyword
  • Title
  • Author
  • Keyword
  • Abstract
Advance search
Advance search

Search

find Keyword "feature fusion" 13 results
  • Psychosis speech recognition algorithm based on deep embedded sparse stacked autoencoder and manifold ensemble

    Speech feature learning is the core and key of speech recognition method for mental illness. Deep feature learning can automatically extract speech features, but it is limited by the problem of small samples. Traditional feature extraction (original features) can avoid the impact of small samples, but it relies heavily on experience and is poorly adaptive. To solve this problem, this paper proposes a deep embedded hybrid feature sparse stack autoencoder manifold ensemble algorithm. Firstly, based on the prior knowledge, the psychotic speech features are extracted, and the original features are constructed. Secondly, the original features are embedded in the sparse stack autoencoder (deep network), and the output of the hidden layer is filtered to enhance the complementarity between the deep features and the original features. Third, the L1 regularization feature selection mechanism is designed to compress the dimensions of the mixed feature set composed of deep features and original features. Finally, a weighted local preserving projection algorithm and an ensemble learning mechanism are designed, and a manifold projection classifier ensemble model is constructed, which further improves the classification stability of feature fusion under small samples. In addition, this paper designs a medium-to-large-scale psychotic speech collection program for the first time, collects and constructs a large-scale Chinese psychotic speech database for the verification of psychotic speech recognition algorithms. The experimental results show that the main innovation of the algorithm is effective, and the classification accuracy is better than other representative algorithms, and the maximum improvement is 3.3%. In conclusion, this paper proposes a new method of psychotic speech recognition based on embedded mixed sparse stack autoencoder and manifold ensemble, which effectively improves the recognition rate of psychotic speech.

    Release date:2021-10-22 02:07 Export PDF Favorites Scan
  • Research progress and prospect of collaborative brain-computer interface for group brain collaboration

    As the most common active brain-computer interaction paradigm, motor imagery brain-computer interface (MI-BCI) suffers from the bottleneck problems of small instruction set and low accuracy, and its information transmission rate (ITR) and practical application are severely limited. In this study, we designed 6-class imagination actions, collected electroencephalogram (EEG) signals from 19 subjects, and studied the effect of collaborative brain-computer interface (cBCI) collaboration strategy on MI-BCI classification performance, the effects of changes in different group sizes and fusion strategies on group multi-classification performance are compared. The results showed that the most suitable group size was 4 people, and the best fusion strategy was decision fusion. In this condition, the classification accuracy of the group reached 77%, which was higher than that of the feature fusion strategy under the same group size (77.31% vs. 56.34%), and was significantly higher than that of the average single user (77.31% vs. 44.90%). The research in this paper proves that the cBCI collaboration strategy can effectively improve the MI-BCI classification performance, which lays the foundation for MI-cBCI research and its future application.

    Release date:2021-08-16 04:59 Export PDF Favorites Scan
  • Selection and Classification of Elastic Net Feature with Fused Electroencephalogram Features

    Signal classification is a key of brain-computer interface (BCI). In this paper, we present a new method for classifying the electroencephalogram (EEG) signals of which the features are heterogeneous. This method is called wrapped elastic net feature selection and classification. Firstly, we used the joint application of time-domain statistic, power spectral density (PSD), common spatial pattern (CSP) and autoregressive (AR) model to extract high-dimensional fused features of the preprocessed EEG signals. Then we used the wrapped method for feature selection. We fitted the logistic regression model penalized with elastic net on the training data, and obtained the parameter estimation by coordinate descent method. Then we selected best feature subset by using 10-fold cross-validation. Finally, we classified the test sample using the trained model. Data used in the experiment were the EEG data from international BCI Competition Ⅳ. The results showed that the method proposed was suitable for fused feature selection with high-dimension. For identifying EEG signals, it is more effective and faster, and can single out a more relevant subset to obtain a relatively simple model. The average test accuracy reached 81.78%.

    Release date:2017-01-17 06:17 Export PDF Favorites Scan
  • Intelligent fault diagnosis of medical equipment based on long short term memory network

    In order to solve the current problems in medical equipment maintenance, this study proposed an intelligent fault diagnosis method for medical equipment based on long short term memory network(LSTM). Firstly, in the case of no circuit drawings and unknown circuit board signal direction, the symptom phenomenon and port electrical signal of 7 different fault categories were collected, and the feature coding, normalization, fusion and screening were preprocessed. Then, the intelligent fault diagnosis model was built based on LSTM, and the fused and screened multi-modal features were used to carry out the fault diagnosis classification and identification experiment. The results were compared with those using port electrical signal, symptom phenomenon and the fusion of the two types. In addition, the fault diagnosis algorithm was compared with BP neural network (BPNN), recurrent neural network (RNN) and convolution neural network (CNN). The results show that based on the fused and screened multi-modal features, the average classification accuracy of LSTM algorithm model reaches 0.970 9, which is higher than that of using port electrical signal alone, symptom phenomenon alone or the fusion of the two types. It also has higher accuracy than BPNN, RNN and CNN, which provides a relatively feasible new idea for intelligent fault diagnosis of similar equipment.

    Release date:2021-06-18 04:50 Export PDF Favorites Scan
  • Dynamic continuous emotion recognition method based on electroencephalography and eye movement signals

    Existing emotion recognition research is typically limited to static laboratory settings and has not fully handle the changes in emotional states in dynamic scenarios. To address this problem, this paper proposes a method for dynamic continuous emotion recognition based on electroencephalography (EEG) and eye movement signals. Firstly, an experimental paradigm was designed to cover six dynamic emotion transition scenarios including happy to calm, calm to happy, sad to calm, calm to sad, nervous to calm, and calm to nervous. EEG and eye movement data were collected simultaneously from 20 subjects to fill the gap in current multimodal dynamic continuous emotion datasets. In the valence-arousal two-dimensional space, emotion ratings for stimulus videos were performed every five seconds on a scale of 1 to 9, and dynamic continuous emotion labels were normalized. Subsequently, frequency band features were extracted from the preprocessed EEG and eye movement data. A cascade feature fusion approach was used to effectively combine EEG and eye movement features, generating an information-rich multimodal feature vector. This feature vector was input into four regression models including support vector regression with radial basis function kernel, decision tree, random forest, and K-nearest neighbors, to develop the dynamic continuous emotion recognition model. The results showed that the proposed method achieved the lowest mean square error for valence and arousal across the six dynamic continuous emotions. This approach can accurately recognize various emotion transitions in dynamic situations, offering higher accuracy and robustness compared to using either EEG or eye movement signals alone, making it well-suited for practical applications.

    Release date: Export PDF Favorites Scan
  • A heart sound segmentation method based on multi-feature fusion network

    Objective To propose a heart sound segmentation method based on multi-feature fusion network. Methods Data were obtained from the CinC/PhysioNet 2016 Challenge dataset (a total of 3 153 recordings from 764 patients, about 91.93% of whom were male, with an average age of 30.36 years). Firstly the features were extracted in time domain and time-frequency domain respectively, and reduced redundant features by feature dimensionality reduction. Then, we selected optimal features separately from the two feature spaces that performed best through feature selection. Next, the multi-feature fusion was completed through multi-scale dilated convolution, cooperative fusion, and channel attention mechanism. Finally, the fused features were fed into a bidirectional gated recurrent unit (BiGRU) network to heart sound segmentation results. Results The proposed method achieved precision, recall and F1 score of 96.70%, 96.99%, and 96.84% respectively. Conclusion The multi-feature fusion network proposed in this study has better heart sound segmentation performance, which can provide high-accuracy heart sound segmentation technology support for the design of automatic analysis of heart diseases based on heart sounds.

    Release date: Export PDF Favorites Scan
  • Breast cancer lesion segmentation based on co-learning feature fusion and Transformer

    The PET/CT imaging technology combining positron emission tomography (PET) and computed tomography (CT) is the most advanced imaging examination method currently, and is mainly used for tumor screening, differential diagnosis of benign and malignant tumors, staging and grading. This paper proposes a method for breast cancer lesion segmentation based on PET/CT bimodal images, and designs a dual-path U-Net framework, which mainly includes three modules: encoder module, feature fusion module and decoder module. Among them, the encoder module uses traditional convolution for feature extraction of single mode image; The feature fusion module adopts collaborative learning feature fusion technology and uses Transformer to extract the global features of the fusion image; The decoder module mainly uses multi-layer perceptron to achieve lesion segmentation. This experiment uses actual clinical PET/CT data to evaluate the effectiveness of the algorithm. The experimental results show that the accuracy, recall and accuracy of breast cancer lesion segmentation are 95.67%, 97.58% and 96.16%, respectively, which are better than the baseline algorithm. Therefore, it proves the rationality of the single and bimodal feature extraction method combining convolution and Transformer in the experimental design of this article, and provides reference for feature extraction methods for tasks such as multimodal medical image segmentation or classification.

    Release date:2024-04-24 09:50 Export PDF Favorites Scan
  • Multi-task motor imagery electroencephalogram classification based on adaptive time-frequency common spatial pattern combined with convolutional neural network

    The effective classification of multi-task motor imagery electroencephalogram (EEG) is helpful to achieve accurate multi-dimensional human-computer interaction, and the high frequency domain specificity between subjects can improve the classification accuracy and robustness. Therefore, this paper proposed a multi-task EEG signal classification method based on adaptive time-frequency common spatial pattern (CSP) combined with convolutional neural network (CNN). The characteristics of subjects' personalized rhythm were extracted by adaptive spectrum awareness, and the spatial characteristics were calculated by using the one-versus-rest CSP, and then the composite time-domain characteristics were characterized to construct the spatial-temporal frequency multi-level fusion features. Finally, the CNN was used to perform high-precision and high-robust four-task classification. The algorithm in this paper was verified by the self-test dataset containing 10 subjects (33 ± 3 years old, inexperienced) and the dataset of the 4th 2018 Brain-Computer Interface Competition (BCI competition Ⅳ-2a). The average accuracy of the proposed algorithm for the four-task classification reached 93.96% and 84.04%, respectively. Compared with other advanced algorithms, the average classification accuracy of the proposed algorithm was significantly improved, and the accuracy range error between subjects was significantly reduced in the public dataset. The results show that the proposed algorithm has good performance in multi-task classification, and can effectively improve the classification accuracy and robustness.

    Release date:2023-02-24 06:14 Export PDF Favorites Scan
  • An automatic pulmonary nodules detection algorithm with multi-scale information fusion

    Lung nodules are the main manifestation of early lung cancer. So accurate detection of lung nodules is of great significance for early diagnosis and treatment of lung cancer. However, the rapid and accurate detection of pulmonary nodules is a challenging task due to the complex background, large detection range of pulmonary computed tomography (CT) images and the different sizes and shapes of pulmonary nodules. Therefore, this paper proposes a multi-scale feature fusion algorithm for the automatic detection of pulmonary nodules to achieve accurate detection of pulmonary nodules. Firstly, a three-layer modular lung nodule detection model was designed on the deep convolutional network (VGG16) for large-scale image recognition. The first-tier module of the network is used to extract the features of pulmonary nodules in CT images and roughly estimate the location of pulmonary nodules. Then the second-tier module of the network is used to fuse multi-scale image features to further enhance the details of pulmonary nodules. The third-tier module of the network was fused to analyze the features of the first-tier and the second-tier module of the network, and the candidate box of pulmonary nodules in multi-scale was obtained. Finally, the candidate box of pulmonary nodules under multi-scale was analyzed with the method of non-maximum suppression, and the final location of pulmonary nodules was obtained. The algorithm is validated by the data of pulmonary nodules on LIDC-IDRI common data set. The average detection accuracy is 90.9%.

    Release date:2020-08-21 07:07 Export PDF Favorites Scan
  • Research on mild cognitive impairment diagnosis based on Bayesian optimized long-short-term neural network model

    The recurrent neural network architecture improves the processing ability of time-series data. However, issues such as exploding gradients and poor feature extraction limit its application in the automatic diagnosis of mild cognitive impairment (MCI). This paper proposed a research approach for building an MCI diagnostic model using a Bayesian-optimized bidirectional long short-term memory network (BO-BiLSTM) to address this problem. The diagnostic model was based on a Bayesian algorithm and combined prior distribution and posterior probability results to optimize the BO-BiLSTM network hyperparameters. It also used multiple feature quantities that fully reflected the cognitive state of the MCI brain, such as power spectral density, fuzzy entropy, and multifractal spectrum, as the input of the diagnostic model to achieve automatic MCI diagnosis. The results showed that the feature-fused Bayesian-optimized BiLSTM network model achieved an MCI diagnostic accuracy of 98.64% and effectively completed the diagnostic assessment of MCI. In conclusion, based on this optimization, the long short-term neural network model has achieved automatic diagnostic assessment of MCI, providing a new diagnostic model for intelligent diagnosis of MCI.

    Release date:2023-08-23 02:45 Export PDF Favorites Scan
2 pages Previous 1 2 Next

Format

Content