Magnetic resonance imaging(MRI) can obtain multi-modal images with different contrast, which provides rich information for clinical diagnosis. However, some contrast images are not scanned or the quality of the acquired images cannot meet the diagnostic requirements due to the difficulty of patient's cooperation or the limitation of scanning conditions. Image synthesis techniques have become a method to compensate for such image deficiencies. In recent years, deep learning has been widely used in the field of MRI synthesis. In this paper, a synthesis network based on multi-modal fusion is proposed, which firstly uses a feature encoder to encode the features of multiple unimodal images separately, and then fuses the features of different modal images through a feature fusion module, and finally generates the target modal image. The similarity measure between the target image and the predicted image in the network is improved by introducing a dynamic weighted combined loss function based on the spatial domain and K-space domain. After experimental validation and quantitative comparison, the multi-modal fusion deep learning network proposed in this paper can effectively synthesize high-quality MRI fluid-attenuated inversion recovery (FLAIR) images. In summary, the method proposed in this paper can reduce MRI scanning time of the patient, as well as solve the clinical problem of missing FLAIR images or image quality that is difficult to meet diagnostic requirements.
In the extraction of fetal electrocardiogram (ECG) signal, due to the unicity of the scale of the U-Net same-level convolution encoder, the size and shape difference of the ECG characteristic wave between mother and fetus are ignored, and the time information of ECG signals is not used in the threshold learning process of the encoder’s residual shrinkage module. In this paper, a method of extracting fetal ECG signal based on multi-scale residual shrinkage U-Net model is proposed. First, the Inception and time domain attention were introduced into the residual shrinkage module to enhance the multi-scale feature extraction ability of the same level convolution encoder and the utilization of the time domain information of fetal ECG signal. In order to maintain more local details of ECG waveform, the maximum pooling in U-Net was replaced by Softpool. Finally, the decoder composed of the residual module and up-sampling gradually generated fetal ECG signals. In this paper, clinical ECG signals were used for experiments. The final results showed that compared with other fetal ECG extraction algorithms, the method proposed in this paper could extract clearer fetal ECG signals. The sensitivity, positive predictive value, and F1 scores in the 2013 competition data set reached 93.33%, 99.36%, and 96.09%, respectively, indicating that this method can effectively extract fetal ECG signals and has certain application values for perinatal fetal health monitoring.
The diagnosis of hypertrophic cardiomyopathy (HCM) is of great significance for the early risk classification of sudden cardiac death and the screening of family genetic diseases. This research proposed a HCM automatic detection method based on convolution neural network (CNN) model, using single-lead electrocardiogram (ECG) signal as the research object. Firstly, the R-wave peak locations of single-lead ECG signal were determined, followed by the ECG signal segmentation and resample in units of heart beats, then a CNN model was built to automatically extract the deep features in the ECG signal and perform automatic classification and HCM detection. The experimental data is derived from 108 ECG records extracted from three public databases provided by PhysioNet, the database established in this research consists of 14,459 heartbeats, and each heartbeat contains 128 sampling points. The results revealed that the optimized CNN model could effectively detect HCM, the accuracy, sensitivity and specificity were 95.98%, 98.03% and 95.79% respectively. In this research, the deep learning method was introduced for the analysis of single-lead ECG of HCM patients, which could not only overcome the technical limitations of conventional detection methods based on multi-lead ECG, but also has important application value for assisting doctor in fast and convenient large-scale HCM preliminary screening.
In recent years, epileptic seizure detection based on electroencephalogram (EEG) has attracted the widespread attention of the academic. However, it is difficult to collect data from epileptic seizure, and it is easy to cause over fitting phenomenon under the condition of few training data. In order to solve this problem, this paper took the CHB-MIT epilepsy EEG dataset from Boston Children's Hospital as the research object, and applied wavelet transform for data augmentation by setting different wavelet transform scale factors. In addition, by combining deep learning, ensemble learning, transfer learning and other methods, an epilepsy detection method with high accuracy for specific epilepsy patients was proposed under the condition of insufficient learning samples. In test, the wavelet transform scale factors 2, 4 and 8 were set for experimental comparison and verification. When the wavelet scale factor was 8, the average accuracy, average sensitivity and average specificity was 95.47%, 93.89% and 96.48%, respectively. Through comparative experiments with recent relevant literatures, the advantages of the proposed method were verified. Our results might provide reference for the clinical application of epilepsy detection.
[Abstract]Automatic and accurate segmentation of lung parenchyma is essential for assisted diagnosis of lung cancer. In recent years, researchers in the field of deep learning have proposed a number of improved lung parenchyma segmentation methods based on U-Net. However, the existing segmentation methods ignore the complementary fusion of semantic information in the feature map between different layers and fail to distinguish the importance of different spaces and channels in the feature map. To solve this problem, this paper proposes the double scale parallel attention (DSPA) network (DSPA-Net) architecture, and introduces the DSPA module and the atrous spatial pyramid pooling (ASPP) module in the “encoder-decoder” structure. Among them, the DSPA module aggregates the semantic information of feature maps of different levels while obtaining accurate space and channel information of feature map with the help of cooperative attention (CA). The ASPP module uses multiple parallel convolution kernels with different void rates to obtain feature maps containing multi-scale information under different receptive fields. The two modules address multi-scale information processing in feature maps of different levels and in feature maps of the same level, respectively. We conducted experimental verification on the Kaggle competition dataset. The experimental results prove that the network architecture has obvious advantages compared with the current mainstream segmentation network. The values of dice similarity coefficient (DSC) and intersection on union (IoU) reached 0.972 ± 0.002 and 0.945 ± 0.004, respectively. This paper achieves automatic and accurate segmentation of lung parenchyma and provides a reference for the application of attentional mechanisms and multi-scale information in the field of lung parenchyma segmentation.
Due to the high dimensionality and complexity of the data, the analysis of spatial transcriptome data has been a challenging problem. Meanwhile, cluster analysis is the core issue of the analysis of spatial transcriptome data. In this article, a deep learning approach is proposed based on graph attention networks for clustering analysis of spatial transcriptome data. Our method first enhances the spatial transcriptome data, then uses graph attention networks to extract features from nodes, and finally uses the Leiden algorithm for clustering analysis. Compared with the traditional non-spatial and spatial clustering methods, our method has better performance in data analysis through the clustering evaluation index. The experimental results show that the proposed method can effectively cluster spatial transcriptome data and identify different spatial domains, which provides a new tool for studying spatial transcriptome data.
Cardiac three-dimensional electrophysiological labeling technology is the prerequisite and foundation of atrial fibrillation (AF) ablation surgery, and invasive labeling is the current clinical method, but there are many shortcomings such as large trauma, long procedure duration, and low success rate. In recent years, because of its non-invasive and convenient characteristics, ex vivo labeling has become a new direction for the development of electrophysiological labeling technology. With the rapid development of computer hardware and software as well as the accumulation of clinical database, the application of deep learning technology in electrocardiogram (ECG) data is becoming more extensive and has made great progress, which provides new ideas for the research of ex vivo cardiac mapping and intelligent labeling of AF substrates. This paper reviewed the research progress in the fields of ECG forward problem, ECG inverse problem, and the application of deep learning in AF labeling, discussed the problems of ex vivo intelligent labeling of AF substrates and the possible approaches to solve them, prospected the challenges and future directions for ex vivo cardiac electrophysiology labeling.
Heart rate is a crucial indicator of human health with significant physiological importance. Traditional contact methods for measuring heart rate, such as electrocardiograph or wristbands, may not always meet the need for convenient health monitoring. Remote photoplethysmography (rPPG) provides a non-contact method for measuring heart rate and other physiological indicators by analyzing blood volume pulse signals. This approach is non-invasive, does not require direct contact, and allows for long-term healthcare monitoring. Deep learning has emerged as a powerful tool for processing complex image and video data, and has been increasingly employed to extract heart rate signals remotely. This article reviewed the latest research advancements in rPPG-based heart rate measurement using deep learning, summarized available public datasets, and explored future research directions and potential advancements in non-contact heart rate measurement.
The electroencephalogram (EEG) signal is a general reflection of the neurophysiological activity of the brain, which has the advantages of being safe, efficient, real-time and dynamic. With the development and advancement of machine learning research, automatic diagnosis of Alzheimer’s diseases based on deep learning is becoming a research hotspot. Started from feedforward neural networks, this paper compared and analysed the structural properties of neural network models such as recurrent neural networks, convolutional neural networks and deep belief networks and their performance in the diagnosis of Alzheimer’s disease. It also discussed the possible challenges and research trends of this research in the future, expecting to provide a valuable reference for the clinical application of neural networks in the EEG diagnosis of Alzheimer’s disease.
The deep learning-based automatic detection of epilepsy electroencephalogram (EEG), which can avoid the artificial influence, has attracted much attention, and its effectiveness mainly depends on the deep neural network model. In this paper, an attention-based multi-scale residual network (AMSRN) was proposed in consideration of the multiscale, spatio-temporal characteristics of epilepsy EEG and the information flow among channels, and it was combined with multiscale principal component analysis (MSPCA) to realize the automatic epilepsy detection. Firstly, MSPCA was used for noise reduction and feature enhancement of original epilepsy EEG. Then, we designed the structure and parameters of AMSRN. Among them, the attention module (AM), multiscale convolutional module (MCM), spatio-temporal feature extraction module (STFEM) and classification module (CM) were applied successively to signal reexpression with attention weighted mechanism as well as extraction, fusion and classification for multiscale and spatio-temporal features. Based on the Children’s Hospital Boston-Massachusetts Institute of Technology (CHB-MIT) public dataset, the AMSRN model achieved good results in sensitivity (98.56%), F1 score (98.35%), accuracy (98.41%) and precision (98.43%). The results show that AMSRN can make good use of brain network information flow caused by seizures to enhance the difference among channels, and effectively capture the multiscale and spatio-temporal features of EEG to improve the performance of epilepsy detection.