Sleep apnea (SA) detection method based on traditional machine learning needs a lot of efforts in feature engineering and classifier design. We constructed a one-dimensional convolutional neural network (CNN) model, which consists in four convolution layers, four pooling layers, two full connection layers and one classification layer. The automatic feature extraction and classification were realized by the structure of the proposed CNN model. The model was verified by the whole night single-channel sleep electrocardiogram (ECG) signals of 70 subjects from the Apnea-ECG dataset. Our results showed that the accuracy of per-segment SA detection was ranged from 80.1% to 88.0%, using the input signals of single-channel ECG signal, RR interval (RRI) sequence, R peak sequence and RRI sequence + R peak sequence respectively. These results indicated that the proposed CNN model was effective and can automatically extract and classify features from the original single-channel ECG signal or its derived signal RRI and R peak sequence. When the input signals were RRI sequence + R peak sequence, the CNN model achieved the best performance. The accuracy, sensitivity and specificity of per-segment SA detection were 88.0%, 85.1% and 89.9%, respectively. And the accuracy of per-recording SA diagnosis was 100%. These findings indicated that the proposed method can effectively improve the accuracy and robustness of SA detection and outperform the methods reported in recent years. The proposed CNN model can be applied to portable screening diagnosis equipment for SA with remote server.
Coronavirus disease 2019 (COVID-19) has spread rapidly around the world. In order to diagnose COVID-19 more quickly, in this paper, a depthwise separable DenseNet was proposed. The paper constructed a deep learning model with 2 905 chest X-ray images as experimental dataset. In order to enhance the contrast, the contrast limited adaptive histogram equalization (CLAHE) algorithm was used to preprocess the X-ray image before network training, then the images were put into the training network and the parameters of the network were adjusted to the optimal. Meanwhile, Leaky ReLU was selected as the activation function. VGG16, ResNet18, ResNet34, DenseNet121 and SDenseNet models were used to compare with the model proposed in this paper. Compared with ResNet34, the proposed classification model of pneumonia had improved 2.0%, 2.3% and 1.5% in accuracy, sensitivity and specificity respectively. Compared with the SDenseNet network without depthwise separable convolution, number of parameters of the proposed model was reduced by 43.9%, but the classification effect did not decrease. It can be found that the proposed DWSDenseNet has a good classification effect on the COVID-19 chest X-ray images dataset. Under the condition of ensuring the accuracy as much as possible, the depthwise separable convolution can effectively reduce number of parameters of the model.
Attention deficit/hyperactivity disorder (ADHD) is a behavioral disorder syndrome found mainly in school-age population. At present, the diagnosis of ADHD mainly depends on the subjective methods, leading to the high rate of misdiagnosis and missed-diagnosis. To solve these problems, we proposed an algorithm for classifying ADHD objectively based on convolutional neural network. At first, preprocessing steps, including skull stripping, Gaussian kernel smoothing, et al., were applied to brain magnetic resonance imaging (MRI). Then, coarse segmentation was used for selecting the right caudate nucleus, left precuneus, and left superior frontal gyrus region. Finally, a 3 level convolutional neural network was used for classification. Experimental results showed that the proposed algorithm was capable of classifying ADHD and normal groups effectively, the classification accuracies obtained by the right caudate nucleus and the left precuneus brain regions were greater than the highest classification accuracy (62.52%) in the ADHD-200 competition, and among 3 brain regions in ADHD and the normal groups, the classification accuracy from the right caudate nucleus was the highest. It is well concluded that the method for classification of ADHD and normal groups proposed in this paper utilizing the coarse segmentation and deep learning is a useful method for the purpose. The classification accuracy of the proposed method is high, and the calculation is simple. And the method is able to extract the unobvious image features better, and can overcome the shortcomings of traditional methods of MRI brain area segmentation, which are time-consuming and highly complicate. The method provides an objective diagnosis approach for ADHD.
When applying deep learning to the automatic segmentation of organs at risk in medical images, we combine two network models of Dense Net and V-Net to develop a Dense V-network for automatic segmentation of three-dimensional computed tomography (CT) images, in order to solve the problems of degradation and gradient disappearance of three-dimensional convolutional neural networks optimization as training samples are insufficient. This algorithm is applied to the delineation of pelvic endangered organs and we take three representative evaluation parameters to quantitatively evaluate the segmentation effect. The clinical result showed that the Dice similarity coefficient values of the bladder, small intestine, rectum, femoral head and spinal cord were all above 0.87 (average was 0.9); Jaccard distance of these were within 2.3 (average was 0.18). Except for the small intestine, the Hausdorff distance of other organs were less than 0.9 cm (average was 0.62 cm). The Dense V-Network has been proven to achieve the accurate segmentation of pelvic endangered organs.
The application of minimally invasive surgical tool detection and tracking technology based on deep learning in minimally invasive surgery is currently a research hotspot. This paper firstly expounds the relevant technical content of the minimally invasive surgery tool detection and tracking, which mainly introduces the advantages based on deep learning algorithm. Then, this paper summarizes the algorithm for detection and tracking surgical tools based on fully supervised deep neural network and the emerging algorithm for detection and tracking surgical tools based on weakly supervised deep neural network. Several typical algorithm frameworks and their flow charts based on deep convolutional and recurrent neural networks are summarized emphatically, so as to enable researchers in relevant fields to understand the current research progress more systematically and provide reference for minimally invasive surgeons to select navigation technology. In the end, this paper provides a general direction for the further research of minimally invasive surgical tool detection and tracking technology based on deep learning.
With the rapid improvement of the perception and computing capacity of mobile devices such as smart phones, human activity recognition using mobile devices as the carrier has been a new research hot-spot. The inertial information collected by the acceleration sensor in the smart mobile device is used for human activity recognition. Compared with the common computer vision recognition, it has the following advantages: convenience, low cost, and better reflection of the essence of human motion. Based on the WISDM data set collected by smart phones, the inertial navigation information and the deep learning algorithm-convolutional neural network (CNN) were adopted to build a human activity recognition model in this paper. The K nearest neighbor algorithm (KNN) and the random forest algorithm were compared with the CNN network in the recognition accuracy to evaluate the performance of the CNN network. The classification accuracy of CNN model reached 92.73%, which was much higher than KNN and random forest. Experimental results show that the CNN algorithm model can achieve more accurate human activity recognition and has broad application prospects in predicting and promoting human health.
A method was proposed to detect pulmonary nodules in low-dose computed tomography (CT) images by two-dimensional convolutional neural network under the condition of fine image preprocessing. Firstly, CT image preprocessing was carried out by image clipping, normalization and other algorithms. Then the positive samples were expanded to balance the number of positive and negative samples in convolutional neural network. Finally, the model with the best performance was obtained by training two-dimensional convolutional neural network and constantly optimizing network parameters. The model was evaluated in Lung Nodule Analysis 2016(LUNA16) dataset by means of five-fold cross validation, and each group's average model experiment results were obtained with the final accuracy of 92.3%, sensitivity of 92.1% and specificity of 92.6%.Compared with other existing automatic detection and classification methods for pulmonary nodules, all indexes were improved. Subsequently, the model perturbation experiment was carried out on this basis. The experimental results showed that the model is stable and has certain anti-interference ability, which could effectively identify pulmonary nodules and provide auxiliary diagnostic advice for early screening of lung cancer.
Ultrasound is the best way to diagnose thyroid nodules. To discriminate benign and malignant nodules, calcification is an important characteristic. However, calcification in ultrasonic images cannot be extracted accurately because of capsule wall and other internal tissue. In this paper, deep learning was first proposed to extract calcification, and two improved methods were proposed on the basis of Alexnet convolutional neural network. First, adding the corresponding anti-pooling (unpooling) and deconvolution layers (deconv2D) made the network to be trained for the required features and finally extract the calcification feature. Second, modifying the number of convolution templates and full connection layer nodes made feature extraction more refined. The final network was the combination of two improved methods above. To verify the method presented in this article, we got 8 416 images with calcification, and 10 844 without calcification. The result showed that the accuracy of the calcification extraction was 86% by using the improved Alexnet convolutional neural network. Compared with traditional methods, it has been improved greatly, which provides effective means for the identification of benign and malignant thyroid nodules.
Alzheimer’s disease (AD) is a progressive and irreversible neurodegenerative disease. Neuroimaging based on magnetic resonance imaging (MRI) is one of the most intuitive and reliable methods to perform AD screening and diagnosis. Clinical head MRI detection generates multimodal image data, and to solve the problem of multimodal MRI processing and information fusion, this paper proposes a structural and functional MRI feature extraction and fusion method based on generalized convolutional neural networks (gCNN). The method includes a three-dimensional residual U-shaped network based on hybrid attention mechanism (3D HA-ResUNet) for feature representation and classification for structural MRI, and a U-shaped graph convolutional neural network (U-GCN) for node feature representation and classification of brain functional networks for functional MRI. Based on the fusion of the two types of image features, the optimal feature subset is selected based on discrete binary particle swarm optimization, and the prediction results are output by a machine learning classifier. The validation results of multimodal dataset from the AD Neuroimaging Initiative (ADNI) open-source database show that the proposed models have superior performance in their respective data domains. The gCNN framework combines the advantages of these two models and further improves the performance of the methods using single-modal MRI, improving the classification accuracy and sensitivity by 5.56% and 11.11%, respectively. In conclusion, the gCNN-based multimodal MRI classification method proposed in this paper can provide a technical basis for the auxiliary diagnosis of Alzheimer’s disease.
The convolutional neural network (CNN) could be used on computer-aided diagnosis of lung tumor with positron emission tomography (PET)/computed tomography (CT), which can provide accurate quantitative analysis to compensate for visual inertia and defects in gray-scale sensitivity, and help doctors diagnose accurately. Firstly, parameter migration method is used to build three CNNs (CT-CNN, PET-CNN, and PET/CT-CNN) for lung tumor recognition in CT, PET, and PET/CT image, respectively. Then, we aimed at CT-CNN to obtain the appropriate model parameters for CNN training through analysis the influence of model parameters such as epochs, batchsize and image scale on recognition rate and training time. Finally, three single CNNs are used to construct ensemble CNN, and then lung tumor PET/CT recognition was completed through relative majority vote method and the performance between ensemble CNN and single CNN was compared. The experiment results show that the ensemble CNN is better than single CNN on computer-aided diagnosis of lung tumor.