• 제목/요약/키워드: Fully convolutional network model

검색결과 60건 처리시간 0.028초

Modeling of Convolutional Neural Network-based Recommendation System

  • Kim, Tae-Yeun
    • 통합자연과학논문집
    • /
    • 제14권4호
    • /
    • pp.183-188
    • /
    • 2021
  • Collaborative filtering is one of the commonly used methods in the web recommendation system. Numerous researches on the collaborative filtering proposed the numbers of measures for enhancing the accuracy. This study suggests the movie recommendation system applied with Word2Vec and ensemble convolutional neural networks. First, user sentences and movie sentences are made from the user, movie, and rating information. Then, the user sentences and movie sentences are input into Word2Vec to figure out the user vector and movie vector. The user vector is input on the user convolutional model while the movie vector is input on the movie convolutional model. These user and movie convolutional models are connected to the fully-connected neural network model. Ultimately, the output layer of the fully-connected neural network model outputs the forecasts for user, movie, and rating. The test result showed that the system proposed in this study showed higher accuracy than the conventional cooperative filtering system and Word2Vec and deep neural network-based system suggested in the similar researches. The Word2Vec and deep neural network-based recommendation system is expected to help in enhancing the satisfaction while considering about the characteristics of users.

Multi-focus Image Fusion using Fully Convolutional Two-stream Network for Visual Sensors

  • Xu, Kaiping;Qin, Zheng;Wang, Guolong;Zhang, Huidi;Huang, Kai;Ye, Shuxiong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제12권5호
    • /
    • pp.2253-2272
    • /
    • 2018
  • We propose a deep learning method for multi-focus image fusion. Unlike most existing pixel-level fusion methods, either in spatial domain or in transform domain, our method directly learns an end-to-end fully convolutional two-stream network. The framework maps a pair of different focus images to a clean version, with a chain of convolutional layers, fusion layer and deconvolutional layers. Our deep fusion model has advantages of efficiency and robustness, yet demonstrates state-of-art fusion quality. We explore different parameter settings to achieve trade-offs between performance and speed. Moreover, the experiment results on our training dataset show that our network can achieve good performance with subjective visual perception and objective assessment metrics.

A Triple Residual Multiscale Fully Convolutional Network Model for Multimodal Infant Brain MRI Segmentation

  • Chen, Yunjie;Qin, Yuhang;Jin, Zilong;Fan, Zhiyong;Cai, Mao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제14권3호
    • /
    • pp.962-975
    • /
    • 2020
  • The accurate segmentation of infant brain MR image into white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) is very important for early studying of brain growing patterns and morphological changes in neurodevelopmental disorders. Because of inherent myelination and maturation process, the WM and GM of babies (between 6 and 9 months of age) exhibit similar intensity levels in both T1-weighted (T1w) and T2-weighted (T2w) MR images in the isointense phase, which makes brain tissue segmentation very difficult. We propose a deep network architecture based on U-Net, called Triple Residual Multiscale Fully Convolutional Network (TRMFCN), whose structure exists three gates of input and inserts two blocks: residual multiscale block and concatenate block. We solved some difficulties and completed the segmentation task with the model. Our model outperforms the U-Net and some cutting-edge deep networks based on U-Net in evaluation of WM, GM and CSF. The data set we used for training and testing comes from iSeg-2017 challenge (http://iseg2017.web.unc.edu).

CCTV 영상과 합성곱 신경망을 활용한 해무 탐지 기법 연구 (Study on Detection Technique for Sea Fog by using CCTV Images and Convolutional Neural Network)

  • 김나경;박수호;정민지;황도현;앵흐자리갈 운자야;박미소;김보람;윤홍주
    • 한국전자통신학회논문지
    • /
    • 제15권6호
    • /
    • pp.1081-1088
    • /
    • 2020
  • 본 논문에서는 합성곱 신경망을 기반으로 CCTV 이미지를 통한 해무 탐지 방법을 제안한다. 학습에 필요한 자료로 시정 1km 기준으로 총 11개의 항만 또는 해수욕장(부산항, 부산신항, 평택항, 인천항, 군산항, 대산항, 목포항, 여수광양항, 울산항, 포항항, 해운대해수욕장)에서 수집된 해무와 해무가 아닌 이미지 10004장을 랜덤 추출하였다. 전체 10004장의 데이터셋 중에 80%를 추출하여 합성곱 신경망 모델 학습에 사용하였다. 사용된 모델은 16개의 합성곱층과 3개의 완전 연결층을 가지고 있으며, 마지막 완전 연결층에서 Softmax 분류를 수행하는 합성곱 신경망을 활용하였다. 나머지 20%를 이용하여 모델 정확도 평가를 수행하였고 정확도 평가 결과 약 96%의 분류 정확도를 보였다.

Speech Emotion Recognition Using 2D-CNN with Mel-Frequency Cepstrum Coefficients

  • Eom, Youngsik;Bang, Junseong
    • Journal of information and communication convergence engineering
    • /
    • 제19권3호
    • /
    • pp.148-154
    • /
    • 2021
  • With the advent of context-aware computing, many attempts were made to understand emotions. Among these various attempts, Speech Emotion Recognition (SER) is a method of recognizing the speaker's emotions through speech information. The SER is successful in selecting distinctive 'features' and 'classifying' them in an appropriate way. In this paper, the performances of SER using neural network models (e.g., fully connected network (FCN), convolutional neural network (CNN)) with Mel-Frequency Cepstral Coefficients (MFCC) are examined in terms of the accuracy and distribution of emotion recognition. For Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) dataset, by tuning model parameters, a two-dimensional Convolutional Neural Network (2D-CNN) model with MFCC showed the best performance with an average accuracy of 88.54% for 5 emotions, anger, happiness, calm, fear, and sadness, of men and women. In addition, by examining the distribution of emotion recognition accuracies for neural network models, the 2D-CNN with MFCC can expect an overall accuracy of 75% or more.

Railroad Surface Defect Segmentation Using a Modified Fully Convolutional Network

  • Kim, Hyeonho;Lee, Suchul;Han, Seokmin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제14권12호
    • /
    • pp.4763-4775
    • /
    • 2020
  • This research aims to develop a deep learning-based method that automatically detects and segments the defects on railroad surfaces to reduce the cost of visual inspection of the railroad. We developed our segmentation model by modifying a fully convolutional network model [1], a well-known segmentation model used for machine learning, to detect and segment railroad surface defects. The data used in this research are images of the railroad surface with one or more defect regions. Railroad images were cropped to a suitable size, considering the long height and relatively narrow width of the images. They were also normalized based on the variance and mean of the data images. Using these images, the suggested model was trained to segment the defect regions. The proposed method showed promising results in the segmentation of defects. We consider that the proposed method can facilitate decision-making about railroad maintenance, and potentially be applied for other analyses.

초음파 B-모드 영상에서 FCN(fully convolutional network) 모델을 이용한 간 섬유화 단계 분류 알고리즘 (A Fully Convolutional Network Model for Classifying Liver Fibrosis Stages from Ultrasound B-mode Images)

  • 강성호;유선경;이정은;안치영
    • 대한의용생체공학회:의공학회지
    • /
    • 제41권1호
    • /
    • pp.48-54
    • /
    • 2020
  • In this paper, we deal with a liver fibrosis classification problem using ultrasound B-mode images. Commonly representative methods for classifying the stages of liver fibrosis include liver biopsy and diagnosis based on ultrasound images. The overall liver shape and the smoothness and roughness of speckle pattern represented in ultrasound images are used for determining the fibrosis stages. Although the ultrasound image based classification is used frequently as an alternative or complementary method of the invasive biopsy, it also has the limitations that liver fibrosis stage decision depends on the image quality and the doctor's experience. With the rapid development of deep learning algorithms, several studies using deep learning methods have been carried out for automated liver fibrosis classification and showed superior performance of high accuracy. The performance of those deep learning methods depends closely on the amount of datasets. We propose an enhanced U-net architecture to maximize the classification accuracy with limited small amount of image datasets. U-net is well known as a neural network for fast and precise segmentation of medical images. We design it newly for the purpose of classifying liver fibrosis stages. In order to assess the performance of the proposed architecture, numerical experiments are conducted on a total of 118 ultrasound B-mode images acquired from 78 patients with liver fibrosis symptoms of F0~F4 stages. The experimental results support that the performance of the proposed architecture is much better compared to the transfer learning using the pre-trained model of VGGNet.

Residual Learning Based CNN for Gesture Recognition in Robot Interaction

  • Han, Hua
    • Journal of Information Processing Systems
    • /
    • 제17권2호
    • /
    • pp.385-398
    • /
    • 2021
  • The complexity of deep learning models affects the real-time performance of gesture recognition, thereby limiting the application of gesture recognition algorithms in actual scenarios. Hence, a residual learning neural network based on a deep convolutional neural network is proposed. First, small convolution kernels are used to extract the local details of gesture images. Subsequently, a shallow residual structure is built to share weights, thereby avoiding gradient disappearance or gradient explosion as the network layer deepens; consequently, the difficulty of model optimisation is simplified. Additional convolutional neural networks are used to accelerate the refinement of deep abstract features based on the spatial importance of the gesture feature distribution. Finally, a fully connected cascade softmax classifier is used to complete the gesture recognition. Compared with the dense connection multiplexing feature information network, the proposed algorithm is optimised in feature multiplexing to avoid performance fluctuations caused by feature redundancy. Experimental results from the ISOGD gesture dataset and Gesture dataset prove that the proposed algorithm affords a fast convergence speed and high accuracy.

훈련자료 내 재배지역의 비율에 따른 딥러닝 모델의 성능 평가 (Performance Evaluation of Deep Learning Model according to the Ratio of Cultivation Area in Training Data)

  • 성선경;최재완
    • 대한원격탐사학회지
    • /
    • 제38권6_1호
    • /
    • pp.1007-1014
    • /
    • 2022
  • 차세대중형위성(Compact Advanced Satellite 500, CAS500)은 식생, 산림, 농업 등의 분야를 포함한 다양한 목적을 위하여 사용될 수 있으며, 다양한 영역에 대한 빠른 위성영상의 취득이 가능할 것으로 기대되고 있다. 차세대중형위성을 통하여 취득된 위성영상을 농업분야에서 활용하기 위해서는 위성영상 기반 작물재배지역 추출 기법에 대한 개발이 필요하다. 특히, 최근 들어 딥러닝 분야에 대한 연구가 활발해짐에 따라서, 작물재배지역 추출을 위한 딥러닝 모델의 개발 및 훈련자료 생성에 관한 연구가 필요한 실정이다. 본 연구에서는 PlanetScope 위성영상과 팜맵을 이용하여 합천군 지역의 양파 및 마늘 재배지역을 분류하고자 하였다. 특히, 효과적인 모델의 학습을 위하여 작물재배지역의 비율에 따른 모델 성능을 분석하고자 하였다. 실험에 사용한 딥러닝 모델은 Fully Convolutional Densely Connected Convolutional Network (FC-DenseNet)을 작물재배지역 분류의 목적에 맞도록 재구성하여 활용하였다. 실험결과, 훈련자료 내 작물재배지역의 비율이 딥러닝 모델의 성능에 영향을 미치는 것을 확인하였다.

폐 결절 검출을 위한 합성곱 신경망의 성능 개선 (Performance Improvement of Convolutional Neural Network for Pulmonary Nodule Detection)

  • 김한웅;김병남;이지은;장원석;유선국
    • 대한의용생체공학회:의공학회지
    • /
    • 제38권5호
    • /
    • pp.237-241
    • /
    • 2017
  • Early detection of the pulmonary nodule is important for diagnosis and treatment of lung cancer. Recently, CT has been used as a screening tool for lung nodule detection. And, it has been reported that computer aided detection(CAD) systems can improve the accuracy of the radiologist in detection nodules on CT scan. The previous study has been proposed a method using Convolutional Neural Network(CNN) in Lung CAD system. But the proposed model has a limitation in accuracy due to its sparse layer structure. Therefore, we propose a Deep Convolutional Neural Network to overcome this limitation. The model proposed in this work is consist of 14 layers including 8 convolutional layers and 4 fully connected layers. The CNN model is trained and tested with 61,404 regions-of-interest (ROIs) patches of lung image including 39,760 nodules and 21,644 non-nodules extracted from the Lung Image Database Consortium(LIDC) dataset. We could obtain the classification accuracy of 91.79% with the CNN model presented in this work. To prevent overfitting, we trained the model with Augmented Dataset and regularization term in the cost function. With L1, L2 regularization at Training process, we obtained 92.39%, 92.52% of accuracy respectively. And we obtained 93.52% with data augmentation. In conclusion, we could obtain the accuracy of 93.75% with L2 Regularization and Data Augmentation.