• Title/Summary/Keyword: AlexNet

Search Result 71, Processing Time 0.026 seconds

Performance Analysis of Feature Extractor for Transfer Learning of a Small Sample of Medical Images (소표본 의료 영상의 전이 학습을 위한 Feature Extractor 기법의 성능 비교 및 분석)

  • Lee, Dong-Ho;Hong, Dae-Yong;Lee, Yeon;Shin, Byeong-Seok
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2018.05a
    • /
    • pp.405-406
    • /
    • 2018
  • 본 논문은 소표본 의료용 영상 분석의 정확도 향상을 위해 전이학습 모델을 feature extractor로 구축하여 학습시키는 방법을 연구하였으며 성능 평가를 위해 선학습모델로 AlexNet, ResNet, DenseNet을 사용하여 fine tuning 기법을 적용하였을 때와의 성능을 비교 분석하였다. 그 결과 실험에 사용된 3개의 모델에서 fine tuning 기법보다 향상된 정확도를 보임을 확인하였고, 또한 ImageNet으로 학습된 AlexNet, ResNet, DenseNet이 소표본 의료용 X-Ray 영상에 적용될 수 있음을 보였다.

Convolutional Neural Networks for Character-level Classification

  • Ko, Dae-Gun;Song, Su-Han;Kang, Ki-Min;Han, Seong-Wook
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.6 no.1
    • /
    • pp.53-59
    • /
    • 2017
  • Optical character recognition (OCR) automatically recognizes text in an image. OCR is still a challenging problem in computer vision. A successful solution to OCR has important device applications, such as text-to-speech conversion and automatic document classification. In this work, we analyze character recognition performance using the current state-of-the-art deep-learning structures. One is the AlexNet structure, another is the LeNet structure, and the other one is the SPNet structure. For this, we have built our own dataset that contains digits and upper- and lower-case characters. We experiment in the presence of salt-and-pepper noise or Gaussian noise, and report the performance comparison in terms of recognition error. Experimental results indicate by five-fold cross-validation that the SPNet structure (our approach) outperforms AlexNet and LeNet in recognition error.

A Study of Kernel Characteristics of CNN Deep Learning for Effective Fire Detection Based on Video (영상기반의 화재 검출에 효과적인 CNN 심층학습의 커널 특성에 대한 연구)

  • Son, Geum-Young;Park, Jang-Sik
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.6
    • /
    • pp.1257-1262
    • /
    • 2018
  • In this paper, a deep learning method is proposed to detect the fire effectively by using video of surveillance camera. Based on AlexNet model, classification performance is compared according to kernel size and stride of convolution layer. Dataset for learning and interfering are classified into two classes such as normal and fire. Normal images include clouds, and foggy images, and fire images include smoke and flames images, respectively. As results of simulations, it is shown that the larger kernel size and smaller stride shows better performance.

Evaluation of Deep-Learning Feature Based COVID-19 Classifier in Various Neural Network (코로나바이러스 감염증19 데이터베이스에 기반을 둔 인공신경망 모델의 특성 평가)

  • Hong, Jun-Yong;Jung, Young-Jin
    • Journal of radiological science and technology
    • /
    • v.43 no.5
    • /
    • pp.397-404
    • /
    • 2020
  • Coronavirus disease(COVID-19) is highly infectious disease that directly affects the lungs. To observe the clinical findings from these lungs, the Chest Radiography(CXR) can be used in a fast manner. However, the diagnostic performance via CXR needs to be improved, since the identifying these findings are highly time-consuming and prone to human error. Therefore, Artificial Intelligence(AI) based tool may be useful to aid the diagnosis of COVID-19 via CXR. In this study, we explored various Deep learning(DL) approach to classify COVID-19, other viral pneumonia and normal. For the original dataset and lung-segmented dataset, the pre-trained AlexNet, SqueezeNet, ResNet18, DenseNet201 were transfer-trained and validated for 3 class - COVID-19, viral pneumonia, normal. In the results, AlexNet showed the highest mean accuracy of 99.15±2.69% and fastest training time of 1.61±0.56 min among 4 pre-trained neural networks. In this study, we demonstrated the performance of 4 pre-trained neural networks in COVID-19 diagnosis with CXR images. Further, we plotted the class activation map(CAM) of each network and demonstrated that the lung-segmentation pre-processing improve the performance of COVID-19 classifier with CXR images by excluding background features.

Deep Learning Based Floating Macroalgae Classification Using Gaofen-1 WFV Images (Gaofen-1 WFV 영상을 이용한 딥러닝 기반 대형 부유조류 분류)

  • Kim, Euihyun;Kim, Keunyong;Kim, Soo Mee;Cui, Tingwei;Ryu, Joo-Hyung
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.2_2
    • /
    • pp.293-307
    • /
    • 2020
  • Every year, the floating macroalgae, green and golden tide, are massively detected at the Yellow Sea and East China Sea. After influx of them to the aquaculture facility or beach, it occurs enormous economic losses to remove them. Currently, remote sensing is used effectively to detect the floating macroalgae flowed into the coast. But it has difficulties to detect the floating macroalgae exactly because of the wavelength overlapped with other targets in the ocean. Also, it is difficult to distinguish between green and golden tide because they have similar spectral characteristics. Therefore, we tried to distinguish between green and golden tide applying the Deep learning method to the satellite images. To determine the network, the optimal training conditions were searched to train the AlexNet. Also, Gaofen-1 WFV images were used as a dataset to train and validate the network. Under these conditions, the network was determined after training, and used to confirm the test data. As a result, the accuracy of test data is 88.89%, and it can be possible to distinguish between green and golden tide with precision of 66.67% and 100%, respectively. It is interpreted that the AlexNet can be pick up on the subtle differences between green and golden tide. Through this study, it is expected that the green and golden tide can be effectively classified from various objects in the ocean and distinguished each other.

Performance Enhancement and Evaluation of a Deep Learning Framework on Embedded Systems using Unified Memory (통합메모리를 이용한 임베디드 환경에서의 딥러닝 프레임워크 성능 개선과 평가)

  • Lee, Minhak;Kang, Woochul
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.7
    • /
    • pp.417-423
    • /
    • 2017
  • Recently, many embedded devices that have the computing capability required for deep learning have become available; hence, many new applications using these devices are emerging. However, these embedded devices have an architecture different from that of PCs and high-performance servers. In this paper, we propose a method that improves the performance of deep-learning framework by considering the architecture of an embedded device that shares memory between the CPU and the GPU. The proposed method is implemented in Caffe, an open-source deep-learning framework, and is evaluated on an NVIDIA Jetson TK1 embedded device. In the experiment, we investigate the image recognition performance of several state-of-the-art deep-learning networks, including AlexNet, VGGNet, and GoogLeNet. Our results show that the proposed method can achieve significant performance gain. For instance, in AlexNet, we could reduce image recognition latency by about 33% and energy consumption by about 50%.

Mushroom Image Recognition using Convolutional Neural Network and Transfer Learning (컨볼루션 신경망과 전이 학습을 이용한 버섯 영상 인식)

  • Kang, Euncheol;Han, Yeongtae;Oh, Il-Seok
    • KIISE Transactions on Computing Practices
    • /
    • v.24 no.1
    • /
    • pp.53-57
    • /
    • 2018
  • A poisoning accident is often caused by a situation in which people eat poisonous mushrooms because they cannot distinguish between edible mushrooms and poisonous mushrooms. In this paper, we propose an automatic mushroom recognition system by using the convolutional neural network. We collected 1478 mushroom images of 38 species using image crawling, and used the dataset for learning the convolutional neural network. A comparison experiment using AlexNet, VGGNet, and GoogLeNet was performed using the collected datasets, and a comparison experiment using a class number expansion and a fine-tuning technique for transfer learning were performed. As a result of our experiment, we achieve 82.63% top-1 accuracy and 96.84% top-5 accuracy on test set of our dataset.

A Study on Detection Performance Comparison of Bone Plates Using Parallel Convolution Neural Networks (병렬형 합성곱 신경망을 이용한 골절합용 판의 탐지 성능 비교에 관한 연구)

  • Lee, Song Yeon;Huh, Yong Jeong
    • Journal of the Semiconductor & Display Technology
    • /
    • v.21 no.3
    • /
    • pp.63-68
    • /
    • 2022
  • In this study, we produced defect detection models using parallel convolution neural networks. If convolution neural networks are constructed parallel type, the model's detection accuracy will increase and detection time will decrease. We produced parallel-type defect detection models using 4 types of convolutional algorithms. The performance of models was evaluated using evaluation indicators. The model's performance is detection accuracy and detection time. We compared the performance of each parallel model. The detection accuracy of the model using AlexNet is 97 % and the detection time is 0.3 seconds. We confirmed that when AlexNet algorithm is constructed parallel type, the model has the highest performance.

Glaucoma Detection of Fundus Images Using Convolution Neural Network (CNN을 이용한 안저 영상의 녹내장 검출)

  • Shin, B.S.
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.636-638
    • /
    • 2022
  • This paper is a study to apply CNN(Convolution Neural Network) to fundus images for identifying glaucoma. Fundus images are evaluated in the field of medical diagnosis detection, which are diagnosing of blood vessels and nerve tissues, retina damage, various cardiovascular diseases and dementia. For the experiment, using normal image set and glaucoma image set, two types of image set are classifed by using AlexNet. The result performs that glaucoma with abnormalities are activated and characterized in feature map.

  • PDF

Emotion Transfer with Strength Control for End-to-End TTS (감정 제어 가능한 종단 간 음성합성 시스템)

  • Jeon, Yejin;Lee, Gary Geunbae
    • Annual Conference on Human and Language Technology
    • /
    • 2021.10a
    • /
    • pp.423-426
    • /
    • 2021
  • 본 논문은 전역 스타일 토큰(Global Style Token)을 기준으로 하여 감정의 세기를 조절할 수 있는 방법을 소개한다. 기존의 전역 스타일 토큰 연구에서는 원하는 스타일이 포함된 참조 오디오(reference audio)을 사용하여 음성을 합성하였다. 그러나, 참조 오디오의 스타일대로만 음성합성이 가능하기 때문에 세밀한 감정 조절에 어려움이 있었다. 이 문제를 해결하기 위해 본 논문에서는 전역 스타일 토큰의 레퍼런스 인코더 부분을 잔여 블록(residual block)과 컴퓨터 비전 분야에서 사용되는 AlexNet으로 대체하였다. AlexNet은 5개의 함성곱 신경망(convolutional neural networks) 으로 구성되어 있지만, 본 논문에서는 1개의 신경망을 제외한 4개의 레이어만 사용했다. 청취 평가(Mean Opinion Score)를 통해 제시된 방법으로 감정 세기의 조절 가능성을 보여준다.

  • PDF