• Title/Summary/Keyword: AlexNet

Search Result 71, Processing Time 0.023 seconds

Application of CNN for Fish Species Classification (어종 분류를 위한 CNN의 적용)

  • Park, Jin-Hyun;Hwang, Kwang-Bok;Park, Hee-Mun;Choi, Young-Kiu
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.1
    • /
    • pp.39-46
    • /
    • 2019
  • In this study, before system development for the elimination of foreign fish species, we propose an algorithm to classify fish species by training fish images with CNN. The raw data for CNN learning were directly captured images for each species, Dataset 1 increases the number of images to improve the classification of fish species and Dataset 2 realizes images close to natural environment are constructed and used as training and test data. The classification performance of four CNNs are over 99.97% for dataset 1 and 99.5% for dataset 2, in particular, we confirm that the learned CNN using Data Set 2 has satisfactory performance for fish images similar to the natural environment. And among four CNNs, AlexNet achieves satisfactory performance, and this has also the shortest execution time and training time, we confirm that it is the most suitable structure to develop the system for the elimination of foreign fish species.

A Study of Facial Organs Classification System Based on Fusion of CNN Features and Haar-CNN Features

  • Hao, Biao;Lim, Hye-Youn;Kang, Dae-Seong
    • The Journal of Korean Institute of Information Technology
    • /
    • v.16 no.11
    • /
    • pp.105-113
    • /
    • 2018
  • In this paper, we proposed a method for effective classification of eye, nose, and mouth of human face. Most recent image classification uses Convolutional Neural Network(CNN). However, the features extracted by CNN are not sufficient and the classification effect is not too high. We proposed a new algorithm to improve the classification effect. The proposed method can be roughly divided into three parts. First, the Haar feature extraction algorithm is used to construct the eye, nose, and mouth dataset of face. The second, the model extracts CNN features of image using AlexNet. Finally, Haar-CNN features are extracted by performing convolution after Haar feature extraction. After that, CNN features and Haar-CNN features are fused and classify images using softmax. Recognition rate using mixed features could be increased about 4% than CNN feature. Experiments have demonstrated the performance of the proposed algorithm.

CNN-based Sign Language Translation Program for the Deaf (CNN기반의 청각장애인을 위한 수화번역 프로그램)

  • Hong, Kyeong-Chan;Kim, Hyung-Su;Han, Young-Hwan
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.22 no.4
    • /
    • pp.206-212
    • /
    • 2021
  • Society is developing more and more, and communication methods are developing in many ways. However, developed communication is a way for the non-disabled and has no effect on the deaf. Therefore, in this paper, a CNN-based sign language translation program is designed and implemented to help deaf people communicate. Sign language translation programs translate sign language images entered through WebCam according to meaning based on data. The sign language translation program uses 24,000 pieces of Korean vowel data produced directly and conducts U-Net segmentation to train effective classification models. In the implemented sign language translation program, 'ㅋ' showed the best performance among all sign language data with 97% accuracy and 99% F1-Score, while 'ㅣ' showed the highest performance among vowel data with 94% accuracy and 95.5% F1-Score.

Transfer Learning Using Convolutional Neural Network Architectures for Glioma Classification from MRI Images

  • Kulkarni, Sunita M.;Sundari, G.
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.2
    • /
    • pp.198-204
    • /
    • 2021
  • Glioma is one of the common types of brain tumors starting in the brain's glial cell. These tumors are classified into low-grade or high-grade tumors. Physicians analyze the stages of brain tumors and suggest treatment to the patient. The status of the tumor has an importance in the treatment. Nowadays, computerized systems are used to analyze and classify brain tumors. The accurate grading of the tumor makes sense in the treatment of brain tumors. This paper aims to develop a classification of low-grade glioma and high-grade glioma using a deep learning algorithm. This system utilizes four transfer learning algorithms, i.e., AlexNet, GoogLeNet, ResNet18, and ResNet50, for classification purposes. Among these algorithms, ResNet18 shows the highest classification accuracy of 97.19%.

Interworking technology of neural network and data among deep learning frameworks

  • Park, Jaebok;Yoo, Seungmok;Yoon, Seokjin;Lee, Kyunghee;Cho, Changsik
    • ETRI Journal
    • /
    • v.41 no.6
    • /
    • pp.760-770
    • /
    • 2019
  • Based on the growing demand for neural network technologies, various neural network inference engines are being developed. However, each inference engine has its own neural network storage format. There is a growing demand for standardization to solve this problem. This study presents interworking techniques for ensuring the compatibility of neural networks and data among the various deep learning frameworks. The proposed technique standardizes the graphic expression grammar and learning data storage format using the Neural Network Exchange Format (NNEF) of Khronos. The proposed converter includes a lexical, syntax, and parser. This NNEF parser converts neural network information into a parsing tree and quantizes data. To validate the proposed system, we verified that MNIST is immediately executed by importing AlexNet's neural network and learned data. Therefore, this study contributes an efficient design technique for a converter that can execute a neural network and learned data in various frameworks regardless of the storage format of each framework.

Obstacle Detection and Recognition System for Autonomous Driving Vehicle (자율주행차를 위한 장애물 탐지 및 인식 시스템)

  • Han, Ju-Chan;Koo, Bon-Cheol;Cheoi, Kyung-Joo
    • Journal of Convergence for Information Technology
    • /
    • v.7 no.6
    • /
    • pp.229-235
    • /
    • 2017
  • In recent years, research has been actively carried out to recognize and recognize objects based on a large amount of data. In this paper, we propose a system that extracts objects that are thought to be obstacles in road driving images and recognizes them by car, man, and motorcycle. The objects were extracted using Optical Flow in consideration of the direction and size of the moving objects. The extracted objects were recognized using Alexnet, one of CNN (Convolutional Neural Network) recognition models. For the experiment, various images on the road were collected and experimented with black box. The result of the experiment showed that the object extraction accuracy was 92% and the object recognition accuracy was 96%.

Model Type Inference Attack Using Output of Black-Box AI Model (블랙 박스 모델의 출력값을 이용한 AI 모델 종류 추론 공격)

  • An, Yoonsoo;Choi, Daeseon
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.32 no.5
    • /
    • pp.817-826
    • /
    • 2022
  • AI technology is being successfully introduced in many fields, and models deployed as a service are deployed with black box environment that does not expose the model's information to protect intellectual property rights and data. In a black box environment, attackers try to steal data or parameters used during training by using model output. This paper proposes a method of inferring the type of model to directly find out the composition of layer of the target model, based on the fact that there is no attack to infer the information about the type of model from the deep learning model. With ResNet, VGGNet, AlexNet, and simple convolutional neural network models trained with MNIST datasets, we show that the types of models can be inferred using the output values in the gray box and black box environments of the each model. In addition, we inferred the type of model with approximately 83% accuracy in the black box environment if we train the big and small relationship feature that proposed in this paper together, the results show that the model type can be infrerred even in situations where only partial information is given to attackers, not raw probability vectors.

An Approximate DRAM Architecture for Energy-efficient Deep Learning

  • Nguyen, Duy Thanh;Chang, Ik-Joon
    • Journal of Semiconductor Engineering
    • /
    • v.1 no.1
    • /
    • pp.31-37
    • /
    • 2020
  • We present an approximate DRAM architecture for energy-efficient deep learning. Our key premise is that by bounding memory errors to non-critical information, we can significantly reduce DRAM refresh energy without compromising recognition accuracy of deep neural networks. To validate the key premise, we make extensive Monte-Carlo simulations for several well-known convolutional neural networks such as LeNet, ConvNet and AlexNet with the input of MINIST, CIFAR-10, and ImageNet, respectively. We assume that the highest-order 8-bits (in single precision) and 4-bits (in half precision) are protected from retention errors under the proposed architecture and then, randomly inject bit-errors to unprotected bits with various bit-error-rates. Here, recognition accuracies of the above convolutional neural networks are successfully maintained up to the 10-5-order bit-error-rate. We simulate DRAM energy during inference of the above convolutional neural networks, where the proposed architecture shows the possibility of considerable energy saving up to 10 ~ 37.5% of total DRAM energy.

Image Segmentation Using SqueezeNet based on CUDA C (CUDA C기반 SqueezeNet을 이용한 영상 분할)

  • Jeon, Sae-Yun;Wang, Jin-Yeong;Lee, Sang-Hwan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2018.10a
    • /
    • pp.631-633
    • /
    • 2018
  • 최근 영상처리 분야에서 딥러닝(Deep learning)을 이용한 기술이 좋은 성능을 보이면서 이에 대한 관심과 연구가 증가하고 있다. 본 연구에서는 최근 딥러닝 네트워크 중 적은 파라미터 수로 AlexNet수준의 성능을 보인 SquezeNet을 영상 분할(Image segmentation)의 특징 추출(feature extraction)영역으로 사용하고, CUDA C기반으로 코드를 작성하여 정확도를 유지하면서 계산 속도 면에서도 좋은 성능을 얻을 수 있었다.

Comparison of CNN-based models for apple pest classification (사과 병해충 분류를 위한 CNN 기반 모델 비교)

  • Lee, Su-min;Lee, Yu-hyeon;Lee, Eun-sol;Han, Se-yun
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.05a
    • /
    • pp.460-463
    • /
    • 2022
  • 세계에서 가장 중요한 온대 과일 작물 중 하나인 사과의 생산성과 품질은 병해충 여부에 큰 영향을 받는다. 이를 진단하기 위해서는 효율적이고 많은 전문 지식과 상당한 시간이 필요하다. 그러므로 이를 해결하기 위해 효율적이고 정확하게 다양한 병해충을 진단하는 시스템이 필요하다. 본 논문에서는 이미지 분석에 큰 효율을 보인 딥러닝 기반 CNN 들을 비교 분석하여 사과의 병해충 여부를 판별하고 최적의 모델을 제시한다. 딥러닝 기반 CNN 구조를 가진 AlexNet, VGGNet, Inception-ResNet-v2, DenseNet 을 채택해 사과 병해충 분류 성능 평가를 진행했다. 그 결과 DenseNet 이 가장 우수한 성능을 보여주었다.