• Title/Summary/Keyword: ImageNet

Search Result 773, Processing Time 0.026 seconds

The Combined Effect and Therapeutic Effects of Color (변환학습을 이용한 장면 분류)

  • Shin, Seong-Yoon;Shin, Kwang-Seong;Nam, Soo-Tai
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.338-339
    • /
    • 2021
  • In this paper, we proposed a multiclass image scene classification method based on transform learning. The method using the Residual Network (ResNet) model which pre-trained on the large image dataset ImageNet for image classification. Compared with the image classification method of the CNN model, it can greatly improve the classification accuracy and efficiency

  • PDF

Image Scene Classification of Multiclass (다중 클래스의 이미지 장면 분류)

  • Shin, Seong-Yoon;Lee, Hyun-Chang;Shin, Kwang-Seong;Kim, Hyung-Jin;Lee, Jae-Wan
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.551-552
    • /
    • 2021
  • In this paper, we present a multi-class image scene classification method based on transformation learning. ImageNet classifies multiple classes of natural scene images by relying on pre-trained network models on large image datasets. In the experiment, we obtained excellent results by classifying the optimized ResNet model on Kaggle's Intel Image Classification data set.

  • PDF

Image generation and classification using GAN-based Semi Supervised Learning (GAN기반의 Semi Supervised Learning을 활용한 이미지 생성 및 분류)

  • Doyoon Jung;Gwangmi Choi;NamHo Kim
    • Smart Media Journal
    • /
    • v.13 no.3
    • /
    • pp.27-35
    • /
    • 2024
  • This study deals with a method of combining image generation using Semi Supervised Learning based on GAN (Generative Adversarial Network) and image classification using ResNet50. Through this, a new approach was proposed to obtain more accurate and diverse results by integrating image generation and classification. The generator and discriminator are trained to distinguish generated images from actual images, and image classification is performed using ResNet50. In the experimental results, it was confirmed that the quality of the generated images changes depending on the epoch, and through this, we aim to improve the accuracy of industrial accident prediction. In addition, we would like to present an efficient method to improve the quality of image generation and increase the accuracy of image classification through the combination of GAN and ResNet50.

U-Net Based Plant Image Segmentation (U-Net 기반의 식물 영상 분할 기법)

  • Lee, Sang-Ho;Kim, Tae-Hyeon;Kim, Jong-Ok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.81-83
    • /
    • 2021
  • In this paper, we propose a method to segment a plant from a plant image using U-Net. The network is an end-to-end fully convolutional network that is mainly used for image segmentation. When training the network, we used a binary image that is acquired by the manual segmentation of a plant from the background. Experimental results show that the U-Net based segmentation network can extract a plant from a digital image accurately.

  • PDF

GeoNet: Web-based Renotely Sensed Image Processing System (GeoNet: 웹 기반 위성영상 처리)

  • Ahn, Chung-Hyun;Kim, Kyung-Ok
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.8 no.2 s.16
    • /
    • pp.109-116
    • /
    • 2000
  • GeoNet is java-based remotely sensed image processing system. It is based on java Ibject-oriented paradigm and features cross-platform, web-based execution and extensibility to client/server remotely sensed image processing model. Remotely sensed image processing softwares made by java programming language can suggest alternatives to meet readily demand on remotely sensed image processing in proportion to increasement of remotely sensed data.

  • PDF

Image Classification Method Using Learning (학습을 이용한 영상 분류 방법)

  • Shin, Seong-Yoon;Lee, Hyun-Chang;Shin, Kwang-Seong
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.01a
    • /
    • pp.285-286
    • /
    • 2021
  • 본 논문에서는 변환 학습을 기반으로 한 다중 클래스 이미지 장면 분류 방법을 제안하도록 한다. ImageNet 대형 이미지 데이터 세트에서 사전 훈련 된 네트워크 모델을 사용하여 다중 클래스의 자연 장면 이미지를 분류하였다. 실험에서 최적화 된 ResNet 모델은 Kaggle의 Intel Image Classification 데이터 세트에 분류되어 우수한 결과를 얻었다.

  • PDF

DP-LinkNet: A convolutional network for historical document image binarization

  • Xiong, Wei;Jia, Xiuhong;Yang, Dichun;Ai, Meihui;Li, Lirong;Wang, Song
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.5
    • /
    • pp.1778-1797
    • /
    • 2021
  • Document image binarization is an important pre-processing step in document analysis and archiving. The state-of-the-art models for document image binarization are variants of encoder-decoder architectures, such as FCN (fully convolutional network) and U-Net. Despite their success, they still suffer from three limitations: (1) reduced feature map resolution due to consecutive strided pooling or convolutions, (2) multiple scales of target objects, and (3) reduced localization accuracy due to the built-in invariance of deep convolutional neural networks (DCNNs). To overcome these three challenges, we propose an improved semantic segmentation model, referred to as DP-LinkNet, which adopts the D-LinkNet architecture as its backbone, with the proposed hybrid dilated convolution (HDC) and spatial pyramid pooling (SPP) modules between the encoder and the decoder. Extensive experiments are conducted on recent document image binarization competition (DIBCO) and handwritten document image binarization competition (H-DIBCO) benchmark datasets. Results show that our proposed DP-LinkNet outperforms other state-of-the-art techniques by a large margin. Our implementation and the pre-trained models are available at https://github.com/beargolden/DP-LinkNet.

Image-Based Application Testing Method Using Faster D2-Net for Identification of the Same Image (동일 이미지 판별을 위해 Faster D2-Net을 이용한 이미지 기반의 애플리케이션 테스트 방법)

  • Chun, Hye-Won;Jo, Min-Seok;Han, Sung-Soo;Jeong, Chang-Sung
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.2
    • /
    • pp.87-92
    • /
    • 2022
  • Image-based application testing proposes an application testing method via image structure comparison. This test method allows testing on various devices without relying on various types of device operating systems or GUI. Traditional studies required the creation of a tester for each variant in the existing case, because it differs from the correct image for operating system changes, screen animation execution, and resolution changes. The study determined that the screen is the same for variations. The tester compares the underlying structure of the objects in the two images and extracts the regions in which the differences exist in the images, and compares image similarity as characteristic points of the Faster D2-Net. The development of the Faster D2-Net reduced the number of operations and spatial losses compared to the D2-Net, making it suitable for extracting features from application images and reducing test performance time.

GRAYSCALE IMAGE COLORIZATION USING A CONVOLUTIONAL NEURAL NETWORK

  • JWA, MINJE;KANG, MYUNGJOO
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.25 no.2
    • /
    • pp.26-38
    • /
    • 2021
  • Image coloration refers to adding plausible colors to a grayscale image or video. Image coloration has been used in many modern fields, including restoring old photographs, as well as reducing the time spent painting cartoons. In this paper, a method is proposed for colorizing grayscale images using a convolutional neural network. We propose an encoder-decoder model, adapting FusionNet to our purpose. A proper loss function is defined instead of the MSE loss function to suit the purpose of coloring. The proposed model was verified using the ImageNet dataset. We quantitatively compared several colorization models with ours, using the peak signal-to-noise ratio (PSNR) metric. In addition, to qualitatively evaluate the results, our model was applied to images in the test dataset and compared to images applied to various other models. Finally, we applied our model to a selection of old black and white photographs.

A new lightweight network based on MobileNetV3

  • Zhao, Liquan;Wang, Leilei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.1
    • /
    • pp.1-15
    • /
    • 2022
  • The MobileNetV3 is specially designed for mobile devices with limited memory and computing power. To reduce the network parameters and improve the network inference speed, a new lightweight network is proposed based on MobileNetV3. Firstly, to reduce the computation of residual blocks, a partial residual structure is designed by dividing the input feature maps into two parts. The designed partial residual structure is used to replace the residual block in MobileNetV3. Secondly, a dual-path feature extraction structure is designed to further reduce the computation of MobileNetV3. Different convolution kernel sizes are used in the two paths to extract feature maps with different sizes. Besides, a transition layer is also designed for fusing features to reduce the influence of the new structure on accuracy. The CIFAR-100 dataset and Image Net dataset are used to test the performance of the proposed partial residual structure. The ResNet based on the proposed partial residual structure has smaller parameters and FLOPs than the original ResNet. The performance of improved MobileNetV3 is tested on CIFAR-10, CIFAR-100 and ImageNet image classification task dataset. Comparing MobileNetV3, GhostNet and MobileNetV2, the improved MobileNetV3 has smaller parameters and FLOPs. Besides, the improved MobileNetV3 is also tested on CPU and Raspberry Pi. It is faster than other networks