• Title/Summary/Keyword: Image Learning

Search Result 3,176, Processing Time 0.033 seconds

Object detection technology trend and development direction using deep learning

  • Kwak, NaeJoung;Kim, DongJu
    • International Journal of Advanced Culture Technology
    • /
    • v.8 no.4
    • /
    • pp.119-128
    • /
    • 2020
  • Object detection is an important field of computer vision and is applied to applications such as security, autonomous driving, and face recognition. Recently, as the application of artificial intelligence technology including deep learning has been applied in various fields, it has become a more powerful tool that can learn meaningful high-level, deeper features, solving difficult problems that have not been solved. Therefore, deep learning techniques are also being studied in the field of object detection, and algorithms with excellent performance are being introduced. In this paper, a deep learning-based object detection algorithm used to detect multiple objects in an image is investigated, and future development directions are presented.

A Review of 3D Object Tracking Methods Using Deep Learning (딥러닝 기술을 이용한 3차원 객체 추적 기술 리뷰)

  • Park, Hanhoon
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.22 no.1
    • /
    • pp.30-37
    • /
    • 2021
  • Accurate 3D object tracking with camera images is a key enabling technology for augmented reality applications. Motivated by the impressive success of convolutional neural networks (CNNs) in computer vision tasks such as image classification, object detection, image segmentation, recent studies for 3D object tracking have focused on leveraging deep learning. In this paper, we review deep learning approaches for 3D object tracking. We describe key methods in this field and discuss potential future research directions.

GAN-based Color Palette Extraction System by Chroma Fine-tuning with Reinforcement Learning

  • Kim, Sanghyuk;Kang, Suk-Ju
    • Journal of Semiconductor Engineering
    • /
    • v.2 no.1
    • /
    • pp.125-129
    • /
    • 2021
  • As the interest of deep learning, techniques to control the color of images in image processing field are evolving together. However, there is no clear standard for color, and it is not easy to find a way to represent only the color itself like the color-palette. In this paper, we propose a novel color palette extraction system by chroma fine-tuning with reinforcement learning. It helps to recognize the color combination to represent an input image. First, we use RGBY images to create feature maps by transferring the backbone network with well-trained model-weight which is verified at super resolution convolutional neural networks. Second, feature maps are trained to 3 fully connected layers for the color-palette generation with a generative adversarial network (GAN). Third, we use the reinforcement learning method which only changes chroma information of the GAN-output by slightly moving each Y component of YCbCr color gamut of pixel values up and down. The proposed method outperforms existing color palette extraction methods as given the accuracy of 0.9140.

Synthetic Image Generation for Military Vehicle Detection (군용물체탐지 연구를 위한 가상 이미지 데이터 생성)

  • Se-Yoon Oh;Hunmin Yang
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.26 no.5
    • /
    • pp.392-399
    • /
    • 2023
  • This research paper investigates the effectiveness of using computer graphics(CG) based synthetic data for deep learning in military vehicle detection. In particular, we explore the use of synthetic image generation techniques to train deep neural networks for object detection tasks. Our approach involves the generation of a large dataset of synthetic images of military vehicles, which is then used to train a deep learning model. The resulting model is then evaluated on real-world images to measure its effectiveness. Our experimental results show that synthetic training data alone can achieve effective results in object detection. Our findings demonstrate the potential of CG-based synthetic data for deep learning and suggest its value as a tool for training models in a variety of applications, including military vehicle detection.

Machine Learning Techniques for Diabetic Retinopathy Detection: A Review

  • Rachna Kumari;Sanjeev Kumar;Sunila Godara
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.4
    • /
    • pp.67-76
    • /
    • 2024
  • Diabetic retinopathy is a threatening complication of diabetes, caused by damaged blood vessels of light sensitive areas of retina. DR leads to total or partial blindness if left untreated. DR does not give any symptoms at early stages so earlier detection of DR is a big challenge for proper treatment of diseases. With advancement of technology various computer-aided diagnostic programs using image processing and machine learning approaches are designed for early detection of DR so that proper treatment can be provided to the patients for preventing its harmful effects. Now a day machine learning techniques are widely applied for image processing. These techniques also provide amazing result in this field also. In this paper we discuss various machine learning and deep learning based techniques developed for automatic detection of Diabetic Retinopathy.

The Optimal GSD and Image Size for Deep Learning Semantic Segmentation Training of Drone Images of Winter Vegetables (드론 영상으로부터 월동 작물 분류를 위한 의미론적 분할 딥러닝 모델 학습 최적 공간 해상도와 영상 크기 선정)

  • Chung, Dongki;Lee, Impyeong
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.6_1
    • /
    • pp.1573-1587
    • /
    • 2021
  • A Drone image is an ultra-high-resolution image that is several or tens of times higher in spatial resolution than a satellite or aerial image. Therefore, drone image-based remote sensing is different from traditional remote sensing in terms of the level of object to be extracted from the image and the amount of data to be processed. In addition, the optimal scale and size of data used for model training is different depending on the characteristics of the applied deep learning model. However, moststudies do not consider the size of the object to be found in the image, the spatial resolution of the image that reflects the scale, and in many cases, the data specification used in the model is applied as it is before. In this study, the effect ofspatial resolution and image size of drone image on the accuracy and training time of the semantic segmentation deep learning model of six wintering vegetables was quantitatively analyzed through experiments. As a result of the experiment, it was found that the average accuracy of dividing six wintering vegetablesincreases asthe spatial resolution increases, but the increase rate and convergence section are different for each crop, and there is a big difference in accuracy and time depending on the size of the image at the same resolution. In particular, it wasfound that the optimal resolution and image size were different from each crop. The research results can be utilized as data for getting the efficiency of drone images acquisition and production of training data when developing a winter vegetable segmentation model using drone images.

A case study on high school students' mental image in the process of solving regular polyhedron problems (정다면체 문제 해결 과정에서 나타나는 고등학교 학생들의 심상에 관한 사례연구)

  • Hong, Gap Lyung;Kim, Won Kyung
    • The Mathematical Education
    • /
    • v.53 no.4
    • /
    • pp.493-507
    • /
    • 2014
  • The purpose of this study is to analyze how high school students form and interpret the mental image in the process of solving regular polyhedron problems. For this purpose, a set of problems about the regular polyhedron's vertex is developed on the base of the regular polyhedron's duality and circulation. and applied to 2 students of the 12th graders in D high school. After 2 hours of teaching and learning and another 2 hours of mental image-analysis process, the following research findings are obtained. Fisrt, a student who recorded medium high-level grade in the national scholastic test can build the dynamic image or the patten image in the process of solving regular polyhedron's vertex problems by utilizing the 3D geometry program. However, the other student who recorded low-level grade can build the concrete-pictorial image. Second, pattern image or dynamic image can help students solve the regular polyhedron's vertex problems by proper transformation of informations and the mental images while the concrete-pictorial image does not help. Hence, it is recommended that the mathematics teachers should develop teaching and learning materials about the regular polyhedron's duality and circulation and also give students suitable questions to build the various mental images.

GAN-based Image-to-image Translation using Multi-scale Images (다중 스케일 영상을 이용한 GAN 기반 영상 간 변환 기법)

  • Chung, Soyoung;Chung, Min Gyo
    • The Journal of the Convergence on Culture Technology
    • /
    • v.6 no.4
    • /
    • pp.767-776
    • /
    • 2020
  • GcGAN is a deep learning model to translate styles between images under geometric consistency constraint. However, GcGAN has a disadvantage that it does not properly maintain detailed content of an image, since it preserves the content of the image through limited geometric transformation such as rotation or flip. Therefore, in this study, we propose a new image-to-image translation method, MSGcGAN(Multi-Scale GcGAN), which improves this disadvantage. MSGcGAN, an extended model of GcGAN, performs style translation between images in a direction to reduce semantic distortion of images and maintain detailed content by learning multi-scale images simultaneously and extracting scale-invariant features. The experimental results showed that MSGcGAN was better than GcGAN in both quantitative and qualitative aspects, and it translated the style more naturally while maintaining the overall content of the image.

GAN System Using Noise for Image Generation (이미지 생성을 위해 노이즈를 이용한 GAN 시스템)

  • Bae, Sangjung;Kim, Mingyu;Jung, Hoekyung
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.6
    • /
    • pp.700-705
    • /
    • 2020
  • Generative adversarial networks are methods of generating images by opposing two neural networks. When generating the image, randomly generated noise is rearranged to generate the image. The image generated by this method is not generated well depending on the noise, and it is difficult to generate a proper image when the number of pixels of the image is small In addition, the speed and size of data accumulation in data classification increases, and there are many difficulties in labeling them. In this paper, to solve this problem, we propose a technique to generate noise based on random noise using real data. Since the proposed system generates an image based on the existing image, it is confirmed that it is possible to generate a more natural image, and if it is used for learning, it shows a higher hit rate than the existing method using the hostile neural network respectively.

Anomaly Detection Methodology Based on Multimodal Deep Learning (멀티모달 딥 러닝 기반 이상 상황 탐지 방법론)

  • Lee, DongHoon;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.2
    • /
    • pp.101-125
    • /
    • 2022
  • Recently, with the development of computing technology and the improvement of the cloud environment, deep learning technology has developed, and attempts to apply deep learning to various fields are increasing. A typical example is anomaly detection, which is a technique for identifying values or patterns that deviate from normal data. Among the representative types of anomaly detection, it is very difficult to detect a contextual anomaly that requires understanding of the overall situation. In general, detection of anomalies in image data is performed using a pre-trained model trained on large data. However, since this pre-trained model was created by focusing on object classification of images, there is a limit to be applied to anomaly detection that needs to understand complex situations created by various objects. Therefore, in this study, we newly propose a two-step pre-trained model for detecting abnormal situation. Our methodology performs additional learning from image captioning to understand not only mere objects but also the complicated situation created by them. Specifically, the proposed methodology transfers knowledge of the pre-trained model that has learned object classification with ImageNet data to the image captioning model, and uses the caption that describes the situation represented by the image. Afterwards, the weight obtained by learning the situational characteristics through images and captions is extracted and fine-tuning is performed to generate an anomaly detection model. To evaluate the performance of the proposed methodology, an anomaly detection experiment was performed on 400 situational images and the experimental results showed that the proposed methodology was superior in terms of anomaly detection accuracy and F1-score compared to the existing traditional pre-trained model.