• Title/Summary/Keyword: AI image analysis

Search Result 175, Processing Time 0.025 seconds

Analysis Method of influence of input for Image recognition result of machine learning (기계습의 영상인식결과에 대한 입력영상의 영향도 분석 기법)

  • Kim, Do-Wan;Kim, Woo-seong;Lee, Eun-hun;Kim, Hyeoncheol
    • Proceedings of The KACE
    • /
    • 2017.08a
    • /
    • pp.209-211
    • /
    • 2017
  • 기계학습은 인공지능(AI, Artificial Intelligence)의 일종으로 다른 인공지능 알고리즘이 정해진 규칙을 기반으로 주어진 임무(Task)를 해결하는 것과는 달리, 기계학습은 수집된 Data를 기반으로 최적의 솔루션을 학습한 후 미래의 값들을 예측하거나 해석하는 방법을 사용하고 있다. 더욱이 인터넷을 통한 연결성의 확대와 컴퓨터의 연산능력 발전으로 가능하게 된 Big-Data를 기반으로 하고 있어 이전의 인공지능 알고리즘에 비해 월등한 성능을 보여주고 있다. 그러나 기계학습 알고리즘이 Data를 학습할 때 학습 결과를 사람이 해석하기에 너무 복잡하여 사람이 그 내부 구조를 이해하는 것은 사실상 불가능하고, 이에 따라 학습된 기계학습 모델의 단점 또는 한계 등을 알지 못하는 문제가 있다. 본 연구에서는 이러한 블랙박스화된 기계학습 알고리즘의 특성을 이해하기 위해, 기계학습 알고리즘이 특정 입력에 대한 결과를 예측할 때 어떤 입력들로 부터 영향을 많이 받는지 그리고 어떤 입력으로부터 영향을 적게 받는지를 알아보는 방법을 소개하고 기존 연구의 단점을 개선하기 위한 방법을 제시한다.

  • PDF

Real-time Camera and Video Streaming Through Optimized Settings of Ethernet AVB in Vehicle Network System

  • An, Byoungman;Kim, Youngseop
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.8
    • /
    • pp.3025-3047
    • /
    • 2021
  • This paper presents the latest Ethernet standardization of in-vehicle network and the future trends of automotive Ethernet technology. The proposed system provides design and optimization algorithms for automotive networking technology related to AVB (Audio Video Bridge) technology. We present a design of in-vehicle network system as well as the optimization of AVB for automotive. A proposal of Reduced Latency of Machine to Machine (RLMM) plays an outstanding role in reducing the latency among devices. RLMM's approach to real-world experimental cases indicates a reduction in latency of around 41.2%. The setup optimized for the automotive network environment is expected to significantly reduce the time in the development and design process. The results obtained in the study of image transmission latency are trustworthy because average values were collected over a long period of time. It is necessary to analyze a latency between multimedia devices within limited time which will be of considerable benefit to the industry. Furthermore, the proposed reliable camera and video streaming through optimized AVB device settings would provide a high level of support in the real-time comprehension and analysis of images with AI (Artificial Intelligence) algorithms in autonomous driving.

Comparison of Classification Rules Regarding SaMD Between the Regulation EU 2017/745 and the Directive 93/42/EEC

  • Ryu, Gyuha;Lee, Jiyoon
    • Journal of Biomedical Engineering Research
    • /
    • v.42 no.6
    • /
    • pp.277-286
    • /
    • 2021
  • The global market size of AI based SaMD for medical image in 2023 will be anticipated to reach around 620 billion won (518 million dollars). In order for Korean manufacturers to efficiently obtain CE marking for marketing in the EU countries, the paper is to introduce the recommendation and suggestion of how to reclassify SaMD based on classification rules of MDR because, after introducing the Regulation EU 2017/745, classification rules are quite modified and newly added compared to the Directive 93/42/EEC. In addition, the paper is to provide several rules of MDR that may be applicable to decide the classification of SaMD. Lastly, the paper is to examine and demonstrate various secondary data supported by qualitative data because the paper focuses on the suggestion and recommendation with a public trust on the basis of various secondary data conducted by the analysis of field data. In conclusion, the paper found that the previous classification of SaMD followed by the rule of MDD should be reclassified based on the Regulation EU 2017/745. Therefore, the suggestion and recommendation are useful for Korean manufacturers to comprehend the classification of SaMD for marketing in the EU countries.

Automatic Poster Generation System Using Protagonist Face Analysis

  • Yeonhwi You;Sungjung Yong;Hyogyeong Park;Seoyoung Lee;Il-Young Moon
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.4
    • /
    • pp.287-293
    • /
    • 2023
  • With the rapid development of domestic and international over-the-top markets, a large amount of video content is being created. As the volume of video content increases, consumers tend to increasingly check data concerning the videos before watching them. To address this demand, video summaries in the form of plot descriptions, thumbnails, posters, and other formats are provided to consumers. This study proposes an approach that automatically generates posters to effectively convey video content while reducing the cost of video summarization. In the automatic generation of posters, face recognition and clustering are used to gather and classify character data, and keyframes from the video are extracted to learn the overall atmosphere of the video. This study used the facial data of the characters and keyframes as training data and employed technologies such as DreamBooth, a text-to-image generation model, to automatically generate video posters. This process significantly reduces the time and cost of video-poster production.

Real time instruction classification system

  • Sang-Hoon Lee;Dong-Jin Kwon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.3
    • /
    • pp.212-220
    • /
    • 2024
  • A recently the advancement of society, AI technology has made significant strides, especially in the fields of computer vision and voice recognition. This study introduces a system that leverages these technologies to recognize users through a camera and relay commands within a vehicle based on voice commands. The system uses the YOLO (You Only Look Once) machine learning algorithm, widely used for object and entity recognition, to identify specific users. For voice command recognition, a machine learning model based on spectrogram voice analysis is employed to identify specific commands. This design aims to enhance security and convenience by preventing unauthorized access to vehicles and IoT devices by anyone other than registered users. We converts camera input data into YOLO system inputs to determine if it is a person, Additionally, it collects voice data through a microphone embedded in the device or computer, converting it into time-domain spectrogram data to be used as input for the voice recognition machine learning system. The input camera image data and voice data undergo inference tasks through pre-trained models, enabling the recognition of simple commands within a limited space based on the inference results. This study demonstrates the feasibility of constructing a device management system within a confined space that enhances security and user convenience through a simple real-time system model. Finally our work aims to provide practical solutions in various application fields, such as smart homes and autonomous vehicles.

Weather Recognition Based on 3C-CNN

  • Tan, Ling;Xuan, Dawei;Xia, Jingming;Wang, Chao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.8
    • /
    • pp.3567-3582
    • /
    • 2020
  • Human activities are often affected by weather conditions. Automatic weather recognition is meaningful to traffic alerting, driving assistance, and intelligent traffic. With the boost of deep learning and AI, deep convolutional neural networks (CNN) are utilized to identify weather situations. In this paper, a three-channel convolutional neural network (3C-CNN) model is proposed on the basis of ResNet50.The model extracts global weather features from the whole image through the ResNet50 branch, and extracts the sky and ground features from the top and bottom regions by two CNN5 branches. Then the global features and the local features are merged by the Concat function. Finally, the weather image is classified by Softmax classifier and the identification result is output. In addition, a medium-scale dataset containing 6,185 outdoor weather images named WeatherDataset-6 is established. 3C-CNN is used to train and test both on the Two-class Weather Images and WeatherDataset-6. The experimental results show that 3C-CNN achieves best on both datasets, with the average recognition accuracy up to 94.35% and 95.81% respectively, which is superior to other classic convolutional neural networks such as AlexNet, VGG16, and ResNet50. It is prospected that our method can also work well for images taken at night with further improvement.

The Design of Application Model using Manufacturing Data in Protection Film Process for Smart Manufacturing Innovation (스마트 제조혁신을 위한 보호필름 공정 제조데이터의 활용모델 설계)

  • Cha, ByungRae;Park, Sun;Lee, Seong-ho;Shin, Byeong-Chun;Kim, JongWon
    • Smart Media Journal
    • /
    • v.8 no.3
    • /
    • pp.95-103
    • /
    • 2019
  • The global manufacturing industry has reached the limit to growth due to a long-term recession, the rise of labor cost and raw material. As a solution to these difficulties, we promote the 4th Industry Revolution based on ICT and sensor technology. Following this trend, this paper proposes the design of a model using manufacturing data in the protection film process for smart manufacturing innovation. In the protective film process, the manufacturing data of temperature, pressure, humidity, and motion and thermal image are acquired by various sensors for the raw material blending, stirring, extrusion, and inspection processes. While the acquired manufacturing data is stored in mass storage, A.I. platform provides time-series image analysis and its visualization.

Study of Black Ice Detection Method through Color Image Analysis (컬러 이미지 분석을 통한 블랙 아이스 검출 방법 연구)

  • Park, Pill-Won;Han, Seong-Soo
    • Journal of Platform Technology
    • /
    • v.9 no.4
    • /
    • pp.90-96
    • /
    • 2021
  • Most of the vehicles currently under development and in operation are equipped with various IoT sensors, but some of the factors that cause car accidents are relatively difficult to detect. One of the major risk factors among these factors is black ice. Black ice is one of the factors most likely to cause major accidents, as it can affect all vehicles passing through areas covered with black ice. Therefore, black ice detection technique is essential to prevent major accidents. For this purpose, some studies have been carried out in the past, but unrealistic factors have been reflected in some parts, so research to supplement this is needed. In this paper, we tried to detect black ice by analyzing color images using the CNN technique, and we succeeded in detecting black ice to a certain level. However, there were differences from previous studies, and the reason was analyzed.

Escape Route Prediction and Tracking System using Artificial Intelligence (인공지능을 활용한 도주경로 예측 및 추적 시스템)

  • Yang, Bum-Suk;Park, Dea-Woo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.8
    • /
    • pp.1130-1135
    • /
    • 2022
  • In Seoul, about 75,000 CCTVs are installed in 25 district offices. Each ward office has built a control center for CCTV control and is performing 24-hour CCTV video control for the safety of citizens. Seoul Metropolitan Government is building a smart city integrated platform that is safe for citizens by providing CCTV images of the ward office to enable rapid response to emergency/emergency situations by signing an MOU with related organizations. In this paper, when an incident occurs at the Seoul Metropolitan Government Office, the escape route is predicted by discriminating people and vehicles using the AI DNN-based Template Matching technology, MLP algorithm and CNN-based YOLO SPP DNN model for CCTV images. In addition, it is designed to automatically disseminate image information and situation information to adjacent ward offices when vehicles and people escape from the competent ward office. The escape route prediction and tracking system using artificial intelligence can expand the smart city integrated platform nationwide.

A Research on Image Metadata Extraction through YCrCb Color Model Analysis for Media Hyper-personalization Recommendation (미디어 초개인화 추천을 위한 YCrCb 컬러 모델 분석을 통한 영상의 메타데이터 추출에 대한 연구)

  • Park, Hyo-Gyeong;Yong, Sung-Jung;You, Yeon-Hwi;Moon, Il-Young
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.277-280
    • /
    • 2021
  • Recently as various contents are mass produced based on high accessibility, the media contents market is more active. Users want to find content that suits their taste, and each platform is competing for personalized recommendations for content. For an efficient recommendation system, high-quality metadata is required. Existing platforms take a method in which the user directly inputs the metadata of an image. This will waste time and money processing large amounts of data. In this paper, for media hyperpersonalization recommendation, keyframes are extracted based on the YCrCb color model of the video based on movie trailers, movie genres are distinguished through supervised learning of artificial intelligence and In the future, we would like to propose a utilization plan for generating metadata.

  • PDF