• Title/Summary/Keyword: Deep Learning AI

Search Result 622, Processing Time 0.029 seconds

Detection of video editing points using facial keypoints (얼굴 특징점을 활용한 영상 편집점 탐지)

  • Joshep Na;Jinho Kim;Jonghyuk Park
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.4
    • /
    • pp.15-30
    • /
    • 2023
  • Recently, various services using artificial intelligence(AI) are emerging in the media field as well However, most of the video editing, which involves finding an editing point and attaching the video, is carried out in a passive manner, requiring a lot of time and human resources. Therefore, this study proposes a methodology that can detect the edit points of video according to whether person in video are spoken by using Video Swin Transformer. First, facial keypoints are detected through face alignment. To this end, the proposed structure first detects facial keypoints through face alignment. Through this process, the temporal and spatial changes of the face are reflected from the input video data. And, through the Video Swin Transformer-based model proposed in this study, the behavior of the person in the video is classified. Specifically, after combining the feature map generated through Video Swin Transformer from video data and the facial keypoints detected through Face Alignment, utterance is classified through convolution layers. In conclusion, the performance of the image editing point detection model using facial keypoints proposed in this paper improved from 87.46% to 89.17% compared to the model without facial keypoints.

Conventional Versus Artificial Intelligence-Assisted Interpretation of Chest Radiographs in Patients With Acute Respiratory Symptoms in Emergency Department: A Pragmatic Randomized Clinical Trial

  • Eui Jin Hwang;Jin Mo Goo;Ju Gang Nam;Chang Min Park;Ki Jeong Hong;Ki Hong Kim
    • Korean Journal of Radiology
    • /
    • v.24 no.3
    • /
    • pp.259-270
    • /
    • 2023
  • Objective: It is unknown whether artificial intelligence-based computer-aided detection (AI-CAD) can enhance the accuracy of chest radiograph (CR) interpretation in real-world clinical practice. We aimed to compare the accuracy of CR interpretation assisted by AI-CAD to that of conventional interpretation in patients who presented to the emergency department (ED) with acute respiratory symptoms using a pragmatic randomized controlled trial. Materials and Methods: Patients who underwent CRs for acute respiratory symptoms at the ED of a tertiary referral institution were randomly assigned to intervention group (with assistance from an AI-CAD for CR interpretation) or control group (without AI assistance). Using a commercial AI-CAD system (Lunit INSIGHT CXR, version 2.0.2.0; Lunit Inc.). Other clinical practices were consistent with standard procedures. Sensitivity and false-positive rates of CR interpretation by duty trainee radiologists for identifying acute thoracic diseases were the primary and secondary outcomes, respectively. The reference standards for acute thoracic disease were established based on a review of the patient's medical record at least 30 days after the ED visit. Results: We randomly assigned 3576 participants to either the intervention group (1761 participants; mean age ± standard deviation, 65 ± 17 years; 978 males; acute thoracic disease in 472 participants) or the control group (1815 participants; 64 ± 17 years; 988 males; acute thoracic disease in 491 participants). The sensitivity (67.2% [317/472] in the intervention group vs. 66.0% [324/491] in the control group; odds ratio, 1.02 [95% confidence interval, 0.70-1.49]; P = 0.917) and false-positive rate (19.3% [249/1289] vs. 18.5% [245/1324]; odds ratio, 1.00 [95% confidence interval, 0.79-1.26]; P = 0.985) of CR interpretation by duty radiologists were not associated with the use of AI-CAD. Conclusion: AI-CAD did not improve the sensitivity and false-positive rate of CR interpretation for diagnosing acute thoracic disease in patients with acute respiratory symptoms who presented to the ED.

The Study on Implementation of Crime Terms Classification System for Crime Issues Response

  • Jeong, Inkyu;Yoon, Cheolhee;Kang, Jang Mook
    • International Journal of Advanced Culture Technology
    • /
    • v.8 no.3
    • /
    • pp.61-72
    • /
    • 2020
  • The fear of crime, discussed in the early 1960s in the United States, is a psychological response, such as anxiety or concern about crime, the potential victim of a crime. These anxiety factors lead to the burden of the individual in securing the psychological stability and indirect costs of the crime against the society. Fear of crime is not a good thing, and it is a part that needs to be adjusted so that it cannot be exaggerated and distorted by the policy together with the crime coping and resolution. This is because fear of crime has as much harm as damage caused by criminal act. Eric Pawson has argued that the popular impression of violent crime is not formed because of media reports, but by official statistics. Therefore, the police should watch and analyze news related to fear of crime to reduce the social cost of fear of crime and prepare a preemptive response policy before the people have 'fear of crime'. In this paper, we propose a deep - based news classification system that helps police cope with crimes related to crimes reported in the media efficiently and quickly and precisely. The goal is to establish a system that can quickly identify changes in security issues that are rapidly increasing by categorizing news related to crime among news articles. To construct the system, crime data was learned so that news could be classified according to the type of crime. Deep learning was applied by using Google tensor flow. In the future, it is necessary to continue research on the importance of keyword according to early detection of issues that are rapidly increasing by crime type and the power of the press, and it is also necessary to constantly supplement crime related corpus.

AI Analysis Method Utilizing Ingestible Bio-Sensors for Bovine Calving Predictions

  • Kim, Heejin;Min, Younjeong;Choi, Changhyuk;Choi, Byoungju
    • The Journal of Korean Institute of Information Technology
    • /
    • v.16 no.12
    • /
    • pp.127-137
    • /
    • 2018
  • Parturition is an important event for farmers as it provides economic gains for the farms. Thus, the effective management of parturition is essential to farm management. In particular, the unit price of cattle is higher than other livestock and the productivity of cattle is closely associated to farm income. In addition, 42% of calving occurs in the nighttime so accurate parturition predictions are all the more important. In this paper, we propose a method that accurately predicts the calving date by applying core body temperature of cattle to deep learning. The body temperature of cattle can be measured without being influenced by the ambient environment by applying an ingestible bio-sensor in the cattle's rumen. By experiment on cattle, we confirmed this method to be more accurate for predicting calving dates than existing parturition prediction methods, showing an average of 3 hour 40 minute error. This proposed method is expected to reduce the economic damages of farms by accurately predicting calving times and assisting in successful parturitions.

CNN-based Building Recognition Method Robust to Image Noises (이미지 잡음에 강인한 CNN 기반 건물 인식 방법)

  • Lee, Hyo-Chan;Park, In-hag;Im, Tae-ho;Moon, Dai-Tchul
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.3
    • /
    • pp.341-348
    • /
    • 2020
  • The ability to extract useful information from an image, such as the human eye, is an interface technology essential for AI computer implementation. The building recognition technology has a lower recognition rate than other image recognition technologies due to the various building shapes, the ambient noise images according to the season, and the distortion by angle and distance. The computer vision based building recognition algorithms presented so far has limitations in discernment and expandability due to manual definition of building characteristics. This paper introduces the deep learning CNN (Convolutional Neural Network) model, and proposes new method to improve the recognition rate even by changes of building images caused by season, illumination, angle and perspective. This paper introduces the partial images that characterize the building, such as windows or wall images, and executes the training with whole building images. Experimental results show that the building recognition rate is improved by about 14% compared to the general CNN model.

Performance Analysis of Cloud-Net with Cross-sensor Training Dataset for Satellite Image-based Cloud Detection

  • Kim, Mi-Jeong;Ko, Yun-Ho
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.1
    • /
    • pp.103-110
    • /
    • 2022
  • Since satellite images generally include clouds in the atmosphere, it is essential to detect or mask clouds before satellite image processing. Clouds were detected using physical characteristics of clouds in previous research. Cloud detection methods using deep learning techniques such as CNN or the modified U-Net in image segmentation field have been studied recently. Since image segmentation is the process of assigning a label to every pixel in an image, precise pixel-based dataset is required for cloud detection. Obtaining accurate training datasets is more important than a network configuration in image segmentation for cloud detection. Existing deep learning techniques used different training datasets. And test datasets were extracted from intra-dataset which were acquired by same sensor and procedure as training dataset. Different datasets make it difficult to determine which network shows a better overall performance. To verify the effectiveness of the cloud detection network such as Cloud-Net, two types of networks were trained using the cloud dataset from KOMPSAT-3 images provided by the AIHUB site and the L8-Cloud dataset from Landsat8 images which was publicly opened by a Cloud-Net author. Test data from intra-dataset of KOMPSAT-3 cloud dataset were used for validating the network. The simulation results show that the network trained with KOMPSAT-3 cloud dataset shows good performance on the network trained with L8-Cloud dataset. Because Landsat8 and KOMPSAT-3 satellite images have different GSDs, making it difficult to achieve good results from cross-sensor validation. The network could be superior for intra-dataset, but it could be inferior for cross-sensor data. It is necessary to study techniques that show good results in cross-senor validation dataset in the future.

Real-time traffic light information recognition based on object detection models (객체 인식 모델 기반 실시간 교통신호 정보 인식)

  • Joo, eun-oh;Kim, Min-Soo
    • Journal of Cadastre & Land InformatiX
    • /
    • v.52 no.1
    • /
    • pp.81-93
    • /
    • 2022
  • Recently, there have been many studies on object recognition around the vehicle and recognition of traffic signs and traffic lights in autonomous driving. In particular, such the recognition of traffic lights is one of the core technologies in autonomous driving. Therefore, many studies for such the recognition of traffic lights have been performed, the studies based on various deep learning models have increased significantly in recent. In addition, as a high-quality AI training data set for voice, vision, and autonomous driving is released on AIHub, it makes it possible to develop a recognition model for traffic lights suitable for the domestic environment using the data set. In this study, we developed a recognition model for traffic lights that can be used in Korea using the AIHub's training data set. In particular, in order to improve the recognition performance, we used various models of YOLOv4 and YOLOv5, and performed our recognition experiments by defining various classes for the training data. In conclusion, we could see that YOLOv5 shows better performance in the recognition than YOLOv4 and could confirm the reason from the architecture comparison of the two models.

A Detailed Review on Recognition of Plant Disease Using Intelligent Image Retrieval Techniques

  • Gulbir Singh;Kuldeep Kumar Yogi
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.9
    • /
    • pp.77-90
    • /
    • 2023
  • Today, crops face many characteristics/diseases. Insect damage is one of the main characteristics/diseases. Insecticides are not always effective because they can be toxic to some birds. It will also disrupt the natural food chain for animals. A common practice of plant scientists is to visually assess plant damage (leaves, stems) due to disease based on the percentage of disease. Plants suffer from various diseases at any stage of their development. For farmers and agricultural professionals, disease management is a critical issue that requires immediate attention. It requires urgent diagnosis and preventive measures to maintain quality and minimize losses. Many researchers have provided plant disease detection techniques to support rapid disease diagnosis. In this review paper, we mainly focus on artificial intelligence (AI) technology, image processing technology (IP), deep learning technology (DL), vector machine (SVM) technology, the network Convergent neuronal (CNN) content Detailed description of the identification of different types of diseases in tomato and potato plants based on image retrieval technology (CBIR). It also includes the various types of diseases that typically exist in tomato and potato. Content-based Image Retrieval (CBIR) technologies should be used as a supplementary tool to enhance search accuracy by encouraging you to access collections of extra knowledge so that it can be useful. CBIR systems mainly use colour, form, and texture as core features, such that they work on the first level of the lowest level. This is the most sophisticated methods used to diagnose diseases of tomato plants.

SCLC-Edge Detection Algorithm for Skin Cancer Classification (피부암 병변 분류를 위한 SCLC-Edge 검출 알고리즘)

  • June-Young Park;Chang-Min Kim;Roy C. Park
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.23 no.4
    • /
    • pp.256-263
    • /
    • 2022
  • Skin cancer is one of the most common diseases in the world, and the incidence rate in Korea has increased by about 100% over the past five years. In the United States, more than 5 million people are diagnosed with skin cancer every year. Skin cancer mainly occurs when skin tissue is damaged for a long time due to exposure to ultraviolet rays. Melanoma, a malignant tumor of skin cancer, is similar in appearance to Atypical melanocytic nevus occurring on the skin, making it difficult for the general public to be aware of it unless secondary signs occur. In this paper, we propose a skin cancer lesion edge detection algorithm and a deep learning model, CRNN, which performs skin cancer lesion classification for early detection and classification of these skin cancers. As a result of the experiment, when using the contour detection algorithm proposed in this paper, the classification accuracy was the highest at 97%. For the Canny algorithm, 78% was shown, 55% for Sobel, and 46% for Laplacian.

Prediction of rainfall abstraction based on deep learning considering watershed and rainfall characteristic factors (유역 및 강우 특성인자를 고려한 딥러닝 기반의 강우손실 예측)

  • Jeong, Minyeob;Kim, Dae-Hong;Kim, Seokgyun
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.37-37
    • /
    • 2022
  • 유효우량 산정을 위하여 국내에서 주로 사용되는 모형은 NRCS-CN(Natural Resources Conservation Service - curve number) 모형으로, 유역의 유출 능력을 나타내는 유출곡선지수(runoff curve number, CN)와 같은 NRCS-CN 모형의 매개변수들은 관측 강우-유출자료 또는 토양도, 토지피복지도 등을 이용하여 유역마다 결정된 값이 사용되고 있다. 그러나 유역의 CN값은 유역의 토양 상태와 같은 환경적 조건에 따라 달라질 수 있으며, 이를 반영하기 위하여 선행토양함수조건(antecedent moisture condition, AMC)을 이용하여 CN값을 조정하는 방법이 사용되고 있으나, AMC 조건에 따른 CN 값의 갑작스런 변화는 유출량의 극단적인 변화를 가져올 수 있다. NRCS-CN 모형과 더불어 강우 손실량 산정에 많이 사용되는 모형으로 Green-Ampt 모형이 있다. Green-Ampt 모형은 유역에서 발생하는 침투현상의 물리적 과정을 고려하는 모형이라는 장점이 있으나, 모형에 활용되는 다양한 물리적인 매개변수들을 산정하기 위해서는 유역에 대한 많은 조사가 선행되어야 한다. 또한 이렇게 산정된 매개변수들은 유역 내 토양이나 식생 조건 등에 따른 여러 불확실성을 내포하고 있어 실무적용에 어려움이 있다. 따라서 본 연구에서는, 현재 사용되고 있는 강우손실 모형들의 매개변수를 추정하기 위한 방법을 제시하고자 하였다. 본 연구에서 제시하는 방법은 인공지능(AI) 기술 중 하나인 딥러닝(deep-learning) 기법을 기반으로 하고 있으며, 딥러닝 모형으로는 장단기 메모리(Long Short-Term Memory, LSTM) 모형이 활용되었다. 딥러닝 모형의 입력 데이터는 유역에서의 강우특성이나 토양수분, 증발산, 식생 특성들을 나타내는 인자이며, 모의 결과는 유역에서 발생한 총 유출량으로 강우손실 모형들의 매개변수 값들은 이들을 활용하여 도출될 수 있다. 산정된 매개변수 값들을 강우손실 모형에 적용하여 실제 유역들에서의 유효우량 산정에 활용해보았으며, 동역학파 기반의 강우-유출 모형을 사용하여 유출을 예측해보았다. 예측된 유출수문곡선을 관측 자료와 비교 시 NSE=0.5 이상으로 산정되어 유출이 적절히 예측되었음을 확인했다.

  • PDF