• Title/Summary/Keyword: vision artificial intelligence

Search Result 172, Processing Time 0.029 seconds

Guidelines for Data Construction when Estimating Traffic Volume based on Artificial Intelligence using Drone Images (드론영상과 인공지능 기반 교통량 추정을 위한 데이터 구축 가이드라인 도출 연구)

  • Han, Dongkwon;Kim, Doopyo;Kim, Sungbo
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.40 no.3
    • /
    • pp.147-157
    • /
    • 2022
  • Recently, many studies have been conducted to analyze traffic or object recognition that classifies vehicles through artificial intelligence-based prediction models using CCTV (Closed Circuit TeleVision)or drone images. In order to develop an object recognition deep learning model for accurate traffic estimation, systematic data construction is required, and related standardized guidelines are insufficient. In this study, previous studies were analyzed to derive guidelines for establishing artificial intelligence-based training data for traffic estimation using drone images, and business reports or training data for artificial intelligence and quality management guidelines were referenced. The guidelines for data construction are divided into data acquisition, preprocessing, and validation, and guidelines for notice and evaluation index for each item are presented. The guidelines for data construction aims to provide assistance in the development of a robust and generalized artificial intelligence model in analyzing the estimation of road traffic based on drone image artificial intelligence.

Application of artificial intelligence-based technologies to the construction sites (이미지 기반 인공지능을 활용한 현장 적용성 연구)

  • Na, Seunguk;Heo, Seokjae;Roh, Youngsook
    • Proceedings of the Korean Institute of Building Construction Conference
    • /
    • 2022.04a
    • /
    • pp.225-226
    • /
    • 2022
  • The construction industry, which has a labour-intensive and conservative nature, is exclusive to adopt new technologies. However, the construction industry is viably introducing the 4th Industrial Revolution technologies represented by artificial intelligence, Internet of Things, robotics and unmanned transportation to promote change into a smart industry. An image-based artificial intelligence technology is a field of computer vision technology that refers to machines mimicking human visual recognition of objects from pictures or videos. The purpose of this article is to explore image-based artificial intelligence technologies which would be able to apply to the construction sites. In this study, we show two examples which is one for a construction waste classification model and another for cast in-situ anchor bolts defection detection model. Image-based intelligence technologies would be used for various measurement, classification, and detection works that occur in the construction projects.

  • PDF

Digital Modelling of Visual Perception in Architectural Environment

  • Seo, Dong-Yeon;Lee, Kyung-Hoi
    • KIEAE Journal
    • /
    • v.3 no.2
    • /
    • pp.59-66
    • /
    • 2003
  • To be the design method supporting aesthetic ability of human, CAAD system should essentially recognize architectural form in the same way of human. In this study, visual perception process of human was analyzed to search proper computational method performing similar step of perception of it. Through the analysis of visual perception, vision was separated to low-level vision and high-level vision. Edge detection and neural network were selected to model after low-level vision and high-level vision. The 24 images of building, tree and landscape were processed by edge detection and trained by neural network. And 24 new images were used to test trained network. The test shows that trained network gives right perception result toward each images with low error rate. This study is on the meaning of artificial intelligence in design process rather than on the design automation strategy through artificial intelligence.

Crowd Activity Recognition using Optical Flow Orientation Distribution

  • Kim, Jinpyung;Jang, Gyujin;Kim, Gyujin;Kim, Moon-Hyun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.8
    • /
    • pp.2948-2963
    • /
    • 2015
  • In the field of computer vision, visual surveillance systems have recently become an important research topic. Growth in this area is being driven by both the increase in the availability of inexpensive computing devices and image sensors as well as the general inefficiency of manual surveillance and monitoring. In particular, the ultimate goal for many visual surveillance systems is to provide automatic activity recognition for events at a given site. A higher level of understanding of these activities requires certain lower-level computer vision tasks to be performed. So in this paper, we propose an intelligent activity recognition model that uses a structure learning method and a classification method. The structure learning method is provided as a K2-learning algorithm that generates Bayesian networks of causal relationships between sensors for a given activity. The statistical characteristics of the sensor values and the topological characteristics of the generated graphs are learned for each activity, and then a neural network is designed to classify the current activity according to the features extracted from the multiple sensor values that have been collected. Finally, the proposed method is implemented and tested by using PETS2013 benchmark data.

A review of space perception applicable to artificial intelligence robots (인공지능 로봇에 적용할 수 있는 공간지각에 대한 종설)

  • Lee, Young-Lim
    • Journal of Digital Convergence
    • /
    • v.17 no.10
    • /
    • pp.233-242
    • /
    • 2019
  • Numerous space perception studies have shown that Euclidean 3-D structure cannot be recovered from binocular stereopsis, motion, combination of stereopsis and motion, or even with combined multiple sources of optical information. Humans, however, have no difficulties to perform the task-specific action despite of poor shape perception. We have applied humans skill and capabilities to artificial intelligence and computer vision but those machines are still far behind from humans abilities. Thus, we need to understand how we perceive depth in space and what information we use to perceive 3-D structure accurately to perform. The purpose of this paper was to review space perception literatures to apply humans abilities to artificial intelligence robots more advanced in future.

Facial Manipulation Detection with Transformer-based Discriminative Features Learning Vision (트랜스포머 기반 판별 특징 학습 비전을 통한 얼굴 조작 감지)

  • Van-Nhan Tran;Minsu Kim;Philjoo Choi;Suk-Hwan Lee;Hoanh-Su Le;Ki-Ryong Kwon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.540-542
    • /
    • 2023
  • Due to the serious issues posed by facial manipulation technologies, many researchers are becoming increasingly interested in the identification of face forgeries. The majority of existing face forgery detection methods leverage powerful data adaptation ability of neural network to derive distinguishing traits. These deep learning-based detection methods frequently treat the detection of fake faces as a binary classification problem and employ softmax loss to track CNN network training. However, acquired traits observed by softmax loss are insufficient for discriminating. To get over these limitations, in this study, we introduce a novel discriminative feature learning based on Vision Transformer architecture. Additionally, a separation-center loss is created to simply compress intra-class variation of original faces while enhancing inter-class differences in the embedding space.

Analysis of Korea's Artificial Intelligence Competitiveness Based on Patent Data: Focusing on Patent Index and Topic Modeling (특허데이터 기반 한국의 인공지능 경쟁력 분석 : 특허지표 및 토픽모델링을 중심으로)

  • Lee, Hyun-Sang;Qiao, Xin;Shin, Sun-Young;Kim, Gyu-Ri;Oh, Se-Hwan
    • Informatization Policy
    • /
    • v.29 no.4
    • /
    • pp.43-66
    • /
    • 2022
  • With the development of artificial intelligence technology, competition for artificial intelligence technology patents around the world is intensifying. During the period 2000 ~ 2021, artificial intelligence technology patent applications at the US Patent and Trademark Office have been steadily increasing, and the growth rate has been steeper since the 2010s. As a result of analyzing Korea's artificial intelligence technology competitiveness through patent indices, it is evaluated that patent activity, impact, and marketability are superior in areas such as auditory intelligence and visual intelligence. However, compared to other countries, overall Korea's artificial intelligence technology patents are good in terms of activity and marketability, but somewhat inferior in technological impact. While noise canceling and voice recognition have recently decreased as topics for artificial intelligence, growth is expected in areas such as model learning optimization, smart sensors, and autonomous driving. In the case of Korea, efforts are required as there is a slight lack of patent applications in areas such as fraud detection/security and medical vision learning.

Prompt Tuning for Facial Action Unit Detection in the Wild

  • Vu Ngoc Tu;Huynh Van Thong;Aera Kim;Soo-Hyung Kim
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.05a
    • /
    • pp.732-734
    • /
    • 2023
  • Facial Action Units Detection (FAUs) problem focuses on identifying various detail units expressing on the human face, as defined by the Facial Action Coding System, which constitutes a fine-grained classification problem. This is a challenging task in computer vision. In this study, we propose a Prompt Tuning approach to address this problem, involving a 2-step training process. Our method demonstrates its effectiveness on the Affective in the Wild dataset, surpassing other existing methods in terms of both accuracy and efficiency.

DiLO: Direct light detection and ranging odometry based on spherical range images for autonomous driving

  • Han, Seung-Jun;Kang, Jungyu;Min, Kyoung-Wook;Choi, Jungdan
    • ETRI Journal
    • /
    • v.43 no.4
    • /
    • pp.603-616
    • /
    • 2021
  • Over the last few years, autonomous vehicles have progressed very rapidly. The odometry technique that estimates displacement from consecutive sensor inputs is an essential technique for autonomous driving. In this article, we propose a fast, robust, and accurate odometry technique. The proposed technique is light detection and ranging (LiDAR)-based direct odometry, which uses a spherical range image (SRI) that projects a three-dimensional point cloud onto a two-dimensional spherical image plane. Direct odometry is developed in a vision-based method, and a fast execution speed can be expected. However, applying LiDAR data is difficult because of the sparsity. To solve this problem, we propose an SRI generation method and mathematical analysis, two key point sampling methods using SRI to increase precision and robustness, and a fast optimization method. The proposed technique was tested with the KITTI dataset and real environments. Evaluation results yielded a translation error of 0.69%, a rotation error of 0.0031°/m in the KITTI training dataset, and an execution time of 17 ms. The results demonstrated high precision comparable with state-of-the-art and remarkably higher speed than conventional techniques.

A Study on Teaching of Convolution in Engineering Mathematics and Artificial Intelligence (인공지능에 활용되는 공학수학 합성곱(convolution) 교수·학습자료 연구)

  • Lee, Sang-Gu;Nam, Yun;Lee, Jae Hwa;Kim, Eung-Ki
    • Communications of Mathematical Education
    • /
    • v.37 no.2
    • /
    • pp.277-297
    • /
    • 2023
  • In mathematics, the concept of convolution is widely used. The convolution operation is required for understanding computer vision and deep learning in artificial intelligence. Therefore, it is vital for this concept to be explained in college mathematics education. In this paper, we present our new teaching and learning materials on convolution available for engineering mathematics. We provide the knowledge and applications on convolution with Python-based code, and introduce Convolutional Neural Network (CNN) used for image classification as an example. These materials can be utilized in class for the teaching of convolution and help students have a good understanding of the related knowledge in artificial intelligence.