• Title/Summary/Keyword: vision artificial intelligence

Search Result 183, Processing Time 0.038 seconds

A Deep Learning Method for Brain Tumor Classification Based on Image Gradient

  • Long, Hoang;Lee, Suk-Hwan;Kwon, Seong-Geun;Kwon, Ki-Ryong
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.8
    • /
    • pp.1233-1241
    • /
    • 2022
  • Tumors of the brain are the deadliest, with a life expectancy of only a few years for those with the most advanced forms. Diagnosing a brain tumor is critical to developing a treatment plan to help patients with the disease live longer. A misdiagnosis of brain tumors will lead to incorrect medical treatment, decreasing a patient's chance of survival. Radiologists classify brain tumors via biopsy, which takes a long time. As a result, the doctor will need an automatic classification system to identify brain tumors. Image classification is one application of the deep learning method in computer vision. One of the deep learning's most powerful algorithms is the convolutional neural network (CNN). This paper will introduce a novel deep learning structure and image gradient to classify brain tumors. Meningioma, glioma, and pituitary tumors are the three most popular forms of brain cancer represented in the Figshare dataset, which contains 3,064 T1-weighted brain images from 233 patients. According to the numerical results, our method is more accurate than other approaches.

AI-BASED Monitoring Of New Plant Growth Management System Design

  • Seung-Ho Lee;Seung-Jung Shin
    • International journal of advanced smart convergence
    • /
    • v.12 no.3
    • /
    • pp.104-108
    • /
    • 2023
  • This paper deals with research on innovative systems using Python-based artificial intelligence technology in the field of plant growth monitoring. The importance of monitoring and analyzing the health status and growth environment of plants in real time contributes to improving the efficiency and quality of crop production. This paper proposes a method of processing and analyzing plant image data using computer vision and deep learning technologies. The system was implemented using Python language and the main deep learning framework, TensorFlow, PyTorch. A camera system that monitors plants in real time acquires image data and provides it as input to a deep neural network model. This model was used to determine the growth state of plants, the presence of pests, and nutritional status. The proposed system provides users with information on plant state changes in real time by providing monitoring results in the form of visual or notification. In addition, it is also used to predict future growth conditions or anomalies by building data analysis and prediction models based on the collected data. This paper is about the design and implementation of Python-based plant growth monitoring systems, data processing and analysis methods, and is expected to contribute to important research areas for improving plant production efficiency and reducing resource consumption.

Real-time Ball Detection and Tracking with P-N Learning in Soccer Game (P-N 러닝을 이용한 실시간 축구공 검출 및 추적)

  • Huang, Shuai-Jie;Li, Gen;Lee, Yill-Byung
    • Annual Conference of KIPS
    • /
    • 2011.04a
    • /
    • pp.447-450
    • /
    • 2011
  • This paper shows the application of P-N Learning [4] method in the soccer ball detection and improvement for increasing the speed of processing. In the P-N learning, the learning process is guided by positive (P) and negative (N) constraints which restrict the labeling of the unlabeled data, identify examples that have been classified in contradiction with structural constraints and augment the training set with the corrected samples in an iterative process. But for the long-view in the soccer game, P-N learning will produce so many ferns that more time is spent than other methods. We propose that color histogram of each frame is constructed to delete the unnecessary details in order to decreasing the number of feature points. We use the mask to eliminate the gallery region and Line Hough Transform to remove the line and adjust the P-N learning's parameters to optimize accurate and speed.

Artificial Intelligence Plant Doctor: Plant Disease Diagnosis Using GPT4-vision

  • Yoeguang Hue;Jea Hyeoung Kim;Gang Lee;Byungheon Choi;Hyun Sim;Jongbum Jeon;Mun-Il Ahn;Yong Kyu Han;Ki-Tae Kim
    • Research in Plant Disease
    • /
    • v.30 no.1
    • /
    • pp.99-102
    • /
    • 2024
  • Integrated pest management is essential for controlling plant diseases that reduce crop yields. Rapid diagnosis is crucial for effective management in the event of an outbreak to identify the cause and minimize damage. Diagnosis methods range from indirect visual observation, which can be subjective and inaccurate, to machine learning and deep learning predictions that may suffer from biased data. Direct molecular-based methods, while accurate, are complex and time-consuming. However, the development of large multimodal models, like GPT-4, combines image recognition with natural language processing for more accurate diagnostic information. This study introduces GPT-4-based system for diagnosing plant diseases utilizing a detailed knowledge base with 1,420 host plants, 2,462 pathogens, and 37,467 pesticide instances from the official plant disease and pesticide registries of Korea. The AI plant doctor offers interactive advice on diagnosis, control methods, and pesticide use for diseases in Korea and is accessible at https://pdoc.scnu.ac.kr/.

Analysis of the effect of class classification learning on the saliency map of Self-Supervised Transformer (클래스분류 학습이 Self-Supervised Transformer의 saliency map에 미치는 영향 분석)

  • Kim, JaeWook;Kim, Hyeoncheol
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2022.07a
    • /
    • pp.67-70
    • /
    • 2022
  • NLP 분야에서 적극 활용되기 시작한 Transformer 모델을 Vision 분야에서 적용하기 시작하면서 object detection과 segmentation 등 각종 분야에서 기존 CNN 기반 모델의 정체된 성능을 극복하며 향상되고 있다. 또한, label 데이터 없이 이미지들로만 자기지도학습을 한 ViT(Vision Transformer) 모델을 통해 이미지에 포함된 여러 중요한 객체의 영역을 검출하는 saliency map을 추출할 수 있게 되었으며, 이로 인해 ViT의 자기지도학습을 통한 object detection과 semantic segmentation 연구가 활발히 진행되고 있다. 본 논문에서는 ViT 모델 뒤에 classifier를 붙인 모델에 일반 학습한 모델과 자기지도학습의 pretrained weight을 사용해서 전이학습한 모델의 시각화를 통해 각 saliency map들을 비교 분석하였다. 이를 통해, 클래스 분류 학습 기반 전이학습이 transformer의 saliency map에 미치는 영향을 확인할 수 있었다.

  • PDF

A study on the improvement of artificial intelligence-based Parking control system to prevent vehicle access with fake license plates (위조번호판 부착 차량 출입 방지를 위한 인공지능 기반의 주차관제시스템 개선 방안)

  • Jang, Sungmin;Iee, Jeongwoo;Park, Jonghyuk
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.2
    • /
    • pp.57-74
    • /
    • 2022
  • Recently, artificial intelligence parking control systems have increased the recognition rate of vehicle license plates using deep learning, but there is a problem that they cannot determine vehicles with fake license plates. Despite these security problems, several institutions have been using the existing system so far. For example, in an experiment using a counterfeit license plate, there are cases of successful entry into major government agencies. This paper proposes an improved system over the existing artificial intelligence parking control system to prevent vehicles with such fake license plates from entering. The proposed method is to use the degree of matching of the front feature points of the vehicle as a passing criterion using the ORB algorithm that extracts information on feature points characterized by an image, just as the existing system uses the matching of vehicle license plates as a passing criterion. In addition, a procedure for checking whether a vehicle exists inside was included in the proposed system to prevent the entry of the same type of vehicle with a fake license plate. As a result of the experiment, it showed the improved performance in identifying vehicles with fake license plates compared to the existing system. These results confirmed that the methods proposed in this paper could be applied to the existing parking control system while taking the flow of the original artificial intelligence parking control system to prevent vehicles with fake license plates from entering.

A computer vision-based approach for crack detection in ultra high performance concrete beams

  • Roya Solhmirzaei;Hadi Salehi;Venkatesh Kodur
    • Computers and Concrete
    • /
    • v.33 no.4
    • /
    • pp.341-348
    • /
    • 2024
  • Ultra-high-performance concrete (UHPC) has received remarkable attentions in civil infrastructure due to its unique mechanical characteristics and durability. UHPC gains increasingly dominant in essential structural elements, while its unique properties pose challenges for traditional inspection methods, as damage may not always manifest visibly on the surface. As such, the need for robust inspection techniques for detecting cracks in UHPC members has become imperative as traditional methods often fall short in providing comprehensive and timely evaluations. In the era of artificial intelligence, computer vision has gained considerable interest as a powerful tool to enhance infrastructure condition assessment with image and video data collected from sensors, cameras, and unmanned aerial vehicles. This paper presents a computer vision-based approach employing deep learning to detect cracks in UHPC beams, with the aim of addressing the inherent limitations of traditional inspection methods. This work leverages computer vision to discern intricate patterns and anomalies. Particularly, a convolutional neural network architecture employing transfer learning is adopted to identify the presence of cracks in the beams. The proposed approach is evaluated with image data collected from full-scale experiments conducted on UHPC beams subjected to flexural and shear loadings. The results of this study indicate the applicability of computer vision and deep learning as intelligent methods to detect major and minor cracks and recognize various damage mechanisms in UHPC members with better efficiency compared to conventional monitoring methods. Findings from this work pave the way for the development of autonomous infrastructure health monitoring and condition assessment, ensuring early detection in response to evolving structural challenges. By leveraging computer vision, this paper contributes to usher in a new era of effectiveness in autonomous crack detection, enhancing the resilience and sustainability of UHPC civil infrastructure.

Robot Vision to Audio Description Based on Deep Learning for Effective Human-Robot Interaction (효과적인 인간-로봇 상호작용을 위한 딥러닝 기반 로봇 비전 자연어 설명문 생성 및 발화 기술)

  • Park, Dongkeon;Kang, Kyeong-Min;Bae, Jin-Woo;Han, Ji-Hyeong
    • The Journal of Korea Robotics Society
    • /
    • v.14 no.1
    • /
    • pp.22-30
    • /
    • 2019
  • For effective human-robot interaction, robots need to understand the current situation context well, but also the robots need to transfer its understanding to the human participant in efficient way. The most convenient way to deliver robot's understanding to the human participant is that the robot expresses its understanding using voice and natural language. Recently, the artificial intelligence for video understanding and natural language process has been developed very rapidly especially based on deep learning. Thus, this paper proposes robot vision to audio description method using deep learning. The applied deep learning model is a pipeline of two deep learning models for generating natural language sentence from robot vision and generating voice from the generated natural language sentence. Also, we conduct the real robot experiment to show the effectiveness of our method in human-robot interaction.

Vision-based Input-Output System identification for pedestrian suspension bridges

  • Lim, Jeonghyeok;Yoon, Hyungchul
    • Smart Structures and Systems
    • /
    • v.29 no.5
    • /
    • pp.715-728
    • /
    • 2022
  • Recently, numbers of long span pedestrian suspension bridges have been constructed worldwide. While recent tragedies regarding pedestrian suspension bridges have shown how these bridges can wreak havoc on the society, there are no specific guidelines for construction standards nor safety inspections yet. Therefore, a structural health monitoring system that could help ensure the safety of pedestrian suspension bridges are needed. System identification is one of the popular applications for structural health monitoring method, which estimates the dynamic system. Most of the system identification methods for bridges are currently adapting output-only system identification method, which assumes the dynamic load to be a white noise due to the difficulty of measuring the dynamic load. In the case of pedestrian suspension bridges, the pedestrian load is within specific frequency range, resulting in large errors when using the output-only system identification method. Therefore, this study aims to develop a system identification method for pedestrian suspension bridges considering both input and output of the dynamic system. This study estimates the location and the magnitude of the pedestrian load, as well as the dynamic response of the pedestrian bridges by utilizing artificial intelligence and computer vision techniques. A simulation-based validation test was conducted to verify the performance of the proposed system. The proposed method is expected to improve the accuracy and the efficiency of the current inspection and monitoring systems for pedestrian suspension bridges.

A Study on Image Annotation Automation Process using SHAP for Defect Detection (SHAP를 이용한 이미지 어노테이션 자동화 프로세스 연구)

  • Jin Hyeong Jung;Hyun Su Sim;Yong Soo Kim
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.46 no.1
    • /
    • pp.76-83
    • /
    • 2023
  • Recently, the development of computer vision with deep learning has made object detection using images applicable to diverse fields, such as medical care, manufacturing, and transportation. The manufacturing industry is saving time and money by applying computer vision technology to detect defects or issues that may occur during the manufacturing and inspection process. Annotations of collected images and their location information are required for computer vision technology. However, manually labeling large amounts of images is time-consuming, expensive, and can vary among workers, which may affect annotation quality and cause inaccurate performance. This paper proposes a process that can automatically collect annotations and location information for images using eXplainable AI, without manual annotation. If applied to the manufacturing industry, this process is thought to save the time and cost required for image annotation collection and collect relatively high-quality annotation information.