• Title/Summary/Keyword: 비전모델

Search Result 553, Processing Time 0.024 seconds

Gradient Descent Training Method for Optimizing Data Prediction Models (데이터 예측 모델 최적화를 위한 경사하강법 교육 방법)

  • Hur, Kyeong
    • Journal of Practical Engineering Education
    • /
    • v.14 no.2
    • /
    • pp.305-312
    • /
    • 2022
  • In this paper, we focused on training to create and optimize a basic data prediction model. And we proposed a gradient descent training method of machine learning that is widely used to optimize data prediction models. It visually shows the entire operation process of gradient descent used in the process of optimizing parameter values required for data prediction models by applying the differential method and teaches the effective use of mathematical differentiation in machine learning. In order to visually explain the entire operation process of gradient descent, we implement gradient descent SW in a spreadsheet. In this paper, first, a two-variable gradient descent training method is presented, and the accuracy of the two-variable data prediction model is verified by comparison with the error least squares method. Second, a three-variable gradient descent training method is presented and the accuracy of a three-variable data prediction model is verified. Afterwards, the direction of the optimization practice for gradient descent was presented, and the educational effect of the proposed gradient descent method was analyzed through the results of satisfaction with education for non-majors.

Computer Vision-based Automated Adhesive Quality Inspection Model of Exterior Insulation and Finishing System (컴퓨터 비전 기반 외단열 공사의 접착제 도포품질 감리 자동화 모델)

  • Yoon, Sebeen;Kang, Mingyun;Jang, Hyounseung;Kim, Taehoon
    • Journal of the Korea Institute of Building Construction
    • /
    • v.23 no.2
    • /
    • pp.165-173
    • /
    • 2023
  • This research proposed a model for automatically monitoring the quality of insulation adhesive application in external insulation construction. Upon case implementation, the area segmentation model demonstrated a 92.3% accuracy, while the area and distance calculation accuracies of the proposed model were 98.8% and 96.7%, respectively. These findings suggest that the model can effectively prevent the most common insulation defect, insulation failure, while simultaneously minimizing the need for on-site supervisory personnel during external insulation construction. This, in turn, contributes to the enhancement of the external insulation system. Moving forward, we plan to gather construction images of various external insulation methods to refine the image segmentation model's performance and develop a model capable of automatically monitoring scenarios with a considerable number of insulation materials in the image.

Resource-Efficient Object Detector for Low-Power Devices (저전력 장치를 위한 자원 효율적 객체 검출기)

  • Akshay Kumar Sharma;Kyung Ki Kim
    • Transactions on Semiconductor Engineering
    • /
    • v.2 no.1
    • /
    • pp.17-20
    • /
    • 2024
  • This paper presents a novel lightweight object detection model tailored for low-powered edge devices, addressing the limitations of traditional resource-intensive computer vision models. Our proposed detector, inspired by the Single Shot Detector (SSD), employs a compact yet robust network design. Crucially, it integrates an 'enhancer block' that significantly boosts its efficiency in detecting smaller objects. The model comprises two primary components: the Light_Block for efficient feature extraction using Depth-wise and Pointwise Convolution layers, and the Enhancer_Block for enhanced detection of tiny objects. Trained from scratch on the Udacity Annotated Dataset with image dimensions of 300x480, our model eschews the need for pre-trained classification weights. Weighing only 5.5MB with approximately 0.43M parameters, our detector achieved a mean average precision (mAP) of 27.7% and processed at 140 FPS, outperforming conventional models in both precision and efficiency. This research underscores the potential of lightweight designs in advancing object detection for edge devices without compromising accuracy.

Development of AI and IoT-based smart farm pest prediction system: Research on application of YOLOv5 and Isolation Forest models (AI 및 IoT 기반 스마트팜 병충해 예측시스템 개발: YOLOv5 및 Isolation Forest 모델 적용 연구)

  • Mi-Kyoung Park;Hyun Sim
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.19 no.4
    • /
    • pp.771-780
    • /
    • 2024
  • In this study, we implemented a real-time pest detection and prediction system for a strawberry farm using a computer vision model based on the YOLOv5 architecture and an Isolation Forest Classifier. The model performance evaluation showed that the YOLOv5 model achieved a mean average precision (mAP 0.5) of 78.7%, an accuracy of 92.8%, a recall of 90.0%, and an F1-score of 76%, indicating high predictive performance. This system was designed to be applicable not only to strawberry farms but also to other crops and various environments. Based on data collected from a tomato farm, a new AI model was trained, resulting in a prediction accuracy of over 85% for major diseases such as late blight and yellow leaf curl virus. Compared to the previous model, this represented an improvement of more than 10% in prediction accuracy.

ONNX-based Runtime Performance Analysis: YOLO and ResNet (ONNX 기반 런타임 성능 분석: YOLO와 ResNet)

  • Jeong-Hyeon Kim;Da-Eun Lee;Su-Been Choi;Kyung-Koo Jun
    • The Journal of Bigdata
    • /
    • v.9 no.1
    • /
    • pp.89-100
    • /
    • 2024
  • In the field of computer vision, models such as You Look Only Once (YOLO) and ResNet are widely used due to their real-time performance and high accuracy. However, to apply these models in real-world environments, factors such as runtime compatibility, memory usage, computing resources, and real-time conditions must be considered. This study compares the characteristics of three deep model runtimes: ONNX Runtime, TensorRT, and OpenCV DNN, and analyzes their performance on two models. The aim of this paper is to provide criteria for runtime selection for practical applications. The experiments compare runtimes based on the evaluation metrics of time, memory usage, and accuracy for vehicle license plate recognition and classification tasks. The experimental results show that ONNX Runtime excels in complex object detection performance, OpenCV DNN is suitable for environments with limited memory, and TensorRT offers superior execution speed for complex models.

Opportunity Tree Framework Design For Optimization of Software Development Project Performance (소프트웨어 개발 프로젝트 성능의 최적화를 위한 Opportunity Tree 모델 설계)

  • Song Ki-Won;Lee Kyung-Whan
    • The KIPS Transactions:PartD
    • /
    • v.12D no.3 s.99
    • /
    • pp.417-428
    • /
    • 2005
  • Today, IT organizations perform projects with vision related to marketing and financial profit. The objective of realizing the vision is to improve the project performing ability in terms of QCD. Organizations have made a lot of efforts to achieve this objective through process improvement. Large companies such as IBM, Ford, and GE have made over $80\%$ of success through business process re-engineering using information technology instead of business improvement effect by computers. It is important to collect, analyze and manage the data on performed projects to achieve the objective, but quantitative measurement is difficult as software is invisible and the effect and efficiency caused by process change are not visibly identified. Therefore, it is not easy to extract the strategy of improvement. This paper measures and analyzes the project performance, focusing on organizations' external effectiveness and internal efficiency (Qualify, Delivery, Cycle time, and Waste). Based on the measured project performance scores, an OT (Opportunity Tree) model was designed for optimizing the project performance. The process of design is as follows. First, meta data are derived from projects and analyzed by quantitative GQM(Goal-Question-Metric) questionnaire. Then, the project performance model is designed with the data obtained from the quantitative GQM questionnaire and organization's performance score for each area is calculated. The value is revised by integrating the measured scores by area vision weights from all stakeholders (CEO, middle-class managers, developer, investor, and custom). Through this, routes for improvement are presented and an optimized improvement method is suggested. Existing methods to improve software process have been highly effective in division of processes' but somewhat unsatisfactory in structural function to develop and systemically manage strategies by applying the processes to Projects. The proposed OT model provides a solution to this problem. The OT model is useful to provide an optimal improvement method in line with organization's goals and can reduce risks which may occur in the course of improving process if it is applied with proposed methods. In addition, satisfaction about the improvement strategy can be improved by obtaining input about vision weight from all stakeholders through the qualitative questionnaire and by reflecting it to the calculation. The OT is also useful to optimize the expansion of market and financial performance by controlling the ability of Quality, Delivery, Cycle time, and Waste.

A Safety Score Prediction Model in Urban Environment Using Convolutional Neural Network (컨볼루션 신경망을 이용한 도시 환경에서의 안전도 점수 예측 모델 연구)

  • Kang, Hyeon-Woo;Kang, Hang-Bong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.8
    • /
    • pp.393-400
    • /
    • 2016
  • Recently, there have been various researches on efficient and automatic analysis on urban environment methods that utilize the computer vision and machine learning technology. Among many new analyses, urban safety analysis has received a major attention. In order to predict more accurately on safety score and reflect the human visual perception, it is necessary to consider the generic and local information that are most important to human perception. In this paper, we use Double-column Convolutional Neural network consisting of generic and local columns for the prediction of urban safety. The input of generic and local column used re-sized and random cropped images from original images, respectively. In addition, a new learning method is proposed to solve the problem of over-fitting in a particular column in the learning process. For the performance comparison of our Double-column Convolutional Neural Network, we compare two Support Vector Regression and three Convolutional Neural Network models using Root Mean Square Error and correlation analysis. Our experimental results demonstrate that our Double-column Convolutional Neural Network model show the best performance with Root Mean Square Error of 0.7432 and Pearson/Spearman correlation coefficient of 0.853/0.840.

Realtime Facial Expression Recognition from Video Sequences Using Optical Flow and Expression HMM (광류와 표정 HMM에 의한 동영상으로부터의 실시간 얼굴표정 인식)

  • Chun, Jun-Chul;Shin, Gi-Han
    • Journal of Internet Computing and Services
    • /
    • v.10 no.4
    • /
    • pp.55-70
    • /
    • 2009
  • Vision-based Human computer interaction is an emerging field of science and industry to provide natural way to communicate with human and computer. In that sense, inferring the emotional state of the person based on the facial expression recognition is an important issue. In this paper, we present a novel approach to recognize facial expression from a sequence of input images using emotional specific HMM (Hidden Markov Model) and facial motion tracking based on optical flow. Conventionally, in the HMM which consists of basic emotional states, it is considered natural that transitions between emotions are imposed to pass through neutral state. However, in this work we propose an enhanced transition framework model which consists of transitions between each emotional state without passing through neutral state in addition to a traditional transition model. For the localization of facial features from video sequence we exploit template matching and optical flow. The facial feature displacements traced by the optical flow are used for input parameters to HMM for facial expression recognition. From the experiment, we can prove that the proposed framework can effectively recognize the facial expression in real time.

  • PDF

A Stduy on Learning Model for Effective Coding Education (효과적인 코딩교육을 위한 학습 모델에 대한 연구)

  • Kim, Si-Jung;Cho, Do-Eun
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.2
    • /
    • pp.7-12
    • /
    • 2018
  • With our society entering the Fourth Industrial Revolution, there has been heightened interest in coding education, which has led to an increased number of coding classes offered in schools. Once catered to degree holders only, coding courses are now being offered as liberal arts courses to even non-majors. As the importance of computing abilities and creativity-oriented education through software learning becomes increasingly pronounced, the need for research on effective coding learning is growing more urgent. The present study sought an effective coding education model that would encourage and enhance learners' participation and interest in coding. The proposed learning model is designed to invoke learner's recognition of various coding grammars and data search in the process of designing and performing their own unique project. Application of the proposed learning model and analysis of such case studies showed improvement in learning outcomes. One can expect improved performance among learners if the proposed learning model is applied to various coding courses.

Development of a Translator for Automatic Generation of Ubiquitous Metaservice Ontology (유비쿼터스 메타서비스 온톨로지 자동 생성을 위한 번역기 개발)

  • Lee, Mee-Yeon;Lee, Jung-Won;Park, Seung-Soo;Cho, We-Duke
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.1
    • /
    • pp.191-203
    • /
    • 2009
  • To provide dynamic services for users in ubiquitous computing environments by considering context in real-time, in our previous work we proposed Metaservice concept, the description specification and the process for building a Metaservice library. However, our previous process generates separated models - UML, OWL, OWL-S based models - from each step, so it did not provide the established method for translation between models. Moreover, it premises aid of experts in various ontology languages, ontology editing tools and the proposed Metaservice specification. In this paper, we design the translation process from domain ontology in OWL to Metaservice Library in OWL-S and develop a visual tool in order to enable non-experts to generate consistent models and to construct a Metaservice library. The purpose of the Metaservice Library translation process is to maintain consistency in all models and to automatically generate OWL-S code for Metaservice library by integrating existing OWL model and Metaservice model.