• Title/Summary/Keyword: vision model

Search Result 1,349, Processing Time 0.033 seconds

A study on the DGPS data errors correction through real-time coordinates conversion using the vision system (비젼 시스템을 이용한 DGPS 데이터 보정에 관한 연구)

  • Mun, Seong-Ryong;Chae, Jung-Su;Park, Jang-Hun;Lee, Ho-Soon;Rho, Do-Hwan
    • Proceedings of the KIEE Conference
    • /
    • 2003.07d
    • /
    • pp.2310-2312
    • /
    • 2003
  • This paper describes a navigation system for an autonomous vehicle in outdoor environments. The vehicle uses vision system to detect coordinates and DGPS information to determine the vehicles initial position and orientation. The vision system detects coordinates in the environment by referring to an environment model. As the vehicle moves, it estimates its position by conventional DGPS data, and matches up the coordinates with the environment model in order to reduce the error in the vehicles position estimate. The vehicles initial position and orientation are calculated from the coordinate values of the first and second locations, which are acquired by DGPS. Subsequent orientations and positions are derived. Experimental results in real environments have showed the effectiveness of our proposed navigation methods and real-time methods.

  • PDF

On low cost model-based monitoring of industrial robotic arms using standard machine vision

  • Karagiannidisa, Aris;Vosniakos, George C.
    • Advances in robotics research
    • /
    • v.1 no.1
    • /
    • pp.81-99
    • /
    • 2014
  • This paper contributes towards the development of a computer vision system for telemonitoring of industrial articulated robotic arms. The system aims to provide precision real time measurements of the joint angles by employing low cost cameras and visual markers on the body of the robot. To achieve this, a mathematical model that connects image features and joint angles was developed covering rotation of a single joint whose axis is parallel to the visual projection plane. The feature that is examined during image processing is the varying area of given circular target placed on the body of the robot, as registered by the camera during rotation of the arm. In order to distinguish between rotation directions four targets were used placed every $90^{\circ}$ and observed by two cameras at suitable angular distances. The results were deemed acceptable considering camera cost and lighting conditions of the workspace. A computational error analysis explored how deviations from the ideal camera positions affect the measurements and led to appropriate correction. The method is deemed to be extensible to multiple joint motion of a known kinematic chain.

Study on estimation of propeller cavitation using computer vision (컴퓨터 비전을 이용한 프로펠러 캐비테이션 평가 연구)

  • Taegoo, Lee;Ki-Seong, Kim;Ji-Woo, Hong;Byoung-Kwon, Ahn;Kyung-Jun, Lee
    • Journal of the Korean Society of Visualization
    • /
    • v.20 no.3
    • /
    • pp.128-135
    • /
    • 2022
  • Cavitation occurs inevitably in marine propellers rotating at high speed in the water, which is a major cause of underwater radiated noise. Cavitation-induced noise from propellers rotating at a specific frequency not only reduces the sonar detection capability, but also exposes the ship's location, and it causes very fatal consequences for the survivability of the navy vessels. Therefore cavity inception speed (CIS) is one of the important factors determining the special performance of the ship. In this study, we present a method using computer vision that can detect and quantitatively estimate tip vortex cavitation on a propeller rotating at high speed. Based on the model test results performed in a large cavitation tunnel, the effectiveness of this method was verified.

Accuracy Analysis of Construction Worker's Protective Equipment Detection Using Computer Vision Technology (컴퓨터 비전 기술을 이용한 건설 작업자 보호구 검출 정확도 분석)

  • Kang, Sungwon;Lee, Kiseok;Yoo, Wi Sung;Shin, Yoonseok;Lee, Myungdo
    • Journal of the Korea Institute of Building Construction
    • /
    • v.23 no.1
    • /
    • pp.81-92
    • /
    • 2023
  • According to the 2020 industrial accident reports of the Ministry of Employment and Labor, the number of fatal accidents in the construction industry over the past 5 years has been higher than in other industries. Of these more than 50% of fatal accidents are initially caused by fall accidents. The central government is intensively managing falling/jamming protection device and the use of personal protective equipment to eradicate the inappropriate factors disrupting safety at construction sites. In addition, although efforts have been made to prevent safety accidents with the proposal of the Special Act on Construction Safety, fatalities on construction sites are constantly occurring. Therefore, this study developed a model that automatically detects the wearing state of the worker's safety helmet and belt using computer vision technology. In considerations of conditions occurring at construction sites, we suggest an optimization method, which has been verified in terms of the accuracy and operation speed of the proposed model. As a result, it is possible to improve the efficiency of inspection and patrol by construction site managers, which is expected to contribute to reinforcing competency of safety management.

The Factors Influencing the Asthenopia of Myopia with Phoria (사위를 가진 근시안의 안정피로에 영향을 미치는 요인)

  • Kim, Jung-Hee;Kim, Chang-Sik
    • Journal of Korean Ophthalmic Optics Society
    • /
    • v.10 no.4
    • /
    • pp.419-428
    • /
    • 2005
  • The aim of this study was to provide data for the relief of asthenopia during binocular vision by determining the characteristics of ocular function in adults. A total of 260 subjects were between the age of 19-35years. We measured individually the refractive error correction, pupillary distance, optical center distance, phoria, convergence, accommodation and the AC/A as well as the asthenopia during binocular vision using a questionnaire. After analysis of factors affecting asthenopia, we also examined the reductive effect of asthenopia in subjects who had asthenopia using prism. To determine the factors affecting asthenopia during binocular vision, statistic analyses were carried out the multivariate Logistic regression model. The results of this study were as follow. The asthenopia during binocular vision was found 26.9% of subjects. Multivariate logistic regression model was used to determine factors affecting binocular vision of myopia. When the accommodation and convergence were low compared to being high, when subjects had esophoria or there was more exophoria, and when AC/A was lower than the standard, the rate of asthenopia was higher. Therefore the accommodation, convergence and AC/A could be predictive factors for asthenopia. We used prism for subjects who had asthenopia during binocular vision, the results showed that the symptom of asthenopia was eased up to 74.3%.

  • PDF

Jointly Image Topic and Emotion Detection using Multi-Modal Hierarchical Latent Dirichlet Allocation

  • Ding, Wanying;Zhu, Junhuan;Guo, Lifan;Hu, Xiaohua;Luo, Jiebo;Wang, Haohong
    • Journal of Multimedia Information System
    • /
    • v.1 no.1
    • /
    • pp.55-67
    • /
    • 2014
  • Image topic and emotion analysis is an important component of online image retrieval, which nowadays has become very popular in the widely growing social media community. However, due to the gaps between images and texts, there is very limited work in literature to detect one image's Topics and Emotions in a unified framework, although topics and emotions are two levels of semantics that often work together to comprehensively describe one image. In this work, a unified model, Joint Topic/Emotion Multi-Modal Hierarchical Latent Dirichlet Allocation (JTE-MMHLDA) model, which extends previous LDA, mmLDA, and JST model to capture topic and emotion information at the same time from heterogeneous data, is proposed. Specifically, a two level graphical structured model is built to realize sharing topics and emotions among the whole document collection. The experimental results on a Flickr dataset indicate that the proposed model efficiently discovers images' topics and emotions, and significantly outperform the text-only system by 4.4%, vision-only system by 18.1% in topic detection, and outperforms the text-only system by 7.1%, vision-only system by 39.7% in emotion detection.

  • PDF

The Development of a Maturity Evaluation Model on the Technology Management Competence (MEMTMC) (국내 기업의 기술경영 역량수준의 성숙도 평가 모형 개발)

  • Ahn, Yeon S.;Kim, Wha Young
    • Journal of Information Technology Applications and Management
    • /
    • v.22 no.4
    • /
    • pp.77-94
    • /
    • 2015
  • This research aims to develope a maturity evaluation model on the technology management competence (MEMTMC) and to investigate the relationship between technology management competencies and maturity evaluation level in Korean firms. After research framework building through literature review and field study focusing on interview process, a maturity evaluation model was suggested, including technology management competency factors such as vision and strategy, leadership, resources, projects, performance management, and systematic procedures. Through the empirical study on the 111 Korean firms, the MEMTMC model was tested and some hypotheses were examined. As a result of the research, we found that the MEMTMC is significant statistically to evaluate the technology management competencies of firms from a practical point of view. Additionally, these findings suggest that firms can build up their technology management competencies, including the vision and strategy, leadership, resources, projects, performance management, and systematic procedures in order to raise their maturity level of technology management competencies. But this research has some limitations such as this MEMTMC has not considered separately the variety of industry, not enough survey respondents and 5 level group firms etc.

Visual Model of Pattern Design Based on Deep Convolutional Neural Network

  • Jingjing Ye;Jun Wang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.2
    • /
    • pp.311-326
    • /
    • 2024
  • The rapid development of neural network technology promotes the neural network model driven by big data to overcome the texture effect of complex objects. Due to the limitations in complex scenes, it is necessary to establish custom template matching and apply it to the research of many fields of computational vision technology. The dependence on high-quality small label sample database data is not very strong, and the machine learning system of deep feature connection to complete the task of texture effect inference and speculation is relatively poor. The style transfer algorithm based on neural network collects and preserves the data of patterns, extracts and modernizes their features. Through the algorithm model, it is easier to present the texture color of patterns and display them digitally. In this paper, according to the texture effect reasoning of custom template matching, the 3D visualization of the target is transformed into a 3D model. The high similarity between the scene to be inferred and the user-defined template is calculated by the user-defined template of the multi-dimensional external feature label. The convolutional neural network is adopted to optimize the external area of the object to improve the sampling quality and computational performance of the sample pyramid structure. The results indicate that the proposed algorithm can accurately capture the significant target, achieve more ablation noise, and improve the visualization results. The proposed deep convolutional neural network optimization algorithm has good rapidity, data accuracy and robustness. The proposed algorithm can adapt to the calculation of more task scenes, display the redundant vision-related information of image conversion, enhance the powerful computing power, and further improve the computational efficiency and accuracy of convolutional networks, which has a high research significance for the study of image information conversion.

Research on damage detection and assessment of civil engineering structures based on DeepLabV3+ deep learning model

  • Chengyan Song
    • Structural Engineering and Mechanics
    • /
    • v.91 no.5
    • /
    • pp.443-457
    • /
    • 2024
  • At present, the traditional concrete surface inspection methods based on artificial vision have the problems of high cost and insecurity, while the computer vision methods rely on artificial selection features in the case of sensitive environmental changes and difficult promotion. In order to solve these problems, this paper introduces deep learning technology in the field of computer vision to achieve automatic feature extraction of structural damage, with excellent detection speed and strong generalization ability. The main contents of this study are as follows: (1) A method based on DeepLabV3+ convolutional neural network model is proposed for surface detection of post-earthquake structural damage, including surface damage such as concrete cracks, spaling and exposed steel bars. The key semantic information is extracted by different backbone networks, and the data sets containing various surface damage are trained, tested and evaluated. The intersection ratios of 54.4%, 44.2%, and 89.9% in the test set demonstrate the network's capability to accurately identify different types of structural surface damages in pixel-level segmentation, highlighting its effectiveness in varied testing scenarios. (2) A semantic segmentation model based on DeepLabV3+ convolutional neural network is proposed for the detection and evaluation of post-earthquake structural components. Using a dataset that includes building structural components and their damage degrees for training, testing, and evaluation, semantic segmentation detection accuracies were recorded at 98.5% and 56.9%. To provide a comprehensive assessment that considers both false positives and false negatives, the Mean Intersection over Union (Mean IoU) was employed as the primary evaluation metric. This choice ensures that the network's performance in detecting and evaluating pixel-level damage in post-earthquake structural components is evaluated uniformly across all experiments. By incorporating deep learning technology, this study not only offers an innovative solution for accurately identifying post-earthquake damage in civil engineering structures but also contributes significantly to empirical research in automated detection and evaluation within the field of structural health monitoring.

Development of a Lane Sensing Algorithm Using Vision Sensors (비전 센서를 이용한 차선 감지 알고리듬 개발)

  • Park, Yong-Jun;Heo, Geon-Su
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.26 no.8
    • /
    • pp.1666-1671
    • /
    • 2002
  • A lane sensing algorithm using vision sensors is developed based on lane geometry models. The parameters of the lane geometry models are estimated by a Kalman filter and utilized to reconstruct the lane geometry in the global coordinate. The inverse perspective mapping from image plane to global coordinate assumes earth to be flat, but roll and pitch motions of a vehicle are considered from the perspective of the lane sensing. The proposed algorithm shows robust lane sensing performance compared to the conventional algorithms.