• Title/Summary/Keyword: computer vision systems

Search Result 600, Processing Time 0.024 seconds

A review on deep learning-based structural health monitoring of civil infrastructures

  • Ye, X.W.;Jin, T.;Yun, C.B.
    • Smart Structures and Systems
    • /
    • v.24 no.5
    • /
    • pp.567-585
    • /
    • 2019
  • In the past two decades, structural health monitoring (SHM) systems have been widely installed on various civil infrastructures for the tracking of the state of their structural health and the detection of structural damage or abnormality, through long-term monitoring of environmental conditions as well as structural loadings and responses. In an SHM system, there are plenty of sensors to acquire a huge number of monitoring data, which can factually reflect the in-service condition of the target structure. In order to bridge the gap between SHM and structural maintenance and management (SMM), it is necessary to employ advanced data processing methods to convert the original multi-source heterogeneous field monitoring data into different types of specific physical indicators in order to make effective decisions regarding inspection, maintenance and management. Conventional approaches to data analysis are confronted with challenges from environmental noise, the volume of measurement data, the complexity of computation, etc., and they severely constrain the pervasive application of SHM technology. In recent years, with the rapid progress of computing hardware and image acquisition equipment, the deep learning-based data processing approach offers a new channel for excavating the massive data from an SHM system, towards autonomous, accurate and robust processing of the monitoring data. Many researchers from the SHM community have made efforts to explore the applications of deep learning-based approaches for structural damage detection and structural condition assessment. This paper gives a review on the deep learning-based SHM of civil infrastructures with the main content, including a brief summary of the history of the development of deep learning, the applications of deep learning-based data processing approaches in the SHM of many kinds of civil infrastructures, and the key challenges and future trends of the strategy of deep learning-based SHM.

Symbol recognition using vectorial signature matching for building mechanical drawings

  • Cho, Chi Yon;Liu, Xuesong;Akinci, Burcu
    • Advances in Computational Design
    • /
    • v.4 no.2
    • /
    • pp.155-177
    • /
    • 2019
  • Operation and Maintenance (O&M) phase is the main contributor to the total lifecycle cost of a building. Previous studies have described that Building Information Models (BIM), if available with detailed asset information and their properties, can enable rapid troubleshooting and execution of O&M tasks by providing the required information of the facility. Despite the potential benefits, there is still rarely BIM with Mechanical, Electrical and Plumbing (MEP) assets and properties that are available for O&M. BIM is usually not in possession for existing buildings and generating BIM manually is a time-consuming process. Hence, there is a need for an automated approach that can reconstruct the MEP systems in BIM. Previous studies investigated automatic reconstruction of BIM using architectural drawings, structural drawings, or the combination with photos. But most of the previous studies are limited to reconstruct the architectural and structural components. Note that mechanical components in the building typically require more frequent maintenance than architectural or structural components. However, the building mechanical drawings are relatively more complex due to various type of symbols that are used to represent the mechanical systems. In order to address this challenge, this paper proposed a symbol recognition framework that can automatically recognize the different type of symbols in the building mechanical drawings. This study applied vector-based computer vision techniques to recognize the symbols and their properties (e.g., location, type, etc.) in two vector-based input documents: 2D drawings and the symbol description document. The framework not only enables recognizing and locating the mechanical component of interest for BIM reconstruction purpose but opens the possibility of merging the updated information into the current BIM in the future reducing the time of repeated manual creation of BIM after every renovation project.

Comparison of error rates of various stereo matching methods for mobile stereo vision systems (모바일 스테레오 비전 시스템을 위한 다양한 스테레오 정합 기법의 오차율 비교)

  • Joo-Young, Lee;Kwang-yeob, Lee
    • Journal of IKEEE
    • /
    • v.26 no.4
    • /
    • pp.686-692
    • /
    • 2022
  • In this paper, the matching error rates of modified area-based, energy-based algorithms, and learning-based structures were compared for stereo image matching. Census transform (CT) based on region and life propagation (BP) algorithm based on energy were selected, respectively.Existing algorithms have been improved and implemented in an embedded processor environment so that they can be used for stereo image matching in mobile systems. Even in the case of the learning base to be compared, a neural network structure that utilizes small-scale parameters was adopted. To compare the error rates of the three matching methods, Middlebury's Tsukuba was selected as a test image and subdivided into non-occlusion, discontinuous, and disparity error rates for accurate comparison. As a result of the experiment, the error rate of modified CT matching improved by about 11% when compared with the existing algorithm. BP matching was about 87% better than conventional CT in the error rate. Compared to the learning base using neural networks, BP matching was about 31% superior.

High-Speed Maritime Object Detection Scheme for the Protection of the Aid to Navigation

  • Lee, Hyochan;Song, Hyunhak;Cho, Sungyoon;Kwon, Kiwon;Park, Sunghyun;Im, Taeho
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.692-712
    • /
    • 2022
  • Buoys used for Aid to Navigation systems are widely used to guide the sea paths and are powered by batteries, requiring continuous battery replacement. However, since human labor is required to replace the batteries, humans can be exposed to dangerous situation, including even collision with shipping vessels. In addition, Maritime sensors are installed on the route signs, so that these are often damaged by collisions with small and medium-sized ships, resulting in significant financial loss. In order to prevent these accidents, maritime object detection technology is essential to alert ships approaching buoys. Existing studies apply a number of filters to eliminate noise and to detect objects within the sea image. For this process, most studies directly access the pixels and process the images. However, this approach typically takes a long time to process because of its complexity and the requirements of significant amounts of computational power. In an emergent situation, it is important to alarm the vessel's rapid approach to buoys in real time to avoid collisions between vessels and route signs, therefore minimizing computation and speeding up processes are critical operations. Therefore, we propose Fast Connected Component Labeling (FCCL) which can reduce computation to minimize the processing time of filter applications, while maintaining the detection performance of existing methods. The results show that the detection performance of the FCCL is close to 30 FPS - approximately 2-5 times faster, when compared to the existing methods - while the average throughput is the same as existing methods.

Estimation of Urban Traffic State Using Black Box Camera (차량 블랙박스 카메라를 이용한 도시부 교통상태 추정)

  • Haechan Cho;Yeohwan Yoon;Hwasoo Yeo
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.22 no.2
    • /
    • pp.133-146
    • /
    • 2023
  • Traffic states in urban areas are essential to implement effective traffic operation and traffic control. However, installing traffic sensors on numerous road sections is extremely expensive. Accordingly, estimating the traffic state using a vehicle-mounted camera, which shows a high penetration rate, is a more effective solution. However, the previously proposed methodology using object tracking or optical flow has a high computational cost and requires consecutive frames to obtain traffic states. Accordingly, we propose a method to detect vehicles and lanes by object detection networks and set the region between lanes as a region of interest to estimate the traffic density of the corresponding area. The proposed method only uses less computationally expensive object detection models and can estimate traffic states from sampled frames rather than consecutive frames. In addition, the traffic density estimation accuracy was over 90% on the black box videos collected from two buses having different characteristics.

Real-time Tooth Region Detection in Intraoral Scanner Images with Deep Learning (딥러닝을 이용한 구강 스캐너 이미지 내 치아 영역 실시간 검출)

  • Na-Yun, Park;Ji-Hoon Kim;Tae-Min Kim;Kyeong-Jin Song;Yu-Jin Byun;Min-Ju Kang․;Kyungkoo Jun;Jae-Gon Kim
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.46 no.3
    • /
    • pp.1-6
    • /
    • 2023
  • In the realm of dental prosthesis fabrication, obtaining accurate impressions has historically been a challenging and inefficient process, often hindered by hygiene concerns and patient discomfort. Addressing these limitations, Company D recently introduced a cutting-edge solution by harnessing the potential of intraoral scan images to create 3D dental models. However, the complexity of these scan images, encompassing not only teeth and gums but also the palate, tongue, and other structures, posed a new set of challenges. In response, we propose a sophisticated real-time image segmentation algorithm that selectively extracts pertinent data, specifically focusing on teeth and gums, from oral scan images obtained through Company D's oral scanner for 3D model generation. A key challenge we tackled was the detection of the intricate molar regions, common in dental imaging, which we effectively addressed through intelligent data augmentation for enhanced training. By placing significant emphasis on both accuracy and speed, critical factors for real-time intraoral scanning, our proposed algorithm demonstrated exceptional performance, boasting an impressive accuracy rate of 0.91 and an unrivaled FPS of 92.4. Compared to existing algorithms, our solution exhibited superior outcomes when integrated into Company D's oral scanner. This algorithm is scheduled for deployment and commercialization within Company D's intraoral scanner.

Deep-learning performance in identifying and classifying dental implant systems from dental imaging: a systematic review and meta-analysis

  • Akhilanand Chaurasia;Arunkumar Namachivayam;Revan Birke Koca-Unsal;Jae-Hong Lee
    • Journal of Periodontal and Implant Science
    • /
    • v.54 no.1
    • /
    • pp.3-12
    • /
    • 2024
  • Deep learning (DL) offers promising performance in computer vision tasks and is highly suitable for dental image recognition and analysis. We evaluated the accuracy of DL algorithms in identifying and classifying dental implant systems (DISs) using dental imaging. In this systematic review and meta-analysis, we explored the MEDLINE/PubMed, Scopus, Embase, and Google Scholar databases and identified studies published between January 2011 and March 2022. Studies conducted on DL approaches for DIS identification or classification were included, and the accuracy of the DL models was evaluated using panoramic and periapical radiographic images. The quality of the selected studies was assessed using QUADAS-2. This review was registered with PROSPERO (CRDCRD42022309624). From 1,293 identified records, 9 studies were included in this systematic review and meta-analysis. The DL-based implant classification accuracy was no less than 70.75% (95% confidence interval [CI], 65.6%-75.9%) and no higher than 98.19 (95% CI, 97.8%-98.5%). The weighted accuracy was calculated, and the pooled sample size was 46,645, with an overall accuracy of 92.16% (95% CI, 90.8%-93.5%). The risk of bias and applicability concerns were judged as high for most studies, mainly regarding data selection and reference standards. DL models showed high accuracy in identifying and classifying DISs using panoramic and periapical radiographic images. Therefore, DL models are promising prospects for use as decision aids and decision-making tools; however, there are limitations with respect to their application in actual clinical practice.

Annotation Method based on Face Area for Efficient Interactive Video Authoring (효과적인 인터랙티브 비디오 저작을 위한 얼굴영역 기반의 어노테이션 방법)

  • Yoon, Ui Nyoung;Ga, Myeong Hyeon;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.83-98
    • /
    • 2015
  • Many TV viewers use mainly portal sites in order to retrieve information related to broadcast while watching TV. However retrieving information that people wanted needs a lot of time to retrieve the information because current internet presents too much information which is not required. Consequentially, this process can't satisfy users who want to consume information immediately. Interactive video is being actively investigated to solve this problem. An interactive video provides clickable objects, areas or hotspots to interact with users. When users click object on the interactive video, they can see additional information, related to video, instantly. The following shows the three basic procedures to make an interactive video using interactive video authoring tool: (1) Create an augmented object; (2) Set an object's area and time to be displayed on the video; (3) Set an interactive action which is related to pages or hyperlink; However users who use existing authoring tools such as Popcorn Maker and Zentrick spend a lot of time in step (2). If users use wireWAX then they can save sufficient time to set object's location and time to be displayed because wireWAX uses vision based annotation method. But they need to wait for time to detect and track object. Therefore, it is required to reduce the process time in step (2) using benefits of manual annotation method and vision-based annotation method effectively. This paper proposes a novel annotation method allows annotator to easily annotate based on face area. For proposing new annotation method, this paper presents two steps: pre-processing step and annotation step. The pre-processing is necessary because system detects shots for users who want to find contents of video easily. Pre-processing step is as follow: 1) Extract shots using color histogram based shot boundary detection method from frames of video; 2) Make shot clusters using similarities of shots and aligns as shot sequences; and 3) Detect and track faces from all shots of shot sequence metadata and save into the shot sequence metadata with each shot. After pre-processing, user can annotates object as follow: 1) Annotator selects a shot sequence, and then selects keyframe of shot in the shot sequence; 2) Annotator annotates objects on the relative position of the actor's face on the selected keyframe. Then same objects will be annotated automatically until the end of shot sequence which has detected face area; and 3) User assigns additional information to the annotated object. In addition, this paper designs the feedback model in order to compensate the defects which are wrong aligned shots, wrong detected faces problem and inaccurate location problem might occur after object annotation. Furthermore, users can use interpolation method to interpolate position of objects which is deleted by feedback. After feedback user can save annotated object data to the interactive object metadata. Finally, this paper shows interactive video authoring system implemented for verifying performance of proposed annotation method which uses presented models. In the experiment presents analysis of object annotation time, and user evaluation. First, result of object annotation average time shows our proposed tool is 2 times faster than existing authoring tools for object annotation. Sometimes, annotation time of proposed tool took longer than existing authoring tools, because wrong shots are detected in the pre-processing. The usefulness and convenience of the system were measured through the user evaluation which was aimed at users who have experienced in interactive video authoring system. Recruited 19 experts evaluates of 11 questions which is out of CSUQ(Computer System Usability Questionnaire). CSUQ is designed by IBM for evaluating system. Through the user evaluation, showed that proposed tool is useful for authoring interactive video than about 10% of the other interactive video authoring systems.

Research Trends of Health Recommender Systems (HRS): Applying Citation Network Analysis and GraphSAGE (건강추천시스템(HRS) 연구 동향: 인용네트워크 분석과 GraphSAGE를 활용하여)

  • Haryeom Jang;Jeesoo You;Sung-Byung Yang
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.2
    • /
    • pp.57-84
    • /
    • 2023
  • With the development of information and communications technology (ICT) and big data technology, anyone can easily obtain and utilize vast amounts of data through the Internet. Therefore, the capability of selecting high-quality data from a large amount of information is becoming more important than the capability of just collecting them. This trend continues in academia; literature reviews, such as systematic and non-systematic reviews, have been conducted in various research fields to construct a healthy knowledge structure by selecting high-quality research from accumulated research materials. Meanwhile, after the COVID-19 pandemic, remote healthcare services, which have not been agreed upon, are allowed to a limited extent, and new healthcare services such as health recommender systems (HRS) equipped with artificial intelligence (AI) and big data technologies are in the spotlight. Although, in practice, HRS are considered one of the most important technologies to lead the future healthcare industry, literature review on HRS is relatively rare compared to other fields. In addition, although HRS are fields of convergence with a strong interdisciplinary nature, prior literature review studies have mainly applied either systematic or non-systematic review methods; hence, there are limitations in analyzing interactions or dynamic relationships with other research fields. Therefore, in this study, the overall network structure of HRS and surrounding research fields were identified using citation network analysis (CNA). Additionally, in this process, in order to address the problem that the latest papers are underestimated in their citation relationships, the GraphSAGE algorithm was applied. As a result, this study identified 'recommender system', 'wireless & IoT', 'computer vision', and 'text mining' as increasingly important research fields related to HRS research, and confirmed that 'personalization' and 'privacy' are emerging issues in HRS research. The study findings would provide both academic and practical insights into identifying the structure of the HRS research community, examining related research trends, and designing future HRS research directions.

Design and Implementation of the Stop line and Crosswalk Recognition Algorithm for Autonomous UGV (자율 주행 UGV를 위한 정지선과 횡단보도 인식 알고리즘 설계 및 구현)

  • Lee, Jae Hwan;Yoon, Heebyung
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.24 no.3
    • /
    • pp.271-278
    • /
    • 2014
  • In spite of that stop line and crosswalk should be aware of the most basic objects in transportation system, its features extracted are very limited. In addition to image-based recognition technology, laser and RF, GPS/INS recognition technology, it is difficult to recognize. For this reason, the limited research in this area has been done. In this paper, the algorithm to recognize the stop line and crosswalk is designed and implemented using image-based recognition technology with the images input through a vision sensor. This algorithm consists of three functions.; One is to select the area, in advance, needed for feature extraction in order to speed up the data processing, 'Region of Interest', another is to process the images only that white color is detected more than a certain proportion in order to remove the unnecessary operation, 'Color Pattern Inspection', the other is 'Feature Extraction and Recognition', which is to extract the edge features and compare this to the previously-modeled one to identify the stop line and crosswalk. For this, especially by using case based feature comparison algorithm, it can identify either both stop line and crosswalk exist or just one exists. Also the proposed algorithm is to develop existing researches by comparing and analysing effect of in-vehicle camera installation and changes in recognition rate of distance estimation and various constraints such as backlight and shadow.