• Title/Summary/Keyword: Segmentation process

Search Result 634, Processing Time 0.021 seconds

Synthetic data augmentation for pixel-wise steel fatigue crack identification using fully convolutional networks

  • Zhai, Guanghao;Narazaki, Yasutaka;Wang, Shuo;Shajihan, Shaik Althaf V.;Spencer, Billie F. Jr.
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.237-250
    • /
    • 2022
  • Structural health monitoring (SHM) plays an important role in ensuring the safety and functionality of critical civil infrastructure. In recent years, numerous researchers have conducted studies to develop computer vision and machine learning techniques for SHM purposes, offering the potential to reduce the laborious nature and improve the effectiveness of field inspections. However, high-quality vision data from various types of damaged structures is relatively difficult to obtain, because of the rare occurrence of damaged structures. The lack of data is particularly acute for fatigue crack in steel bridge girder. As a result, the lack of data for training purposes is one of the main issues that hinders wider application of these powerful techniques for SHM. To address this problem, the use of synthetic data is proposed in this article to augment real-world datasets used for training neural networks that can identify fatigue cracks in steel structures. First, random textures representing the surface of steel structures with fatigue cracks are created and mapped onto a 3D graphics model. Subsequently, this model is used to generate synthetic images for various lighting conditions and camera angles. A fully convolutional network is then trained for two cases: (1) using only real-word data, and (2) using both synthetic and real-word data. By employing synthetic data augmentation in the training process, the crack identification performance of the neural network for the test dataset is seen to improve from 35% to 40% and 49% to 62% for intersection over union (IoU) and precision, respectively, demonstrating the efficacy of the proposed approach.

Selective labeling using image super resolution for improving the efficiency of object detection in low-resolution oriental paintings

  • Moon, Hyeyoung;Kim, Namgyu
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.9
    • /
    • pp.21-32
    • /
    • 2022
  • Image labeling must be preceded in order to perform object detection, and this task is considered a significant burden in building a deep learning model. Tens of thousands of images need to be trained for building a deep learning model, and human labelers have many limitations in labeling these images manually. In order to overcome these difficulties, this study proposes a method to perform object detection without significant performance degradation, even though labeling some images rather than the entire image. Specifically, in this study, low-resolution oriental painting images are converted into high-quality images using a super-resolution algorithm, and the effect of SSIM and PSNR derived in this process on the mAP of object detection is analyzed. We expect that the results of this study can contribute significantly to constructing deep learning models such as image classification, object detection, and image segmentation that require efficient image labeling.

HMM Based Part of Speech Tagging for Hadith Isnad

  • Abdelkarim Abdelkader
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.3
    • /
    • pp.151-160
    • /
    • 2023
  • The Hadith is the second source of Islamic jurisprudence after Qur'an. Both sources are indispensable for muslims to practice Islam. All Ahadith are collected and are written. But most books of Hadith contain Ahadith that can be weak or rejected. So, quite a long time, scholars of Hadith have defined laws, rules and principles of Hadith to know the correct Hadith (Sahih) from the fair (Hassen) and weak (Dhaif). Unfortunately, the application of these rules, laws and principles is done manually by the specialists or students until now. The work presented in this paper is part of the automatic treatment of Hadith, and more specifically, it aims to automatically process the chain of narrators (Hadith Isnad) to find its different components and affect for each component its own tag using a statistical method: the Hidden Markov Models (HMM). This method is a power abstraction for times series data and a robust tool for representing probability distributions over sequences of observations. In this paper, we describe an important tool in the Hadith isnad processing: A chunker with HMM. The role of this tool is to decompose the chain of narrators (Isnad) and determine the tag of each part of Isnad (POI). First, we have compiled a tagset containing 13 tags. Then, we have used these tags to manually conceive a corpus of 100 chains of narrators from "Sahih Alboukhari" and we have extracted a lexicon from this corpus. This lexicon is a set of XML documents based on HPSG features and it contains the information of 134 narrators. After that, we have designed and implemented an analyzer based on HMM that permit to assign for each part of Isnad its proper tag and for each narrator its features. The system was tested on 2661 not duplicated Isnad from "Sahih Alboukhari". The obtained result achieved F-scores of 93%.

Efficient Semi-automatic Annotation System based on Deep Learning

  • Hyunseok Lee;Hwa Hui Shin;Soohoon Maeng;Dae Gwan Kim;Hyojeong Moon
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.6
    • /
    • pp.267-275
    • /
    • 2023
  • This paper presents the development of specialized software for annotating volume-of-interest on 18F-FDG PET/CT images with the goal of facilitating the studies and diagnosis of head and neck cancer (HNC). To achieve an efficient annotation process, we employed the SE-Norm-Residual Layer-based U-Net model. This model exhibited outstanding proficiency to segment cancerous regions within 18F-FDG PET/CT scans of HNC cases. Manual annotation function was also integrated, allowing researchers and clinicians to validate and refine annotations based on dataset characteristics. Workspace has a display with fusion of both PET and CT images, providing enhance user convenience through simultaneous visualization. The performance of deeplearning model was validated using a Hecktor 2021 dataset, and subsequently developed semi-automatic annotation functionalities. We began by performing image preprocessing including resampling, normalization, and co-registration, followed by an evaluation of the deep learning model performance. This model was integrated into the software, serving as an initial automatic segmentation step. Users can manually refine pre-segmented regions to correct false positives and false negatives. Annotation images are subsequently saved along with their corresponding 18F-FDG PET/CT fusion images, enabling their application across various domains. In this study, we developed a semi-automatic annotation software designed for efficiently generating annotated lesion images, with applications in HNC research and diagnosis. The findings indicated that this software surpasses conventional tools, particularly in the context of HNC-specific annotation with 18F-FDG PET/CT data. Consequently, developed software offers a robust solution for producing annotated datasets, driving advances in the studies and diagnosis of HNC.

Applications of Artificial Intelligence in MR Image Acquisition and Reconstruction (MRI 신호획득과 영상재구성에서의 인공지능 적용)

  • Junghwa Kang;Yoonho Nam
    • Journal of the Korean Society of Radiology
    • /
    • v.83 no.6
    • /
    • pp.1229-1239
    • /
    • 2022
  • Recently, artificial intelligence (AI) technology has shown potential clinical utility in a wide range of MRI fields. In particular, AI models for improving the efficiency of the image acquisition process and the quality of reconstructed images are being actively developed by the MR research community. AI is expected to further reduce acquisition times in various MRI protocols used in clinical practice when compared to current parallel imaging techniques. Additionally, AI can help with tasks such as planning, parameter optimization, artifact reduction, and quality assessment. Furthermore, AI is being actively applied to automate MR image analysis such as image registration, segmentation, and object detection. For this reason, it is important to consider the effects of protocols or devices in MR image analysis. In this review article, we briefly introduced issues related to AI application of MR image acquisition and reconstruction.

A Study on Object-Based Image Analysis Methods for Land Cover Classification in Agricultural Areas (농촌지역 토지피복분류를 위한 객체기반 영상분석기법 연구)

  • Kim, Hyun-Ok;Yeom, Jong-Min
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.15 no.4
    • /
    • pp.26-41
    • /
    • 2012
  • It is necessary to manage, forecast and prepare agricultural production based on accurate and up-to-date information in order to cope with the climate change and its impacts such as global warming, floods and droughts. This study examined the applicability as well as challenges of the object-based image analysis method for developing a land cover image classification algorithm, which can support the fast thematic mapping of wide agricultural areas on a regional scale. In order to test the applicability of RapidEye's multi-temporal spectral information for differentiating agricultural land cover types, the integration of other GIS data was minimized. Under this circumstance, the land cover classification accuracy at the study area of Kimje ($1300km^2$) was 80.3%. The geometric resolution of RapidEye, 6.5m showed the possibility to derive the spatial features of agricultural land use generally cultivated on a small scale in Korea. The object-based image analysis method can realize the expert knowledge in various ways during the classification process, so that the application of spectral image information can be optimized. An additional advantage is that the already developed classification algorithm can be stored, edited with variables in detail with regard to analytical purpose, and may be applied to other images as well as other regions. However, the segmentation process, which is fundamental for the object-based image classification, often cannot be explained quantitatively. Therefore, it is necessary to draw the best results based on expert's empirical and scientific knowledge.

Fast information extraction algorithm for object-based MPEG-4 application from MPEG-2 bit-streamaper (MPEG-2 비트열로부터 객체 기반 MPEG-4 응용을 위한 고속 정보 추출 알고리즘)

  • 양종호;원치선
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.26 no.12A
    • /
    • pp.2109-2119
    • /
    • 2001
  • In this paper, a fast information extraction algorithm for object-based MPEG-4 application from MPEG-2 bit-steam is proposed. For object-based MPEG-4 conversion, we need to extract such information as object-image, shape-image, macro-block motion vector, and header information from MPEG-2 bit-stream. If we use the extracted information, fast conversion for object-based MPEG-4 is possible. The proposed object extraction algorithm has two important steps, namely the motion vectors extraction from MPEG-2 bit-stream and the watershed algorithm. The algorithm extracts objects using user\`s assistance in the intra frame and tracks then in the following inter frames. If we have an unsatisfactory result for a fast moving object, the user can intervene to correct the segmentation. The proposed algorithm consist of two steps, which are intra frame object extracts processing and inter frame tracking processing. Object extracting process is the step in which user extracts a semantic object directly by using the block classification and watersheds. Object tacking process is the step of the following the object in the subsequent frames. It is based on the boundary fitting method using motion vector, object-mask, and modified watersheds. Experimental results show that the proposed method can achieve a fast conversion from the MPEG-2 bit-stream to the object-based MPEG-4 input.

  • PDF

An Automatic Mobile Cell Counting System for the Analysis of Biological Image (생물학적 영상 분석을 위한 자동 모바일 셀 계수 시스템)

  • Seo, Jaejoon;Chun, Junchul;Lee, Jin-Sung
    • Journal of Internet Computing and Services
    • /
    • v.16 no.1
    • /
    • pp.39-46
    • /
    • 2015
  • This paper presents an automatic method to detect and count the cells from microorganism images based on mobile environments. Cell counting is an important process in the field of biological and pathological image analysis. In the past, cell counting is done manually, which is known as tedious and time consuming process. Moreover, the manual cell counting can lead inconsistent and imprecise results. Therefore, it is necessary to make an automatic method to detect and count cells from biological images to obtain accurate and consistent results. The proposed multi-step cell counting method automatically segments the cells from the image of cultivated microorganism and labels the cells by utilizing topological analysis of the segmented cells. To improve the accuracy of the cell counting, we adopt watershed algorithm in separating agglomerated cells from each other and morphological operation in enhancing the individual cell object from the image. The system is developed by considering the availability in mobile environments. Therefore, the cell images can be obtained by a mobile phone and the processed statistical data of microorganism can be delivered by mobile devices in ubiquitous smart space. From the experiments, by comparing the results between manual and the proposed automatic cell counting we can prove the efficiency of the developed system.

Fast information extraction algorithm for object-based MPEG-4 conversion from MPEG-1,2 (MPEG-1,2로부터 객체 기반 MPEG-4 변환을 위한 고속 정보 추출 알고리즘)

  • 양종호;박성욱
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.3
    • /
    • pp.91-102
    • /
    • 2004
  • In this paper, a fast information extraction algorithm for object-based MPEG-4 application from MPEG-1,2 is proposed. For object-based MPEG-4 conversion, we need to extract such information as object-image, shape-image, macro-block motion vector, and header information from MPEG-1,2 bit-stream. If we use the extracted information, fast conversion for object-based MPEG-4 is possible. The proposed object extraction algerian has two important steps, namely the motion vector extraction from MPEG-1,2 bit-stream and the watershed algerian The algorithm extracts objects using user's assistance in the intra frame and tracks then in the following inter frames. If we have an unsatisfactory result for a fast moving object the user can intervene to connect the segmentation. The proposed algorithm consist of two steps, which are intra frame object extracting processing and inter frame tracking processing. Object extracting process is the step in which user extracts a semantic object directly by using the block classification and watersheds. Object tracking process is the step of the following the object in the subsequent frames. It is based on the boundary fitting method using motion vector, object-mask and modified watersheds. Experimental results show that the proposed method can achieve a fast conversion from the MPEG-1,2 bit-stream to the object-based MPEG-4 input.

A Study on Class Sample Extraction Technique Using Histogram Back-Projection for Object-Based Image Classification (객체 기반 영상 분류를 위한 히스토그램 역투영을 이용한 클래스 샘플 추출 기법에 관한 연구)

  • Chul-Soo Ye
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.2
    • /
    • pp.157-168
    • /
    • 2023
  • Image segmentation and supervised classification techniques are widely used to monitor the ground surface using high-resolution remote sensing images. In order to classify various objects, a process of defining a class corresponding to each object and selecting samples belonging to each class is required. Existing methods for extracting class samples should select a sufficient number of samples having similar intensity characteristics for each class. This process depends on the user's visual identification and takes a lot of time. Representative samples of the class extracted are likely to vary depending on the user, and as a result, the classification performance is greatly affected by the class sample extraction result. In this study, we propose an image classification technique that minimizes user intervention when extracting class samples by applying the histogram back-projection technique and has consistent intensity characteristics of samples belonging to classes. The proposed classification technique using histogram back-projection showed improved classification accuracy in both the experiment using hue subchannels of the hue saturation value transformed image from Compact Advanced Satellite 500-1 imagery and the experiment using the original image compared to the technique that did not use histogram back-projection.