• Title/Summary/Keyword: image segmentation technique

Search Result 346, Processing Time 0.032 seconds

Facial Features and Motion Recovery using multi-modal information and Paraperspective Camera Model (다양한 형식의 얼굴정보와 준원근 카메라 모델해석을 이용한 얼굴 특징점 및 움직임 복원)

  • Kim, Sang-Hoon
    • The KIPS Transactions:PartB
    • /
    • v.9B no.5
    • /
    • pp.563-570
    • /
    • 2002
  • Robust extraction of 3D facial features and global motion information from 2D image sequence for the MPEG-4 SNHC face model encoding is described. The facial regions are detected from image sequence using multi-modal fusion technique that combines range, color and motion information. 23 facial features among the MPEG-4 FDP (Face Definition Parameters) are extracted automatically inside the facial region using color transform (GSCD, BWCD) and morphological processing. The extracted facial features are used to recover the 3D shape and global motion of the object using paraperspective camera model and SVD (Singular Value Decomposition) factorization method. A 3D synthetic object is designed and tested to show the performance of proposed algorithm. The recovered 3D motion information is transformed into global motion parameters of FAP (Face Animation Parameters) of the MPEG-4 to synchronize a generic face model with a real face.

Fiber Classification and Detection Technique Proposed for Applying on the PVA-ECC Sectional Image (PVA-ECC단면 이미지의 섬유 분류 및 검출 기법)

  • Kim, Yun-Yong;Lee, Bang-Yeon;Kim, Jin-Keun
    • Journal of the Korea Concrete Institute
    • /
    • v.20 no.4
    • /
    • pp.513-522
    • /
    • 2008
  • The fiber dispersion performance in fiber-reinforced cementitious composites is a crucial factor with respect to achieving desired mechanical performance. However, evaluation of the fiber dispersion performance in the composite PVA-ECC (Polyvinyl alcohol-Engineered Cementitious Composite) is extremely challenging because of the low contrast of PVA fibers with the cement-based matrix. In the present work, an enhanced fiber detection technique is developed and demonstrated. Using a fluorescence technique on the PVA-ECC, PVA fibers are observed as green dots in the cross-section of the composite. After capturing the fluorescence image with a Charged Couple Device (CCD) camera through a microscope. The fibers are more accurately detected by employing a series of process based on a categorization, watershed segmentation, and morphological reconstruction.

Deep learning algorithm of concrete spalling detection using focal loss and data augmentation (Focal loss와 데이터 증강 기법을 이용한 콘크리트 박락 탐지 심층 신경망 알고리즘)

  • Shim, Seungbo;Choi, Sang-Il;Kong, Suk-Min;Lee, Seong-Won
    • Journal of Korean Tunnelling and Underground Space Association
    • /
    • v.23 no.4
    • /
    • pp.253-263
    • /
    • 2021
  • Concrete structures are damaged by aging and external environmental factors. This type of damage is to appear in the form of cracks, to proceed in the form of spalling. Such concrete damage can act as the main cause of reducing the original design bearing capacity of the structure, and negatively affect the stability of the structure. If such damage continues, it may lead to a safety accident in the future, thus proper repair and reinforcement are required. To this end, an accurate and objective condition inspection of the structure must be performed, and for this inspection, a sensor technology capable of detecting damage area is required. For this reason, we propose a deep learning-based image processing algorithm that can detect spalling. To develop this, 298 spalling images were obtained, of which 253 images were used for training, and the remaining 45 images were used for testing. In addition, an improved loss function and data augmentation technique were applied to improve the detection performance. As a result, the detection performance of concrete spalling showed a mean intersection over union of 80.19%. In conclusion, we developed an algorithm to detect concrete spalling through a deep learning-based image processing technique, with an improved loss function and data augmentation technique. This technology is expected to be utilized for accurate inspection and diagnosis of structures in the future.

Tillage boundary detection based on RGB imagery classification for an autonomous tractor

  • Kim, Gookhwan;Seo, Dasom;Kim, Kyoung-Chul;Hong, Youngki;Lee, Meonghun;Lee, Siyoung;Kim, Hyunjong;Ryu, Hee-Seok;Kim, Yong-Joo;Chung, Sun-Ok;Lee, Dae-Hyun
    • Korean Journal of Agricultural Science
    • /
    • v.47 no.2
    • /
    • pp.205-217
    • /
    • 2020
  • In this study, a deep learning-based tillage boundary detection method for autonomous tillage by a tractor was developed, which consisted of image cropping, object classification, area segmentation, and boundary detection methods. Full HD (1920 × 1080) images were obtained using a RGB camera installed on the hood of a tractor and were cropped to 112 × 112 size images to generate a dataset for training the classification model. The classification model was constructed based on convolutional neural networks, and the path boundary was detected using a probability map, which was generated by the integration of softmax outputs. The results show that the F1-score of the classification was approximately 0.91, and it had a similar performance as the deep learning-based classification task in the agriculture field. The path boundary was determined with edge detection and the Hough transform, and it was compared to the actual path boundary. The average lateral error was approximately 11.4 cm, and the average angle error was approximately 8.9°. The proposed technique can perform as well as other approaches; however, it only needs low cost memory to execute the process unlike other deep learning-based approaches. It is possible that an autonomous farm robot can be easily developed with this proposed technique using a simple hardware configuration.

Development of the Visualization Prototype of Radar Rainfall Data Using the Unity 3D Engine (Unity 3D 엔진을 활용한 강우레이더 자료 시각화 프로토타입 개발)

  • CHOI, Hyeoung-Wook;KANG, Soo-Myung;KIM, Kyung-Jun;KIM, Dong-Young;CHOUNG, Yun-Jae
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.18 no.4
    • /
    • pp.131-144
    • /
    • 2015
  • This research proposes a prototype for visualizing radar rainfall data using the unity 3D engine. The mashup of radar data with topographic information is necessary for the 3D visualization of the radar data with high quality. However, the mashup of a huge amount of radar data and topographic data causes the overload of data processing and low quality of the visualization results. This research utilized the Unitiy 3D engine, a widely used engine in the game industry, for visualizing the 3D topographic data such as the satellite imagery/the DEM(Digital Elevation Model) and radar rainfall data. The satellite image segmentation technique and the image texture layer mashup technique are employed to construct the 3D visualization system prototype based on the topographic information. The developed protype will be applied to the disaster-prevention works by providing the radar rainfall data with the 3D visualization based on the topographic information.

A Study on the Deep Neural Network based Recognition Model for Space Debris Vision Tracking System (심층신경망 기반 우주파편 영상 추적시스템 인식모델에 대한 연구)

  • Lim, Seongmin;Kim, Jin-Hyung;Choi, Won-Sub;Kim, Hae-Dong
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.45 no.9
    • /
    • pp.794-806
    • /
    • 2017
  • It is essential to protect the national space assets and space environment safely as a space development country from the continuously increasing space debris. And Active Debris Removal(ADR) is the most active way to solve this problem. In this paper, we studied the Artificial Neural Network(ANN) for a stable recognition model of vision-based space debris tracking system. We obtained the simulated image of the space environment by the KARICAT which is the ground-based space debris clearing satellite testbed developed by the Korea Aerospace Research Institute, and created the vector which encodes structure and color-based features of each object after image segmentation by depth discontinuity. The Feature Vector consists of 3D surface area, principle vector of point cloud, 2D shape and color information. We designed artificial neural network model based on the separated Feature Vector. In order to improve the performance of the artificial neural network, the model is divided according to the categories of the input feature vectors, and the ensemble technique is applied to each model. As a result, we confirmed the performance improvement of recognition model by ensemble technique.

Detecting high-resolution usage status of individual parcel of land using object detecting deep learning technique (객체 탐지 딥러닝 기법을 활용한 필지별 조사 방안 연구)

  • Jeon, Jeong-Bae
    • Journal of Cadastre & Land InformatiX
    • /
    • v.54 no.1
    • /
    • pp.19-32
    • /
    • 2024
  • This study examined the feasibility of image-based surveys by detecting objects in facilities and agricultural land using the YOLO algorithm based on drone images and comparing them with the land category by law. As a result of detecting objects through the YOLO algorithm, buildings showed a performance of detecting objects corresponding to 96.3% of the buildings provided in the existing digital map. In addition, the YOLO algorithm developed in this study detected 136 additional buildings that were not located in the digital map. Plastic greenhouses detected a total of 297 objects, but the detection rate was low for some plastic greenhouses for fruit trees. Also, agricultural land had the lowest detection rate. This result is because agricultural land has a larger area and irregular shape than buildings, so the accuracy is lower than buildings due to the inconsistency of training data. Therefore, segmentation detection, rather than box-shaped detection, is likely to be more effective for agricultural fields. Comparing the detected objects with the land category by law, it was analyzed that some buildings exist in agricultural and forest areas where it is difficult to locate buildings. It seems that it is necessary to link with administrative information to understand that these buildings are used illegally. Therefore, at the current level, it is possible to objectively determine the existence of buildings in fields where it is difficult to locate buildings.

Object Extraction Technique using Extension Search Algorithm based on Bidirectional Stereo Matching (양방향 스테레오 정합 기반 확장탐색 알고리즘을 이용한 물체추출 기법)

  • Choi, Young-Seok;Kim, Seung-Geun;Kang, Hyun-Soo
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.2
    • /
    • pp.1-9
    • /
    • 2008
  • In this paper, to extract object regions in stereo image, we propose an enhanced algorithm that extracts objects combining both of brightness information and disparity information. The approach that extracts objects using both has been studied by Ping and Chaohui. In their algorithm, the segmentation for an input image is carried out using the brightness, and integration of segmented regions in consideration of disparity information within the previously segmented regions. In the regions where the brightness values between object regions and background regions are similar, however, the segmented regions probably include both of object regions and background regions. It may cause incorrect object extraction in the merging process executed in the unit of the segmented region. To solve this problem, in proposed method, we adopt the merging process which is performed in pixel unit. In addition, we perform the bi-directional stereo matching process to enhance reliability of the disparity information and supplement the disparity information resulted from a single directional matching process. Further searching for disparity is decided by edge information of the input image. The proposed method gives good performance in the object extraction since we find the disparity information that is not extracted in the traditional methods. Finally, we evaluate our method by experiments for the pictures acquired from a real stereoscopic camera.

Front Classification using Back Propagation Algorithm (오류 역전파 알고리즘을 이용한 영문자의 폰트 분류 방법에 관한 연구)

  • Jung Minchul
    • Journal of Intelligence and Information Systems
    • /
    • v.10 no.2
    • /
    • pp.65-77
    • /
    • 2004
  • This paper presents a priori and the local font classification method. The font classification uses ascenders, descenders, and serifs extracted from a word image. The gradient features of those sub-images are extracted, and used as an input to a neural network classifier to produce font classification results. The font classification determines 2 font styles (upright or slant), 3 font groups (serif sans-serif or typewriter), and 7-font names (Postscript fonts such as Avant Garde, Helvetica, Bookman, New Century Schoolbook, Palatine, Times, and Courier). The proposed a priori and local font classification method allows an OCR system consisting of various font-specific character segmentation tools and various mono-font character recognizers. Experiments have shown font classification accuracies reach high performance levels of about 95.4 percent even with severely touching characters. The technique developed for tile selected 7 fonts in this paper can be applied to any other fonts.

  • PDF

Indirect Volume Rendering of Hepatobiliary System from CT and MRI Images (CT와 MRI 영상을 이용한 간담도계 간접볼륨렌더링)

  • Jin, Gye-Hwan;Lee, Tae-Soo
    • Journal of the Korean Society of Radiology
    • /
    • v.1 no.2
    • /
    • pp.23-30
    • /
    • 2007
  • This paper presents a method of generating 3-dimensional images by preprocessing 2-dimensional abdominal images obtained using CT (computed tomography) and MRI (magnetic resonance imaging) through segmentation, threshold technique, etc. and apply the method to virtual endoscopy. Three-dimensional images were visualized using indirect volume rendering, which can render at high speed using a general-purpose graphic accelerator used in personal computers. The algorithm used in the rendering is Marching Cubes, which has only a small volume of calculation. In addition, we suggested a method of producing 3-dimensional images in VRML (virtual reality modeling language) running on the Web browser without a workstation or an exclusive program. The number of nodes, the number of triangles and the size of a 3-dimensional image file from CT were 85,367, 174,150 and 10,124, respectively, and those from MRI were 34,029, 67,824 and 3,804, respectively.

  • PDF