• Title/Summary/Keyword: 에지 병합

Search Result 44, Processing Time 0.019 seconds

Development of CPLD technology mapping algorithm improving run-time under Time Constraint (시간적 조건에서 실행시간을 개선한 CPLD 기술 매핑 알고리즘 개발)

  • 윤충모;김희석
    • Journal of the Korea Society of Computer and Information
    • /
    • v.4 no.3
    • /
    • pp.35-46
    • /
    • 1999
  • In this paper, we propose a new CPLD technology mapping algorithm improving run-time under time constraint. In our technology mapping algorithm. a given logic equation is constructed as the DAG type. then the DAG is reconstructed by replicating the node that outdegree is more than or equal to 2. As a result, it makes delay time and the number of CLBs, run-time to be minimized. Also. after the number of multi-level is defined and cost of each nodes is calculated, the graph is partitioned in order to fit to k that is the number of OR term within Cl.B. The partitioned nodes are merged through collapsing and bin packing is performed in order to fit to the number of OR term within CLB. In the results of experiments to MCNC circuits for logic synthesis benchmark, we can shows that proposed technology mapping algorithm reduces run-time much more than the TMCPLD.

Detection of Artificial Caption using Temporal and Spatial Information in Video (시·공간 정보를 이용한 동영상의 인공 캡션 검출)

  • Joo, SungIl;Weon, SunHee;Choi, HyungIl
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.2
    • /
    • pp.115-126
    • /
    • 2012
  • The artificial captions appearing in videos include information that relates to the videos. In order to obtain the information carried by captions, many methods for caption extraction from videos have been studied. Most traditional methods of detecting caption region have used one frame. However video include not only spatial information but also temporal information. So we propose a method of detection caption region using temporal and spatial information. First, we make improved Text-Appearance-Map and detect continuous candidate regions through matching between candidate-regions. Second, we detect disappearing captions using disappearance test in candidate regions. In case of captions disappear, the caption regions are decided by a merging process which use temporal and spatial information. Final, we decide final caption regions through ANNs using edge direction histograms for verification. Our proposed method was experienced on many kinds of captions with a variety of sizes, shapes, positions and the experiment result was evaluated through Recall and Precision.

Block-based Color Image Segmentation Using Y/C Bit-Plane Sum]nation Image (Y/C 비트 평면합 영상을 이용한 블록 기반 칼라 영상 분할)

  • Kwak, No-Yoon
    • Journal of Digital Contents Society
    • /
    • v.1 no.1
    • /
    • pp.53-64
    • /
    • 2000
  • This paper is related to color image segmentation scheme which makes it possible to achieve the excellent segmented results by block-based segmentation using Y/C bit-plane summation image. First, normalized chrominance summation image is obtained by normalizing the image which is summed up the absolutes of color-differential values between R, G, B images. Secondly, upper 2 bits of the luminance image and upper 6bits of and the normalized chrominance summation image are bitwise operated by the pixel to generate the Y/C bit-plane summation image. Next, the Y/C bit-plane summation image divided into predetermined block size, is classified into monotone blocks, texture blocks and edge blocks, and then each classified block is merged to the regions including one more blocks in the individual block type, and each region is selectively allocated to unique marker according to predetermined marker allocation rules. Finally, fine segmented results are obtained by applying the watershed algorithm to each pixel in the unmarked blocks. As shown in computer simulation, the main advantage of the proposed method is that it suppresses the over-segmentation in the texture regions and reduces computational load. Furthermore, it is able to apply global parameters to various images with different pixel distribution properties because they are nonsensitive for pixel distribution. Especially, the proposed method offers reasonable segmentation results in edge areas with lower contrast owing to the regional characteristics of the color components reflected in the Y/C bit-plane summation image.

  • PDF

Scene Text Extraction in Natural Images using Hierarchical Feature Combination and Verification (계층적 특징 결합 및 검증을 이용한 자연이미지에서의 장면 텍스트 추출)

  • 최영우;김길천;송영자;배경숙;조연희;노명철;이성환;변혜란
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.4
    • /
    • pp.420-438
    • /
    • 2004
  • Artificially or naturally contained texts in the natural images have significant and detailed information about the scenes. If we develop a method that can extract and recognize those texts in real-time, the method can be applied to many important applications. In this paper, we suggest a new method that extracts the text areas in the natural images using the low-level image features of color continuity. gray-level variation and color valiance and that verifies the extracted candidate regions by using the high-level text feature such as stroke. And the two level features are combined hierarchically. The color continuity is used since most of the characters in the same text lesion have the same color, and the gray-level variation is used since the text strokes are distinctive in their gray-values to the background. Also, the color variance is used since the text strokes are distinctive in their gray-values to the background, and this value is more sensitive than the gray-level variations. The text level stroke features are extracted using a multi-resolution wavelet transforms on the local image areas and the feature vectors are input to a SVM(Support Vector Machine) classifier for the verification. We have tested the proposed method using various kinds of the natural images and have confirmed that the extraction rates are very high even in complex background images.