• Title/Summary/Keyword: 최적 후보

Search Result 344, Processing Time 0.025 seconds

Development of polygon object set matching algorithm between heterogeneous digital maps - using the genetic algorithm based on the shape similarities (형상 유사도 기반의 유전 알고리즘을 활용한 이종 수치지도 간의 면 객체 집합 정합 알고리즘 개발)

  • Huh, Yong;Lee, Jeabin
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.31 no.1
    • /
    • pp.1-9
    • /
    • 2013
  • This paper proposes a matching algorithm to find corresponding polygon feature sets between heterogeneous digital maps. The algorithm finds corresponding sets in terms of optimizing their shape similarities based on the assumption that the feature sets describing the same entities in the real world are represented in similar shapes. Then, by using a binary code, it is represented that a polygon feature is chosen for constituting a corresponding set or not. These codes are combined into a binary string as a candidate solution of the matching problem. Starting from initial candidate solutions, a genetic algorithm iteratively optimizes the candidate solutions until it meets a termination condition. Finally, it presents the solution with the highest similarity. The proposed method is applied for the topographical and cadastral maps of an urban region in Suwon, Korea to find corresponding polygon feature sets for block areas, and the results show its feasibility. The results were assessed with manual detection results, and showed overall accuracy of 0.946.

Fast Motion Estimation Algorithm for Efficient MPEG-2 Video Transcoding with Scan Format Conversion (스캔 포맷 변환이 있는 효율적인 MPEG-2 동영상 트랜스코딩을 위한 고속 움직임 추정 기법)

  • 송병철;천강욱
    • Journal of Broadcast Engineering
    • /
    • v.8 no.3
    • /
    • pp.288-296
    • /
    • 2003
  • ATSC (Advanced Television System Committee) has specified 18 video formats for DTV (Digital Television), e.g., scan format, size format, and frame rate format conversion. Effective MPEG-2 video transcoders should support any conversion between the above-mentioned formats. Scan format conversion Is hard to Implement because it may often induce frame rate and size format conversion together. Especially. because of picture type conversion caused by scan format conversion, the computational burden of motion estimation (ME) in transcoding becomes serious. This paper proposes a fast ME algorithm for MPEG-2 video transcoding supporting scan format conversion. Firstly, we extract and compose a set of candidate motion vectors (MVs) from the input bit-stream to comply with the re-encoding format. Secondly, the best MV is chosen among several candidate MVs by using a weighted median selector. Simulation results show that the proposed ME algorithm provides outstanding PSNR performance close to full search ME, while reducing the transcoding complexity significantly.

Real-time Hand Pose Recognition Using HLF (HLF(Haar-like Feature)를 이용한 실시간 손 포즈 인식)

  • Kim, Jang-Woon;Kim, Song-Gook;Hong, Seok-Ju;Jang, Han-Byul;Lee, Chil-Woo
    • 한국HCI학회:학술대회논문집
    • /
    • 2007.02a
    • /
    • pp.897-902
    • /
    • 2007
  • 인간과 컴퓨터간의 전통적인 인터페이스는 인간이 요구하는 다양한 인터페이스를 제공하지 못한다는 점에서 점차 사용하기 불편하게 되었고 이는 새로운 형태의 인터페이스에 대한 요구로 이어지게 되었다. 본 논문에서는 이러한 추세에 맞추어 카메라를 통해 인간의 손 제스처를 인식하는 새로운 인터페이스를 연구하였다. 손은 자유도가 높고 3차원의 view direction에 의해 형상이 매우 심하게 변한다. 따라서 윤곽선 기반방법과 같은 2차원으로 투영된 영상에서 contour나 edge의 정보로 손 제스처를 인식하는 데는 한계가 있다. 그러나 모델기반 방법은 3차원 정보를 이용하기 때문에 손 제스처를 인식하는데 좋으나 계산량이 많아 실시간으로 처리하기가 쉽지 않다. 이러한 문제점을 해결하기 위해 손 형상에 대한 대규모 데이터베이스를 구성하고 정규화된 공간에서 Feature 간의 연관성을 파악하여 훈련 데이터 모델을 구성하여 비교함으로써 실시간으로 손 포즈를 구별할 수 있다. 이러한 통계적 학습 기반의 알고리즘은 다양한 데이터와 좋은 feature의 검출이 최적의 성능을 구현하는 것과 연관된다. 따라서 배경으로부터 노이즈를 최대한 줄이기 위해 피부의 색상 정보를 이용하여 손 후보 영역을 검출하고 검출된 후보 영역으로부터 HLF(Haar-like Feature)를 이용하여 손 영역을 검출한다. 검출된 손 영역으로부터 패턴 분류 과정을 거쳐 손 포즈를 인식 하게 된다. 패턴 분류 과정은 HLF를 이용하여 손 포즈를 인식하게 되는데 미리 학습된 각 포즈에 대한 HLF를 이용하여 손 포즈를 인식하게 된다. HLF는 Violar가 얼굴 검출에 적용한 것으로 얼굴 검출에 좋은 결과를 보여 주었으며, 이는 적분 이미지로부터 추출한 HLF를 이용한 Adaboost 학습 알고리즘을 사용하였다. 본 논문에서는 피부색의 색상 정보를 이용 배경과 손 영상을 최대한 분리하여 배경의 대부분이 Adaboost-Haar Classifier의 첫 번째 스테이지에서 제거되는 방법을 이용하여 그 성능을 더 향상 시켜 손 형상 인식에 적용하였다.

  • PDF

Goal-Directed Reinforcement Learning System (목표지향적 강화학습 시스템)

  • Lee, Chang-Hoon
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.10 no.5
    • /
    • pp.265-270
    • /
    • 2010
  • Reinforcement learning performs learning through interacting with trial-and-error in dynamic environment. Therefore, in dynamic environment, reinforcement learning method like TD-learning and TD(${\lambda}$)-learning are faster in learning than the conventional stochastic learning method. However, because many of the proposed reinforcement learning algorithms are given the reinforcement value only when the learning agent has reached its goal state, most of the reinforcement algorithms converge to the optimal solution too slowly. In this paper, we present GDRLS algorithm for finding the shortest path faster in a maze environment. GDRLS is select the candidate states that can guide the shortest path in maze environment, and learn only the candidate states to find the shortest path. Through experiments, we can see that GDRLS can search the shortest path faster than TD-learning and TD(${\lambda}$)-learning in maze environment.

Complexity Reduction of HEVC SAO Intra Modes By Adjustment of Offset Values (HEVC SAO 인트라 모드 오프셋 값 조정을 통한 복잡도 감소)

  • Mun, Ji-Hun;Choi, Jung-Ah;Ho, Yo-Sung
    • Journal of Broadcast Engineering
    • /
    • v.19 no.3
    • /
    • pp.355-361
    • /
    • 2014
  • In this paper, we propose a complexity reduction method of sample adaptive offset (SAO), which is an in-loop filter in high-efficiency video coding (HEVC). In the conventional SAO, an offset value is calculated for each coding tree block (CTB) to minimize the error between the original and reconstructed images. In order to determine the optimal offset value, all offset candidates are examined and the offset value that leads to the smallest rate-distortion cost is chosen. Thus, SAO occupies a significant amount of the computational complexity in the HEVC encoder. In the proposed method, we determine the least-used band (LUB) by considering the statistical characteristics of offset values and without processing the offset value included in the LUB. Also, in the offset value decision stage, we check only a certain number of candidates rather than all of them. Experimental results show that the proposed method reduces the encoding time by approximately 8.15% without yielding a significant loss in terms of coding efficiency.

Natural Scene Text Binarization using Tensor Voting and Markov Random Field (텐서보팅과 마르코프 랜덤 필드를 이용한 자연 영상의 텍스트 이진화)

  • Choi, Hyun Su;Lee, Guee Sang
    • Smart Media Journal
    • /
    • v.4 no.4
    • /
    • pp.18-23
    • /
    • 2015
  • In this paper, we propose a method for detecting the number of clusters. This method can improve the performance of a gaussian mixture model function in conventional markov random field method by using the tensor voting. The key point of the proposed method is that extracts the number of the center through the continuity of saliency map of the input data of the tensor voting token. At first, we separate the foreground and background region candidate in a given natural images. After that, we extract the appropriate cluster number for each separate candidate regions by applying the tensor voting. We can make accurate modeling a gaussian mixture model by using a detected number of cluster. We can return the result of natural binary text image by calculating the unary term and the pairwise term of markov random field. After the experiment, we can confirm that the proposed method returns the optimal cluster number and text binarization results are improved.

Selection Method of Fuzzy Partitions in Fuzzy Rule-Based Classification Systems (퍼지 규칙기반 분류시스템에서 퍼지 분할의 선택방법)

  • Son, Chang-S.;Chung, Hwan-M.;Kwon, Soon-H.
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.18 no.3
    • /
    • pp.360-366
    • /
    • 2008
  • The initial fuzzy partitions in fuzzy rule-based classification systems are determined by considering the domain region of each attribute with the given data, and the optimal classification boundaries within the fuzzy partitions can be discovered by tuning their parameters using various learning processes such as neural network, genetic algorithm, and so on. In this paper, we propose a selection method for fuzzy partition based on statistical information to maximize the performance of pattern classification without learning processes where statistical information is used to extract the uncertainty regions (i.e., the regions which the classification boundaries in pattern classification problems are determined) in each input attribute from the numerical data. Moreover the methods for extracting the candidate rules which are associated with the partition intervals generated by statistical information and for minimizing the coupling problem between the candidate rules are additionally discussed. In order to show the effectiveness of the proposed method, we compared the classification accuracy of the proposed with those of conventional methods on the IRIS and New Thyroid Cancer data. From experimental results, we can confirm the fact that the proposed method only considering statistical information of the numerical patterns provides equal to or better classification accuracy than that of the conventional methods.

Automatic Generation of Aimpoints Using 3D Target Shapes and CEP (3차원 모델 형상과 원형공산오차를 고려한 목표점 자동 생성 기법)

  • Kang, Yuna
    • Journal of the Korea Society for Simulation
    • /
    • v.28 no.2
    • /
    • pp.71-79
    • /
    • 2019
  • This research proposes an automatic generation system of aimpoints with considering 3D target shapes and CEP(circular error probability) values after determining a target and an azimuth value. In the past, users decided aimpoints manually based on experience or just chose a middle point of the 3D model. Futhermore, it was not possible to select a proper position with consideration of azimuth because users should decide aimpoints before the azimuth value was determined. To solve this problem, this research provides a automatic system to compute proper aimpoints with 3D target shapes, azimuth values and CEP. This article contains the explanation of 3 steps for generating aimpoints automatically: first, generating a reference plane and candidate entry points, then computing scores of each entry points and finally determining an aimpoint from the entry point with the highest score. Users can easily determine a final aimpoint with high probability of success using this system.

Lip Contour Detection by Multi-Threshold (다중 문턱치를 이용한 입술 윤곽 검출 방법)

  • Kim, Jeong Yeop
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.9 no.12
    • /
    • pp.431-438
    • /
    • 2020
  • In this paper, the method to extract lip contour by multiple threshold is proposed. Spyridonos et. el. proposed a method to extract lip contour. First step is get Q image from transform of RGB into YIQ. Second step is to find lip corner points by change point detection and split Q image into upper and lower part by corner points. The candidate lip contour can be obtained by apply threshold to Q image. From the candidate contour, feature variance is calculated and the contour with maximum variance is adopted as final contour. The feature variance 'D' is based on the absolute difference near the contour points. The conventional method has 3 problems. The first one is related to lip corner point. Calculation of variance depends on much skin pixels and therefore the accuracy decreases and have effect on the split for Q image. Second, there is no analysis for color systems except YIQ. YIQ is a good however, other color systems such as HVS, CIELUV, YCrCb would be considered. Final problem is related to selection of optimal contour. In selection process, they used maximum of average feature variance for the pixels near the contour points. The maximum of variance causes reduction of extracted contour compared to ground contours. To solve the first problem, the proposed method excludes some of skin pixels and got 30% performance increase. For the second problem, HSV, CIELUV, YCrCb coordinate systems are tested and found there is no relation between the conventional method and dependency to color systems. For the final problem, maximum of total sum for the feature variance is adopted rather than the maximum of average feature variance and got 46% performance increase. By combine all the solutions, the proposed method gives 2 times in accuracy and stability than conventional method.

Fast Multi-Resolution Exhaustive Search Algorithm Based on Clustering for Efficient Image Retrieval (효율적인 영상 검색을 위한 클러스터링 기반 고속 다 해상도 전역 탐색 기법)

  • Song, Byeong-Cheol;Kim, Myeong-Jun;Ra, Jong-Beom
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.38 no.2
    • /
    • pp.117-128
    • /
    • 2001
  • In order to achieve optimal retrieval, i.e., to find the best match to a query according to a certain similarity measure, the exhaustive search should be performed literally for all the images in a database. However, the straightforward exhaustive search algorithm is computationally expensive in large image databases. To reduce its heavy computational cost, this paper presents a fast exhaustive multi-resolution search algorithm based on image database clustering. Firstly, the proposed algorithm partitions the whole image data set into a pre-defined number of clusters having similar feature contents. Next, for a given query, it checks the lower bound of distances in each cluster, eliminating disqualified clusters. Then, it only examines the candidates in the remaining clusters. To alleviate unnecessary feature matching operations in the search procedure, the distance inequality property is employed based on a multi-resolution data structure. The proposed algorithm realizes a fast exhaustive multi-resolution search for either the best match or multiple best matches to the query. Using luminance histograms as a feature, we prove that the proposed algorithm guarantees optimal retrieval with high searching speed.

  • PDF