• Title/Summary/Keyword: Candidate Vector Sets

Search Result 9, Processing Time 0.025 seconds

Shape and Appearance Repair for Incomplete Point Surfaces (결함이 있는 점집합 곡면의 형상 및 외관 수정)

  • Park, Se-Youn;Guo, Xiaohu;Shin, Ha-Yong;Qin, Hong
    • Korean Journal of Computational Design and Engineering
    • /
    • v.12 no.5
    • /
    • pp.330-343
    • /
    • 2007
  • In this paper, we present a new surface content completion system that can effectively repair both shape and appearance from scanned, incomplete point set inputs. First, geometric holes can be robustly identified from noisy and defective data sets without the need for any normal or orientation information. The geometry and texture information of the holes can then be determined either automatically from the models' context, or manually from users' selection. After identifying the patch that most resembles each hole region, the geometry and texture information can be completed by warping the candidate region and gluing it onto the hole area. The displacement vector field for the exact alignment process is computed by solving a Poisson equation with boundary conditions. Out experiments show that the unified framework, founded upon the techniques of deformable models and PDE modeling, can provide a robust and elegant solution for content completion of defective, complex point surfaces.

Motion Vector Recovery Scheme for H.264/AVC (H.264/AVC을 위한 움직임 벡터 복원 방법)

  • Son, Nam-Rye
    • The Journal of the Korea Contents Association
    • /
    • v.8 no.5
    • /
    • pp.29-37
    • /
    • 2008
  • To transmit video bit stream over low bandwidth such as wireless channel, high compression algorithm like H.264 codec is exploited. In transmitting high compressed video bit-stream over low bandwidth, packet loss causes severe degradation in image quality. In this paper, a new algorithm for recovery of missing or erroneous motion vector is proposed. Considering that the missing or erroneous motion vectors in blocks are closely correlated with those of neighboring blocks. Motion vector of neighboring blocks are clustered according to average linkage algorithm clustering and a representative value for each cluster is determined to obtain the candidate motion vector sets. As a result, simulation results show that the proposed method dramatically improves processing time compared to existing H.264/AVC. Also the proposed method is similar to existing H.264/AVC in terms of visual quality.

A partially occluded object recognition technique using a probabilistic analysis in the feature space (특징 공간상에서 의 확률적 해석에 기반한 부분 인식 기법에 관한 연구)

  • 박보건;이경무;이상욱;이진학
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.26 no.11A
    • /
    • pp.1946-1956
    • /
    • 2001
  • In this paper, we propose a novel 2-D partial matching algorithm based on model-based stochastic analysis of feature correspondences in a relation vector space, which is quite robust to shape variations as well as invariant to geometric transformations. We represent an object using the ARG (Attributed Relational Graph) model with features of a set of relation vectors. In addition, we statistically model the partial occlusion or noise as the distortion of the relation vector distribution in the relation vector space. Our partial matching algorithm consists of two-phases. First, a finite number of candidate sets areselected by using logical constraint embedding local and structural consistency Second, the feature loss detection is done iteratively by error detection and voting scheme thorough the error analysis of relation vector space. Experimental results on real images demonstrate that the proposed algorithm is quite robust to noise and localize target objects correctly even inseverely noisy and occluded scenes.

  • PDF

Modulation Recognition of BPSK/QPSK Signals based on Features in the Graph Domain

  • Yang, Li;Hu, Guobing;Xu, Xiaoyang;Zhao, Pinjiao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.11
    • /
    • pp.3761-3779
    • /
    • 2022
  • The performance of existing recognition algorithms for binary phase shift keying (BPSK) and quadrature phase shift keying (QPSK) signals degrade under conditions of low signal-to-noise ratios (SNR). Hence, a novel recognition algorithm based on features in the graph domain is proposed in this study. First, the power spectrum of the squared candidate signal is truncated by a rectangular window. Thereafter, the graph representation of the truncated spectrum is obtained via normalization, quantization, and edge construction. Based on the analysis of the connectivity difference of the graphs under different hypotheses, the sum of degree (SD) of the graphs is utilized as a discriminate feature to classify BPSK and QPSK signals. Moreover, we prove that the SD is a Schur-concave function with respect to the probability vector of the vertices (PVV). Extensive simulations confirm the effectiveness of the proposed algorithm, and its superiority to the listed model-driven-based (MDB) algorithms in terms of recognition performance under low SNRs and computational complexity. As it is confirmed that the proposed method reduces the computational complexity of existing graph-based algorithms, it can be applied in modulation recognition of radar or communication signals in real-time processing, and does not require any prior knowledge about the training sets, channel coefficients, or noise power.

Extracting Alternative Word Candidates for Patent Information Search (특허 정보 검색을 위한 대체어 후보 추출 방법)

  • Baik, Jong-Bum;Kim, Seong-Min;Lee, Soo-Won
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.4
    • /
    • pp.299-303
    • /
    • 2009
  • Patent information search is used for checking existence of earlier works. In patent information search, there are many reasons that fails to get appropriate information. This research proposes a method extracting alternative word candidates in order to minimize search failure due to keyword mismatch. Assuming that two words have similar meaning if they have similar co-occurrence words, the proposed method uses the concept of concentration, association word set, cosine similarity between association word sets and a ranking modification technique. Performance of the proposed method is evaluated using a manually extracted alternative word candidate list. Evaluation results show that the proposed method outperforms the document vector space model in recall.

A New Approach to Automatic Keyword Generation Using Inverse Vector Space Model (키워드 자동 생성에 대한 새로운 접근법: 역 벡터공간모델을 이용한 키워드 할당 방법)

  • Cho, Won-Chin;Rho, Sang-Kyu;Yun, Ji-Young Agnes;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.21 no.1
    • /
    • pp.103-122
    • /
    • 2011
  • Recently, numerous documents have been made available electronically. Internet search engines and digital libraries commonly return query results containing hundreds or even thousands of documents. In this situation, it is virtually impossible for users to examine complete documents to determine whether they might be useful for them. For this reason, some on-line documents are accompanied by a list of keywords specified by the authors in an effort to guide the users by facilitating the filtering process. In this way, a set of keywords is often considered a condensed version of the whole document and therefore plays an important role for document retrieval, Web page retrieval, document clustering, summarization, text mining, and so on. Since many academic journals ask the authors to provide a list of five or six keywords on the first page of an article, keywords are most familiar in the context of journal articles. However, many other types of documents could not benefit from the use of keywords, including Web pages, email messages, news reports, magazine articles, and business papers. Although the potential benefit is large, the implementation itself is the obstacle; manually assigning keywords to all documents is a daunting task, or even impractical in that it is extremely tedious and time-consuming requiring a certain level of domain knowledge. Therefore, it is highly desirable to automate the keyword generation process. There are mainly two approaches to achieving this aim: keyword assignment approach and keyword extraction approach. Both approaches use machine learning methods and require, for training purposes, a set of documents with keywords already attached. In the former approach, there is a given set of vocabulary, and the aim is to match them to the texts. In other words, the keywords assignment approach seeks to select the words from a controlled vocabulary that best describes a document. Although this approach is domain dependent and is not easy to transfer and expand, it can generate implicit keywords that do not appear in a document. On the other hand, in the latter approach, the aim is to extract keywords with respect to their relevance in the text without prior vocabulary. In this approach, automatic keyword generation is treated as a classification task, and keywords are commonly extracted based on supervised learning techniques. Thus, keyword extraction algorithms classify candidate keywords in a document into positive or negative examples. Several systems such as Extractor and Kea were developed using keyword extraction approach. Most indicative words in a document are selected as keywords for that document and as a result, keywords extraction is limited to terms that appear in the document. Therefore, keywords extraction cannot generate implicit keywords that are not included in a document. According to the experiment results of Turney, about 64% to 90% of keywords assigned by the authors can be found in the full text of an article. Inversely, it also means that 10% to 36% of the keywords assigned by the authors do not appear in the article, which cannot be generated through keyword extraction algorithms. Our preliminary experiment result also shows that 37% of keywords assigned by the authors are not included in the full text. This is the reason why we have decided to adopt the keyword assignment approach. In this paper, we propose a new approach for automatic keyword assignment namely IVSM(Inverse Vector Space Model). The model is based on a vector space model. which is a conventional information retrieval model that represents documents and queries by vectors in a multidimensional space. IVSM generates an appropriate keyword set for a specific document by measuring the distance between the document and the keyword sets. The keyword assignment process of IVSM is as follows: (1) calculating the vector length of each keyword set based on each keyword weight; (2) preprocessing and parsing a target document that does not have keywords; (3) calculating the vector length of the target document based on the term frequency; (4) measuring the cosine similarity between each keyword set and the target document; and (5) generating keywords that have high similarity scores. Two keyword generation systems were implemented applying IVSM: IVSM system for Web-based community service and stand-alone IVSM system. Firstly, the IVSM system is implemented in a community service for sharing knowledge and opinions on current trends such as fashion, movies, social problems, and health information. The stand-alone IVSM system is dedicated to generating keywords for academic papers, and, indeed, it has been tested through a number of academic papers including those published by the Korean Association of Shipping and Logistics, the Korea Research Academy of Distribution Information, the Korea Logistics Society, the Korea Logistics Research Association, and the Korea Port Economic Association. We measured the performance of IVSM by the number of matches between the IVSM-generated keywords and the author-assigned keywords. According to our experiment, the precisions of IVSM applied to Web-based community service and academic journals were 0.75 and 0.71, respectively. The performance of both systems is much better than that of baseline systems that generate keywords based on simple probability. Also, IVSM shows comparable performance to Extractor that is a representative system of keyword extraction approach developed by Turney. As electronic documents increase, we expect that IVSM proposed in this paper can be applied to many electronic documents in Web-based community and digital library.

A New Error Concealment Based on Edge Detection (에지검출을 기반으로 한 새로운 에러 은닉 기법)

  • Yang, Yo-Jin;Son, Nam-Rye;Lee, Guee-Sang
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.39 no.6
    • /
    • pp.623-629
    • /
    • 2002
  • In transmitting compressed video bit-stream over Internet, packet losses cause error propagations in both spatial and temporal domains, which in turn leads to severe degradation I image quality. In this paper, a new error concealment algorithm, called EBMA(Edge Detection based Boundary Matching Algorithm), is proposed to repair damaged portions of the video frames in the receiver. Conventional BMA(Boundary Matching Algorithm) assumes that the pixels on the boundary of the missing block and its neighboring blocks are very similar, but has no consideration of edges across the boundary. In our approach, the edges are detected across the boundary of the lost or erroneous block. Once the orientation of each edge is found, only the pixel difference along the expected edges across the boundary is measured instead of the calculation of difference along the expected edges across the boundary is measured instead of the calculation of differences between all adjacent pixels on the boundary Therefore, the proposed approach needs very few computations and the experiment shows and improvement of the performance over the conventional BMA in terms of both subjective and objective quality of video sequences.

Tuple Pruning Using Bloom Filter for Packet Classification (패킷 분류를 위한 블룸 필터 이용 튜플 제거 알고리즘)

  • Kim, So-Yeon;Lim, Hye-Sook
    • Journal of KIISE:Information Networking
    • /
    • v.37 no.3
    • /
    • pp.175-186
    • /
    • 2010
  • Due to the emergence of new application programs and the fast growth of Internet users, Internet routers are required to provide the quality of services according to the class of input packets, which is identified by wire-speed packet classification. For a pre-defined rule set, by performing multi-dimensional search using various header fields of an input packet, packet classification determines the highest priority rule matching to the input packet. Efficient packet classification algorithms have been widely studied. Tuple pruning algorithm provides fast classification performance using hash-based search against the candidate tuples that may include matching rules. Bloom filter is an efficient data structure composed of a bit vector which represents the membership information of each element included in a given set. It is used as a pre-filter determining whether a specific input is a member of a set or not. This paper proposes new tuple pruning algorithms using Bloom filters, which effectively remove unnecessary tuples which do not include matching rules. Using the database known to be similar to actual rule sets used in Internet routers, simulation results show that the proposed tuple pruning algorithm provides faster packet classification as well as consumes smaller memory amount compared with the previous tuple pruning algorithm.

A DB Pruning Method in a Large Corpus-Based TTS with Multiple Candidate Speech Segments (대용량 복수후보 TTS 방식에서 합성용 DB의 감량 방법)

  • Lee, Jung-Chul;Kang, Tae-Ho
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.6
    • /
    • pp.572-577
    • /
    • 2009
  • Large corpus-based concatenating Text-to-Speech (TTS) systems can generate natural synthetic speech without additional signal processing. To prune the redundant speech segments in a large speech segment DB, we can utilize a decision-tree based triphone clustering algorithm widely used in speech recognition area. But, the conventional methods have problems in representing the acoustic transitional characteristics of the phones and in applying context questions with hierarchic priority. In this paper, we propose a new clustering algorithm to downsize the speech DB. Firstly, three 13th order MFCC vectors from first, medial, and final frame of a phone are combined into a 39 dimensional vector to represent the transitional characteristics of a phone. And then the hierarchically grouped three question sets are used to construct the triphone trees. For the performance test, we used DTW algorithm to calculate the acoustic similarity between the target triphone and the triphone from the tree search result. Experimental results show that the proposed method can reduce the size of speech DB by 23% and select better phones with higher acoustic similarity. Therefore the proposed method can be applied to make a small sized TTS.