• Title/Summary/Keyword: 편집 알고리즘

Search Result 152, Processing Time 0.028 seconds

A Method for Spelling Error Correction in Korean Using a Hangul Edit Distance Algorithm (한글 편집거리 알고리즘을 이용한 한국어 철자오류 교정방법)

  • Bak, Seung Hyeon;Lee, Eun Ji;Kim, Pan Koo
    • Smart Media Journal
    • /
    • v.6 no.1
    • /
    • pp.16-21
    • /
    • 2017
  • Long time has passed since computers which used to be a means of research were commercialized and available for the general public. People used writing instruments to write before computer was commercialized. However, today a growing number of them are using computers to write instead. Computerized word processing helps write faster and reduces fatigue of hands than writing instruments, making it better fit to making long texts. However, word processing programs are more likely to cause spelling errors by the mistake of users. Spelling errors distort the shape of words, making it easy for the writer to find and correct directly, but those caused due to users' lack of knowledge or those hard to find may make it almost impossible to produce a document free of spelling errors. However, spelling errors in important documents such as theses or business proposals may lead to falling reliability. Consequently, it is necessary to conduct research on high-level spelling error correction programs for the general public. This study was designed to produce a system to correct sentence-level spelling errors to normal words with Korean alphabet similarity algorithm. On the basis of findings reported in related literatures that corrected words are significantly similar to misspelled words in form, spelling errors were extracted from a corpus. Extracted corrected words were replaced with misspelled ones to correct spelling errors with spelling error detection algorithm.

Musical Score Recognition Using Hierarchical ART2 Algorithm (Hierarchical ART2 알고리즘을 이용한 악보 인식)

  • Kim, Kwang-Baek;Woo, Young-Woon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.13 no.10
    • /
    • pp.1997-2003
    • /
    • 2009
  • Methods for effective musical score recognition and efficient editing of musical scores are demanded because functions of computers for researches on musical activities become more and more important parts in recent days. In the conventional methods for handling musical scores manually, there are weak points such as incorrect score symbols in input process and requirement of much time to adjust the incorrect symbols. And also there is another weak point that the scores edited by each application program can be remodified by a specific application program only. In this paper, we proposed a method for automatic musical score recognition of printed musical scores in order to make up for the weak points. In the proposed method, staffs in a scanned score image are eliminated by horizontal histogram, noises are removed by 4 directional edge tracking algorithm, and then musical score symbols are extracted by using Grassfire algorithm. The extracted symbols are recognized by hierarchical ART2 algorithm. In order to evaluate the performance of the proposed method, we used 100 musical scores for experiment. In the experiment, we verified that the proposed method using hierarchical ART2 algorithm is efficient.

Context-Weighted Metrics for Example Matching (문맥가중치가 반영된 문장 유사 척도)

  • Kim, Dong-Joo;Kim, Han-Woo
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.43 no.6 s.312
    • /
    • pp.43-51
    • /
    • 2006
  • This paper proposes a metrics for example matching under the example-based machine translation for English-Korean machine translation. Our metrics served as similarity measure is based on edit-distance algorithm, and it is employed to retrieve the most similar example sentences to a given query. Basically it makes use of simple information such as lemma and part-of-speech information of typographically mismatched words. Edit-distance algorithm cannot fully reflect the context of matched word units. In other words, only if matched word units are ordered, it is considered that the contribution of full matching context to similarity is identical to that of partial matching context for the sequence of words in which mismatching word units are intervened. To overcome this drawback, we propose the context-weighting scheme that uses the contiguity information of matched word units to catch the full context. To change the edit-distance metrics representing dissimilarity to similarity metrics, to apply this context-weighted metrics to the example matching problem and also to rank by similarity, we normalize it. In addition, we generalize previous methods using some linguistic information to one representative system. In order to verify the correctness of the proposed context-weighted metrics, we carry out the experiment to compare it with generalized previous methods.

Sentence Similarity Measurement Method Using a Set-based POI Data Search (집합 기반 POI 검색을 이용한 문장 유사도 측정 기법)

  • Ko, EunByul;Lee, JongWoo
    • KIISE Transactions on Computing Practices
    • /
    • v.20 no.12
    • /
    • pp.711-716
    • /
    • 2014
  • With the gradual increase of interest in plagiarism and intelligent file content search, the demand for similarity measuring between two sentences is increasing. There is a lot of researches for sentence similarity measurement methods in various directions such as n-gram, edit-distance and LSA. However, these methods have their own advantages and disadvantages. In this paper, we propose a new sentence similarity measurement method approaching from another direction. The proposed method uses the set-based POI data search that improves search performance compared to the existing hard matching method when data includes the inverse, omission, insertion and revision of characters. Using this method, we are able to measure the similarity between two sentences more accurately and more quickly. We modified the data loading and text search algorithm of the set-based POI data search. We also added a word operation algorithm and a similarity measure between two sentences expressed as a percentage. From the experimental results, we observe that our sentence similarity measurement method shows better performance than n-gram and the set-based POI data search.

Video Copy Detection Algorithm Against Online Piracy of DTV Broadcast Program (DTV 방송프로그램의 온라인 불법전송 차단을 위한 비디오 복사본 검출 알고리즘)

  • Kim, Joo-Sub;Nam, Je-Ho
    • Journal of Broadcast Engineering
    • /
    • v.13 no.5
    • /
    • pp.662-676
    • /
    • 2008
  • This paper presents a video copy detection algorithm that blocks online transfer of illegally copied DTV broadcast programs. Particularly, the proposed algorithm establishes a set of keyframes by detecting abrupt changes of luminance, and then exploits the spatio-temporal features of keyframes. Comparing with the preregistered features stored in the database of DTV broadcast programs, the proposed scheme performs a function of video filtering in order to distinguish whether an uploaded video is illegally copied or not. Note that we analyze only a set of keyframes instead of an entire video frame. Thus, it is highly efficient to identify illegal copied video when we deal with a vast size of broadcast programs. Also, we confirm that the proposed technique is robust to a variety of video edit-effects that are often applied by online video redistribution, such as apsect-ratio change, logo insertion, caption insertion, visual quality degradation, and resolution change (downscaling). In addition, we perform a benchmark test in which the proposed scheme outperforms previous techniques.

A Study on the Effective Algorithms for tine Generalization (선형성 지형자료의 일반화에 대한 효율적인 알고리즘에 관한 연구)

  • 김감래;이호남
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.12 no.1
    • /
    • pp.43-52
    • /
    • 1994
  • This paper outlines a new approach to the line generalization when preparing small scale map on the basis of existing large scale digital map. Line generalizations are conducted based on Douglas algorithm using 1/25,000 scale topographic maps of southeastern JEJU island which produced by National Geographic Institute to analyze the fitness to the original and problems of graphical representation. Compare to the same scale map which was generated by manual method, a verity of small, but sometimes significant errors & modification of topological relationship have been detected. The research gives full details of three algorithms that operationalize the smallest visible object method, together with some empirical results. A comparison of the results produced by the new algorithms with those produced by manual generalization and Douglas method of data reduction is provided. Also this paper presents the preliminary results of an relationships between the size of smallest visual object and requiring data storages for each algorithms.

  • PDF

Scenario-based 3D Objects Reuse Algorithm Scheme (시나리오 기반의 3D 객체 재사용 알고리즘)

  • Kang, Mi-Young;Lee, Hyung-Ok;Son, Seung-Chul;Heo, Kwon;Kim, Bong-Tae;Nam, Ji-Seung
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.11
    • /
    • pp.302-309
    • /
    • 2006
  • This paper propose a practical algorithm to reuse and expand the objects. This algorithm is based on the Motion Path Modification rules. We focus on reusing of the existing motions for synthesizing new motions for the objects. Both the linear and the nonlinear curve-fitting algorithm are applied to modify an animation by keyframe interpolation and to make the motion appear realistic. We also proposes a framework of the scenario-based 3D image synthesizing system that allows common users, who envision a scenario in their minds, to realize it into segments of a cool animation. The framework is useful in building a 3D animation in game programming with a limited set of 3D objects.

  • PDF

Construction of Linearly Aliened Corpus Using Unsupervised Learning (자율 학습을 이용한 선형 정렬 말뭉치 구축)

  • Lee, Kong-Joo;Kim, Jae-Hoon
    • The KIPS Transactions:PartB
    • /
    • v.11B no.3
    • /
    • pp.387-394
    • /
    • 2004
  • In this paper, we propose a modified unsupervised linear alignment algorithm for building an aligned corpus. The original algorithm inserts null characters into both of two aligned strings (source string and target string), because the two strings are different from each other in length. This can cause some difficulties like the search space explosion for applications using the aligned corpus with null characters and no possibility of applying to several machine learning algorithms. To alleviate these difficulties, we modify the algorithm not to contain null characters in the aligned source strings. We have shown the usability of our approach by applying it to different areas such as Korean-English back-trans literation, English grapheme-phoneme conversion, and Korean morphological analysis.

A Theoretical Study on Indexing Methods using the Metadata for the Automatic Construction of a Thesaurus Browser (시소러스 브라우저 자동구현을 위한 Metadata를 이용한 색인어 처리방안에 대한 연구)

  • Seo , Whee
    • Journal of Korean Library and Information Science Society
    • /
    • v.35 no.4
    • /
    • pp.451-467
    • /
    • 2004
  • This paper is intended to present the theoretical analyses on automatic indexing, which is vital in the process of constructing a thesaurus browser, and clustering algorithms to construct hierarchical relations among terms as well as the methods for the automatic construction of a thesaurus browser. The methods to select the index term automatically in the web documents are studied by surveying the methods for analyzing and processing metadata which conforms to bibliographical roles of traditional paper documents in web documents. Also, the result of the study suggests to adding or involving the metadata in web documents, using the metadata automatic editor because metadata is not listed in most of the web documents.

  • PDF

Fast Image Splicing Detection Algorithm Using Markov Features (마코프 특징을 이용하는 고속 위조 영상 검출 알고리즘)

  • Kim, Soo-min;Park, Chun-Su
    • Journal of IKEEE
    • /
    • v.22 no.2
    • /
    • pp.227-232
    • /
    • 2018
  • Nowadays, image manipulation is enormously popular and easier than ever with tons of convenient images editing tools. After several simple operations, users can get visually attractive images which easily trick viewers. In this paper, we propose a fast algorithm which can detect the image splicing using the Markov features. The proposed algorithm reduces the computational complexity by removing unnecessary Markov features which are not used in the image splicing detection process. The performance of the proposed algorithm is evaluated using a famous image splicing dataset which is publicly available. The experimental results show that the proposed technique outperforms the state-of-the-art splicing detection methods.