• Title/Summary/Keyword: Overlapping Window

Search Result 24, Processing Time 0.027 seconds

Elliptic Curve Scalar Multiplication Resistant against Side Channel Attacks (부채널 공격에 안전한 타원곡선 스칼라 곱셈 알고리즘)

  • Kim Tae Hyun;Jang Sang-Woon;Kim Woong Hee;Park Young-Ho
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.14 no.6
    • /
    • pp.125-134
    • /
    • 2004
  • When cryptosystem designers implement devices that computing power or memory is limited such as smart cards, PDAs and so on, not only he/she has to be careful side channel attacks(SCA) but also the cryptographic algorithms within the device has to be efficient using small memory. For this purpose, countermeasures such as Moiler's method, Okeya-Takagi's one and overlapping window method, based on window method to prevent SCA were proposed. However, Moiler's method and Okeya-Talngi's one require additional cost to prevent other SCA such as DPA, Second-Order DPA, Address-DPA, and so on since they are immune to only SPA. Also, overlapping window method has a drawback that requires big memory. In this paper, we analyze existing countermeasures and propose an efficient and secure countermeasure that is immune to all existing SCA using advantages of each countermeasure. Moreover, the proposed countermeasure can enhance the efficiency using mixed coordinate systems.

An Efficient Join Algorithm for Data Streams with Overlapping Window (중첩 윈도우를 가진 데이터 스트링을 위한 효율적인 조인 알고리즘)

  • Kim, Hyeon-Gyu;Kang, Woo-Lam;Kim, Myoung-Ho
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.5
    • /
    • pp.365-369
    • /
    • 2009
  • Overlapping windows are generally used for queries to process continuous data streams. Nevertheless, existing approaches discussed join algorithms only for basic types of windows such as tumbling windows and tuple-driven windows. In this paper, we propose an efficient join algorithm for overlapping windows, which are considered as a more general type of windows. The proposed algorithm is based on an incremental window join. It focuses on producing join results continuously when the memory overflow frequently occurs. It consists of (1) a method to use both of the incremental and full joins selectively, (2) a victim selection algorithm to minimize latency of join processing and (3) an idle time professing algorithm. We show through our experiments that the selective use of incremental and full joins provides better performance than using one of them only.

Effective Reconstruction of Stereoscopic Image Pair by using Regularized Adaptive Window Matching Algorithm

  • Ko, Jung-Hwan;Lee, Sang-Tae;Kim, Eun-Soo
    • Journal of Information Display
    • /
    • v.5 no.4
    • /
    • pp.31-37
    • /
    • 2004
  • In this paper, an effective method for reconstruction of stereoscopic image pair through the regularized adaptive disparity estimation is proposed. Although the conventional adaptive disparity window matching can sharply improve the PSNR of a reconstructed stereo image, but there still exist some problems of overlapping between the matching windows and disallocation of the matching windows, because the size of the matching window tend to changes adaptively in accordance with the magnitude of the feature values. In the proposed method, the problems relating to the conventional adaptive disparity estimation scheme can be solved and the predicted stereo image can be more effectively reconstructed by regularizing the extimated disparity vector with the neighboring disparity vectors. From the experimental results, it is found that the proposed algorithm show improvements the PSNR of the reconstructed right image by about 2.36${\sim}$2.76 dB, on average, compared with that of conventional algorithms.

Toward High Utilization of Heterogeneous Computing Resources in SNP Detection

  • Lim, Myungeun;Kim, Minho;Jung, Ho-Youl;Kim, Dae-Hee;Choi, Jae-Hun;Choi, Wan;Lee, Kyu-Chul
    • ETRI Journal
    • /
    • v.37 no.2
    • /
    • pp.212-221
    • /
    • 2015
  • As the amount of re-sequencing genome data grows, minimizing the execution time of an analysis is required. For this purpose, recent computing systems have been adopting both high-performance coprocessors and host processors. However, there are few applications that efficiently utilize these heterogeneous computing resources. This problem equally refers to the work of single nucleotide polymorphism (SNP) detection, which is one of the bottlenecks in genome data processing. In this paper, we propose a method for speeding up an SNP detection by enhancing the utilization of heterogeneous computing resources often used in recent high-performance computing systems. Through the measurement of workload in the detection procedure, we divide the SNP detection into several task groups suitable for each computing resource. These task groups are scheduled using a window overlapping method. As a result, we improved upon the speedup achieved by previous open source applications by a magnitude of 10.

Image Path Searching using Auto and Cross Correlations

  • Kim, Young-Bin;Ryu, Kwang-Ryol
    • Journal of information and communication convergence engineering
    • /
    • v.9 no.6
    • /
    • pp.747-752
    • /
    • 2011
  • The position detection of overlapping area in the interframe for image stitching using auto and cross correlation function (ACCF) and compounding one image with the stitching algorithm is presented in this paper. ACCF is used by autocorrelation to the featured area to extract the filter mask in the reference (previous) image and the comparing (current) image is used by crosscorrelation. The stitching is detected by the position of high correlation, and aligns and stitches the image in shifting the current image based on the moving vector. The ACCF technique results in a few computations and simplicity because the filter mask is given by the featuring block, and the position is enabled to detect a bit movement. Input image captured from CMOS is used to be compared with the performance between the ACCF and the window correlation. The results of ACCF show that there is no seam and distortion at the joint parts in the stitched image, and the detection performance of the moving vector is improved to 12% in comparison with the window correlation method.

Development of an X-window Program, XFAP, for Assembling Contigs from DNA Fragment Data (DNA 염기 서열로부터 contig 구성을 위한 프로그램 XFAP의 개발)

  • Lee, Byung-Uk;Park, Kie-Jung;Kim, Seung-Moak
    • Korean Journal of Microbiology
    • /
    • v.34 no.1_2
    • /
    • pp.58-63
    • /
    • 1998
  • Fragment assembly problem is to reconstruct DNA sequence contigs from a collection of fragment sequences. We have developed an efficient X-window program, XFAP, for assembling DNA fragments. In the XFAP, the dimer frequency comparison method is used to quickly eliminate pairs of fragments that can not overlap. This method takes advantage of the difference of dimer frequencies within the minimum acceptable overlap length in each fragment pair. Hirschberg algorithm is applied to compute the maximal-scoring overlapping alignment in linear space. The perfomance of XFAP was tested on a set of DNA fragment sequences extracted from long DNA sequences of GenBank by a fragmentation program and showed a great improvement in execution time, especially as the number of fragments increases.

  • PDF

Automatic Construction of Alternative Word Candidates to Improve Patent Information Search Quality (특허 정보 검색 품질 향상을 위한 대체어 후보 자동 생성 방법)

  • Baik, Jong-Bum;Kim, Seong-Min;Lee, Soo-Won
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.10
    • /
    • pp.861-873
    • /
    • 2009
  • There are many reasons that fail to get appropriate information in information retrieval. Allomorph is one of the reasons for search failure due to keyword mismatch. This research proposes a method to construct alternative word candidates automatically in order to minimize search failure due to keyword mismatch. Assuming that two words have similar meaning if they have similar co-occurrence words, the proposed method uses the concept of concentration, association word set, cosine similarity between association word sets and a filtering technique using confidence. Performance of the proposed method is evaluated using a manually extracted alternative list. Evaluation results show that the proposed method outperforms the context window overlapping in precision and recall.

Automatic Text Categorization Using Passage-based Weight Function and Passage Type (문단 단위 가중치 함수와 문단 타입을 이용한 문서 범주화)

  • Joo, Won-Kyun;Kim, Jin-Suk;Choi, Ki-Seok
    • The KIPS Transactions:PartB
    • /
    • v.12B no.6 s.102
    • /
    • pp.703-714
    • /
    • 2005
  • Researches in text categorization have been confined to whole-document-level classification, probably due to lacks of full-text test collections. However, full-length documents availably today in large quantities pose renewed interests in text classification. A document is usually written in an organized structure to present its main topic(s). This structure can be expressed as a sequence of sub-topic text blocks, or passages. In order to reflect the sub-topic structure of a document, we propose a new passage-level or passage-based text categorization model, which segments a test document into several Passages, assigns categories to each passage, and merges passage categories to document categories. Compared with traditional document-level categorization, two additional steps, passage splitting and category merging, are required in this model. By using four subsets of Routers text categorization test collection and a full-text test collection of which documents are varying from tens of kilobytes to hundreds, we evaluated the proposed model, especially the effectiveness of various passage types and the importance of passage location in category merging. Our results show simple windows are best for all test collections tested in these experiments. We also found that passages have different degrees of contribution to main topic(s), depending on their location in the test document.

A design of FFT processor for EEG signal analysis (뇌전기파 분석용 FFT 프로세서 설계)

  • Kim, Eun-Suk;Shin, Kyung-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.14 no.11
    • /
    • pp.2548-2554
    • /
    • 2010
  • This paper describes a design of fast Fourier transform(FFT) processor for EEG(electroencephalogram) signal analysis for health care services. Hamming window function with 1/2 overlapping is adopted to perform short-time FFT(ST-FFT) of a long period EEG signal occurred in real-time. In order to analyze efficiently EEG signals which have frequency characteristics in the range of 0 Hz to 100 Hz, a 256-point FFT processor is designed, which is based on a single-memory bank architecture and the radix-4 algorithm. The designed FFT processor has been verified by FPGA implementation, and has high accuracy with arithmetic error less than 2%.

A design of FFT processor for EEG signal analysis (뇌전기파 분석용 FFT 프로세서 설계)

  • Kim, Eun-Suk;Kim, Hae-Ju;Na, Young-Heon;Shin, Kyung-Wook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2010.10a
    • /
    • pp.88-91
    • /
    • 2010
  • This paper describes a design of fast Fourier transform(FFT) processor for EEG(electroencephalogram) signal analysis for health care services. Hamming window function with 1/2 overlapping is adopted to perform short-time FFT(ST-FFT) of a long period EEG signal occurred in real-time. In order to analyze efficiently EEG signals which have frequency characteristics in the range of 0 Hz to 100 Hz, a 256-point FFT processor based on single-memory bank architecture and radix-4 algorithm is designed. The designed FFT processor has high accuracy with arithmetic error less than 3%.

  • PDF