• Title/Summary/Keyword: Data Partitioning

Search Result 388, Processing Time 0.027 seconds

Partitioning Bimodal Spectrum Peak in Raw Data of UHF Wind Profiler (UHF 윈드프로파일러 원시 자료의 이중 스펙트럼 첨두 분리)

  • Jo, Won-Gi;Kwon, Byung-Hyuk;Yoon, Hong-Joo
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.14 no.1
    • /
    • pp.61-68
    • /
    • 2019
  • In addition to non-meteorological echoes, meteorological echoes with large scattering effects, such as precipitation, cause errors in wind data measured by wind profiler. In the rainfall situation, the Doppler spectrum of wind profiler shows both the rainfall signal and the atmospheric signal as two peaks. The vertical radial velocity is very large due to the falling rain drop. The radial velocity contaminated by rainfall decreases the accuracy of the horizontal wind vector and leads to inaccurate weather analysis. In this study, we developed an algorithm to process raw data of wind profiler and distinguished rainfall signal and wind signal by partitioning bimodal peak for Doppler spectrum in rainfall environment.

Data Partitioning on MapReduce by Leveraging Data Utility (맵리듀스에서 데이터의 유용성을 이용한 데이터 분할 기법)

  • Kim, Jong Wook
    • Journal of Korea Multimedia Society
    • /
    • v.16 no.5
    • /
    • pp.657-666
    • /
    • 2013
  • Today, many aspects of our lives are characterized by the rapid influx of large amounts of data from various application domains. The applications that produce this massive of data span a large spectrum, from social media to business intelligence or biology. This massive influx of data necessitates large scale parallelism for efficiently supporting a large class of analysis tasks. Recently, there have been extensive studies in using MapReduce framework to support large parallelism. While this technique has produced impressive results in diverse applications, the same can not be said for multimedia applications where most of users are interested in a small number of results having high or low score. Thus, in this paper, we develop the data partitioning algorithm which is able to efficiently process large data set having different data utility. The experiment results show that the proposed technique provides significant execution time gains over the existing solution.

Improving Data Availability by Data Partitioning and Partial Overlapping on Multiple Cloud Storages (다수 클라우드 스토리지로의 데이터 분할 및 부분 중복을 통한 데이터 가용성 향상)

  • Park, Jun-Cheol
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.12B
    • /
    • pp.1498-1508
    • /
    • 2011
  • A cloud service customer has no other way but to wait for his lost data to be recovered by the cloud service provider when the data was lost or not accessible for a while due to the provider's system failure, cracking attempt, malfunction, or outage. We consider a solution to address this problem that can be implemented in the cloud client's domain, rather than in the cloud service provider's domain. We propose a high level architecture and scheme for successfully retrieving data units even when several cloud storages are not accessible at the same time. The scheme is based on a clever way of partitioning and partial overlapping of data for being stored on multiple cloud storages. In addition to providing a high level of data availability, the scheme makes it possible to re-encrypt data units with new keys in a user transparent way, and can produce the complete log of every user's data units accessed, for assessing data disclosure, if needed.

POSTERIOR COMPUTATION OF SURVIVAL MODEL WITH DISCRETE APPROXIMATION

  • Lee, Jae-Yong;Kwon, Yong-Chan
    • Journal of the Korean Statistical Society
    • /
    • v.36 no.2
    • /
    • pp.321-333
    • /
    • 2007
  • In the proportional hazard model with the beta process prior, the posterior computation with the discrete approximation is considered. The time period of interest is partitioned by small intervals. On each partitioning interval, the likelihood is approximated by that of a binomial experiment and the beta process prior is by a beta distribution. Consequently, the posterior is approximated by that of many independent binomial model with beta priors. The analysis of the leukemia remission data is given as an example. It is illustrated that the length of the partitioning interval affects the posterior and one needs to be careful in choosing it.

Improved Algorithm for Haplotype Block Partitioning : Application to Human Chromosome 21

  • Na, Kyoung-Rak;Kim, Sang-Jun;Kim, Sung-Kwon
    • Proceedings of the Korean Society for Bioinformatics Conference
    • /
    • 2003.10a
    • /
    • pp.229-235
    • /
    • 2003
  • Research of basis technology to construct the human haplotype map is one of active areas in SNP post-genomics research. Identification of haplotype block structure from haplotype data is key step in the haplotype map project. Several algorithms have been proposed for the block identification, including the greedy algorithm, and the dynamic programming based algorithm. This paper analyzed block partitioning method of several algorithm which has been proposed in recent years. HapBlock and HaploBlockFinder are programs used in our experiment.

  • PDF

A Study on Effect of Domain-Decomposition Method on Parallel Efficiency in 2-D Flow Computations (2차원 유동장 해석에서 영역분할법에 따른 병렬효율성 검토)

  • Lee Sangyeul;Hur Nahmkeon
    • 한국전산유체공학회:학술대회논문집
    • /
    • 1998.11a
    • /
    • pp.147-152
    • /
    • 1998
  • 2-D flow fields are studied by using a shared memory parallel computer with a parallel flow analysis program which uses domain decomposition method and MPI library for data exchange at overlapped interface. Especially, effects of directional domain decomposition on parallel efficiency are studied for 2-D Lid-Driven cavity flow and flow through square cavity. It is known from the present study that domain decomposition along the main flow direction gives better parallel efficiency in 1-D partitioning than along the other direction. 2-D partitioning, however, is less sensitive to flow directions and gives good parallel efficiency for most of the cases considered.

  • PDF

Flexible Partitioning of CDFGs for Compact Asynchronous Controllers

  • Sretasereekul, Nattha;Okuyama, Yuichi;Saito, Hiroshi;Imai, Masashi;Kuroda, Kenichi;Nanya, Takashi
    • Proceedings of the IEEK Conference
    • /
    • 2002.07c
    • /
    • pp.1724-1727
    • /
    • 2002
  • Asynchronous circuits have the potential to solve the problems related to parameter variations such as gate delays in deep sub-micron technologies. However, current CAD tools for large-scale asyn-chronous circuits partition specification irrelevantly, because these tools cannot control the granularity of circuit decomposition. In this paper we propose a hierarchical Control/Data Flow Graph (CDFG) containing nodes that are flexibly partitioned or merged into other nodes. We show a partitioning algorithm for such CDFGs to generate handleable Signal Transition Graphs (STGs) for asynchronous synthesis tools. The algorithm a1lows designers to assign the maximum number of signals of partitioned nodes considering of timality. From an experiment, this algorithm can flexibly partition and result in more compact asynchronous controllers.

  • PDF

A Study on Data Allocation Problems of Distributed Databases (분산 데이타 베이스 설계시의 자료 배정문제에 관한 연구)

  • Sin, Gi-Tae;Park, Jin-Woo
    • Asia pacific journal of information systems
    • /
    • v.1 no.1
    • /
    • pp.49-62
    • /
    • 1991
  • This paper examines the problems of database partitioning and file allocation in a fixed topology distributed computer network. The design objective is to make files as collections of attributes and to allocate these files to network nodes so that a minimum total transmission cost is achieved subject to storage capacity constraints. A mathematical model for solving the problem is formulated and, the resulting optimization problem is shown to fall in a class of NP-complete problems. A new heuristic algorithm is developed which uses the idea of allocating attributes according to the transaction requirements at each computer node and then making files using the allocated attributes. Numerical results indicate that the heuristic algorithm yields practicable low cost solutions in comparison with the existing methods which deal with the file allocation problems and database partitioning problems independently.

  • PDF

Considering Barrier Overhead in Parallelizing AES-CCM (동기화 오버헤드를 고려한 AES-CCM의 병렬 처리)

  • Chung, Yong-Wha;Kim, Sang-Choon
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.21 no.3
    • /
    • pp.3-9
    • /
    • 2011
  • In this paper, we propose workload partitioning methods in parallelizing AES-CCM which is proposed as the wireless encryption and message integrity standard IEEE 802.11i. In parallelizing AES-CCM having data dependency, synchronizations among processors are required, and multi-core processors have a very large range of synchronization performance. We propose and compare the performance of various workload partitioning methods by considering both the computational characteristics of AES-CCM and the synchronization overhead.

Verification of Logic Gate Interconnection (논리회로 상호간의 연결도 검증)

  • Jung, Ja Choon;Kyung, Chong Min
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.24 no.2
    • /
    • pp.338-346
    • /
    • 1987
  • This paper describes a method for verifying whether a given geometrical layout correcdtly reflects the original logic level description. The logic description extracted from layout data was directly compadred with the original logic diagram generated at logic level design stage where the logic diagram is represented as a weighted multi-place graph. The comparison is based on graph isomorphism and error messages(error categories and locations)are invoked if any difference is found between the two logic descriptions. An efficient partitioning algorithm which consists of two steps, candidate selection and equal weight partitioning procedure, enables the entire verification process to occur in O(n log n) time.

  • PDF