• Title/Summary/Keyword: Data partitioning

Search Result 388, Processing Time 0.048 seconds

Parallel 2D-DWT Hardware Architecture for Image Compression Using the Lifting Scheme (이미지 압축을 위한 Lifting Scheme을 이용한 병렬 2D-DWT 하드웨어 구조)

  • Kim, Jong-Woog;Chong, Jong-Wha
    • Journal of IKEEE
    • /
    • v.6 no.1 s.10
    • /
    • pp.80-86
    • /
    • 2002
  • This paper presents a fast hardware architecture to implement a 2-D DWT(Discrete Wavelet Transform) computed by lifting scheme framework. The conventional 2-D DWT hardware architecture has problem in internal memory, hardware resource, and latency. The proposed architecture was based on the 4-way partitioned data set. This architecture is configured with a pipelining parallel architecture for 4-way partitioning method. Due to the use of this architecture, total latency was improved by 50%, and memory size was reduced by using lifting scheme.

  • PDF

The Application of Quantum Yield of Nitrate Uptake to Estimate New Production in Well-Mixed Waters of the Yellow Sea: A Preliminary Result

  • Park, Myung-Gil;Shim, Jae-Hyung;Yang, Sung-Ryull
    • Journal of the korean society of oceanography
    • /
    • v.37 no.1
    • /
    • pp.45-50
    • /
    • 2002
  • New production (NP) values in well-mixed waters of the Yellow Sea were estimated using two different methods and were compared with each other; one is from the quantum yield model of nitrate uptake and chlorophyll ${\alpha}$-specific light absorption coefficient, and the other is from a traditional $^{15}N$-labelled stable isotope uptake technique. The quantum yields of nitrate uptake were highly variable, ranging from 0.0001 to 0.04 mol $NO_3Ein^{-1}$, and the small values in this study might have resulted from either the partitioning into nitrate uptake of little portions of light energy absorbed by phytoplankton or that phytoplankton may predominantly utilize other N sources (E. G. ammonium and/or urea) than nitrate. The estimates (0.54-8.47 nM $h^{-1}$) of NP from the quantum yield model correlated well ($r^2$=0.67, p<0.1) with those (0.01-4.93 nM $h^{-1}$) obtained using the $^{15}NO_3$ uptake technique. To improve the ability of estimating NP values using this model in the Yellow Sea, more data need to be accumulated in the future over a variety of time and space scales.

Combining Distributed Word Representation and Document Distance for Short Text Document Clustering

  • Kongwudhikunakorn, Supavit;Waiyamai, Kitsana
    • Journal of Information Processing Systems
    • /
    • v.16 no.2
    • /
    • pp.277-300
    • /
    • 2020
  • This paper presents a method for clustering short text documents, such as news headlines, social media statuses, or instant messages. Due to the characteristics of these documents, which are usually short and sparse, an appropriate technique is required to discover hidden knowledge. The objective of this paper is to identify the combination of document representation, document distance, and document clustering that yields the best clustering quality. Document representations are expanded by external knowledge sources represented by a Distributed Representation. To cluster documents, a K-means partitioning-based clustering technique is applied, where the similarities of documents are measured by word mover's distance. To validate the effectiveness of the proposed method, experiments were conducted to compare the clustering quality against several leading methods. The proposed method produced clusters of documents that resulted in higher precision, recall, F1-score, and adjusted Rand index for both real-world and standard data sets. Furthermore, manual inspection of the clustering results was conducted to observe the efficacy of the proposed method. The topics of each document cluster are undoubtedly reflected by members in the cluster.

Wage Determinants Analysis by Quantile Regression Tree

  • Chang, Young-Jae
    • Communications for Statistical Applications and Methods
    • /
    • v.19 no.2
    • /
    • pp.293-301
    • /
    • 2012
  • Quantile regression proposed by Koenker and Bassett (1978) is a statistical technique that estimates conditional quantiles. The advantage of using quantile regression is the robustness in response to large outliers compared to ordinary least squares(OLS) regression. A regression tree approach has been applied to OLS problems to fit flexible models. Loh (2002) proposed the GUIDE algorithm that has a negligible selection bias and relatively low computational cost. Quantile regression can be regarded as an analogue of OLS, therefore it can also be applied to GUIDE regression tree method. Chaudhuri and Loh (2002) proposed a nonparametric quantile regression method that blends key features of piecewise polynomial quantile regression and tree-structured regression based on adaptive recursive partitioning. Lee and Lee (2006) investigated wage determinants in the Korean labor market using the Korean Labor and Income Panel Study(KLIPS). Following Lee and Lee, we fit three kinds of quantile regression tree models to KLIPS data with respect to the quantiles, 0.05, 0.2, 0.5, 0.8, and 0.95. Among the three models, multiple linear piecewise quantile regression model forms the shortest tree structure, while the piecewise constant quantile regression model has a deeper tree structure with more terminal nodes in general. Age, gender, marriage status, and education seem to be the determinants of the wage level throughout the quantiles; in addition, education experience appears as the important determinant of the wage level in the highly paid group.

The Silver Cycle and Fluxes in the Ocean

  • Ju, Se-Jong
    • Journal of the korean society of oceanography
    • /
    • v.32 no.3
    • /
    • pp.156-161
    • /
    • 1997
  • The biogeochemical cycle of silver has rarely been reviewed, even though the silver ion (Ag$^{\times}$) is extremly toxic to some organisms. Its concentration is still rising sharply because of increased anthropogenic activity, specifically the discharge from the film industry (mainly, silver thiosulfate: Ag (S$_2$O$_3$)${^3-}_2$). Recently, a number of researchers have quantified the major fluxes and reservoirs of silver in the open ocean, bays, and estuaries. A review of the available information for Ag cycling in the open ocean shows that the riverine input (from human activity and weathering processes: 7${\times}$10$^6$ kg/yr and 5${\times}$10$^6$ kg/yr, respectively) is the dominant source of Ag to estuarine and coastal regions. Most of the silver (90% of riverine input silver) is removed in coastal sediments by the physical-chemical character of silver due to its high partitioning with particulate matter. On the other hand, in the open ocean the atmospheric input (wet and dry deposition: 1.48${\times}$10$^6$ kg/yr and 1.94${\times}$ 10$^5$ kg/yr, respectively) becomes more important as a source of silver than riverine input. The residence time of silver calculated from available data is 1250 yrs in the deep ocean below 500 m, but only 3 yrs in the surface ocean.

  • PDF

Subnetwork-based Segment Restoration for fast fault Recovery in the MPLS network (MPLS 통신망에서의 신속한 장애복구를 위한 서브네트워크 기반의 세그먼트 단위 자동복구 기법)

  • 신해준;장재준;김영탁
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.11C
    • /
    • pp.1046-1054
    • /
    • 2002
  • In this paper, we propose a subnetwork-based segment restoration scheme to reduce the restoration time and restoration resources. And we compare and analyze the restoration performance according to the size of divided subnetworks. Segment restoration is based on network partitioning where a large network is divided into several small subnetworks and the end-to-end data path is divided into multiple segments according to the subnetworks. In segment restoration, the link/node failure is restored by segment instead of end-to-end path. Because most faults are restored within the subnetwork, the restoration performance can be improved. From the simulation analysis, we verified that the proposed segment restoration has advantage of restoration time and backup resource utilization.

Region Classification and Image Based on Region-Based Prediction (RBP) Model

  • Cassio-M.Yorozuya;Yu-Liu;Masayuki-Nakajima
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1998.06b
    • /
    • pp.165-170
    • /
    • 1998
  • This paper presents a new prediction method RBP region-based prediction model where the context used for prediction contains regions instead of individual pixels. There is a meaningful property that RBP can partition a cartoon image into two distinctive types of regions, one containing full-color backgrounds and the other containing boundaries, edges and home-chromatic areas. With the development of computer techniques, synthetic images created with CG (computer graphics) becomes attactive. Like the demand on data compression, it is imperative to efficiently compress synthetic images such as cartoon animation generated with CG for storage of finite capacity and transmission of narrow bandwidth. This paper a lossy compression method to full-color regions and a lossless compression method to homo-chromatic and boundaries regions. Two criteria for partitioning are described, constant criterion and variable criterion. The latter criterion, in form of a linear function, gives the different threshold for classification in terms of contents of the image of interest. We carry out experiments by applying our method to a sequence of cartoon animation. We carry out experiments by applying our method to a sequence of cartoon animation. Compared with the available image compression standard MPEG-1, our method gives the superior results in both compression ratio and complexity.

  • PDF

Design and Implementation of Distribution in Distributed Object-Oriented Databases (분산 객체지향 데이타베이스에서 분산 설계 및 구현)

  • Lee, Soon-Mi;Park, Hea-Sook;Ha, Yan
    • The KIPS Transactions:PartB
    • /
    • v.11B no.5
    • /
    • pp.611-618
    • /
    • 2004
  • This paper addresses the design and implementation of class distribution in distributed object-oriented databases. The proposed strategy of distribution consists of two-step design of fragments. One is class fragmentation and the other is allocation of fragments. In step of class fragmentation, we have defined partitioning algorithms to reflect the characteristics of object-oriented databases such as method, inheritance and composite-object. In step of allocation, we have defined the objective function for allocation considering system operating cost including storage, query processing and communication and implemented it using Genetic Algorithm.

Parallel Finite Element Simulation of the Incompressible Navier-stokes Equations (병렬 유한요소 해석기법을 이용한 유동장 해석)

  • Choi H. G.;Kim B. J.;Kang S. W.;Yoo J. Y.
    • 한국전산유체공학회:학술대회논문집
    • /
    • 2002.05a
    • /
    • pp.8-15
    • /
    • 2002
  • For the large scale computation of turbulent flows around an arbitrarily shaped body, a parallel LES (large eddy simulation) code has been recently developed in which domain decomposition method is adopted. METIS and MPI (message Passing interface) libraries are used for domain partitioning and data communication between processors, respectively. For unsteady computation of the incompressible Wavier-Stokes equation, 4-step splitting finite element algorithm [1] is adopted and Smagorinsky or dynamic LES model can be chosen fur the modeling of small eddies in turbulent flows. For the validation and performance-estimation of the parallel code, a three-dimensional laminar flow generated by natural convection inside a cube has been solved. Then, we have solved the turbulent flow around MIRA (Motor Industry Research Association) model at $Re = 2.6\times10^6$, which is based on the model height and inlet free stream velocity, using 32 processors on IBM SMP cluster and compared with the existing experiment.

  • PDF

An Attribute Replicating Vertical Partition Method by Genetic Algorithm in the Physical Design of Relational Database (관계형 데이터베이스의 물리적 설계에서 유전해법을 이용한 속성 중복 수직분할 방법)

  • 유종찬;김재련
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.21 no.46
    • /
    • pp.33-49
    • /
    • 1998
  • In order to improve the performance of relational databases, one has to reduce the number of disk accesses necessary to transfer data from disk to main memory. The paper proposes to reduce the number of disk I/O accesses by vertically partitioning relation into fragments and allowing attribute replication to fragments if necessary. When zero-one integer programming model is solved by the branch-and-bound method, it requires much computing time to solve a large sized problem. Therefore, heuristic solutions using genetic algorithm(GA) are presented. GA in this paper adapts a few ideas which are different from traditional genetic algorithms, for examples, a rank-based sharing fitness function, elitism and so on. In order to improve performance of GA, a set of optimal parameter levels is determined by the experiment and makes use of it. As relations are vertically partitioned allowing attribute replications and saved in disk, an attribute replicating vertical partition method by GA can attain less access cost than non-attribute-replication one and require less computing time than the branch-and-bound method in large-sized problems. Also, it can acquire a good solution similar to the optimum solution in small-sized problem.

  • PDF