• Title/Summary/Keyword: large database

Search Result 1,454, Processing Time 0.033 seconds

A Prototyping Framework of the Documentation Retrieval System for Enhancing Software Development Quality

  • Chang, Wen-Kui;Wang, Tzu-Po
    • International Journal of Quality Innovation
    • /
    • v.2 no.2
    • /
    • pp.93-100
    • /
    • 2001
  • This paper illustrates a prototyping framework of the documentation-standards retrieval system via the data mining approach for enhancing software development quality. We first present an approach for designing a retrieval algorithm based on data mining, with the three basic technologies of machine learning, statistics and database management, applied to this system to speed up the searching time and increase the fitness. This approach derives from the observation that data mining can discover unsuspected relationships among elements in large databases. This observation suggests that data mining can be used to elicit new knowledge about the design of a subject system and that it can be applied to large legacy systems for efficiency. Finally, software development quality will be improved at the same time when the project managers retrieving for the documentation standards.

  • PDF

An efficient method for directory management of the partitioned signature file (분할 시그너춰 화일을 위한 효율적인 디렉토리 관리 기법)

  • 김상욱;황환규;최황규;윤용익
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.35C no.3
    • /
    • pp.32-45
    • /
    • 1998
  • A partitioned signature file is an enhancement of the signature file that divides all the signatures into blocks in such a way that each block contains the signatures with the same key. Its directory storesall the keys as meta information for avoiding unnecessary block accesses by examming them first before the acture searching of the blocks.. Efficient directory management is very important in large databasse environments since ist size gets larger proportionally to that of the database. In this paper, we first point out the problems in the directory management methods of the previous partitioned signature files, and then present a new one solving them. OUr method offers good features in the followint three aspects: (1) suitability for large database environments, (2) adaptability to dynamic situations, and (3) storage overhead for the directory. Moreover, we can seamlessly integrate it as a subcomponent into previously-developed general-purpose storage engines. These features show that our method is applicableto signature-based access structures for the content-based retrieval in various multimedia applications such as hypermedia systems, digital library systems, multimedia document systems, multimedia mailing systems, and so on.

  • PDF

A Vertical File Partitioning Method Using SOFM in Database Design (데이터베이스 설계에서 SOFM 을 이용한 화일 수직분할 방법)

  • Shin, K.H.;Kim, J.Y.
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.24 no.4
    • /
    • pp.661-671
    • /
    • 1998
  • It is important to minimize the number of disk accesses which is necessary to transfer data in disk into main memory when processing transactions in physical database design. A vertical file partitioning method is used to reduce the number of disk accesses by partitioning relations vertically and accessing only necessay fragments. In this paper, SOFM(Self-Organizing Feature Maps) network is used to solve vertical partitioning problems. This paper shows that SOFM network is efficient in solving vertical partitioning problem by comparing approximate solution of SOFM network with optimal solution of N-ary branch and bound method. And this paper presents a heuristic algorithm for allocating duplicate attributes to vertically partitioned fragments. As branch and bound method requires particularly much computing time to solve large-sized problems, it is shown that SOFM network is able to overcome this limitation of branch and bound method and solve large-sized problems efficiently in a short time.

  • PDF

EFFICIENT OPEN SOURCE DISTRIBUTED ERP SYSTEM FOR LARGE SCALE ENTERPRISE

  • ELMASSRY, MOHAMED;AL-AHAMADI, SAAD
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.12
    • /
    • pp.280-292
    • /
    • 2021
  • Enterprise Resource Planning (ERP) is a software that manages and automate the internal processes of an organization. Process speed and quality can be increased, and cost reduced by process automation. Odoo is an open source ERP platform including more than 15000 apps. ERP systems such as Odoo are all-in-one management systems. Odoo can be suitable for small and medium organizations, but duo to efficiency limitations, Odoo is not suitable for the large ones. Furthermore, Odoo can be implemented on both local or public servers in which each has some advantages and disadvantages such as; the speed of internet, synced data or anywhere access. In many cases, there is a persistent need to have more than one synchronized Odoo instance in several physical places. We modified Odoo to support this kind of requirements and improve its efficiency by replacing its standard database with a distributed one, namely CockroachDB.

New Construction of Order-Preserving Encryption Based on Order-Revealing Encryption

  • Kim, Kee Sung
    • Journal of Information Processing Systems
    • /
    • v.15 no.5
    • /
    • pp.1211-1217
    • /
    • 2019
  • Developing methods to search over an encrypted database (EDB) have received a lot of attention in the last few years. Among them, order-revealing encryption (OREnc) and order-preserving encryption (OPEnc) are the core parts in the case of range queries. Recently, some ideally-secure OPEnc schemes whose ciphertexts reveal no additional information beyond the order of the underlying plaintexts have been proposed. However, these schemes either require a large round complexity or a large persistent client-side storage of size O(n) where n denotes the number of encrypted items stored in EDB. In this work, we propose a new construction of an efficient OPEnc scheme based on an OREnc scheme. Security of our construction inherits the security of the underlying OREnc scheme. Moreover, we also show that the construction of a non-interactive ideally-secure OPEnc scheme with a constant client-side storage is theoretically possible from our construction.

Optimization for Large-Scale n-ary Family Tree Visualization

  • Kyoungju, Min;Jeongyun, Cho;Manho, Jung;Hyangbae, Lee
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.1
    • /
    • pp.54-61
    • /
    • 2023
  • The family tree is one of the key elements of humanities classics research and is very important for accurately understanding people or families. In this paper, we introduce a method for automatically generating a family tree using information on interpersonal relationships (IIPR) from the Korean Classics Database (KCDB) and visualize interpersonal searches within a family tree using data-driven document JavaScript (d3.js). To date, researchers of humanities classics have wasted considerable time manually drawing family trees to understand people's influence relationships. An automatic family tree builder analyzes a database that visually expresses the desired family tree. Because a family tree contains a large amount of data, we analyze the performance and bottlenecks according to the amount of data for visualization and propose an optimal way to construct a family tree. To this end, we create an n-ary tree with fake data, visualize it, and analyze its performance using simulation results.

Implementation of Human Spine Database System for Realizing e-Spine

  • Roh, Hyun Sook;Choi, Boongki;Bae, Kook-jin;Park, Jinseo
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2014.06a
    • /
    • pp.193-194
    • /
    • 2014
  • This paper provides the market opportunity analysis process and technology providing matching process based on the efficient quantitative analysis of a large amount of papers and network analysis. In the future, it is required to analyze and complement the quantitative analysis result by commercialization experts and to build emerging issue database.

  • PDF

Directions for Developing Database Schema of Records in Archives Management Systems (영구기록물관리를 위한 기록물 데이터베이스 스키마 개발 방향)

  • Yim, Jin-Hee;Lee, Dae-Wook;Kim, Eun-Sil;Kim, Ik-Han
    • The Korean Journal of Archival Studies
    • /
    • no.34
    • /
    • pp.57-105
    • /
    • 2012
  • The CAMS(Central Archives Management System) of NAK(National Archives of Korea) is an important system which receives and manages large amount of electronic records annually from 2015. From the point of view in database design, this paper analyzes the database schema of CAMS and discusses the direction of overall improvement of the CAMS. Firstly this research analyzes the tables for records and folders in the CAMS database which are core tables for the electronic records management. As a result, researchers notice that it is difficult to trust the quality of the records in the CAMS, because two core tables are entirely not normalized and have many columns whose roles are unknown. Secondly, this study suggests directions of normalization for the tables for records and folders in the CAMS database like followings: First, redistributing the columns into proper tables to reduce the duplication. Second, separating the columns about the classification scheme into separate tables. Third, separating the columns about the records types and sorts into separate tables. Lastly, separating metadata information related to the acquisition, takeover and preservation into separate tables. Thirdly, this paper suggests considerations to design and manage the database schema in each phase of archival management. In the ingest phase, the system should be able to process large amount of records as batch jobs in time annually. In the preservation phase, the system should be able to keep the management histories in the CAMS as audit trails including the reclassification, revaluation, and preservation activities related to the records. In the access phase, the descriptive metadata sets for the access should be selected and confirmed in various ways. Lastly, this research also shows the prototype of conceptual database schema for the CAMS which fulfills the metadata standards for records.

A Subsequence Matching Technique that Supports Time Warping Efficiently (타임 워핑을 지원하는 효율적인 서브시퀀스 매칭 기법)

  • Park, Sang-Hyun;Kim, Sang-Wook;Cho, June-Suh;Lee, Hoen-Gil
    • Journal of Industrial Technology
    • /
    • v.21 no.A
    • /
    • pp.167-179
    • /
    • 2001
  • This paper discusses an index-based subsequence matching that supports time warping in large sequence databases. Time warping enables finding sequences with similar patterns even when they are of different lengths. In earlier work, we suggested an efficient method for whole matching under time warping. This method constructs a multidimensional index on a set of feature vectors, which are invariant to time warping, from data sequences. For filtering at feature space, it also applies a lower-bound function, which consistently underestimates the time warping distance as well as satisfies the triangular inequality. In this paper, we incorporate the prefix-querying approach based on sliding windows into the earlier approach. For indexing, we extract a feature vector from every subsequence inside a sliding window and construct a multi-dimensional index using a feature vector as indexing attributes. For query precessing, we perform a series of index searches using the feature vectors of qualifying query prefixes. Our approach provides effective and scalable subsequence matching even with a large volume of a database. We also prove that our approach does not incur false dismissal. To verily the superiority of our method, we perform extensive experiments. The results reseal that our method achieves significant speedup with real-world S&P 500 stock data and with very large synthetic data.

  • PDF

DNS and Analysis on the Interscale Interactions of the Turbulent Flow past a Circular Cylinder for Large Eddy Simulation (원형 실린더를 지나는 난류 유동장의 직접수치해석과 큰 에디모사를 위한 스케일 간 상호작용 연구)

  • Kim, Taek-Keun;Park, No-Ma;Yoo, Jung-Yul
    • Proceedings of the KSME Conference
    • /
    • 2004.04a
    • /
    • pp.1801-1806
    • /
    • 2004
  • Stochastic nature of subgrid-scale stress causes the predictability problem in large eddy simulation (LES) by which the LES solution field decorrelates with field from filtered directnumerical simulation (DNS). In order to evaluate the predictability limit in a priori sense, the information on the interplay between resolved scale and subgrid-scale (SGS) is required. In this study, the analysis on the inter-scale interaction is performed by applying tophat and cutoff filters to DNS database of flow over a circular cylinder at Reynolds number of 3900. The effect of filter shape is investigated on the interpretation of correlation between scales. A critique is given on the use of tophat filter for SGS analysis using DNS database. It is shown that correlations between Karman vortex and SGS kinetic energy drastically decrease when the cutoff filter is used, which implies that the small scale universality holds even in the presence of the large scale coherent structure.

  • PDF