• Title/Summary/Keyword: DBMS(Database Management System)

Search Result 155, Processing Time 0.023 seconds

An Analysis of the Overhead of Multiple Buffer Pool Scheme on InnoDB-based Database Management Systems (InnoDB 기반 DBMS에서 다중 버퍼 풀 오버헤드 분석)

  • Song, Yongju;Lee, Minho;Eom, Young Ik
    • Journal of KIISE
    • /
    • v.43 no.11
    • /
    • pp.1216-1222
    • /
    • 2016
  • The advent of large-scale web services has resulted in gradual increase in the amount of data used in those services. These big data are managed efficiently by DBMS such as MySQL and MariaDB, which use InnoDB engine as their storage engine, since InnoDB guarantees ACID and is suitable for handling large-scale data. To improve I/O performance, InnoDB caches data and index of its database through a buffer pool. It also supports multiple buffer pools to mitigate lock contentions. However, the multiple buffer pool scheme leads to the additional data consistency overhead. In this paper, we analyze the overhead of the multiple buffer pool scheme. In our experimental results, although multiple buffer pool scheme mitigates the lock contention by up to 46.3%, throughput of DMBS is significantly degraded by up to 50.6% due to increased disk I/O and fsync calls.

Geovisualization of Coastal Ocean Model Data Using Web Services and Smartphone Apps (웹서비스와 스마트폰앱을 이용한 연안해양모델 예측자료의 시각화시스템 구현)

  • Kim, Hyung-Woo;Koo, Bon-Ho;Woo, Seung-Buhm;Lee, Ho-Sang;Lee, Yang-Won
    • Spatial Information Research
    • /
    • v.22 no.2
    • /
    • pp.63-71
    • /
    • 2014
  • Ocean leisure sports have recently emerged as one of so-called blue ocean industries. They are sensitive to diverse environmental conditions such as current, temperature, and salinity, which can increase needs of forecasting data as well as in-situ observations for the ocean. In this context, a Web-based geovisualization system for coastal information produced by model forecasts was implemented for use in supporting various ocean activities. First, FVCOM(Finite Volume Coastal Ocean Model) was selected as a forecasting model, and its data was preprocessed by a spatial interpolation and sampling library. The interpolated raster data for water surface elevation, temperature, and salinity were stored in image files, and the vector data for currents including speed and direction were imported into a distributed DBMS(Database Management System). Web services in REST(Representational State Transfer) API(Application Programming Interface) were composed using Spring Framework and integrated with desktop and mobile applications developed on the basis of hybrid structure, which can realize a cross-platform environment for geovisualization.

XML View Indexing Using an RDBMS based XML Storage System (관계 DBMS 기반 XML 저장시스템 상에서의 XML 뷰 인덱싱)

  • Park Dae-Sung;Kim Young-Sung;Kang Hyunchul
    • Journal of Internet Computing and Services
    • /
    • v.6 no.4
    • /
    • pp.59-73
    • /
    • 2005
  • Caching query results and reusing them in processing of subsequent queries is an important query optimization technique. Materialized view and view indexing are the representative examples of such a technique. The two schemes had received much attention for relational databases, and have been investigated for XML data since XML emerged as the standard for data exchange on the Web. In XML view indexing, XML view xv which is the result of an XML query is represented as an XML view index(XVI), a structure containing the identifiers of xv's underlying XML elements as well as the information on xv. Since XVI for xv stores just the identifiers of the XML elements not the elements themselves, when xv is requested, its XVI should be materialized against xv's underlying XML documents. In this paper, we address the problem of integrating an XML view index management system with an RDBMS based XML storage system. The proposed system was implemented in Java on Windows 2000 Server with each of two different commercial RDBMSs, and used in evaluating performance improvement through XML view indexing as well as its overheads. The experimental results revealed that XML view indexing was very effective with an RDBMS based XML storage system while its overhead was negligible.

  • PDF

GIS Oriented Platform For Solving Real World Logistic Vehicle Routing Problem

  • Md. Shahid Uz Zaman;Chen, Yen-Wei;Hayao Miyagi
    • Proceedings of the IEEK Conference
    • /
    • 2002.07b
    • /
    • pp.1248-1251
    • /
    • 2002
  • Logistics optimization problems related with vehicle routing such as warehouse locating, track scheduling, customer order delivery, wastage pickup etc. are very interesting and important issues to date. Many Vehicle Routing and Scheduling Systems (VRSS) have been developed/proposed to optimize the logistics problems. But majority of them are dedicated to a particular problem and are unable to handle the real world spatial data directly. The system developed for one problem may not be suitable for others due to inter-problem constraint variations. The constraints may include geographical, environmental and road traffic nature of the working region along with other constraints related with the problem. So the developer always needs to modify the original routing algorithm in order to fulfill the purpose. In our study, we propose a general-purpose platform by combining GIS road map and Database Management System (DBMS), so that VRSS can interact with real world spatial data directly to solve different kinds of vehicle routing problems. Using the features of our developed system, the developer can frequently modify the existing algorithm or create a new one to serve the purpose.

  • PDF

Metadata Management Method for Consistency and Recency in Digital Library (디지탈 도서관 환경에서 일관성과 최근성을 고려한 메타데이타 관리 방법)

  • Lee, Hai-Min;Park, Seog
    • Journal of KIISE:Databases
    • /
    • v.27 no.1
    • /
    • pp.22-32
    • /
    • 2000
  • The Digital Library is the integrated system of Information Retrieval System(IRS) and Database Management system(DBMS). In the Digital Library environment where dynamic query and update processes are required, however, the existing transaction management methods cause the following problems. First, since the traditional consistency criteria is too restrictive, it causes increment of query processing time and cannot guarantee the reflection of recency. Second, query result could be unreliable because the consistency criteria between source data and metadata is not defined. This paper models the access to metadata based on Dublin Core as query transactions and update transactions, and gives the efficient method to manage them. Particularly, this paper describes the consistency criteria of metadata which takes it Into consideration the consistency between the result of query transaction and status of source data in the Digital Library, that is different from the consistency criteria in traditional transaction management. It also presents analysis of the view point of query transaction to reflect recency and proposes metadata management to guarantee recency within metadata consistency.

  • PDF

A Framework of Intelligent Middleware for DNA Sequence Analysis in Cloud Computing Environment (DNA 서열 분석을 위한 클라우드 컴퓨팅 기반 지능형 미들웨어 설계)

  • Oh, Junseok;Lee, Yoonjae;Lee, Bong Gyou
    • Journal of Internet Computing and Services
    • /
    • v.15 no.1
    • /
    • pp.29-43
    • /
    • 2014
  • The development of NGS technologies, such as scientific workflows, has reduced the time required for decoding DNA sequences. Although the automated technologies change the genome sequence analysis environment, limited computing resources still pose problems for the analysis. Most scientific workflow systems are pre-built platforms and are highly complex because a lot of the functions are implemented into one system platform. It is also difficult to apply components of pre-built systems to a new system in the cloud environment. Cloud computing technologies can be applied to the systems to reduce analysis time and enable simultaneous analysis of massive DNA sequence data. Web service techniques are also introduced for improving the interoperability between DNA sequence analysis systems. The workflow-based middleware, which supports Web services, DBMS, and cloud computing, is proposed in this paper for expecting to reduceanalysis time and aiding lightweight virtual instances. It uses DBMS for managing the pipeline status and supporting the creation of lightweight virtual instances in the cloud environment. Also, the RESTful Web services with simple URI and XML contents are applied for improving the interoperability. The performance test of the system needs to be conducted by comparing results other developed DNA analysis services at the stabilization stage.

The Design and Performance Analysis of an Effective OWL Storage System Based on the DBMS (데이터베이스 시스템에 기반한 효율적인 OWL 저장시스템 설계 및 성능분석)

  • Cha, Seong-Hwan;Kim, Seong-Sik;Kim, TaeYoung
    • The Journal of Korean Association of Computer Education
    • /
    • v.11 no.5
    • /
    • pp.77-88
    • /
    • 2008
  • Having observed the restriction of the current Web technology, the semantic Web has been developed, and it now has grown up with the core help of the W3C to a level where it recommends the OWL Web ontology language. Besides, in order to deduce the information out of OWL data, several inference systems have been developed such as Jena, Jess, and JTP. Unfortunately, however, quite few systems can effectively handle recently developed OWL data, and further, due to the limitation of file-based operation, the current inference systems cannot meet the requirements for handing huge OWL data. An efficient method for storing and searching ontology data is essential for ensuring stable information inference processes. In this study, firstly, we proposed a model based on the database management system to transform and store OWL data and to enable deduction process from the database. Secondly, we designed and implemented an effective OWL storing system based on our model. Finally, we compare our system with the previous inference systems through experimental performance analysis.

  • PDF

Visualization of Path Expressions with Set Attributes and Methods in Graphical Object Query Languages (그래픽 객체 질의어에서 집합 속성과 메소드를 포함한 경로식의 시각화)

  • 조완섭
    • Journal of KIISE:Databases
    • /
    • v.30 no.2
    • /
    • pp.109-124
    • /
    • 2003
  • Although most commercial relational DBMSs Provide a graphical query language for the user friendly interfaces of the databases, few research has been done for graphical query languages in object databases. Expressing complex query conditions in a concise and intuitive way has been an important issue in the design of graphical query languages. Since the object data model and object query languages are more complex than those of the relational ones, the graphical object query language should have a concise and intuitive representation method. We propose a graphical object query language called GOQL (Graphical Object Query Language) for object databases. By employing simple graphical notations, advanced features of the object queries such as path expressions including set attributes, quantifiers, and/or methods can be represented in a simple graphical notation. GOQL has an excellent expressive power compared with previous graphical object query languages. We show that path expressions in XSQL(1,2) can be represented by the simple graphical notations in GOQL. We also propose an algorithm that translates a graphical query in GOQL into the textual object query with the same semantics. We finally describe implementation results of GOQL in the Internet environments.

DEEP-South: Automated Scheduler and Data Pipeline

  • Yim, Hong-Suh;Kim, Myung-Jin;Roh, Dong-Goo;Park, Jintae;Moon, Hong-Kyu;Choi, Young-Jun;Bae, Young-Ho;Lee, Hee-Jae;Oh, Young-Seok
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.41 no.1
    • /
    • pp.54.3-55
    • /
    • 2016
  • DEEP-South Scheduling and Data reduction System (DS SDS) consists of two separate software subsystems: Headquarters (HQ) at Korea Astronomy and Space Science Institute (KASI), and SDS Data Reduction (DR) at Korea Institute of Science and Technology Information (KISTI). HQ runs the DS Scheduling System (DSS), DS database (DB), and Control and Monitoring (C&M) designed to monitor and manage overall SDS actions. DR hosts the Moving Object Detection Program (MODP), Asteroid Spin Analysis Package (ASAP) and Data Reduction Control & Monitor (DRCM). MODP and ASAP conduct data analysis while DRCM checks if they are working properly. The functions of SDS is three-fold: (1) DSS plans schedules for three KMTNet stations, (2) DR performs data analysis, and (3) C&M checks whether DSS and DR function properly. DSS prepares a list of targets, aids users in deciding observation priority, calculates exposure time, schedules nightly runs, and archives data using Database Management System (DBMS). MODP is designed to discover moving objects on CCD images, while ASAP performs photometry and reconstructs their lightcurves. Based on ASAP lightcurve analysis and/or MODP astrometry, DSS schedules follow-up runs to be conducted with a part of, or three KMTNet telescopes.

  • PDF

Relational Database SQL Test Auto-scoring System

  • Hur, Tai-Sung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.11
    • /
    • pp.127-133
    • /
    • 2019
  • SQL is the most common language in data processing. Therefore, most of the colleges offer SQL in their curriculum. In this research, an auto scoring SQL test is proposed for the efficient results of SQL education. The system was treated with algorithms instead of using expensive DBMS(Data Base Management System) for automatic scoring, and satisfactory results were produced. For this system, the test question bank was established out of 'personnel management' and 'academic management'. It provides users with different sets of test each time. Scoring was done by dividing tables into two sections. The one that does not change the table(select) and the other that actually changes the table(update, insert, delete). In the case of a search, the answer and response were executed at first and then the results were compared and processed, the user's answers are evaluated by comparing the table with the correct answer. Modification, insertion, and deletion of table actually changes the data table, so data was restored by using ROLLBACK command. This system was implemented and tested 772 times on the 88 students in Computer Information Division of our college. The results of the implementation show that the average scoring time for a test consisting of 10 questions is 0.052 seconds, and the performance of this system is distinguished considering that multiple responses cannot be processed at the same time by a human grader, we want to develop a problem system that takes into account the difficulty of the problem into account near future.