• Title/Summary/Keyword: large database

Search Result 1,454, Processing Time 0.027 seconds

A New Pruning Method for Synthesis Database Reduction Using Weighted Vector Quantization

  • Kim, Sanghun;Lee, Youngjik;Keikichi Hirose
    • The Journal of the Acoustical Society of Korea
    • /
    • v.20 no.4E
    • /
    • pp.31-38
    • /
    • 2001
  • A large-scale synthesis database for a unit selection based synthesis method usually retains redundant synthesis unit instances, which are useless to the synthetic speech quality. In this paper, to eliminate those instances from the synthesis database, we proposed a new pruning method called weighted vector quantization (WVQ). The WVQ reflects relative importance of each synthesis unit instance when clustering the similar instances using vector quantization (VQ) technique. The proposed method was compared with two conventional pruning methods through the objective and subjective evaluations of the synthetic speech quality: one to simply limit maximum number of instance, and the other based on normal VQ-based clustering. The proposed method showed the best performance under 50% reduction rates. Over 50% of reduction rates, the synthetic speech quality is not seriously but perceptibly degraded. Using the proposed method, the synthesis database can be efficiently reduced without serious degradation of the synthetic speech quality.

  • PDF

Development of the Unified Database Design Methodology for Big Data Applications - based on MongoDB -

  • Lee, Junho;Joo, Kyungsoo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.3
    • /
    • pp.41-48
    • /
    • 2018
  • The recent sudden increase of big data has characteristics such as continuous generation of data, large amount, and unstructured format. The existing relational database technologies are inadequate to handle such big data due to the limited processing speed and the significant storage expansion cost. Current implemented solutions are mainly based on relational database that are no longer adapted to these data volume. NoSQL solutions allow us to consider new approaches for data warehousing, especially from the multidimensional data management point of view. In this paper, we develop and propose the integrated design methodology based on MongoDB for big data applications. The proposed methodology is more scalable than the existing methodology, so it is easy to handle big data.

Construction of a Database for Road Images and Geometric Transformation (도로영상 데이터베이스 구축 및 기하학적 변환)

  • Lee, Joon-Woong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.6
    • /
    • pp.534-539
    • /
    • 2013
  • Recently, the number of vehicles equipped with cameras that are generally used to recognize surroundings is increasing. For robust recognition, a huge amount of tests under various environments are performed. Furthermore, the installation position or orientation of the camera is also changed depending on the vehicle. This change also accompanies many tests. Correspondingly, a large cost and a great deal of manpower are required to perform these tests. This paper proposes a method to cut these costs while conducting enough tests through the construction of a database of videos and a geometric transformation of images.

A Compressed Histogram Technique for Spatial Selectivity Estimation (공간 선택률 추정을 위한 압축 히스토그램 기법)

  • Chung, Jae-Du;Chi, Jeong-Hee;Ryu, Keun-Ho
    • 한국공간정보시스템학회:학술대회논문집
    • /
    • 2004.12a
    • /
    • pp.69-74
    • /
    • 2004
  • Selectivity estimation for spatial query is very important process in finding the most efficient execution plan. Many works have been performed to estimate accurately selectivity. Although they deal with some problems such as false-count, multi-count, they require a large amount of memory to retain accurate selectivity, so they can not get good results in little memory environments such as mobile-based small database. In order to solve this problem, we propose a new technique called MW histogram which is able to compress summary data and get reasonable results. It also has a flexible structure to react dynamic update. The experimental results showed that the MW histogram has lower relative error than MinSkew histogram and gets a good selectivity in little memory.

  • PDF

A study on automatic data conversion from electronic drawings to make feature database for GIS system (CAD도면과 GIS구조화 자동변환 방안에 관한연구)

  • Park, Dong-Heui;Kim, Young-Guk;Kang, Yu-Shin;Oh, Ju-Hwan;Choo, Jun-Sup
    • Proceedings of the KSR Conference
    • /
    • 2008.11b
    • /
    • pp.2121-2124
    • /
    • 2008
  • The total length of Korean railway network is about 3,300km. Since it is of great scale in system view point, the systemization of GIS-based information system requires so much cost and time. One of the difficulties is due to the fact that GIS-based information system requires the feature database for GIS, which is generally built manually from many as-built drawing files. In order to build-up the feature database for GIS with ease, this study suggests the automatic data conversion from electronic drawings to make feature database for GIS. The proposed method can be applied to build large-scale railway facility management system at lower cost.

  • PDF

Trends in Data Management Technology Using Artificial Intelligence (인공지능 기술을 활용한 데이터 관리 기술 동향)

  • C.S. Kim;C.S. Park;T.W. Lee;J.Y. Kim
    • Electronics and Telecommunications Trends
    • /
    • v.38 no.6
    • /
    • pp.22-30
    • /
    • 2023
  • Recently, artificial intelligence has been in the spotlight across various fields. Artificial intelligence uses massive amounts of data to train machine learning models and performs various tasks using the trained models. For model training, large, high-quality data sets are essential, and database systems have provided such data. Driven by advances in artificial intelligence, attempts are being made to improve various components of database systems using artificial intelligence. Replacing traditional complex algorithm-based database components with their artificial-intelligence-based counterparts can lead to substantial savings of resources and computation time, thereby improving the system performance and efficiency. We analyze trends in the application of artificial intelligence to database systems.

Modeling with Design Patterns in MongoDB for Public Transportation Data

  • Meekyung Min
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.4
    • /
    • pp.460-465
    • /
    • 2024
  • MongoDB, a document-based database, is suitable for distributed management environments of large-scale databases due to its high scalability, performance, and flexibility. Recently, as MongoDB has been widely used as a new database, many studies have been conducted including data modeling for MongoDB and studies on applying MongoDB to various applications. In this paper, we propose a data modeling method for implementing Seoul public transportation data with MongoDB. Seoul public transportation data is public data provided by the Korea Public Data Portal. In this study, we analyze the target data and find design patterns such as polymorphic pattern, subset pattern, computed pattern, and extended reference pattern in the data. Then, we present data modeling with these patterns. We also show examples of implementation of Seoul public transportation database in MongoDB. The proposed modeling method can improve database performance by leveraging the flexibility and scalability that are characteristics of MongoDB.

Development of Database Supported Graph Library and Graph Algorithms (데이터베이스에 기반한 그래프 라이브러리 및 그래프 알고리즘 개발)

  • 박휴찬;추인경
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.6 no.5
    • /
    • pp.653-660
    • /
    • 2002
  • This paper proposes a method for storing graphs and defining graph algorithms based on the well-developed relational database. In this method, graphs are represented in the form of relations. Each vertex and edge of a graph is represented as tuples of the table and saved in a database. We developed a library of graph operations for the storage and management of graphs and the development of graph applications. Furthermore, we defined graph algorithms in terms of relational algebraic operations such as projection, selection, and join. They can be implemented with the database language such as SQL. This database approach provides an efficient methodology to deal with very large-scale graphs and to support the development of graph applications.

Access Control Mechanism for CouchDB

  • Ashwaq A., Al-otaibi;Reem M., Alotaibi;Nermin, Hamza
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.12
    • /
    • pp.107-115
    • /
    • 2022
  • Recently, big data applications need another database different from the Relation database. NoSQL databases are used to save and handle massive amounts of data. NoSQL databases have many advantages over traditional databases like flexibility, efficiently processing data, scalability, and dynamic schemas. Most of the current applications are based on the web, and the size of data is in increasing. NoSQL databases are expected to be used on a more and large scale in the future. However, NoSQL suffers from many security issues, and one of them is access control. Many recent applications need Fine-Grained Access control (FGAC). The integration of the NoSQL databases with FGAC will increase their usability in various fields. It will offer customized data protection levels and enhance security in NoSQL databases. There are different NoSQL database models, and a document-based database is one type of them. In this research, we choose the CouchDB NoSQL document database and develop an access control mechanism that works at a fain-grained level. The proposed mechanism uses role-based access control of CouchDB and restricts read access to work at the document level. The experiment shows that our mechanism effectively works at the document level in CouchDB with good execution time.

The Effectiveness of Life Cycle Assessment for Building Using Inventory Database "IDEA" in Japan

  • Yosuke TANAKA;Yoshiyuki SUZUKI;kensuke KOBAYASHI
    • International conference on construction engineering and project management
    • /
    • 2024.07a
    • /
    • pp.1137-1144
    • /
    • 2024
  • Reducing the environmental impact in the construction industry is essential for a sustainable future, and life cycle assessment (LCA) should be effectively conducted to reduce the environmental impact. The construction industry is one of the fields that emits a large amount of Greenhouse Gas (GHG). It is also characterized by many material inputs and a one-off single production. Therefore, it took a lot of effort to evaluate all the input materials, and it was difficult to implement a detailed LCA. There is need to solve these problems and to establish a fair and reliable evaluation method. In order to solve this problem, it is proposed to establish a common rule for calculating environmental loads of buildings, such as carbon dioxide emissions. In addition, by effectively utilizing the Inventory Database for Environmental Analysis (IDEA) database, which is an inventory database developed in Japan. It can evaluate not only carbon dioxide but also various environmental substances, and analyze how the environmental impact is correlated with each building and its constituent materials. Furthermore, by analyzing the actual buildings of 83 projects, the differences in the tendency of building type and materiall was clarified. A database was constructed to help reduce the environmental impact during the early stages of construction project and for different types of buildings.