• Title/Summary/Keyword: Query type

Search Result 166, Processing Time 0.024 seconds

Identification of Transformed Image Using the Composition of Features

  • Yang, Won-Keun;Cho, A-Young;Cho, Ik-Hwan;Oh, Weon-Geun;Jeong, Dong-Seok
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.6
    • /
    • pp.764-776
    • /
    • 2008
  • Image identification is the process of checking whether the query image is the transformed version of the specific original image or not. In this paper, image identification method based on feature composition is proposed. Used features include color distance, texture information and average pixel intensity. We extract color characteristics using color distance and texture information by Modified Generalized Symmetry Transform as well as average intensity of each pixel as features. Individual feature is quantized adaptively to be used as bins of histogram. The histogram is normalized according to data type and it is used as the signature in comparing the query image with database images. In matching part, Manhattan distance is used for measuring distance between two signatures. To evaluate the performance of the proposed method, independent test and accuracy test are achieved. In independent test, 60,433 images are used to evaluate the ability of discrimination between different images. And 4,002 original images and its 29 transformed versions are used in accuracy test, which evaluate the ability that the proposed algorithm can find the original image correctly when some transforms was applied in original image. Experiment results show that the proposed identification method has good performance in accuracy test. And the proposed method is very useful in real environment because of its high accuracy and fast matching capacity.

  • PDF

Policy Reorganization Method for Performance Improvements in SELinux using Loadable Module Policy (로드 가능한 모듈 정책을 사용하는 SELinux의 성능 향상을 위한 정책 재구성 방법)

  • Ko, Jae-Yong;Lee, Sanggil;Cho, Kyung-Yeon;Lee, Cheol-Hoon
    • The Journal of the Korea Contents Association
    • /
    • v.18 no.3
    • /
    • pp.309-319
    • /
    • 2018
  • SELinux is used for system level security in various systems using Linux, and is now being used for device security such as IoT. However, since SELinux has inherent problems of execution time degradation, various studies have been conducted to solve this problem. In this paper, we show that performance can be improved through policy reconfiguration in the environment where the loadable module policy method, which is a general method using SELinux, is applied. By reconfiguring the access query table through the Priority-TE policy that gives priority to the type, it is possible to provide faster execution time for types requiring faster access query performance. This paper introduces the differences between SELinux policy configuration method in Monolithic environment and performance analysis. This can be used as a reference by security administrators or developers in applying SELinux.

Efficient Linear Path Query Processing using Information Retrieval Techniques for Large-Scale Heterogeneous XML Documents (정보 검색 기술을 이용한 대규모 이질적인 XML 문서에 대한 효율적인 선형 경로 질의 처리)

  • 박영호;한욱신;황규영
    • Journal of KIISE:Databases
    • /
    • v.31 no.5
    • /
    • pp.540-552
    • /
    • 2004
  • We propose XIR-Linear, a novel method for processing partial match queries on large-scale heterogeneous XML documents using information retrieval (IR) techniques. XPath queries are written in path expressions on a tree structure representing an XML document. An XPath query in its major form is a partial match query. The objective of XIR-Linear is to efficiently support this type of queries for large-scale documents of heterogeneous schemas. XIR-Linear has its basis on the schema-level methods using relational tables and drastically improves their efficiency and scalability using an inverted index technique. The method indexes the labels in label paths as key words in texts, and allows for finding the label paths that match the queries far more efficiently than string match used in conventional methods. We demonstrate the efficiency and scalability of XIR-Linear by comparing it with XRel and XParent using XML documents crawled from the Internet. The results show that XIR-Linear is more efficient than both XRel and XParent by several orders of magnitude for linear path expressions as the number of XML documents increases.

The Cognitive Degree and Its Related Factors about Positive Hepatitis and Hepatosis of 20s Adults

  • Lee, Jae Sik
    • Korean Journal of Clinical Laboratory Science
    • /
    • v.45 no.4
    • /
    • pp.159-163
    • /
    • 2013
  • To investigate the cognitive degree about hepatitis and hepatosis, 916 subjects are examined with query and hepatotitis B, C, E test as well as s-AST and s-ALT as liver function test. Based on results, there are 4.9% of positive hepatitis and 8.9% of hepatosis and 13.8% of liver disorder. Among positive hepatitis, there are 93.3% of type B, 42.2% of type E and 6.7% of type C, respectively. From 45% of positive hepatitis B, they carry hepatitis B and E together. The cognitive degree about positive hepatitis is 64.4%, hepatosis 8.6%. The knowledge degree from cognitive group is higher than that of noncognitive group but there is no difference from hepatosis between two groups (p<0.001). The cognitive degree of liver disorder depends on academic background (p<0.001), mother's academic background (p<0.001), job (p<0.05) and family's income (p<0.001), showing significant difference. In summary, hepatitis carrier aware quite well about liver disorder but very low from hepatosis. Accordingly, the plan to increase a cognitive degree and continuous education as well as policy support to minimize spread of disease and to protect not to be worsen disease will be needed.

  • PDF

Development of Adoption Strategy and Guideline of Business Process Management Standards: Focusing on Business Process Execution Language (비즈니스 프로세스 관리 표준 도입 전략 및 지침 개발: 비즈니스 프로세스 실행 언어를 중심으로)

  • Kim, Dong-Soo
    • Journal of Information Technology Services
    • /
    • v.5 no.2
    • /
    • pp.107-123
    • /
    • 2006
  • The objectives of this study is to develop a strategy for the adoption of BPM(Business Process Management) standards and an implementation guideline of the BPM standard for BPM solution developers focusing on BPEL(Business Process Execution Language) which is regarded as the most important BPM standard. In the heterogeneous and distributed IT environments, every type of enterprise software requires standards to enhance interoperability. BPMS(Business Process Management System), which is a type of enterprise software requires BPM standards such as BPEL(Business Process Execution Language), BPMN(Business Process Modeling and Notation), BPQL(Business Process Query Language) and so on to achieve multi-system interoperability and component interoperability with their BPM solutions. It is quite helpful to provide the adoption strategy concerning BPM standards for each type of BPM solution vendors who need the BPM standards. Since the BPEL is conceived as the most important BPM standard and widely adopted by many BPM vendors, we have proposed a reference architecture for BPEL implementation and also developed the detail implementation guideline of core components of the BPM system supporting the BPEL standard. Using the strategy and implementation guideline proposed in this work, BPM solution vendors can establish their own standard adoption strategy and they can also develop their BPM solutions supporting the BPM standards more efficiently.

A Study of Approximation Method of Spatial Objects (공간 객체의 근사화 방법 연구)

  • Kim, Yong-Hyeon;Lee, Hyeong-Su;Lee, Seong-Su;Kim, Eung-Mo
    • The Transactions of the Korea Information Processing Society
    • /
    • v.3 no.7
    • /
    • pp.1753-1762
    • /
    • 1996
  • Since the number of spatial objects go easily into millions, they are stored on a secondary storage. In order to speed up accessing the geometric objects, we investigate spatial database system. The spatial objects are organized and accessed by spatial access methods(SAMs). But, SAMs are not able to organize polygons directly The most popular approach for handing polygon in SAMs is to use MBR apporximaotion as a geometric key. MBR provides a fast but inaccurate answers to approximation-based query processing. The performance of approximation-based spatial query processing depends on which type of approximation is chosen for the spatial objects. A suitable approximation is crucial for reducing the size of the candidate set. The better the approxmation quality, the fewer accesses to the exact object are necessary. In this paper, I proposed a new multicontainer approximation named Slice decomposition. The performance of proposed method is compared with other method.

  • PDF

The Scalability and the Strategy for EMR Database Encryption Techniques

  • Shin, David;Sahama, Tony;Kim, Steve Jung-Tae;Kim, Ji-Hong
    • Journal of information and communication convergence engineering
    • /
    • v.9 no.5
    • /
    • pp.577-582
    • /
    • 2011
  • EMR(Electronic Medical Record) is an emerging technology that is highly-blended between non-IT and IT area. One of methodology to link non-IT and IT area is to construct databases. Nowadays, it supports before and after-treatment for patients and should satisfy all stakeholders such as practitioners, nurses, researchers, administrators and financial department and so on. In accordance with the database maintenance, DAS (Data as Service) model is one solution for outsourcing. However, there are some scalability and strategy issues when we need to plan to use DAS model properly. We constructed three kinds of databases such as plain-text, MS built-in encryption which is in-house model and custom AES (Advanced Encryption Standard) - DAS model scaling from 5K to 2560K records. To perform custom AES-DAS better, we also devised Bucket Index using Bloom Filter. The simulation showed the response times arithmetically increased in the beginning but after a certain threshold, exponentially increased in the end. In conclusion, if the database model is close to in-house model, then vendor technology is a good way to perform and get query response times in a consistent manner. If the model is DAS model, it is easy to outsource the database, however, some technique like Bucket Index enhances its utilization. To get faster query response times, designing database such as consideration of the field type is also important. This study suggests cloud computing would be a next DAS model to satisfy the scalability and the security issues.

Decision Method of Importance of E-Mail based on User Profiles (사용자 프로파일에 기반한 전자 메일의 중요도 결정)

  • Lee, Samuel Sang-Kon
    • The KIPS Transactions:PartB
    • /
    • v.15B no.5
    • /
    • pp.493-500
    • /
    • 2008
  • Although modern day people gather many data from the network, the users want only the information needed. Using this technology, the users can extract on the data that satisfy the query. As the previous studies use the single data in the document, frequency of the data for example, it cannot be considered as the effective data clustering method. What is needed is the effective clustering technology that can process the electronic network documents such as the e-mail or XML that contain the tags of various formats. This paper describes the study of extracting the information from the user query based on the multi-attributes. It proposes a method of extracting the data such as the sender, text type, time limit syntax in the text, and title from the e-mail and using such data for filtering. It also describes the experiment to verify that the multi-attribute based clustering method is more accurate than the existing clustering methods using only the word frequency.

The XML Compression Algorithm Supporting Query Processing For Compressed Documents (압축된 문서에 대해 질의 처리를 지원하는 XML 압축 알고리즘)

  • 강영준;이석재;유재수
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2003.11a
    • /
    • pp.195-203
    • /
    • 2003
  • With the spread of interment, the digitalization and knowledge-based information are in progress. Specially, numerous users make the various works and use the services on the web. For the most part, these works make use of the XML. The XML shines the reusing of the Documents because it is separated from contents and styles. Also, it can re-define the logic structure of the Document for requirement of the developer. However, the XML document's size is much larger than common text document because it basically handles the document type and adds numerous tags for representing structure of the document. To utilize the limited storage of Palmtop, PDA and so on, it is necessary to compress and handle the documents efficiently. Recently, the compression techniques for efficiently handling and compressing the XML documents are in progress to solve this problem. But the existing research doesn't support the query processing for that. In this paper, we design and implement the XML compression algorithm that compresses the XML document and processes the quay of compressed XML document faster and mote effciently than the previous techniques.

  • PDF

Data Transformation through Mapping between XML and Relation Database (XML과 관계형 데이타베이스 매핑을 통한 자료의 변환)

  • Kim Gil-Choon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.9 no.4 s.32
    • /
    • pp.5-12
    • /
    • 2004
  • The data transformation between XML and Relation Database is made through the Principle of mapping bewtween them. There are two ways to access SQL Server, one is to assign SQL query to URL and the other is to use template file. MS-SQL server takes advantage of OpenXML function to transform the results of executing SQL query into XML documents. That is, OpenXML first makes node tree and then transforms row set data of XML documents into XML data of relation type. In order to insert XML data into database data. data is extracted from parsing XML documents using sp_xml_preparedocument procedure, and then the document structure is mapped into tree structure and stored in a table of database. Consequently, Data transformation between XML and Relation Database is made through mapping bewtween them. This article proposes the principle of mapping between XML and Relation Database and then shows the implementation of transformation between them so that it introduces the possibilty of bringing the extension and efficiency of data and various effects.

  • PDF