• Title/Summary/Keyword: Query Ratio

Search Result 68, Processing Time 0.024 seconds

Spatial Selectivity Estimation for Intersection region Information Using Cumulative Density Histogram

  • Kim byung Cheol;Moon Kyung Do;Ryu Keun Ho
    • Proceedings of the KSRS Conference
    • /
    • 2004.10a
    • /
    • pp.721-725
    • /
    • 2004
  • Multiple-count problem is occurred when rectangle objects span across several buckets. The Cumulative Density (CD) histogram is a technique which solves multiple-count problem by keeping four sub-histograms corresponding to the four points of rectangle. Although it provides exact results with constant response time, there is still a considerable issue. Since it is based on a query window which aligns with a given grid, a number of errors may be occurred when it is applied to real applications. In this paper, we proposed selectivity estimation techniques using the generalized cumulative density histogram based on two probabilistic models: (1) probabilistic model which considers the query window area ratio, (2) probabilistic model which considers intersection area between a given grid and objects. In order to evaluate the proposed methods, we experimented with real dataset and experimental results showed that the proposed technique was superior to the existing selectivity estimation techniques. The proposed techniques can be used to accurately quantify the selectivity of the spatial range query on rectangle objects.

  • PDF

A Location Management Scheme using User Locality in Cellular Communciation Systems (셀룰러 통신 시스템에서 사용자 지역성을 고려한 위치 관리 기법)

  • Yang, Gwon-U;Gil, Jun-Min;Kim, Gi-Beom;Jeong, Sun-Yeong;Hwang, Jong-Seon
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.26 no.9
    • /
    • pp.1135-1144
    • /
    • 1999
  • 본 논문에서 우리는 셀룰러 통신 시스템을 위한 위치갱신과 위치질의 비용을 줄일 수 있는 새로운 위치관리 기법을 제안한다. 위치갱신 및 위치질의 비용을 줄이기 위하여, 제안한 위치관리 기법은 사용자 지역성을 사용한다. 사용자 지역성은 이동 지역성과 호 지역성으로 이루어진다. 이동 지역성을 갖는 이동단말들을 위한 위치갱신은 지역 PVLR(Proxy-Visitor Location Register) 갱신에 의해 위치갱신 비용을 줄일 수 있다. 호 지역성을 갖는 이동단말들을 위한 위치질의는 지역 PVLR 질의에 의해 위치질의 비용을 줄일 수 있다. 분석 모델을 통하여, 제안한 기법의 성능 분석과 IS-41 표준과 제안 기법과의 성능을 비교한다. 사용된 비교 척도는 네트워크 비용, 데이타베이스 비용 그리고 호 설정 지연이다. 이들 모든 척도들에 대해서, 제안 기법이 IS-41 표준에서 정의된 위치관리 기법보다 더 좋은 성능을 보인다. 또한 비교 결과는 CMR(Call-to-Mobility Ratio)에 상관없이 제안 기법이 IS-41 기법보다 성능이 좋음을 보여준다. Abstract In this paper, we propose a new location management scheme for cellular communication systems with the goal of reducing both location update and query cost. In order to reduce both location update and query cost, the proposed scheme uses user locality. User locality consists of movement and call locality. For mobile terminals with movement locality, location update cost is reduced by local PVLR(Proxy-Visitor Location Register) update. Location query cost is reduced by local PVLR query for mobile terminals with call locality. Through analytical models, we analyze the performance of the proposed scheme and compare it with that of the IS-41 scheme. The measures of comparison used are the network cost, the database cost, and the call set up delay. For all these measures, the proposed scheme performs better than the location management scheme defined in the IS-41 standard. Result comparison also shows that the our scheme performs better than the IS-41 scheme, regardless of CMR(Call-to-Mobility Ratio).

The Application and Integration of an Improvement Technique for Layers of NETCONF (NETCONF 계층에 대한 개선 기법 적용 및 통합)

  • Lee, YangMin;Lee, JaeKee
    • Journal of KIISE
    • /
    • v.43 no.2
    • /
    • pp.256-268
    • /
    • 2016
  • Modern networks consisting of various heterogeneous equipment are often installed in a distributed manner. Thus the NETCONF standard was established to manage networks centrally and efficiently. In this paper, we present a method that integrates each NETCONF layer into a single system based on the results of previous studies. In the RPC Layer, an asynchronous communication channel and parallel processes are possible using multi-threading. In the Operation Layer, operational efficiency is increased by using a data group with dependencies between the equipment configuration data and by improving the data structure, enabling efficiently processing of XML queries even with multiple managers. The data modeling techniques and grouping methods in the Content Layer are presented in detail for interoperability between the Operation Layer and the Content Layer. Finally, the GUI program was implemented and its implementation is reported. We performed an experiment comparing the improved NETCONF with the standard NETCONF to measure factors, such as query processing ratio, query processing speed, and CPU utilization. The improved NETCONF demonstrated excellent query processing ratio and query processing speed, whereas the standard NETCONF had excellent CPU utilization.

A Semantic-based Video Retrieval System using Method of Automatic Annotation Update and Multi-Partition Color Histogram (자동 주석 갱신 및 멀티 분할 색상 히스토그램 기법을 이용한 의미기반 비디오 검색 시스템)

  • 이광형;전문석
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.8C
    • /
    • pp.1133-1141
    • /
    • 2004
  • In order to process video data effectively, it is required that the content information of video data is loaded in database and semantic-based retrieval method can be available for various query of users. In this paper, we propose semantic-based video retrieval system which support semantic retrieval of various users by feature-based retrieval and annotation-based retrieval of massive video data. By user's fundamental query and selection of image for key frame that extracted from query, the agent gives the detail shape for annotation of extracted key frame. Also, key frame selected by user become query image and searches the most similar key frame through feature based retrieval method that propose. From experiment, the designed and implemented system showed high precision ratio in performance assessment more than 90 percents.

Skyline Query Algorithm in the Categoric Data (범주형 데이터에 대한 스카이라인 질의 알고리즘)

  • Lee, Woo-Key;Choi, Jung-Ho;Song, Jong-Su
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.7
    • /
    • pp.819-823
    • /
    • 2010
  • The skyline query is one of the effective methods to deal with the large amounts and multi-dimensional data set. By utilizing the concept of 'dominate' the skyline query can pinpoint the target data so that the dominated ones, about 95% of them, can efficiently be excluded as an unnecessary data. Most of the skyline query algorithms, however, have been developed in terms of the numerical data set. This paper pioneers an entirely new domain, the categorical data, on which the corresponding ranking measures for the skyline queries are suggested. In the experiment, the ACM Computing Classification System has been exploited to which our methods are significantly represented with respect to performance thresholds such as the processing time and precision ratio, etc.

Data-Aware Priority-Based Energy Efficient Top-k Query Processing in Sensor Networks (센서 네트워크를 위한 데이터 인지 우선순위 기반의 에너지 효율적인 Top-k 질의 처리)

  • Yeo, Myung-Ho;Seong, Dong-Ook;Yoo, Jae-Soo
    • Journal of KIISE:Databases
    • /
    • v.36 no.3
    • /
    • pp.189-197
    • /
    • 2009
  • Top-k queries are important to many wireless sensor applications. Conventional Top-k query processing algorithms install a filter at each sensor node and suppress unnecessary sensor updates. However, they have some drawbacks that the sensor nodes consume energy extremely to probe sensor reading or update filters. Especially, it becomes worse, when the variation ratio of top-k result is higher. In this paper, we propose a novel Top-k query processing algorithm for energy-efficiency. First, each sensor determines its priority as the order of data gathering. Next, sensor nodes that have higher priority transmit their sensor readings to the base station until gathering k sensor readings. In order to show the superiority of our query processing algorithm, we simulate the performance with the existing query processing algorithms. As a result, our experimental results show that the network lifetime of our method is prolonged largely over the existing method.

A Study on Automatic Text Categorization of Web-Based Query Using Synonymy List (유사어 사전을 이용한 웹기반 질의문의 자동 범주화에 관한 연구)

  • Nam, Young-Joon;Kim, Gyu-Hwan
    • Journal of Information Management
    • /
    • v.35 no.4
    • /
    • pp.81-105
    • /
    • 2004
  • In this study, the way of the automatic text categorization on web-based query was implemented. X2 methods based on the Supported Vector Machine were used to test the efficiency of text categorization on queries. This test is carried out by the model using the Synonymy List. 713 synonyms were extracted manually from the tested documents. As the result of this test, the precision ratio and the recall ratio were decreased by -0.01% and by 8.53%, respectively whether the synonyms were assigned or not. It also shows that the Value of F1 Measure was increased by 4.58%. The standard deviation between the recall and precision ratio was improve by 18.39%.

Analyzing errors in selectivity estimation using the multilevel grid file (계층 그리드 화일을 이용한 선택률 추정에서 발생되는 오차 분석)

  • 김상욱;황환규;황규영
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.33B no.9
    • /
    • pp.24-36
    • /
    • 1996
  • In this paper, we discuss the errors in selectivity estimation using the multilevel grid file (MLGF). We first demonstrate that the estimatio errors stem from the uniformity assumption that records are uniformly distributed in their belonging region represented by an entry in a level of an MLGF directory. Bsed on this demonstration, we then investigate five factors affecting the accuracy of estimation: (1) the data distribution in a region (2) the number of records stored in an MLFG (3) the page size, (4) the query region size, and (5) the level of an MLFG directory. Next we present the tendancy of estimation errors according to the change of values for each factor through experiments. The results show that the errors decrease when (1) the distribution of records in a region becomes closer to the uniform one, (2) the number of records in an MLFG increases, (3) the page size decreases, (4) the query region size increases, and (5) the level of an MLFG directory employed as data distribution information becomes lower. After the definition of the granule ratio, the core formula representing the basic relationship between the estimation errors and the above five factors, we finally examine the change of estimation errors according to the change of the values for the granule ratio through experiments. The results indicate that errors tend to be similar depending on the values for the granule ratio regardless of the various changes of the values for the five factors. factors affecting the accuracy of estimation:

  • PDF

Semantic Query Expansion based on Concept Coverage of a Deep Question Category in QA systems (질의 응답 시스템에서 심층적 질의 카테고리의 개념 커버리지에 기반한 의미적 질의 확장)

  • Kim Hae-Jung;Kang Bo-Yeong;Lee Sang-Jo
    • Journal of KIISE:Databases
    • /
    • v.32 no.3
    • /
    • pp.297-303
    • /
    • 2005
  • When confronted with a query, question answering systems endeavor to extract the most exact answers possible by determining the answer type that fits with the key terms used in the query. However, the efficacy of such systems is limited by the fact that the terms used in a query may be in a syntactic form different to that of the same words in a document. In this paper, we present an efficient semantic query expansion methodology based on a question category concept list comprised of terms that are semantically close to terms used in a query. The semantically close terms of a term in a query may be hypernyms, synonyms, or terms in a different syntactic category. The proposed system constructs a concept list for each question type and then builds the concept list for each question category using a learning algorithm. In the question answering experiments on 42,654 Wall Street Journal documents of the TREC collection, the traditional system showed in 0.223 in MRR and the proposed system showed 0.50 superior to the traditional question answering system. The results of the present experiments suggest the promise of the proposed method.

Spatial View Materialization Technique by using R-Tree Reconstruction (R-tree 재구성 방법을 이용한 공간 뷰 실체화 기법)

  • Jeong, Bo-Heung;Bae, Hae-Yeong
    • The KIPS Transactions:PartD
    • /
    • v.8D no.4
    • /
    • pp.377-386
    • /
    • 2001
  • In spatial database system, spatial view is supported for efficient access method to spatial database and is managed by materialization and non-materialization technique. In non-materialization technique, repeated execution on the same query makes problems such as the bottle-neck effect of server-side and overloads on a network. In materialization technique, view maintenance technique is very difficult and maintenance cost is too high when the base table has been changed. In this paper, the SVMT (Spatial View Materialization Technique) is proposed by using R-tree re-construction. The SVMT is a technique which constructs a spatial index according to the distribution ratio of objects in spatial view. This ratio is computed by using a SVHR (Spatial View Height in R-tree) and SVOC (Spatial View Object Count). If the ratio is higher than the average, a spatial view is materialized and the R-tree index is re-used. In this case, the root node of this index is exchanged a node which has a MBR (Minimum Boundary Rectangle) value that can contains the whole region of spatial view at a minimum size. Otherwise, a spatial view is materialized and the R-tree is re-constructed. In this technique, the information of spatial view is managed by using a SVIT (Spatial View Information Table) and is stored on the record of this table. The proposed technique increases the speed of response time through fast query processing on a materialized view and eliminates additional costs occurred from repeatable query modification on the same query. With these advantages, it can greatly minimize the network overloads and the bottle-neck effect on the server.

  • PDF