• Title/Summary/Keyword: Search Volume Index

Search Result 30, Processing Time 0.023 seconds

An Improved Split Algorithm for Indexing of Moving Object Trajectories (이동 객체 궤적의 색인을 위한 개선된 분할 알고리즘)

  • Jeon, Hyun-Jun;Park, Ju-Hyun;Park, Hee-Suk;Cho, Woo-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.16D no.2
    • /
    • pp.161-168
    • /
    • 2009
  • Recently, use of various position base servicesthat collect position information for moving object and utilize in real life is increasing by the development of wireless network technology. Accordingly, new index structures are required to efficiently retrieve the consecutive positions of moving objects. This paper addresses an improved trajectory split algorithm for the purpose of efficiently supporting spatio-temporal range queries using index structures that use Minimum Bounding Rectangles(MBR) as trajectory approximations. We consider volume of Extended Minimum Bounding Rectangles (EMBR) to be determined by average size of range queries. Also, Use a priority queue to speed up our process. This algorithm gives in general sub-optimal solutions with respect to search space. Our improved trajectory split algorithm is going to derive minimizing volume of EMBRs better than previously proposed split algorithm.

Anomaly Detection Analysis using Repository based on Inverted Index (역방향 인덱스 기반의 저장소를 이용한 이상 탐지 분석)

  • Park, Jumi;Cho, Weduke;Kim, Kangseok
    • Journal of KIISE
    • /
    • v.45 no.3
    • /
    • pp.294-302
    • /
    • 2018
  • With the emergence of the new service industry due to the development of information and communication technology, cyber space risks such as personal information infringement and industrial confidentiality leakage have diversified, and the security problem has emerged as a critical issue. In this paper, we propose a behavior-based anomaly detection method that is suitable for real-time and large-volume data analysis technology. We show that the proposed detection method is superior to existing signature security countermeasures that are based on large-capacity user log data according to in-company personal information abuse and internal information leakage. As the proposed behavior-based anomaly detection method requires a technique for processing large amounts of data, a real-time search engine is used, called Elasticsearch, which is based on an inverted index. In addition, statistical based frequency analysis and preprocessing were performed for data analysis, and the DBSCAN algorithm, which is a density based clustering method, was applied to classify abnormal data with an example for easy analysis through visualization. Unlike the existing anomaly detection system, the proposed behavior-based anomaly detection technique is promising as it enables anomaly detection analysis without the need to set the threshold value separately, and was proposed from a statistical perspective.

Efficient Indexing for Large DNA Sequence Databases (대용량 DNA 시퀀스 데이타베이스를 위한 효율적인 인덱싱)

  • Won Jung-Im;Yoon Jee-Hee;Park Sang-Hyun;Kim Sang-Wook
    • Journal of KIISE:Databases
    • /
    • v.31 no.6
    • /
    • pp.650-663
    • /
    • 2004
  • In molecular biology, DNA sequence searching is one of the most crucial operations. Since DNA databases contain a huge volume of sequences, a fast indexing mechanism is essential for efficient processing of DNA sequence searches. In this paper, we first identify the problems of the suffix tree in aspects of the storage overhead, search performance, and integration with DBMSs. Then, we propose a new index structure that solves those problems. The proposed index consists of two parts: the primary part represents the trie as bit strings without any pointers, and the secondary part helps fast accesses of the leaf nodes of the trio that need to be accessed for post processing. We also suggest an efficient algorithm based on that index for DNA sequence searching. To verify the superiority of the proposed approach, we conducted a performance evaluation via a series of experiments. The results revealed that the proposed approach, which requires smaller storage space, achieves 13 to 29 times performance improvement over the suffix tree.

Volumetric measurement of the tongue and oral cavity with cone-beam computed tomography: A systematic review

  • Kannitha Alina, Aflah;Winny, Yohana;Fahmi, Oscandar
    • Imaging Science in Dentistry
    • /
    • v.52 no.4
    • /
    • pp.333-342
    • /
    • 2022
  • Purpose: The goal of this systematic review was to compare the use of cone-beam computed tomography (CBCT) with that of computed tomography (CT) for volumetric evaluations of the tongue and oral cavity. Materials and Methods: A search for articles was conducted in accordance with the Preferred Reporting Items for Systematic Review and Meta-analyses guidelines. The PubMed, Scopus, ScienceDirect, and SAGE Journals databases were searched for articles published between 2011 and 2021. Articles were screened and assessed for eligibility. Screening involved checking for duplication, reading the title and abstract, and reading the full text. Results: The initial search retrieved 25,780 articles. Application of the eligibility criteria yielded 16 articles for qualitative analysis. Multiple uses of CBCT were identified. In several studies, researchers assessed the volumetric correlation between tongue and oral cavity volumes, as well as other parameters. Post-treatment volumetric evaluations of the oral cavity were also reported, and the reliability of CBCT was assessed. The use of CT resembled that of CBCT. Conclusion: CBCT has been used in the evaluation of tongue and oral cavity volumes to assess correlations between those volumes and with the upper airway. It has also been used for volumetric evaluation after surgical and nonsurgical procedures and to assess the relationships between tongue volume, tooth position, occlusion, and body mass index. Participants with obstructive sleep apnea and malocclusion have been evaluated, and the reliability of CBCT has been assessed. In the included studies, CT was utilized for similar purposes as CBCT, but its reliability was not assessed.

A Study on Ships Optimal Speed, Deadweight and Their Economy (On the Operations of Common Bulk Carriers Under the Various Managerial Circumstances of Shipping Companies) (상선의 최적속력 및 적화중량톤과 경제성에 관한 연구 ( 일반살적화물선에 있어서 해운운영상의 여건변동을 중심으로 ))

  • 양시권
    • Journal of the Korean Institute of Navigation
    • /
    • v.7 no.2
    • /
    • pp.65-113
    • /
    • 1983
  • A lot of studies of ship's economy are on the traditional fields such asreducing propulsion resistance, raising cargo handling rates and lessening building consts, but there are few researches on the merchant ship's economy concerning their deadweights and speeds according to shipping companies managerial cercumstances. Contrary to the contemporary trend that "the bigger, the better, if the cargo handling rate could increased sufficiently to hold down port time to that rate of smmaler vessels", this paper demonstrates the existence of certain limits in ship's size and speed according to the coditions of the freight rates, voyage distances, cargo handing rates, prices of fuel oil, interst rates etc. Fom the curves of criteria contour for various ship's deadweights and speeds which are depicted from the gird search method, one can get the costs and the yearly profit rates under the conditiions of large volume with long term contracts for the transportation of bulk cargoes. In estimating ship's transportation economy, the auther takes the position that the profit rate method is properer than the cost method, and introduces the calculation table of the voyage profit rate index. The use of the criteria contours will be of help to ship owners in determining the size and speed of the ship which will be built or purchased and serve in a certain trade route.

  • PDF

An Index-Based Approach for Subsequence Matching Under Time Warping in Sequence Databases (시퀀스 데이터베이스에서 타임 워핑을 지원하는 효과적인 인덱스 기반 서브시퀀스 매칭)

  • Park, Sang-Hyeon;Kim, Sang-Uk;Jo, Jun-Seo;Lee, Heon-Gil
    • The KIPS Transactions:PartD
    • /
    • v.9D no.2
    • /
    • pp.173-184
    • /
    • 2002
  • This paper discuss an index-based subsequence matching that supports time warping in large sequence databases. Time warping enables finding sequences with similar patterns even when they are of different lengths. In earlier work, Kim et al. suggested an efficient method for whole matching under time warping. This method constructs a multidimensional index on a set of feature vectors, which are invariant to time warping, from data sequences. For filtering at feature space, it also applies a lower-bound function, which consistently underestimates the time warping distance as well as satisfies the triangular inequality. In this paper, we incorporate the prefix-querying approach based on sliding windows into the earlier approach. For indexing, we extract a feature vector from every subsequence inside a sliding window and construct a multidimensional index using a feature vector as indexing attributes. For query processing, we perform a series of index searches using the feature vectors of qualifying query prefixes. Our approach provides effective and scalable subsequence matching even with a large volume of a database. We also prove that our approach does not incur false dismissal. To verify the superiority of our approach, we perform extensive experiments. The results reveal that our approach achieves significant speedup with real-world S&P 500 stock data and with very large synthetic data.

Efficient Methods for Detecting Frame Characteristics and Objects in Video Sequences (내용기반 비디오 검색을 위한 움직임 벡터 특징 추출 알고리즘)

  • Lee, Hyun-Chang;Lee, Jae-Hyun;Jang, Ok-Bae
    • Journal of KIISE:Software and Applications
    • /
    • v.35 no.1
    • /
    • pp.1-11
    • /
    • 2008
  • This paper detected the characteristics of motion vector to support efficient content -based video search of video. Traditionally, the present frame of a video was divided into blocks of equal size and BMA (block matching algorithm) was used, which predicts the motion of each block in the reference frame on the time axis. However, BMA has several restrictions and vectors obtained by BMA are sometimes different from actual motions. To solve this problem, the foil search method was applied but this method is disadvantageous in that it has to make a large volume of calculation. Thus, as an alternative, the present study extracted the Spatio-Temporal characteristics of Motion Vector Spatio-Temporal Correlations (MVSTC). As a result, we could predict motion vectors more accurately using the motion vectors of neighboring blocks. However, because there are multiple reference block vectors, such additional information should be sent to the receiving end. Thus, we need to consider how to predict the motion characteristics of each block and how to define the appropriate scope of search. Based on the proposed algorithm, we examined motion prediction techniques for motion compensation and presented results of applying the techniques.

Analysis of Clinical Research Trends on Quantitative Indicators of Mibyeong in China - Using China National Knowledge Infrastructure - (미병 정량 지표에 관한 중국의 임상연구 동향 분석 - China National Knowledge Infrastructure를 중심으로 -)

  • Yeo, Minkyung;Lee, Youngseop
    • Journal of Society of Preventive Korean Medicine
    • /
    • v.22 no.1
    • /
    • pp.15-28
    • /
    • 2018
  • Objectives : The purpose of this study was to analyze the trend of clinical research on quantitative indicators of Mibyeong in traditional chinese medicine(TCM). Method : The journal search was performed using china national knowledge infrastructure(CNKI) database. Our inclusion criteria were as following: TCM clinical researches for quantitative indicators of Mibyeong. Exclusion criteria were as following: non-TCM clinical researches, used intervention methods. Results : Eleven clinical researches were analyzed in this study. Four of these researches classify the Mibyeong as a type of pattern identification(PI) and studied the characteristics of the PI quantitative indicators. Mibyeong diagnosis was done through guidelines and questionnaires, each was used at a similar rate. Quantitative indicators mentioned in the selected researches were blood indices, nailfold capillary, complexion, color of tongue substance and coating, pulse wave diagrams and heart rate variability. Among them, seven researches related to blood indices were the most. Blood indicators include whole-blood viscosity, plasma viscosity, fibrinogen, packed cell volume(Hct), triglycerides, total cholesterol, HDL-C, LDL-C, glucose, BUN/CREA, luteinzing hormone, estradiol, follicle stimulating hormone, IgA, IgG, etc. Conclusions : Based on this results, in combination with western medicine, it seems necessary to try to interpret the Mibyeong in more various ways. Even if the same Mibyeong, it is necessary to identify the index which changes according to the PI or chief complaint, and to set the Mibyeong standard corresponding thereto.

Predicting the Direction of the Stock Index by Using a Domain-Specific Sentiment Dictionary (주가지수 방향성 예측을 위한 주제지향 감성사전 구축 방안)

  • Yu, Eunji;Kim, Yoosin;Kim, Namgyu;Jeong, Seung Ryul
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.1
    • /
    • pp.95-110
    • /
    • 2013
  • Recently, the amount of unstructured data being generated through a variety of social media has been increasing rapidly, resulting in the increasing need to collect, store, search for, analyze, and visualize this data. This kind of data cannot be handled appropriately by using the traditional methodologies usually used for analyzing structured data because of its vast volume and unstructured nature. In this situation, many attempts are being made to analyze unstructured data such as text files and log files through various commercial or noncommercial analytical tools. Among the various contemporary issues dealt with in the literature of unstructured text data analysis, the concepts and techniques of opinion mining have been attracting much attention from pioneer researchers and business practitioners. Opinion mining or sentiment analysis refers to a series of processes that analyze participants' opinions, sentiments, evaluations, attitudes, and emotions about selected products, services, organizations, social issues, and so on. In other words, many attempts based on various opinion mining techniques are being made to resolve complicated issues that could not have otherwise been solved by existing traditional approaches. One of the most representative attempts using the opinion mining technique may be the recent research that proposed an intelligent model for predicting the direction of the stock index. This model works mainly on the basis of opinions extracted from an overwhelming number of economic news repots. News content published on various media is obviously a traditional example of unstructured text data. Every day, a large volume of new content is created, digitalized, and subsequently distributed to us via online or offline channels. Many studies have revealed that we make better decisions on political, economic, and social issues by analyzing news and other related information. In this sense, we expect to predict the fluctuation of stock markets partly by analyzing the relationship between economic news reports and the pattern of stock prices. So far, in the literature on opinion mining, most studies including ours have utilized a sentiment dictionary to elicit sentiment polarity or sentiment value from a large number of documents. A sentiment dictionary consists of pairs of selected words and their sentiment values. Sentiment classifiers refer to the dictionary to formulate the sentiment polarity of words, sentences in a document, and the whole document. However, most traditional approaches have common limitations in that they do not consider the flexibility of sentiment polarity, that is, the sentiment polarity or sentiment value of a word is fixed and cannot be changed in a traditional sentiment dictionary. In the real world, however, the sentiment polarity of a word can vary depending on the time, situation, and purpose of the analysis. It can also be contradictory in nature. The flexibility of sentiment polarity motivated us to conduct this study. In this paper, we have stated that sentiment polarity should be assigned, not merely on the basis of the inherent meaning of a word but on the basis of its ad hoc meaning within a particular context. To implement our idea, we presented an intelligent investment decision-support model based on opinion mining that performs the scrapping and parsing of massive volumes of economic news on the web, tags sentiment words, classifies sentiment polarity of the news, and finally predicts the direction of the next day's stock index. In addition, we applied a domain-specific sentiment dictionary instead of a general purpose one to classify each piece of news as either positive or negative. For the purpose of performance evaluation, we performed intensive experiments and investigated the prediction accuracy of our model. For the experiments to predict the direction of the stock index, we gathered and analyzed 1,072 articles about stock markets published by "M" and "E" media between July 2011 and September 2011.

The Performance Bottleneck of Subsequence Matching in Time-Series Databases: Observation, Solution, and Performance Evaluation (시계열 데이타베이스에서 서브시퀀스 매칭의 성능 병목 : 관찰, 해결 방안, 성능 평가)

  • 김상욱
    • Journal of KIISE:Databases
    • /
    • v.30 no.4
    • /
    • pp.381-396
    • /
    • 2003
  • Subsequence matching is an operation that finds subsequences whose changing patterns are similar to a given query sequence from time-series databases. This paper points out the performance bottleneck in subsequence matching, and then proposes an effective method that improves the performance of entire subsequence matching significantly by resolving the performance bottleneck. First, we analyze the disk access and CPU processing times required during the index searching and post processing steps through preliminary experiments. Based on their results, we show that the post processing step is the main performance bottleneck in subsequence matching, and them claim that its optimization is a crucial issue overlooked in previous approaches. In order to resolve the performance bottleneck, we propose a simple but quite effective method that processes the post processing step in the optimal way. By rearranging the order of candidate subsequences to be compared with a query sequence, our method completely eliminates the redundancy of disk accesses and CPU processing occurred in the post processing step. We formally prove that our method is optimal and also does not incur any false dismissal. We show the effectiveness of our method by extensive experiments. The results show that our method achieves significant speed-up in the post processing step 3.91 to 9.42 times when using a data set of real-world stock sequences and 4.97 to 5.61 times when using data sets of a large volume of synthetic sequences. Also, the results show that our method reduces the weight of the post processing step in entire subsequence matching from about 90% to less than 70%. This implies that our method successfully resolves th performance bottleneck in subsequence matching. As a result, our method provides excellent performance in entire subsequence matching. The experimental results reveal that it is 3.05 to 5.60 times faster when using a data set of real-world stock sequences and 3.68 to 4.21 times faster when using data sets of a large volume of synthetic sequences compared with the previous one.