• Title/Summary/Keyword: Similarity Distance

Search Result 632, Processing Time 0.023 seconds

Video Content Indexing using Kullback-Leibler Distance

  • Kim, Sang-Hyun
    • International Journal of Contents
    • /
    • v.5 no.4
    • /
    • pp.51-54
    • /
    • 2009
  • In huge video databases, the effective video content indexing method is required. While manual indexing is the most effective approach to this goal, it is slow and expensive. Thus automatic indexing is desirable and recently various indexing tools for video databases have been developed. For efficient video content indexing, the similarity measure is an important factor. This paper presents new similarity measures between frames and proposes a new algorithm to index video content using Kullback-Leibler distance defined between two histograms. Experimental results show that the proposed algorithm using Kullback-Leibler distance gives remarkable high accuracy ratios compared with several conventional algorithms to index video content.

Luminance Projection Model for Efficient Video Similarity Measure (효율적인 비디오 유사도 측정을 위한 휘도 투영모델)

  • Kim, Sang-Hyun
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.10 no.2
    • /
    • pp.132-135
    • /
    • 2009
  • The video similarity measure is very important factor to index and to retrieve for video data. In this paper, we propose the luminance projection model to measure the video similarity efficiently. Most algorithms for video indexing have been commonly used histograms, edges, or motion features, whereas in this paper, the proposed algorithm is employed an efficient measure using the luminance projection. To index effectively the video sequences and to decrease the computational complexity, we calculate video similarity using the key frames extracted by the cumulative measure, and compare the set of key frames using the modified Hausdorff distance. Experimental results show that the proposed luminance projection model yields the remarkable accuracy and performance than the conventional algorithm.

  • PDF

Moving Objects Modeling for Supporting Content and Similarity Searches (내용 및 유사도 검색을 위한 움직임 객체 모델링)

  • 복경수;김미희;신재룡;유재수;조기형
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.5
    • /
    • pp.617-632
    • /
    • 2004
  • Video Data includes moving objects which change spatial positions as time goes by. In this paper, we propose a new modeling method for a moving object contained in the video data. In order to effectively retrieve moving objects, the proposed modeling method represents the spatial position and the size of a moving object. It also represents the visual features and the trajectory by considering direction, distance and speed or moving objects as time goes by. Therefore, It allows various types of retrieval such as visual feature based similarity retrieval, distance based similarity retrieval and trajectory based similarity retrieval and their mixed type of weighted retrieval.

  • PDF

A Semantic Distance Measurement Model using Weights on the LOD Graph in an LOD-based Recommender System (LOD-기반 추천 시스템에서 LOD 그래프에 가중치를 사용한 의미 거리 측정 모델)

  • Huh, Wonwhoi
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.7
    • /
    • pp.53-60
    • /
    • 2021
  • LOD-based recommender systems usually leverage the data available within LOD datasets, such as DBpedia, in order to recommend items(movies, books, music) to the end users. These systems use a semantic similarity algorithm that calculates the degree of matching between pairs of Linked Data resources. In this paper, we proposed a new approach to measuring semantic distance in an LOD-based recommender system by assigning weights converted from user ratings to links in the LOD graph. The semantic distance measurement model proposed in this paper is based on a processing step in which a graph is personalized to a user through weight calculation and a method of applying these weights to LDSD. The Experimental results showed that the proposed method showed higher accuracy compared to other similar methods, and it contributed to the improvement of similarity by expanding the range of semantic distance measurement of the recommender system. As future work, we aim to analyze the impact on the model using different methods of LOD-based similarity measurement.

Color Similarity Definition Based on Quantized Color Histogram for Clothing Identification

  • Choi, Yoo-Joo;Moon, Nam-Mee
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.396-399
    • /
    • 2009
  • In this paper, we present a method to define a color similarity between color images using Octree-based quantization and similar color integration. The proposed method defines major colors from each image using Octree-based quantization. Two color palettes to consist of major colors are compared based on Euclidean distance and similar color bins between palettes are matched. Multiple matched color bins are integrated and major colors are adjusted. Color histogram based on the color palette is constructed for each image and the difference between two histograms is computed by the weighted Euclidean distance between the matched color bins in consideration of the frequency of each bin. As an experiment to validate the usefulness, we discriminated the same clothing from CCD camera images based on the proposed color similarity analysis. We retrieved the same clothing images with the success rate of 88 % using only color analysis without texture analysis.

  • PDF

Computing Semantic Similarity between ECG-Information Concepts Based on an Entropy-Weighted Concept Lattice

  • Wang, Kai;Yang, Shu
    • Journal of Information Processing Systems
    • /
    • v.16 no.1
    • /
    • pp.184-200
    • /
    • 2020
  • Similarity searching is a basic issue in information processing because of the large size of formal contexts and their complicated derivation operators. Recently, some researchers have focused on knowledge reduction methods by using granular computing. In this process, suitable information granules are vital to characterizing the quantities of attributes and objects. To address this problem, a novel approach to obtain an entropy-weighted concept lattice with inclusion degree and similarity distance (ECLisd) has been proposed. The approach aims to compute the combined weights by merging the inclusion degree and entropy degree between two concepts. In addition, another method is utilized to measure the hierarchical distance by considering the different degrees of importance of each attribute. Finally, the rationality of the ECLisd is validated via a comparative analysis.

An Effective Metric for Measuring the Degree of Web Page Changes (효과적인 웹 문서 변경도 측정 방법)

  • Kwon, Shin-Young;Kim, Sung-Jin;Lee, Sang-Ho
    • Journal of KIISE:Databases
    • /
    • v.34 no.5
    • /
    • pp.437-447
    • /
    • 2007
  • A variety of similarity metrics have been used to measure the degree of web page changes. In this paper, we first define criteria for web page changes to evaluate the effectiveness of the similarity metrics in terms of six important types of web page changes. Second, we propose a new similarity metric appropriate for measuring the degree of web page changes. Using real web pages and synthesized pages, we analyze the five existing metrics (i.e., the byte-wise comparison, the TF IDF cosine distance, the word distance, the edit distance, and the shingling) and ours under the proposed criteria. The analysis result shows that our metric represents the changes more effectively than other metrics. We expect that our study can help users select an appropriate metric for particular web applications.

New Approach of Evaluating Poomsae Performance with Inertial Measurement Unit Sensors (관성센서를 활용한 새로운 품새 경기력 평가 방법 연구)

  • Kim, Young-Kwan
    • Korean Journal of Applied Biomechanics
    • /
    • v.31 no.3
    • /
    • pp.199-204
    • /
    • 2021
  • Objective: The purpose of this study was to present a new idea of methodology to evaluate Poomsae performance using inertial measurement unit (IMU) sensors in terms of signal processing techniques. Method: Ten collegian Taekwondo athletes, consisting of five Poomsae elite athletes (age: 21.4 ± 0.9 years, height: 168.4 ± 11.3 cm, weight: 65.0 ± 10.6 kg, experience: 12 ± 0.7 years) and five breaking demonstration athletes (age: 21.0 ± 0.0 years, height: 168.4 ± 4.7 cm, weight: 63.8 ± 8.2 kg, experience: 13.0 ± 2.1 years), voluntarily participated in this study. They performed three different black belt Poomsae such as Goryeo, Geumgang, and Taebaek Poomsae repeatedly twice. Repeated measured motion data on the wrist and ankle were calculated by the methods of cosine similarity and Euclidean distance. Results: The Poomsse athletes showed superior performance in terms of temporal consistency at Goryeo and Taebaek Poomsae, cosine similarity at Geumgang and Taebaek Poomsae, and Euclidian distance at Geumgang Poomsae. Conclusion: IMU sensor would be a useful tool for monitoring and evaluating within-subject temporal variability of Taekwondo Poomsae motions. As well it distinguished spatiotemporal characteristics among three different Poomsae.

Similarity Analysis of Exports Value Added by Country and Implication for Korea's Global Value Added Chains

  • Cho, Jung-Hwan
    • Journal of Korea Trade
    • /
    • v.23 no.4
    • /
    • pp.103-114
    • /
    • 2019
  • Purpose - This paper investigates the structure of exports across countries in terms of value added. Exports value added is examined under two categories, domestic and overseas. Using a statistical classification method by distance based on these two value added categories, this paper estimates the similarity of exports value added across countries including Korea. Design/methodology - The model of study is to employ a generalized distance function and then derive the Manhattan and Euclidean distances. The paper also performs cluster analysis using the Partitioning Around Medoids (PAM) and hierarchical methods to classify the 44 sample countries considered in this study. Findings - Our main findings are as follows. The 44 countries can be classified under 5 groups by their domestic and overseas value added in exports. Korea has a sandwich global value chains (GVCs) position between Japan, China, and Taiwan in the East Asian region. Originality/value - Existing papers point out the double counting problem of trade statistics as the intermediate goods trade across borders increases. This paper addresses the double counting problem by using the World Input-Output Table. The paper shows the need to explore the similarity of value added in exports structure across countries and investigate the GVCs position and role of each country.

Semantic Process Retrieval with Similarity Algorithms (유사도 알고리즘을 활용한 시맨틱 프로세스 검색방안)

  • Lee, Hong-Joo;Klein, Mark
    • Asia pacific journal of information systems
    • /
    • v.18 no.1
    • /
    • pp.79-96
    • /
    • 2008
  • One of the roles of the Semantic Web services is to execute dynamic intra-organizational services including the integration and interoperation of business processes. Since different organizations design their processes differently, the retrieval of similar semantic business processes is necessary in order to support inter-organizational collaborations. Most approaches for finding services that have certain features and support certain business processes have relied on some type of logical reasoning and exact matching. This paper presents our approach of using imprecise matching for expanding results from an exact matching engine to query the OWL(Web Ontology Language) MIT Process Handbook. MIT Process Handbook is an electronic repository of best-practice business processes. The Handbook is intended to help people: (1) redesigning organizational processes, (2) inventing new processes, and (3) sharing ideas about organizational practices. In order to use the MIT Process Handbook for process retrieval experiments, we had to export it into an OWL-based format. We model the Process Handbook meta-model in OWL and export the processes in the Handbook as instances of the meta-model. Next, we need to find a sizable number of queries and their corresponding correct answers in the Process Handbook. Many previous studies devised artificial dataset composed of randomly generated numbers without real meaning and used subjective ratings for correct answers and similarity values between processes. To generate a semantic-preserving test data set, we create 20 variants for each target process that are syntactically different but semantically equivalent using mutation operators. These variants represent the correct answers of the target process. We devise diverse similarity algorithms based on values of process attributes and structures of business processes. We use simple similarity algorithms for text retrieval such as TF-IDF and Levenshtein edit distance to devise our approaches, and utilize tree edit distance measure because semantic processes are appeared to have a graph structure. Also, we design similarity algorithms considering similarity of process structure such as part process, goal, and exception. Since we can identify relationships between semantic process and its subcomponents, this information can be utilized for calculating similarities between processes. Dice's coefficient and Jaccard similarity measures are utilized to calculate portion of overlaps between processes in diverse ways. We perform retrieval experiments to compare the performance of the devised similarity algorithms. We measure the retrieval performance in terms of precision, recall and F measure? the harmonic mean of precision and recall. The tree edit distance shows the poorest performance in terms of all measures. TF-IDF and the method incorporating TF-IDF measure and Levenshtein edit distance show better performances than other devised methods. These two measures are focused on similarity between name and descriptions of process. In addition, we calculate rank correlation coefficient, Kendall's tau b, between the number of process mutations and ranking of similarity values among the mutation sets. In this experiment, similarity measures based on process structure, such as Dice's, Jaccard, and derivatives of these measures, show greater coefficient than measures based on values of process attributes. However, the Lev-TFIDF-JaccardAll measure considering process structure and attributes' values together shows reasonably better performances in these two experiments. For retrieving semantic process, we can think that it's better to consider diverse aspects of process similarity such as process structure and values of process attributes. We generate semantic process data and its dataset for retrieval experiment from MIT Process Handbook repository. We suggest imprecise query algorithms that expand retrieval results from exact matching engine such as SPARQL, and compare the retrieval performances of the similarity algorithms. For the limitations and future work, we need to perform experiments with other dataset from other domain. And, since there are many similarity values from diverse measures, we may find better ways to identify relevant processes by applying these values simultaneously.