• Title/Summary/Keyword: similarity metric

Search Result 111, Processing Time 0.024 seconds

Collaborative Similarity Metric Learning for Semantic Image Annotation and Retrieval

  • Wang, Bin;Liu, Yuncai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.5
    • /
    • pp.1252-1271
    • /
    • 2013
  • Automatic image annotation has become an increasingly important research topic owing to its key role in image retrieval. Simultaneously, it is highly challenging when facing to large-scale dataset with large variance. Practical approaches generally rely on similarity measures defined over images and multi-label prediction methods. More specifically, those approaches usually 1) leverage similarity measures predefined or learned by optimizing for ranking or annotation, which might be not adaptive enough to datasets; and 2) predict labels separately without taking the correlation of labels into account. In this paper, we propose a method for image annotation through collaborative similarity metric learning from dataset and modeling the label correlation of the dataset. The similarity metric is learned by simultaneously optimizing the 1) image ranking using structural SVM (SSVM), and 2) image annotation using correlated label propagation, with respect to the similarity metric. The learned similarity metric, fully exploiting the available information of datasets, would improve the two collaborative components, ranking and annotation, and sequentially the retrieval system itself. We evaluated the proposed method on Corel5k, Corel30k and EspGame databases. The results for annotation and retrieval show the competitive performance of the proposed method.

A Novel Similarity Measure for Sequence Data

  • Pandi, Mohammad. H.;Kashefi, Omid;Minaei, Behrouz
    • Journal of Information Processing Systems
    • /
    • v.7 no.3
    • /
    • pp.413-424
    • /
    • 2011
  • A variety of different metrics has been introduced to measure the similarity of two given sequences. These widely used metrics are ranging from spell correctors and categorizers to new sequence mining applications. Different metrics consider different aspects of sequences, but the essence of any sequence is extracted from the ordering of its elements. In this paper, we propose a novel sequence similarity measure that is based on all ordered pairs of one sequence and where a Hasse diagram is built in the other sequence. In contrast with existing approaches, the idea behind the proposed sequence similarity metric is to extract all ordering features to capture sequence properties. We designed a clustering problem to evaluate our sequence similarity metric. Experimental results showed the superiority of our proposed sequence similarity metric in maximizing the purity of clustering compared to metrics such as d2, Smith-Waterman, Levenshtein, and Needleman-Wunsch. The limitation of those methods originates from some neglected sequence features, which are considered in our proposed sequence similarity metric.

Image Denoising via Fast and Fuzzy Non-local Means Algorithm

  • Lv, Junrui;Luo, Xuegang
    • Journal of Information Processing Systems
    • /
    • v.15 no.5
    • /
    • pp.1108-1118
    • /
    • 2019
  • Non-local means (NLM) algorithm is an effective and successful denoising method, but it is computationally heavy. To deal with this obstacle, we propose a novel NLM algorithm with fuzzy metric (FM-NLM) for image denoising in this paper. A new feature metric of visual features with fuzzy metric is utilized to measure the similarity between image pixels in the presence of Gaussian noise. Similarity measures of luminance and structure information are calculated using a fuzzy metric. A smooth kernel is constructed with the proposed fuzzy metric instead of the Gaussian weighted L2 norm kernel. The fuzzy metric and smooth kernel computationally simplify the NLM algorithm and avoid the filter parameters. Meanwhile, the proposed FM-NLM using visual structure preferably preserves the original undistorted image structures. The performance of the improved method is visually and quantitatively comparable with or better than that of the current state-of-the-art NLM-based denoising algorithms.

Scanline Based Metric for Evaluating the Accuracy of Automatic Fracture Survey Methods (자동 균열 조사기법의 정확도 평가를 위한 조사선 기반의 지표 제안)

  • Kim, Jineon;Song, Jae-Joon
    • Tunnel and Underground Space
    • /
    • v.29 no.4
    • /
    • pp.230-242
    • /
    • 2019
  • While various automatic rock fracture survey methods have been researched, the evaluation of the accuracy of these methods raises issues due to the absence of a metric which fully expresses the similarity between automatic and manual fracture maps. Therefore, this paper proposes a geometry similarity metric which is especially designed to determine the overall similarity of fracture maps and to evaluate the accuracy of rock fracture survey methods by a single number. The proposed metric, Scanline Intersection Similarity (SIS), is derived by conducting a large number of scanline surveys upon two fracture maps using Python code. By comparing the frequency of intersections over a large number of scanlines, SIS is able to express the overall similarity between two fracture maps. The proposed metric was compared with Intersection Over Union (IoU) which is a widely used evaluation metric in computer vision. Results showed that IoU is inappropriate for evaluating the geometry similarity of fracture maps because it is overly sensitive to minor geometry differences of thin elongated objects. The proposed metric, on the other hand, reflected macro-geometry differences rather than micro-geometry differences, showing good agreement with human perception. The metric was further applied to evaluate the accuracy of a deep learning-based automatic fracture surveying method which resulted as 0.674 (SIS). However, the proposed metric is currently limited to 2D fracture maps and requires comparison with rock joint parameters such as RQD.

A Modified Domain Deformation Theory for Signal Classification (함수의 정의역 변형에 의한 신호간의 거리 측정 방법)

  • Kim, Sung-Soo
    • The Transactions of the Korean Institute of Electrical Engineers A
    • /
    • v.48 no.3
    • /
    • pp.342-349
    • /
    • 1999
  • The metric defined on the domain deformation space better measures the similarity between bounded and continuous signals for the purpose of classification via the metric distances between signals. In this paper, a modified domain deformation theory is introduced for one-dimensional signal classification. A new metric defined on a modified domain deformation for measuring the distance between signals is employed. By introducing a newly defined metric space via the newly defined Integra-Normalizer, the assumption that domain deformation is applicable only to continuous signals is removed such that any kind of integrable signal can be classified. The metric on the modified domain deformation has an advantage over the $L^2$ metric as well as the previously introduced domain deformation does.

  • PDF

WHEN ALL PERMUTATIONS ARE COMBINATORIAL SIMILARITIES

  • Viktoriia Bilet;Oleksiy Dovgoshey
    • Bulletin of the Korean Mathematical Society
    • /
    • v.60 no.3
    • /
    • pp.733-746
    • /
    • 2023
  • Let (X, d) be a semimetric space. A permutation Φ of the set X is a combinatorial self similarity of (X, d) if there is a bijective function f : d(X × X) → d(X × X) such that d(x, y) = f(d(Φ(x), Φ(y))) for all x, y ∈ X. We describe the set of all semimetrics ρ on an arbitrary nonempty set Y for which every permutation of Y is a combinatorial self similarity of (Y, ρ).

Semantic Word Categorization using Feature Similarity based K Nearest Neighbor

  • Jo, Taeho
    • Journal of Multimedia Information System
    • /
    • v.5 no.2
    • /
    • pp.67-78
    • /
    • 2018
  • This article proposes the modified KNN (K Nearest Neighbor) algorithm which considers the feature similarity and is applied to the word categorization. The texts which are given as features for encoding words into numerical vectors are semantic related entities, rather than independent ones, and the synergy effect between the word categorization and the text categorization is expected by combining both of them with each other. In this research, we define the similarity metric between two vectors, including the feature similarity, modify the KNN algorithm by replacing the exiting similarity metric by the proposed one, and apply it to the word categorization. The proposed KNN is empirically validated as the better approach in categorizing words in news articles and opinions. The significance of this research is to improve the classification performance by utilizing the feature similarities.

Image Quality Assessment Considering both Computing Speed and Robustness to Distortions (계산 속도와 왜곡 강인성을 동시 고려한 이미지 품질 평가)

  • Kim, Suk-Won;Hong, Seongwoo;Jin, Jeong-Chan;Kim, Young-Jin
    • Journal of KIISE
    • /
    • v.44 no.9
    • /
    • pp.992-1004
    • /
    • 2017
  • To assess image quality accurately, an image quality assessment (IQA) metric is required to reflect the human visual system (HVS) properly. In other words, the structure, color, and contrast ratio of the image should be evaluated in consideration of various factors. In addition, as mobile embedded devices such as smartphone become popular, a fast computing speed is important. In this paper, the proposed IQA metric combines color similarity, gradient similarity, and phase similarity synergistically to satisfy the HVS and is designed by using optimized pooling and quantization for fast computation. The proposed IQA metric is compared against existing 13 methods using 4 kinds of evaluation methods. The experimental results show that the proposed IQA metric ranks the first on 3 evaluation methods and the first on the remaining method, next to VSI which is the most remarkable IQA metric. Its computing speed is on average about 20% faster than VSI's. In addition, we find that the proposed IQA metric has a bigger amount of correlation with the HVS than existing IQA metrics.

Metric Defined by Wavelets and Integra-Normalizer (웨이브렛과 인테그라-노말라이저를 이용한 메트릭)

  • Kim, Sung-Soo;Park, Byoung-Seob
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.50 no.7
    • /
    • pp.350-353
    • /
    • 2001
  • In general, the Least Square Error method is used for signal classification to measure distance in the $l^2$ metric or the $L^2$ metric space. A defect of the Least Square Error method is that it does not classify properly some waveforms, which is due to the property of the Least Square Error method: the global analysis. This paper proposes a new linear operator, the Integra-Normalizer, that removes the problem. The Integra-Normalizer possesses excellent property that measures the degree of relative similarity between signals by expanding the functional space with removing the restriction on the functional space inherited by the Least Square Error method. The Integra-Normalizer shows superiority to the Least Square Error method in measuring the relative similarity among one dimensional waveforms.

  • PDF

An Effective Metric for Measuring the Degree of Web Page Changes (효과적인 웹 문서 변경도 측정 방법)

  • Kwon, Shin-Young;Kim, Sung-Jin;Lee, Sang-Ho
    • Journal of KIISE:Databases
    • /
    • v.34 no.5
    • /
    • pp.437-447
    • /
    • 2007
  • A variety of similarity metrics have been used to measure the degree of web page changes. In this paper, we first define criteria for web page changes to evaluate the effectiveness of the similarity metrics in terms of six important types of web page changes. Second, we propose a new similarity metric appropriate for measuring the degree of web page changes. Using real web pages and synthesized pages, we analyze the five existing metrics (i.e., the byte-wise comparison, the TF IDF cosine distance, the word distance, the edit distance, and the shingling) and ours under the proposed criteria. The analysis result shows that our metric represents the changes more effectively than other metrics. We expect that our study can help users select an appropriate metric for particular web applications.