• Title/Summary/Keyword: Random clustering

Search Result 152, Processing Time 0.022 seconds

A Single Mobile Target Tracking in Voronoi-based Clustered Wireless Sensor Network

  • Chen, Jiehui;Salim, Mariam B.;Matsumoto, Mitsuji
    • Journal of Information Processing Systems
    • /
    • v.7 no.1
    • /
    • pp.17-28
    • /
    • 2011
  • Despite the fact that the deployment of sensor networks and target tracking could both be managed by taking full advantage of Voronoi diagrams, very little few have been made in this regard. In this paper, we designed an optimized barrier coverage and an energy-efficient clustering algorithm for forming Vonoroi-based Wireless Sensor Networks(WSN) in which we proposed a mobile target tracking scheme (CTT&MAV) that takes full advantage of Voronoi-diagram boundary to improve detectability. Simulations verified that CTT&MAV outperforms random walk, random waypoint, random direction and Gauss-Markov in terms of both the average hop distance that the mobile target moved before being detected and lower sensor death rate. Moreover, we demonstrate that our results are robust as realistic sensing models and also validate our observations through extensive simulations.

Magnifying Block Diagonal Structure for Spectral Clustering (스펙트럼 군집화에서 블록 대각 형태의 유사도 행렬 구성)

  • Heo, Gyeong-Yong;Kim, Kwang-Baek;Woo, Young-Woon
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.9
    • /
    • pp.1302-1309
    • /
    • 2008
  • Traditional clustering methods, like k-means or fuzzy clustering, are prototype-based methods which are applicable only to convex clusters. On the other hand, spectral clustering tries to find clusters only using local similarity information. Its ability to handle concave clusters has gained the popularity recent years together with support vector machine (SVM) which is a kernel-based classification method. However, as is in SVM, the kernel width plays an important role and has a great impact on the result. Several methods are proposed to decide it automatically, it is still determined based on heuristics. In this paper, we proposed an adaptive method deciding the kernel width based on distance histogram. The proposed method is motivated by the fact that the affinity matrix should be formed into a block diagonal matrix to generate the best result. We use the tradition Euclidean distance together with the random walk distance, which make it possible to form a more apparent block diagonal affinity matrix. Experimental results show that the proposed method generates more clear block structured affinity matrix than the existing one does.

  • PDF

Wavelet을 이용한 K-means clustering algorithm의 초기화

  • Kim Guk-Hwan;Jang U-Jin;Lee Jun-Seok
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2006.05a
    • /
    • pp.305-312
    • /
    • 2006
  • K-means clustering algorithm 에서 주로 이루어지는 랜덤 초기화 (random initialization) 방법은 전역 최적화된 해(global minimum)를 찾아내기에 문제점을 지니고 있다. 즉, 여러 횟수의 알고리듬 반복(iteration)을 실행하더라도 전역 최적화된 해를 찾아내기가 매우 힘들며 주어진 자료의 크기(data size)가 큰 경우에 있어서 이는 거의 불가능하다. 본 논문은 이러한 문제점들을 극복하기 위한 방안으로, wavelet을 이용하여 최적의 초기 군집 중심점(initial clustering center)들을 선택하는 방법을 제시한다. 즉, 웨이블릿을 이용한 효과적인 초기화 (initialization)를 통해서 작은 알고리듬 반복 횟수만으로도 전역 최적화에 도달하는 초기화 방법을 기술한다. 이런 초기화 방법이 군집 알고리즘에 사용될 경우, 온라인상에서 실시간 이루어지는 군집 분석에 큰 도움이 된 수 있다.

  • PDF

STATISTICAL NOISE BAND REMOVAL FOR SURFACE CLUSTERING OF HYPERSPECTRAL DATA

  • Huan, Nguyen Van;Kim, Hak-Il
    • Proceedings of the KSRS Conference
    • /
    • 2008.10a
    • /
    • pp.111-114
    • /
    • 2008
  • The existence of noise bands may deform the typical shape of the spectrum, making the accuracy of clustering degraded. This paper proposes a statistical approach to remove noise bands in hyperspectral data using the correlation coefficient of bands as an indicator. Considering each band as a random variable, two adjacent signal bands in hyperspectral data are highly correlative. On the contrary, existence of a noise band will produce a low correlation. For clustering, the unsupervised ${\kappa}$-nearest neighbor clustering method is implemented in accordance with three well-accepted spectral matching measures, namely ED, SAM and SID. Furthermore, this paper proposes a hierarchical scheme of combining those measures. Finally, a separability assessment based on the between-class and the within-class scatter matrices is followed to evaluate the applicability of the proposed noise band removal method. Also, the paper brings out a comparison for spectral matching measures.

  • PDF

K-means based Clustering Method with a Fixed Number of Cluster Members

  • Yi, Faliu;Moon, Inkyu
    • Journal of Korea Multimedia Society
    • /
    • v.17 no.10
    • /
    • pp.1160-1170
    • /
    • 2014
  • Clustering methods are very useful in many fields such as data mining, classification, and object recognition. Both the supervised and unsupervised grouping approaches can classify a series of sample data with a predefined or automatically assigned cluster number. However, there is no constraint on the number of elements for each cluster. Numbers of cluster members for each cluster obtained from clustering schemes are usually random. Thus, some clusters possess a large number of elements whereas others only have a few members. In some areas such as logistics management, a fixed number of members are preferred for each cluster or logistic center. Consequently, it is necessary to design a clustering method that can automatically adjust the number of group elements. In this paper, a k-means based clustering method with a fixed number of cluster members is proposed. In the proposed method, first, the data samples are clustered using the k-means algorithm. Then, the number of group elements is adjusted by employing a greedy strategy. Experimental results demonstrate that the proposed clustering scheme can classify data samples efficiently for a fixed number of cluster members.

Fast Outlier Removal for Image Registration based on Modified K-means Clustering

  • Soh, Young-Sung;Qadir, Mudasar;Kim, In-Taek
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.16 no.1
    • /
    • pp.9-14
    • /
    • 2015
  • Outlier detection and removal is a crucial step needed for various image processing applications such as image registration. Random Sample Consensus (RANSAC) is known to be the best algorithm so far for the outlier detection and removal. However RANSAC requires a cosiderable computation time. To drastically reduce the computation time while preserving the comparable quality, a outlier detection and removal method based on modified K-means is proposed. The original K-means was conducted first for matching point pairs and then cluster merging and member exclusion step are performed in the modification step. We applied the methods to various images with highly repetitive patterns under several geometric distortions and obtained successful results. We compared the proposed method with RANSAC and showed that the proposed method runs 3~10 times faster than RANSAC.

An Optimization Approach to Data Clustering

  • Kim, Ju-Mi;Olafsson, Sigurdur
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2005.05a
    • /
    • pp.621-628
    • /
    • 2005
  • Scalability of clustering algorithms is critical issues facing the data mining community. This is particularly true for computationally intense tasks such as data clustering. Random sampling of instances is one possible means of achieving scalability but a pervasive problem with this approach is how to deal with the noise that this introduces in the evaluation of the learning algorithm. This paper develops a new optimization based clustering approach using an algorithms specifically designed for noisy performance. Numerical results illustrate that with this algorithm substantial benefits can be achieved in terms of computational time without sacrificing solution quality.

  • PDF

Development of Random fracture network for discontinuity plane (불연속면의 확률절리망 알고리즘의 개발)

  • Ko, Wang-Kyung
    • Journal of the Korean Data and Information Science Society
    • /
    • v.11 no.2
    • /
    • pp.189-199
    • /
    • 2000
  • A major deficiency of laboratory testing of rock structure is that the structures are limited in size and therefore present a very small and highly selective sample of the rock mass from which were removed. In a typical engineering project, the samples tested in the laboratory represent only a very small fraction of one percent of the volume of the rock mass. In this paper, we calculate the representative orientation of the resultant vector, the measure of the degree of clustering, the volume of rock mass, the trace length of discontinuity spacing under underlying distributions. And we generate the random fracture networks using real data. We propose the calculating the trace length.

  • PDF

On the Clustering Networks using the Kohonen's Elf-Organization Architecture (코호넨의 자기조직화 구조를 이용한 클러스터링 망에 관한 연구)

  • Lee, Ji-Young
    • The Journal of Information Technology
    • /
    • v.8 no.1
    • /
    • pp.119-124
    • /
    • 2005
  • Learning procedure in the neural network is updating of weights between neurons. Unadequate initial learning coefficient causes excessive iterations of learning process or incorrect learning results and degrades learning efficiency. In this paper, adaptive learning algorithm is proposed to increase the efficient in the learning algorithms of Kohonens Self-Organization Neural networks. The algorithm updates the weights adaptively when learning procedure runs. To prove the efficiency the algorithm is experimented to clustering of the random weight. The result shows improved learning rate about 42~55% ; less iteration counts with correct answer.

  • PDF

Volatility clustering in data breach counts

  • Shim, Hyunoo;Kim, Changki;Choi, Yang Ho
    • Communications for Statistical Applications and Methods
    • /
    • v.27 no.4
    • /
    • pp.487-500
    • /
    • 2020
  • Insurers face increasing demands for cyber liability; entailed in part by a variety of new forms of risk of data breaches. As data breach occurrences develop, our understanding of the volatility in data breach counts has also become important as well as its expected occurrences. Volatility clustering, the tendency of large changes in a random variable to cluster together in time, are frequently observed in many financial asset prices, asset returns, and it is questioned whether the volatility of data breach occurrences are also clustered in time. We now present volatility analysis based on INGARCH models, i.e., integer-valued generalized autoregressive conditional heteroskedasticity time series model for frequency counts due to data breaches. Using the INGARCH(1, 1) model with data breach samples, we show evidence of temporal volatility clustering for data breaches. In addition, we present that the firms' volatilities are correlated between some they belong to and that such a clustering effect remains even after excluding the effect of financial covariates such as the VIX and the stock return of S&P500 that have their own volatility clustering.