• Title/Summary/Keyword: Improved k-Nearest Neighbor

Search Result 47, Processing Time 0.033 seconds

An Improved Preliminary Cut-off Indoor Positioning Scheme in Case of No Neighborhood Reference Point (이웃 참조 위치가 없는 경우를 개선한 실내 위치 추정 사전 컷-오프 방식)

  • Park, Byoungkwan;Kim, Dongjun;Son, Jooyoung;Choi, Jongmin
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.1
    • /
    • pp.74-81
    • /
    • 2017
  • In learning stage of the preliminary Cut-off indoor positioning scheme, RSSI and UUID data received from beacons at each reference point(RP) are stored in fingerprint map. The fingerprint map and real-time beacon information are compared to identify the nearest K reference points through which the user position is estimated. If the number of K is zero, this scheme cannot estimate user position. We have improved the preliminary Cut-off scheme to get the estimated user position even in the case. The improved scheme excludes the beacon of the weakest signal received by user mobile device and identifies neighborhood reference points using the other beacon information. This procedure are performed repetitively until K > 0. The simulation results confirm that the proposed scheme outperforms K-Nearest-Neighbor (KNN), Cluster KNN and the conventional Cut-off scheme in terms of accuracy while the constraints are guaranteed to be satisfied.

Appearance-based Object Recognition Using Higher Order Local Auto Correlation Feature Information (고차 국소 자동 상관 특징 정보를 이용한 외관 기반 객체 인식)

  • Kang, Myung-A
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.7
    • /
    • pp.1439-1446
    • /
    • 2011
  • This paper describes the algorithm that lowers the dimension, maintains the object recognition and significantly reduces the eigenspace configuration time by combining the higher correlation feature information and Principle Component Analysis. Since the suggested method doesn't require a lot of computation than the method using existing geometric information or stereo image, the fact that it is very suitable for building the real-time system has been proved through the experiment. In addition, since the existing point to point method which is a simple distance calculation has many errors, in this paper to improve recognition rate the recognition error could be reduced by using several successive input images as a unit of recognition with K-Nearest Neighbor which is the improved Class to Class method.

HD-Tree: High performance Lock-Free Nearest Neighbor Search KD-Tree (HD-Tree: 고성능 Lock-Free NNS KD-Tree)

  • Lee, Sang-gi;Jung, NaiHoon
    • Journal of Korea Game Society
    • /
    • v.20 no.5
    • /
    • pp.53-64
    • /
    • 2020
  • Supporting NNS method in KD-Tree algorithm is essential in multidimensional data applications. In this paper, we propose HD-Tree, a high-performance Lock-Free KD-Tree that supports NNS in situations where reads and writes occurs concurrently. HD-Tree reduced the number of synchronization nodes used in NNS and requires less atomic operations during Lock-Free method execution. Comparing with existing algorithms, in a multi-core system with 8 core 16 thread, HD-Tree's performance has improved up to 95% on NNS and 15% on modifying in oversubscription situation.

A study on neighbor selection methods in k-NN collaborative filtering recommender system (근접 이웃 선정 협력적 필터링 추천시스템에서 이웃 선정 방법에 관한 연구)

  • Lee, Seok-Jun
    • Journal of the Korean Data and Information Science Society
    • /
    • v.20 no.5
    • /
    • pp.809-818
    • /
    • 2009
  • Collaborative filtering approach predicts the preference of active user about specific items transacted on the e-commerce by using others' preference information. To improve the prediction accuracy through collaborative filtering approach, it must be needed to gain enough preference information of users' for predicting preference. But, a bit much information of users' preference might wrongly affect on prediction accuracy, and also too small information of users' preference might make bad effect on the prediction accuracy. This research suggests the method, which decides suitable numbers of neighbor users for applying collaborative filtering algorithm, improved by existing k nearest neighbors selection methods. The result of this research provides useful methods for improving the prediction accuracy and also refines exploratory data analysis approach for deciding appropriate numbers of nearest neighbors.

  • PDF

An Improved Algorithm of Searching Neighbor Agents in a Large Flocking Behavior (대규모 무리 짓기에서 이웃 에이전트 탐색의 개선된 알고리즘)

  • Lee, Jae-Moon;Jung, In-Hwan
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.5
    • /
    • pp.763-770
    • /
    • 2010
  • This paper proposes an algorithm to enhance the performance of the spatial partitioning method for a flocking behavior. One of the characteristics in a flocking behavior is that two agents may share many common neighbors if they are spatially close to each other. This paper improves the spatial partitioning method by applying this characteristic. While the conventional spatial partitioning method computes the k-nearest neighbors of an agent one by one, the proposed method computes simultaneously the k-nearest neighbors of agents if they are spatially close to each other. The proposed algorithm was implemented and its performance was experimentally compared with the original spatial partitioning method. The results of the comparison showed that the proposed algorithm outperformed the original method by about 33% in average.

Comparison of Forest Growing Stock Estimates by Distance-Weighting and Stratification in k-Nearest Neighbor Technique (거리 가중치와 층화를 이용한 최근린기반 임목축적 추정치의 정확도 비교)

  • Yim, Jong Su;Yoo, Byung Oh;Shin, Man Yong
    • Journal of Korean Society of Forest Science
    • /
    • v.101 no.3
    • /
    • pp.374-380
    • /
    • 2012
  • The k-Nearest Neighbor (kNN) technique is popularly applied to assess forest resources at the county level and to provide its spatial information by combining large area forest inventory data and remote sensing data. In this study, two approaches such as distance-weighting and stratification of training dataset, were compared to improve kNN-based forest growing stock estimates. When compared with five distance weights (0 to 2 by 0.5), the accuracy of kNN-based estimates was very similar ranged ${\pm}0.6m^3/ha$ in mean deviation. The training dataset were stratified by horizontal reference area (HRA) and forest cover type, which were applied by separately and combined. Even though the accuracy of estimates by combining forest cover type and HRA- 100 km was slightly improved, that by forest cover type was more efficient with sufficient number of training data. The mean of forest growing stock based kNN with HRA-100 and stratification by forest cover type when k=7 were somewhat underestimated ($5m^3/ha$) compared to statistical yearbook of forestry at 2011.

Design of an Efficient Parallel High-Dimensional Index Structure (효율적인 병렬 고차원 색인구조 설계)

  • Park, Chun-Seo;Song, Seok-Il;Sin, Jae-Ryong;Yu, Jae-Su
    • Journal of KIISE:Databases
    • /
    • v.29 no.1
    • /
    • pp.58-71
    • /
    • 2002
  • Generally, multi-dimensional data such as image and spatial data require large amount of storage space. There is a limit to store and manage those large amount of data in single workstation. If we manage the data on parallel computing environment which is being actively researched these days, we can get highly improved performance. In this paper, we propose a parallel high-dimensional index structure that exploits the parallelism of the parallel computing environment. The proposed index structure is nP(processor)-n$\times$mD(disk) architecture which is the hybrid type of nP-nD and lP-nD. Its node structure increases fan-out and reduces the height of a index tree. Also, A range search algorithm that maximizes I/O parallelism is devised, and it is applied to K-nearest neighbor queries. Through various experiments, it is shown that the proposed method outperforms other parallel index structures.

K Nearest Neighbor Joins for Big Data Processing based on Spark (Spark 기반 빅데이터 처리를 위한 K-최근접 이웃 연결)

  • JIAQI, JI;Chung, Yeongjee
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.9
    • /
    • pp.1731-1737
    • /
    • 2017
  • K Nearest Neighbor Join (KNN Join) is a simple yet effective method in machine learning. It is widely used in small dataset of the past time. As the number of data increases, it is infeasible to run this model on an actual application by a single machine due to memory and time restrictions. Nowadays a popular batch process model called MapReduce which can run on a cluster with a large number of computers is widely used for large-scale data processing. Hadoop is a framework to implement MapReduce, but its performance can be further improved by a new framework named Spark. In the present study, we will provide a KNN Join implement based on Spark. With the advantage of its in-memory calculation capability, it will be faster and more effective than Hadoop. In our experiments, we study the influence of different factors on running time and demonstrate robustness and efficiency of our approach.

Relation Based Bayesian Network for NBNN

  • Sun, Mingyang;Lee, YoonSeok;Yoon, Sung-eui
    • Journal of Computing Science and Engineering
    • /
    • v.9 no.4
    • /
    • pp.204-213
    • /
    • 2015
  • Under the conditional independence assumption among local features, the Naive Bayes Nearest Neighbor (NBNN) classifier has been recently proposed and performs classification without any training or quantization phases. While the original NBNN shows high classification accuracy without adopting an explicit training phase, the conditional independence among local features is against the compositionality of objects indicating that different, but related parts of an object appear together. As a result, the assumption of the conditional independence weakens the accuracy of classification techniques based on NBNN. In this work, we look into this issue, and propose a novel Bayesian network for an NBNN based classification to consider the conditional dependence among features. To achieve our goal, we extract a high-level feature and its corresponding, multiple low-level features for each image patch. We then represent them based on a simple, two-level layered Bayesian network, and design its classification function considering our Bayesian network. To achieve low memory requirement and fast query-time performance, we further optimize our representation and classification function, named relation-based Bayesian network, by considering and representing the relationship between a high-level feature and its low-level features into a compact relation vector, whose dimensionality is the same as the number of low-level features, e.g., four elements in our tests. We have demonstrated the benefits of our method over the original NBNN and its recent improvement, and local NBNN in two different benchmarks. Our method shows improved accuracy, up to 27% against the tested methods. This high accuracy is mainly due to consideration of the conditional dependences between high-level and its corresponding low-level features.

Supervised learning and frequency domain averaging-based adaptive channel estimation scheme for filterbank multicarrier with offset quadrature amplitude modulation

  • Singh, Vibhutesh Kumar;Upadhyay, Nidhi;Flanagan, Mark;Cardiff, Barry
    • ETRI Journal
    • /
    • v.43 no.6
    • /
    • pp.966-977
    • /
    • 2021
  • Filterbank multicarrier with offset quadrature amplitude modulation (FBMC-OQAM) is an attractive alternative to the orthogonal frequency division multiplexing (OFDM) modulation technique. In comparison with OFDM, the FBMC-OQAM signal has better spectral confinement and higher spectral efficiency and tolerance to synchronization errors, primarily due to per-subcarrier filtering using a frequency-time localized prototype filter. However, the filtering process introduces intrinsic interference among the symbols and complicates channel estimation (CE). An efficient way to improve the CE in FBMC-OQAM is using a technique known as windowed frequency domain averaging (FDA); however, it requires a priori knowledge of the window length parameter which is set based on the channel's frequency selectivity (FS). As the channel's FS is not fixed and not a priori known, we propose a k-nearest neighbor-based machine learning algorithm to classify the FS and decide on the FDA's window length. A comparative theoretical analysis of the mean-squared error (MSE) is performed to prove the proposed CE scheme's effectiveness, validated through extensive simulations. The adaptive CE scheme is shown to yield a reduction in CE-MSE and improved bit error rates compared with the popular preamble-based CE schemes for FBMC-OQAM, without a priori knowledge of channel's frequency selectivity.