• Title/Summary/Keyword: Nearest Neighbor Algorithm

Search Result 334, Processing Time 0.021 seconds

GLSL based Additional Learning Nearest Neighbor Algorithm suitable for Locating Unpaved Road (추가 학습이 빈번히 필요한 비포장도로에서 주행로 탐색에 적합한 GLSL 기반 ALNN Algorithm)

  • Ku, Bon Woo;Kim, Jun kyum;Rhee, Eun Joo
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.12 no.1
    • /
    • pp.29-36
    • /
    • 2019
  • Unmanned Autonomous Vehicle's driving road in the national defense includes not only paved roads, but also unpaved roads which have rough and unexpected changes. This Unmanned Autonomous Vehicles monitor and recon rugged or remote areas, and defend own position, they frequently encounter environments roads of various and unpredictable. Thus, they need additional learning to drive in this environment, we propose a Additional Learning Nearest Neighbor (ALNN) which is modified from Approximate Nearest Neighbor to allow for quick learning while avoiding the 'Forgetting' problem. In addition, since the Execution speed of the ALNN algorithm decreases as the learning data accumulates, we also propose a solution to this problem using GPU parallel processing based on OpenGL Shader Language. The ALNN based on GPU algorithm can be used in the field of national defense and other similar fields, which require frequent and quick application of additional learning in real-time without affecting the existing learning data.

A Efficient Query Processing of Constrained Nearest Neighbor Search for Moving Query Point (제약을 가진 최소근접을 찾는 이동질의의 효율적인 수행)

  • Ban, Chae-Hoon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2003.11c
    • /
    • pp.1429-1432
    • /
    • 2003
  • This paper addresses the problem of finding a constrained nearest neighbor for moving query point(we call it CNNMP) The Nearest neighbor problem is classified by existence of a constrained region, the number of query result and movement of query point and target. The problem assumes that the query point is not static, as 1-nearest neighbor problem, but varies its position over time to the constrained region. The parameters as NC, NCMBR, CQR and QL for the algorithm are also presented. We suggest the query optimization algorithm in consideration of topological relationship among them

  • PDF

k-Nearest Neighbor Classifier using Local Values of k (지역적 k값을 사용한 k-Nearest Neighbor Classifier)

  • 이상훈;오경환
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2003.10a
    • /
    • pp.193-195
    • /
    • 2003
  • 본 논문에서는 k-Nearest Neighbor(k-NN) 알고리즘을 최적화하기 위해 지역적으로 다른 k(고려할 neighbor의 개수)를 사용하는 새로운 방법을 제안한다. 인스턴스 공간(instance space)에서 노이즈(noise)의 분포가 지역적(local)으로 다를 경우, 각 지점에서 고려해야 할 최적의 이웃 인스턴스(neighbor)의 수는 해당 지점에서의 국부적인 노이즈 분포에 따라 다르다. 그러나 기존의 방법은 전체 인스턴스 공간에 대해 동일한 k를 사용하기 때문에 이러한 인스턴스 공간의 지역적인 특성을 고려하지 못한다. 따라서 본 논문에서는 지역적으로 분포가 다른 노이즈 문제를 해결하기 위해 인스턴스 공간을 여러 개의 부분으로 나누고, 각 부분에 최적화된 k의 값을 사용하여 kNN을 수행하는 새로운 방법인 Local-k Nearest Neighbor 알고리즘(LkNN Algorithm)을 제안한다. LkNN을 통해 생성된 k의 집합은 인스턴스 공간의 각 부분을 대표하는 값으로, 해당 지역의 인스턴스가 고려해야 할 이웃(neighbor)의 수를 결정지어준다. 제안한 알고리즘에 적합한 데이터의 도메인(domain)과 그것의 향상된 성능은 UCI ML Data Repository 데이터를 사용한 실험을 통해 검증하였다.

  • PDF

Fuzzy K-Nearest Neighbor Algorithm based on Kernel Method (커널 기반의 퍼지 K-Nearest Neighbor 알고리즘)

  • Choi Byung-In;Rhee Frank Chung-Hoon
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2005.11a
    • /
    • pp.267-270
    • /
    • 2005
  • 커널 함수는 데이터를 high dimension 상의 속성 공간으로 mapping함으로써 복잡한 분포를 가지는 데이터에 대하여 기존의 선형 분류 알고리즘들의 성능을 향상시킬 수 있다. 본 논문에서는 기존의 유클리디안 거리측정방법 대신에 커널 함수에 의한 속성 공간의 거리측정방법을 fuzzy K-nearest neighbor 알고리즘에 적용한 fuzzy kernel K-nearest neighbor(FKKNN) 알고리즘을 제안한다. 제시한 알고리즘은 데이터에 대한 적절한 커널 함수의 선택으로 기존 알고리즘의 성능을 향상 시킬 수 있다. 제시한 알고리즘의 타당성을 보이기 위하여 여러 데이터 집합에 대한 실험결과를 분석한다.

  • PDF

The Processing Method for a Reverse Nearest Neighbor Queries in a Search Space with the Presence of Obstacles (장애물이 존재하는 검색공간에서 역최대근접질의 처리방법에 관한 연구)

  • Seon, Hwi Joon;Kim, Hong Ki
    • Convergence Security Journal
    • /
    • v.17 no.2
    • /
    • pp.81-88
    • /
    • 2017
  • It is occurred frequently the reverse nearest neighbor queries to find objects where a query point can be the nearest neighbor object in recently applications like the encrypted spatial database. In a search space of the real world, however, there are many physical obstacles(e.g., rivers, lakes, highways, etc.). It is necessary the accurate measurement of distances considered the obstacles to increase the retrieval performance such as this circumstance. In this study, we present the algorithm and the measurement of distance to optimize the processing performance of reverse nearest neighbor queries in a search space with the presence of obstacles.

The Method to Process Nearest Neighbor Queries Using an Optimal Search Distance (최적탐색거리를 이용한 최근접질의의 처리 방법)

  • Seon, Hwi-Joon;Hwang, Bu-Hyun;Ryu, Keun-Ho
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.9
    • /
    • pp.2173-2184
    • /
    • 1997
  • Among spatial queries handled in spatial database systems, nearest neighbor queries to find the nearest spatial object from the given locaion occur frequently. The number of searched nodes in an index must be minimized in order to increase the performance of nearest neighbor queries. An Existing approach considered only the processing of an nearest neighbor query in a two-dimensional search space and could not optimize the number of searched nodes accurately. In this paper, we propose the optimal search distance and prove its properties. The proposed optimal search distance is the measurement of a new search distance for accurately selecting the nodes which will be searched in processing nearest neighbor queries. We present an algorithm for processing the nearest neighbor query by applying the optimal search distance to R-trees and prove that the result of query processing is correcter than the existing approach.

  • PDF

Range and k-Nearest Neighbor Query Processing Algorithms using Materialization Techniques in Spatial Network Databases (공간 네트워크 데이터베이스에서 실체화 기법을 이용한 범위 및 k-최근접 질의처리 알고리즘)

  • Kim, Yong-Ki;Chowdhury, Nihad Karim;Lee, Hyun-Jo;Chang, Jae-Woo
    • Journal of Korea Spatial Information System Society
    • /
    • v.9 no.2
    • /
    • pp.67-79
    • /
    • 2007
  • Recently, to support LBS(location-based services) and telematics applications efficiently, there have been many researches which consider the spatial network instead of Euclidean space. However, existing range query and k-nearest neighbor query algorithms show a linear decrease in performance as the value of radius and k is increased. In this paper, to increase the performance of query processing algorithm, we propose materialization-based range and k-nearest neighbor algorithms. In addition, we make the performance comparison to show the proposed algorithm achieves better retrieval performance than the existing algorithm.

  • PDF

Calculating Attribute Weights in K-Nearest Neighbor Algorithms using Information Theory (정보이론을 이용한 K-최근접 이웃 알고리즘에서의 속성 가중치 계산)

  • Lee Chang-Hwan
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.9
    • /
    • pp.920-926
    • /
    • 2005
  • Nearest neighbor algorithms classify an unseen input instance by selecting similar cases and use the discovered membership to make predictions about the unknown features of the input instance. The usefulness of the nearest neighbor algorithms have been demonstrated sufficiently in many real-world domains. In nearest neighbor algorithms, it is an important issue to assign proper weights to the attributes. Therefore, in this paper, we propose a new method which can automatically assigns to each attribute a weight of its importance with respect to the target attribute. The method has been implemented as a computer program and its effectiveness has been tested on a number of machine learning databases publicly available.

An Approach of Dimension Reduction in k-Nearest Neighbor Based Short-term Load Forecasting

  • Chu, FaZheng;Jung, Sung-Hwan
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.9
    • /
    • pp.1567-1573
    • /
    • 2017
  • The k-nearest neighbor (k-NN) algorithm is one of the most widely used benchmark algorithm in classification. Nowadays it has been further applied to predict time series. However, one of the main concerns of the algorithm applied on short-term electricity load forecasting is high computational burden. In the paper, we propose an approach of dimension reduction that follows the principles of highlighting the temperature effect on electricity load data series. The results show the proposed approach is able to reduce the dimension of the data around 30%. Moreover, with temperature effect highlighting, the approach will contribute to finding similar days accurately, and then raise forecasting accuracy slightly.

Performance Improvement of Nearest-neighbor Classification Learning through Prototype Selections (프로토타입 선택을 이용한 최근접 분류 학습의 성능 개선)

  • Hwang, Doo-Sung
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.49 no.2
    • /
    • pp.53-60
    • /
    • 2012
  • Nearest-neighbor classification predicts the class of an input data with the most frequent class among the near training data of the input data. Even though nearest-neighbor classification doesn't have a training stage, all of the training data are necessary in a predictive stage and the generalization performance depends on the quality of training data. Therefore, as the training data size increase, a nearest-neighbor classification requires the large amount of memory and the large computation time in prediction. In this paper, we propose a prototype selection algorithm that predicts the class of test data with the new set of prototypes which are near-boundary training data. Based on Tomek links and distance metric, the proposed algorithm selects boundary data and decides whether the selected data is added to the set of prototypes by considering classes and distance relationships. In the experiments, the number of prototypes is much smaller than the size of original training data and we takes advantages of storage reduction and fast prediction in a nearest-neighbor classification.