• Title/Summary/Keyword: nearest neighbor

Search Result 832, Processing Time 0.033 seconds

A Efficient Query Processing of Constrained Nearest Neighbor Search for Moving Query Point (제약을 가진 최소근접을 찾는 이동질의의 효율적인 수행)

  • Ban, Chae-Hoon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2003.11c
    • /
    • pp.1429-1432
    • /
    • 2003
  • This paper addresses the problem of finding a constrained nearest neighbor for moving query point(we call it CNNMP) The Nearest neighbor problem is classified by existence of a constrained region, the number of query result and movement of query point and target. The problem assumes that the query point is not static, as 1-nearest neighbor problem, but varies its position over time to the constrained region. The parameters as NC, NCMBR, CQR and QL for the algorithm are also presented. We suggest the query optimization algorithm in consideration of topological relationship among them

  • PDF

A Distributed Nearest Neighbor Heuristic with Bounding Function (분기 함수를 적용한 분산 최근접 휴리스틱)

  • Kim, Jung-Sook
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.7
    • /
    • pp.377-383
    • /
    • 2002
  • The TSP(Traveling Salesman Problem) has been known as NP-complete, there have been various studies to find the near optimal solution. The nearest neighbor heuristic is more simple than the other algorithms which are to find the optimal solution. This paper designs and implements a new distributed nearest neighbor heuristic with bounding function for the TSP using the master/slave model of PVM(Parallel Virtual Machine). Distributed genetic algorithm obtains a near optimal solution and distributed nearest neighbor heuristic finds an optimal solution for the TSP using the near optimal value obtained by distributed genetic algorithm as the initial bounding value. Especially, we get more speedup using a new genetic operator in the genetic algorithm.

Performance Improvement of Nearest-neighbor Classification Learning through Prototype Selections (프로토타입 선택을 이용한 최근접 분류 학습의 성능 개선)

  • Hwang, Doo-Sung
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.49 no.2
    • /
    • pp.53-60
    • /
    • 2012
  • Nearest-neighbor classification predicts the class of an input data with the most frequent class among the near training data of the input data. Even though nearest-neighbor classification doesn't have a training stage, all of the training data are necessary in a predictive stage and the generalization performance depends on the quality of training data. Therefore, as the training data size increase, a nearest-neighbor classification requires the large amount of memory and the large computation time in prediction. In this paper, we propose a prototype selection algorithm that predicts the class of test data with the new set of prototypes which are near-boundary training data. Based on Tomek links and distance metric, the proposed algorithm selects boundary data and decides whether the selected data is added to the set of prototypes by considering classes and distance relationships. In the experiments, the number of prototypes is much smaller than the size of original training data and we takes advantages of storage reduction and fast prediction in a nearest-neighbor classification.

GLSL based Additional Learning Nearest Neighbor Algorithm suitable for Locating Unpaved Road (추가 학습이 빈번히 필요한 비포장도로에서 주행로 탐색에 적합한 GLSL 기반 ALNN Algorithm)

  • Ku, Bon Woo;Kim, Jun kyum;Rhee, Eun Joo
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.12 no.1
    • /
    • pp.29-36
    • /
    • 2019
  • Unmanned Autonomous Vehicle's driving road in the national defense includes not only paved roads, but also unpaved roads which have rough and unexpected changes. This Unmanned Autonomous Vehicles monitor and recon rugged or remote areas, and defend own position, they frequently encounter environments roads of various and unpredictable. Thus, they need additional learning to drive in this environment, we propose a Additional Learning Nearest Neighbor (ALNN) which is modified from Approximate Nearest Neighbor to allow for quick learning while avoiding the 'Forgetting' problem. In addition, since the Execution speed of the ALNN algorithm decreases as the learning data accumulates, we also propose a solution to this problem using GPU parallel processing based on OpenGL Shader Language. The ALNN based on GPU algorithm can be used in the field of national defense and other similar fields, which require frequent and quick application of additional learning in real-time without affecting the existing learning data.

On the Use of Weighted k-Nearest Neighbors for Missing Value Imputation (Weighted k-Nearest Neighbors를 이용한 결측치 대치)

  • Lim, Chanhui;Kim, Dongjae
    • The Korean Journal of Applied Statistics
    • /
    • v.28 no.1
    • /
    • pp.23-31
    • /
    • 2015
  • A conventional missing value problem in the statistical analysis k-Nearest Neighbor(KNN) method are used for a simple imputation method. When one of the k-nearest neighbors is an extreme value or outlier, the KNN method can create a bias. In this paper, we propose a Weighted k-Nearest Neighbors(WKNN) imputation method that can supplement KNN's faults. A Monte-Carlo simulation study is also adapted to compare the WKNN method and KNN method using real data set.

VLSI design of a FNNPDS encoder for vector quantization (벡터양자화를 위한 FNNPDS 인코더의 VLSI 설계)

  • Kim Hyeung-Cheol;Shim Jeong-Bo;Jo Je-Hwang
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.42 no.2 s.332
    • /
    • pp.83-88
    • /
    • 2005
  • We propose the design method for the VLSI architecture of FNNPDS combined PDS(partial distance search) and FNNS(fast nearest neighbor search), which are used to fast encoding in vector quantization, and obtain the results that FNNPDS(fast nearest neighbor partial distance search) is faster method than the conventional methods by simulation. In simulations, we investigate timing diagrams described searching time of the nearest codevector for an input vector, and compare the average clock cycles per input vector for Lena and Peppers images. According to the result of simulations, the number of the clock cycle of FNNPDS was reduced to $79.2\%\~11.7\%$ as compared with the number using the conventional techniques.

Nearest Neighbor Query Processing in the Mobile Environment

  • Choi Hyun Mi;Jung Young Jin;Lee Eung Jae;Ryu Keun Ho
    • Proceedings of the KSRS Conference
    • /
    • 2004.10a
    • /
    • pp.677-680
    • /
    • 2004
  • In the mobile environment, according to the movement of the object, the query finds the nearest special object or place from object position. However, because query object moves continuously in the mobile environment, query demand changes according to the direction attribute of query object. Also, in the case of moving of query object and simply the minimum distance value of query result, sometimes we find the result against the query object direction. Especially, in most road condition, as user has to return after reaching U-turn area, user rather spends time and cost. Therefore, in order to solve those problems, in this paper we propose the nearest neighbor method considering moving object position and direction for mobile recommendation system.

  • PDF

A KD-Tree-Based Nearest Neighbor Search for Large Quantities of Data

  • Yen, Shwu-Huey;Hsieh, Ya-Ju
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.3
    • /
    • pp.459-470
    • /
    • 2013
  • The discovery of nearest neighbors, without training in advance, has many applications, such as the formation of mosaic images, image matching, image retrieval and image stitching. When the quantity of data is huge and the number of dimensions is high, the efficient identification of a nearest neighbor (NN) is very important. This study proposes a variation of the KD-tree - the arbitrary KD-tree (KDA) - which is constructed without the need to evaluate variances. Multiple KDAs can be constructed efficiently and possess independent tree structures, when the amount of data is large. Upon testing, using extended synthetic databases and real-world SIFT data, this study concludes that the KDA method increases computational efficiency and produces satisfactory accuracy, when solving NN problems.

Flexible Nearest Neighbor Search for Grouping kNN (그룹핑 k-NN을 위한 유연한 최근접 객체 검색)

  • Song, Doohee;Park, Kwangjin
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2015.10a
    • /
    • pp.469-470
    • /
    • 2015
  • 우리는 그룹핑 k-최근접 (Grouping k Nearest Neighbor; GkNN)질의를 지원하기 위하여 유연한 최근접객체(Flexible Nearest Neighbor; FNN)검색 방법을 제안한다. GkNN이란 기존에 제안된 kNN과 다르게 질의자가 요청한 k개의 객체를 모두 확인한 후에 이동 경로의 총합이 가장 작은 k개의 객체를 검색하는 방법이다. 기존 연구에서 제안된 최근접 객체들 (Nearest Neighborhood; NNH) 또한 이 문제를 해결하기 위하여 제안되었다. 그러나 NNH의 문제점은 객체 k와 p가 고정되어 있기 때문에 이동 환경에서 q에서 C까지의 거리가 증가하는 것이다. FNN의 환경은 NNH의 환경과 유사하다. 우리는 NNH의 q에서 집합 C 중 거리 중 가장 짧은 $c_i$ 선택한 후 q에서 $c_i$에 포함된 객체들 모두 검색하는 이동 경로의 총합과 FNN의 이동경로의 총 합을 비교하여 NNH의 문제점을 해결하였다.