• Title/Summary/Keyword: K-Nearest Neighbor

Search Result 640, Processing Time 0.031 seconds

Stormwater Quality simulation with KNNR Method based on Depth function

  • Lee, Taesam;Park, Daeryong
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2015.05a
    • /
    • pp.557-557
    • /
    • 2015
  • To overcome main drawbacks of parametric models, k-nearest neighbor resampling (KNNR) is suggested for water quality analysis involving geographic information. However, with KNNR nonparametric model, Geographic information is not properly handled. In the current study, to manipulate geographic information properly, we introduce a depth function which is a novel statistical concept in the classical KNNR model for stormwater quality simulation. An application is presented for a case study of the total suspended solids throughout the entire United States. Total suspended solids concentration data of stormwater demonstrated that the proposed model significantly improves the simulation performance rather than the existing KNNR model.

  • PDF

Assembly Performance Evaluation for Prefabricated Steel Structures Using k-nearest Neighbor and Vision Sensor (k-근접 이웃 및 비전센서를 활용한 프리팹 강구조물 조립 성능 평가 기술)

  • Bang, Hyuntae;Yu, Byeongjun;Jeon, Haemin
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.35 no.5
    • /
    • pp.259-266
    • /
    • 2022
  • In this study, we developed a deep learning and vision sensor-based assembly performance evaluation method isfor prefabricated steel structures. The assembly parts were segmented using a modified version of the receptive field block convolution module inspired by the eccentric function of the human visual system. The quality of the assembly was evaluated by detecting the bolt holes in the segmented assembly part and calculating the bolt hole positions. To validate the performance of the evaluation, models of standard and defective assembly parts were produced using a 3D printer. The assembly part segmentation network was trained based on the 3D model images captured from a vision sensor. The sbolt hole positions in the segmented assembly image were calculated using image processing techniques, and the assembly performance evaluation using the k-nearest neighbor algorithm was verified. The experimental results show that the assembly parts were segmented with high precision, and the assembly performance based on the positions of the bolt holes in the detected assembly part was evaluated with a classification error of less than 5%.

Predicting sorptivity and freeze-thaw resistance of self-compacting mortar by using deep learning and k-nearest neighbor

  • Turk, Kazim;Kina, Ceren;Tanyildizi, Harun
    • Computers and Concrete
    • /
    • v.30 no.2
    • /
    • pp.99-111
    • /
    • 2022
  • In this study, deep learning and k-Nearest Neighbor (kNN) models were used to estimate the sorptivity and freeze-thaw resistance of self-compacting mortars (SCMs) having binary and ternary blends of mineral admixtures. Twenty-five environment-friendly SCMs were designed as binary and ternary blends of fly ash (FA) and silica fume (SF) except for control mixture with only Portland cement (PC). The capillary water absorption and freeze-thaw resistance tests were conducted for 91 days. It was found that the use of SF with FA as ternary blends reduced sorptivity coefficient values compared to the use of FA as binary blends while the presence of FA with SF improved freeze-thaw resistance of SCMs with ternary blends. The input variables used the models for the estimation of sorptivity were defined as PC content, SF content, FA content, sand content, HRWRA, water/cementitious materials (W/C) and freeze-thaw cycles. The input variables used the models for the estimation of sorptivity were selected as PC content, SF content, FA content, sand content, HRWRA, W/C and predefined intervals of the sample in water. The deep learning and k-NN models estimated the durability factor of SCM with 94.43% and 92.55% accuracy and the sorptivity of SCM was estimated with 97.87% and 86.14% accuracy, respectively. This study found that deep learning model estimated the sorptivity and durability factor of SCMs having binary and ternary blends of mineral admixtures higher accuracy than k-NN model.

Analysis of Urban Distribution Pattern with Satellite Imagery

  • Roh, Young-Hee;Jeong, Jae-Joon
    • Proceedings of the KSRS Conference
    • /
    • 2007.10a
    • /
    • pp.616-619
    • /
    • 2007
  • Nowadays, urbanized area expands its boundary, and distribution of urbanized area is gradually transformed into more complicated pattern. In Korea, SMA(Seoul Metropolitan Area) has outstanding urbanized area since 1950s. But it is ambiguous whether urban distribution is clustered or dispersed. This study aims to show the way in which expansion of urbanized area impacts on spatial distribution pattern of urbanized area. We use quadrat analysis, nearest-neighbor analysis and fractal analysis to know distribution pattern of urbanized area in time-series urban growth. The quadrat analysis indicates that distribution pattern of urbanized area is clustered but the cohesion is gradually weakened. And the nearest-neighbor analysis shows that point patterns are changed that urbanized area distribution pattern is progressively changed from clustered pattern into dispersed pattern. The fractal dimension analysis shows that 1972's distribution dimension is 1.428 and 2000's dimension is 1.777. Therefore, as time goes by, the complexity of urbanized area is more increased through the years. As a result, we can show that the cohesion of the urbanized area is weakened and complicated.

  • PDF

Countinuous k-Nearest Neighbor Query Processing Algorithm for Distributed Grid Scheme (분산 그리드 기법을 위한 연속 k-최근접 질의처리 알고리즘)

  • Kim, Young-Chang;Chang, Jae-Woo
    • Journal of Korea Spatial Information System Society
    • /
    • v.11 no.3
    • /
    • pp.9-18
    • /
    • 2009
  • Recently, due to the advanced technologies of mobile devices and wireless communication, there are many studies on telematics and LBS(location-based service) applications. because moving objects usually move on spatial networks, their locations are updated frequently, leading to the degradation of retrieval performance. To manage the frequent updates of moving objects' locations in an efficient way, a new distributed grid scheme, called DS-GRID (distributed S-GRID), and k-NN(k-nearest neighbor) query processing algorithm was proposed[1]. However, the result of k-NN query processing technique may be invalidated as the location of query and moving objects are changed. Therefore, it is necessary to study on continuous k-NN query processing algorithm. In this paper, we propose both MCE-CKNN and MBP(Monitoring in Border Point)-CKNN algorithmss are S-GRID. The MCE-CKNN algorithm splits a query route into sub-routes based on cell and seproves retrieval performance by processing query in parallel way by. In addition, the MBP-CKNN algorithm stores POIs from the border points of each grid cells and seproves retrieval performance by decreasing the number of accesses to the adjacent cells. Finally, it is shown from the performance analysis that our CKNN algorithms achieves 15-53% better retrieval performance than the Kolahdouzan's algorithm.

  • PDF

The Effect of Data Size on the k-NN Predictability: Application to Samsung Electronics Stock Market Prediction (데이터 크기에 따른 k-NN의 예측력 연구: 삼성전자주가를 사례로)

  • Chun, Se-Hak
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.239-251
    • /
    • 2019
  • Statistical methods such as moving averages, Kalman filtering, exponential smoothing, regression analysis, and ARIMA (autoregressive integrated moving average) have been used for stock market predictions. However, these statistical methods have not produced superior performances. In recent years, machine learning techniques have been widely used in stock market predictions, including artificial neural network, SVM, and genetic algorithm. In particular, a case-based reasoning method, known as k-nearest neighbor is also widely used for stock price prediction. Case based reasoning retrieves several similar cases from previous cases when a new problem occurs, and combines the class labels of similar cases to create a classification for the new problem. However, case based reasoning has some problems. First, case based reasoning has a tendency to search for a fixed number of neighbors in the observation space and always selects the same number of neighbors rather than the best similar neighbors for the target case. So, case based reasoning may have to take into account more cases even when there are fewer cases applicable depending on the subject. Second, case based reasoning may select neighbors that are far away from the target case. Thus, case based reasoning does not guarantee an optimal pseudo-neighborhood for various target cases, and the predictability can be degraded due to a deviation from the desired similar neighbor. This paper examines how the size of learning data affects stock price predictability through k-nearest neighbor and compares the predictability of k-nearest neighbor with the random walk model according to the size of the learning data and the number of neighbors. In this study, Samsung electronics stock prices were predicted by dividing the learning dataset into two types. For the prediction of next day's closing price, we used four variables: opening value, daily high, daily low, and daily close. In the first experiment, data from January 1, 2000 to December 31, 2017 were used for the learning process. In the second experiment, data from January 1, 2015 to December 31, 2017 were used for the learning process. The test data is from January 1, 2018 to August 31, 2018 for both experiments. We compared the performance of k-NN with the random walk model using the two learning dataset. The mean absolute percentage error (MAPE) was 1.3497 for the random walk model and 1.3570 for the k-NN for the first experiment when the learning data was small. However, the mean absolute percentage error (MAPE) for the random walk model was 1.3497 and the k-NN was 1.2928 for the second experiment when the learning data was large. These results show that the prediction power when more learning data are used is higher than when less learning data are used. Also, this paper shows that k-NN generally produces a better predictive power than random walk model for larger learning datasets and does not when the learning dataset is relatively small. Future studies need to consider macroeconomic variables related to stock price forecasting including opening price, low price, high price, and closing price. Also, to produce better results, it is recommended that the k-nearest neighbor needs to find nearest neighbors using the second step filtering method considering fundamental economic variables as well as a sufficient amount of learning data.

Adaptive Nearest Neighbors for Classification (Adaptive Nearest Neighbors를 활용한 판별분류방법)

  • Jhun, Myoung-Shic;Choi, In-Kyung
    • The Korean Journal of Applied Statistics
    • /
    • v.22 no.3
    • /
    • pp.479-488
    • /
    • 2009
  • The ${\kappa}$-Nearest Neighbors Classification(KNNC) is a popular non-parametric classification method which assigns a fixed number ${\kappa}$ of neighbors to every observation without consideration of the local feature of the each observation. In this paper, we propose an Adaptive Nearest Neighbors Classification(ANNC) as an alternative to KNNC. The proposed ANNC method adapts the number of neighbors according to the local feature of the observation such as density of data. To verify characteristics of ANNC, we compare the number of misclassified observation with KNNC by Monte Carlo study and confirm the potential performance of ANNC method.

Image Feature Point Selection Method Using Nearest Neighbor Distance Ratio Matching (최인접 거리 비율 정합을 이용한 영상 특징점 선택 방법)

  • Lee, Jun-Woo;Jeong, Jea-Hyup;Kang, Jong-Wook;Na, Sang-Il;Jeong, Dong-Seok
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.49 no.12
    • /
    • pp.124-130
    • /
    • 2012
  • In this paper, we propose a feature point selection method for MPEG CDVS CE-7 which is processing on International Standard task. Among a large number of extracted feature points, more important feature points which is used in image matching should be selected for the compactness of image descriptor. The proposed method is that remove the feature point in the extraction phase which is filtered by nearest neighbor distance ratio matching in the matching phase. We can avoid the waste of the feature point and employ additional feature points by the proposed method. The experimental results show that our proposed method can obtain true positive rate improvement about 2.3% in pair-wise matching test compared with Test Model.

Mining Proteins Associated with Oral Squamous Cell Carcinoma in Complex Networks

  • Liu, Ying;Liu, Chuan-Xia;Wu, Zhong-Ting;Ge, Lin;Zhou, Hong-Mei
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.14 no.8
    • /
    • pp.4621-4625
    • /
    • 2013
  • The purpose of this study was to construct a protein-protein interaction (PPI) network related to oral squamous cell carcinoma (OSCC). Each protein was ranked and those most associated with OSCC were mined within the network. First, OSCC-related genes were retrieved from the Online Mendelian Inheritance in Man (OMIM) database. Then they were mapped to their protein identifiers and a seed set of proteins was built. The seed proteins were expanded using the nearest neighbor expansion method to construct a PPI network through the Online Predicated Human Interaction Database (OPHID). The network was verified to be statistically significant, the score of each protein was evaluated by algorithm, then the OSCC-related proteins were ranked. 38 OSCC related seed proteins were expanded to 750 protein pairs. A protein-protein interaction nerwork was then constructed and the 30 top-ranked proteins listed. The four highest-scoring seed proteins were SMAD4, CTNNB1, HRAS, NOTCH1, and four non-seed proteins P53, EP300, SMAD3, SRC were mined using the nearest neighbor expansion method. The methods shown here may facilitate the discovery of important OSCC proteins and guide medical researchers in further pertinent studies.

A Fast Fractal Image Compression Using The Normalized Variance (정규화된 분산을 이용한 프랙탈 압축방법)

  • Kim, Jong-Koo;Hamn, Do-Yong;Wee, Young-Cheul;Kimn, Ha-Jine
    • The KIPS Transactions:PartA
    • /
    • v.8A no.4
    • /
    • pp.499-502
    • /
    • 2001
  • Fractal image coding suffers from the long search time of domain pool although it provides many properties including the high compression ratio. We find that the normalized variance of a block is independent of contrast, brightness. Using this observation, we introduce a self similar block searching method employing the d-dimensional nearest neighbor searching. This method takes Ο(log/N) time for searching the self similar domain blocks for each range block where N is the number of domain blocks. PSNR (Peak Signal Noise Ratio) of this method is similar to that of the full search method that requires Ο(N) time for each range block. Moreover, the image quality of this method is independent of the number of edges in the image.

  • PDF