• Title/Summary/Keyword: 최근접 이웃

Search Result 187, Processing Time 0.019 seconds

The Road condition-based Braking Strength Calculation System for a fully autonomous driving vehicle (완전 자율주행을 위한 도로 상태 기반 제동 강도 계산 시스템)

  • Son, Su-Rak;Jeong, Yi-Na
    • Journal of Internet Computing and Services
    • /
    • v.23 no.2
    • /
    • pp.53-59
    • /
    • 2022
  • After the 3rd level autonomous driving vehicle, the 4th and 5th level of autonomous driving technology is trying to maintain the optimal condition of the passengers as well as the perfect driving of the vehicle. However current autonomous driving technology is too dependent on visual information such as LiDAR and front camera, so it is difficult to fully autonomously drive on roads other than designated roads. Therefore this paper proposes a Braking Strength Calculation System (BSCS), in which a vehicle classifies road conditions using data other than visual information and calculates optimal braking strength according to road conditions and driving conditions. The BSCS consists of RCDM (Road Condition Definition Module), which classifies road conditions based on KNN algorithm, and BSCM (Braking Strength Calculation Module), which calculates optimal braking strength while driving based on current driving conditions and road conditions. As a result of the experiment in this paper, it was possible to find the most suitable number of Ks for the KNN algorithm, and it was proved that the RCDM proposed in this paper is more accurate than the unsupervised K-means algorithm. By using not only visual information but also vibration data applied to the suspension, the BSCS of the paper can make the braking of autonomous vehicles smoother in various environments where visual information is limited.

Implementation of DTW-kNN-based Decision Support System for Discriminating Emerging Technologies (DTW-kNN 기반의 유망 기술 식별을 위한 의사결정 지원 시스템 구현 방안)

  • Jeong, Do-Heon;Park, Ju-Yeon
    • Journal of Industrial Convergence
    • /
    • v.20 no.8
    • /
    • pp.77-84
    • /
    • 2022
  • This study aims to present a method for implementing a decision support system that can be used for selecting emerging technologies by applying a machine learning-based automatic classification technique. To conduct the research, the architecture of the entire system was built and detailed research steps were conducted. First, emerging technology candidate items were selected and trend data was automatically generated using a big data system. After defining the conceptual model and pattern classification structure of technological development, an efficient machine learning method was presented through an automatic classification experiment. Finally, the analysis results of the system were interpreted and methods for utilization were derived. In a DTW-kNN-based classification experiment that combines the Dynamic Time Warping(DTW) method and the k-Nearest Neighbors(kNN) classification model proposed in this study, the identification performance was up to 87.7%, and particularly in the 'eventual' section where the trend highly fluctuates, the maximum performance difference was 39.4% points compared to the Euclidean Distance(ED) algorithm. In addition, through the analysis results presented by the system, it was confirmed that this decision support system can be effectively utilized in the process of automatically classifying and filtering by type with a large amount of trend data.

Welding Bead Detection Inspection Using the Brightness Value of Vertical and Horizontal Direction (수직 및 수평 방향의 밝깃값을 이용한 용접 비드 검출 검사)

  • Jae Eun Lee;Jong-Nam Kim
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.23 no.4
    • /
    • pp.241-248
    • /
    • 2022
  • Shear Reinforcement of Dual Anchorage(SRD) is used to reinforce the safety of reinforced concrete structures at construction sites. Welding is used to make shear reinforcement, and welding plays an important role in determining productivity and competitiveness of products. Therefore, a weld bead detection inspection is required. In this paper, we suggest an algorithm for inspecting welding beads using image data of welding beads. First, the proposed algorithm calculates a brightness value in a vertical direction in an image, and then divides a welding bead in a vertical direction by finding a position corresponding to a 50% height point of the brightness value distribution in the image. The welding bead area is also divided in the same way for the horizontal direction, and then the segmentation image is analyzed if there is a welding bead. The proposed algorithm reduced the amount of computation by performing analysis after specifying the region of interest. In addition, accuracy could be improved by using all brightness values in the vertical and horizontal directions using the difference of brightness between the base metal and the welding bead region in the SRD image. The experiment compared the analysis results using five algorithms, such as K-mean and K-neighborhood, as a method to detect if there is a welding bead, and the experimental result proved that the proposed algorithm was the most accurate.

Comparison of Effective Soil Depth Classification Methods Using Topographic Information (지형정보를 이용한 유효토심 분류방법비교)

  • Byung-Soo Kim;Ju-Sung Choi;Ja-Kyung Lee;Na-Young Jung;Tae-Hyung Kim
    • Journal of the Korean Geosynthetics Society
    • /
    • v.22 no.2
    • /
    • pp.1-12
    • /
    • 2023
  • Research on the causes of landslides and prediction of vulnerable areas is being conducted globally. This study aims to predict the effective soil depth, a critical element in analyzing and forecasting landslide disasters, using topographic information. Topographic data from various institutions were collected and assigned as attribute information to a 100 m × 100 m grid, which was then reduced through data grading. The study predicted effective soil depth for two cases: three depths (shallow, normal, deep) and five depths (very shallow, shallow, normal, deep, very deep). Three classification models, including K-Nearest Neighbor, Random Forest, and Deep Artificial Neural Network, were used, and their performance was evaluated by calculating accuracy, precision, recall, and F1-score. Results showed that the performance was in the high 50% to early 70% range, with the accuracy of the three classification criteria being about 5% higher than the five criteria. Although the grading criteria and classification model's performance presented in this study are still insufficient, the application of the classification model is possible in predicting the effective soil depth. This study suggests the possibility of predicting more reliable values than the current effective soil depth, which assumes a large area uniformly.

Performance Comparison of Automatic Classification Using Word Embeddings of Book Titles (단행본 서명의 단어 임베딩에 따른 자동분류의 성능 비교)

  • Yong-Gu Lee
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.4
    • /
    • pp.307-327
    • /
    • 2023
  • To analyze the impact of word embedding on book titles, this study utilized word embedding models (Word2vec, GloVe, fastText) to generate embedding vectors from book titles. These vectors were then used as classification features for automatic classification. The classifier utilized the k-nearest neighbors (kNN) algorithm, with the categories for automatic classification based on the DDC (Dewey Decimal Classification) main class 300 assigned by libraries to books. In the automatic classification experiment applying word embeddings to book titles, the Skip-gram architectures of Word2vec and fastText showed better results in the automatic classification performance of the kNN classifier compared to the TF-IDF features. In the optimization of various hyperparameters across the three models, the Skip-gram architecture of the fastText model demonstrated overall good performance. Specifically, better performance was observed when using hierarchical softmax and larger embedding dimensions as hyperparameters in this model. From a performance perspective, fastText can generate embeddings for substrings or subwords using the n-gram method, which has been shown to increase recall. The Skip-gram architecture of the Word2vec model generally showed good performance at low dimensions(size 300) and with small sizes of negative sampling (3 or 5).

Intellectual Structure Analysis on the Field of Open Data Using Co-word Analysis (동시출현단어 분석을 이용한 오픈 데이터 분야의 지적 구조 분석)

  • HyeKyung Lee;Yong-Gu Lee
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.4
    • /
    • pp.429-450
    • /
    • 2023
  • The purpose of this study is to examine recent trends and intellectual structures in research related to open data. To achieve this, the study conducted a search for the keyword "open data" in Scopus and collected a total of 6,543 papers from 1999 to 2023. After data preprocessing, the study focused on the author keywords of 5,589 papers to perform network analysis and derive centrality in the field of open data research and linked open data research. As a result, the study found that "big data" exhibited the highest centrality in research related to open data. The research in this area mainly focuses on the utilization of open data as a concept of public data, studies on the application of open data in analysis related to big data as an associated concept, and research on topics related to the use of open data, such as the reproduction, utilization, and access of open data. In linked open data research, both triadic centrality and closeness centrality showed that "the semantic web" had the highest centrality. Moreover, it was observed that research emphasizing data linkage and relationship formation, rather than public data policies, was more prevalent in this field.

Analyzing the Main Paths and Intellectual Structure of the Data Literacy Research Domain (데이터 리터러시 연구 분야의 주경로와 지적구조 분석)

  • Jae Yun Lee
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.4
    • /
    • pp.403-428
    • /
    • 2023
  • This study investigates the development path and intellectual structure of data literacy research, aiming to identify emerging topics in the field. A comprehensive search for data literacy-related articles on the Web of Science reveals that the field is primarily concentrated in Education & Educational Research and Information Science & Library Science, accounting for nearly 60% of the total. Citation network analysis, employing the PageRank algorithm, identifies key papers with high citation impact across various topics. To accurately trace the development path of data literacy research, an enhanced PageRank main path algorithm is developed, which overcomes the limitations of existing methods confined to the Education & Educational Research field. Keyword bibliographic coupling analysis is employed to unravel the intellectual structure of data literacy research. Utilizing the PNNC algorithm, the detailed structure and clusters of the derived keyword bibliographic coupling network are revealed, including two large clusters, one with two smaller clusters and the other with five smaller clusters. The growth index and mean publishing year of each keyword and cluster are measured to pinpoint emerging topics. The analysis highlights the emergence of critical data literacy for social justice in higher education amidst the ongoing pandemic and the rise of AI chatbots. The enhanced PageRank main path algorithm, developed in this study, demonstrates its effectiveness in identifying parallel research streams developing across different fields.

Bibliometric Analysis on Health Information-Related Research in Korea (국내 건강정보관련 연구에 대한 계량서지학적 분석)

  • Jin Won Kim;Hanseul Lee
    • Journal of the Korean Society for information Management
    • /
    • v.41 no.1
    • /
    • pp.411-438
    • /
    • 2024
  • This study aims to identify and comprehensively view health information-related research trends using a bibliometric analysis. To this end, 1,193 papers from 2002 to 2023 related to "health information" were collected through the Korea Citation Index (KCI) database and analyzed in diverse aspects: research trends by period, academic fields, intellectual structure, and keyword changes. Results indicated that the number of papers related to health information continued to increase and has been decreasing since 2021. The main academic fields of health information-related research included "biomedical engineering," "preventive medicine/occupational environmental medicine," "law," "nursing," "library and information science," and "interdisciplinary research." Moreover, a co-word analysis was performed to understand the intellectual structure of research related to health information. As a result of applying the parallel nearest neighbor clustering (PNNC) algorithm to identify the structure and cluster of the derived network, four clusters and 17 subgroups belonging to them could be identified, centering on two conglomerates: "medical engineering perspective on health information" and "social science perspective on health information." An inflection point analysis was attempted to track the timing of change in the academic field and keywords, and common changes were observed between 2010 and 2011. Finally, a strategy diagram was derived through the average publication year and word frequency, and high-frequency keywords were presented by dividing them into "promising," "growth," and "mature." Unlike previous studies that mainly focused on content analysis, this study is meaningful in that it viewed the research area related to health information from an integrated perspective using various bibliometric methods.

Investigating Dynamic Mutation Process of Issues Using Unstructured Text Analysis (부도예측을 위한 KNN 앙상블 모형의 동시 최적화)

  • Min, Sung-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.1
    • /
    • pp.139-157
    • /
    • 2016
  • Bankruptcy involves considerable costs, so it can have significant effects on a country's economy. Thus, bankruptcy prediction is an important issue. Over the past several decades, many researchers have addressed topics associated with bankruptcy prediction. Early research on bankruptcy prediction employed conventional statistical methods such as univariate analysis, discriminant analysis, multiple regression, and logistic regression. Later on, many studies began utilizing artificial intelligence techniques such as inductive learning, neural networks, and case-based reasoning. Currently, ensemble models are being utilized to enhance the accuracy of bankruptcy prediction. Ensemble classification involves combining multiple classifiers to obtain more accurate predictions than those obtained using individual models. Ensemble learning techniques are known to be very useful for improving the generalization ability of the classifier. Base classifiers in the ensemble must be as accurate and diverse as possible in order to enhance the generalization ability of an ensemble model. Commonly used methods for constructing ensemble classifiers include bagging, boosting, and random subspace. The random subspace method selects a random feature subset for each classifier from the original feature space to diversify the base classifiers of an ensemble. Each ensemble member is trained by a randomly chosen feature subspace from the original feature set, and predictions from each ensemble member are combined by an aggregation method. The k-nearest neighbors (KNN) classifier is robust with respect to variations in the dataset but is very sensitive to changes in the feature space. For this reason, KNN is a good classifier for the random subspace method. The KNN random subspace ensemble model has been shown to be very effective for improving an individual KNN model. The k parameter of KNN base classifiers and selected feature subsets for base classifiers play an important role in determining the performance of the KNN ensemble model. However, few studies have focused on optimizing the k parameter and feature subsets of base classifiers in the ensemble. This study proposed a new ensemble method that improves upon the performance KNN ensemble model by optimizing both k parameters and feature subsets of base classifiers. A genetic algorithm was used to optimize the KNN ensemble model and improve the prediction accuracy of the ensemble model. The proposed model was applied to a bankruptcy prediction problem by using a real dataset from Korean companies. The research data included 1800 externally non-audited firms that filed for bankruptcy (900 cases) or non-bankruptcy (900 cases). Initially, the dataset consisted of 134 financial ratios. Prior to the experiments, 75 financial ratios were selected based on an independent sample t-test of each financial ratio as an input variable and bankruptcy or non-bankruptcy as an output variable. Of these, 24 financial ratios were selected by using a logistic regression backward feature selection method. The complete dataset was separated into two parts: training and validation. The training dataset was further divided into two portions: one for the training model and the other to avoid overfitting. The prediction accuracy against this dataset was used to determine the fitness value in order to avoid overfitting. The validation dataset was used to evaluate the effectiveness of the final model. A 10-fold cross-validation was implemented to compare the performances of the proposed model and other models. To evaluate the effectiveness of the proposed model, the classification accuracy of the proposed model was compared with that of other models. The Q-statistic values and average classification accuracies of base classifiers were investigated. The experimental results showed that the proposed model outperformed other models, such as the single model and random subspace ensemble model.

Development of Regularized Expectation Maximization Algorithms for Fan-Beam SPECT Data (부채살 SPECT 데이터를 위한 정칙화된 기댓값 최대화 재구성기법 개발)

  • Kim, Soo-Mee;Lee, Jae-Sung;Lee, Soo-Jin;Kim, Kyeong-Min;Lee, Dong-Soo
    • The Korean Journal of Nuclear Medicine
    • /
    • v.39 no.6
    • /
    • pp.464-472
    • /
    • 2005
  • Purpose: SPECT using a fan-beam collimator improves spatial resolution and sensitivity. For the reconstruction from fan-beam projections, it is necessary to implement direct fan-beam reconstruction methods without transforming the data into the parallel geometry. In this study, various fan-beam reconstruction algorithms were implemented and their performances were compared. Materials and Methods: The projector for fan-beam SPECT was implemented using a ray-tracing method. The direct reconstruction algorithms implemented for fan-beam projection data were FBP (filtered backprojection), EM (expectation maximization), OS-EM (ordered subsets EM) and MAP-EM OSL (maximum a posteriori EM using the one-step late method) with membrane and thin-plate models as priors. For comparison, the fan-beam protection data were also rebinned into the parallel data using various interpolation methods, such as the nearest neighbor, bilinear and bicubic interpolations, and reconstructed using the conventional EM algorithm for parallel data. Noiseless and noisy projection data from the digital Hoffman brain and Shepp/Logan phantoms were reconstructed using the above algorithms. The reconstructed images were compared in terms of a percent error metric. Results: for the fan-beam data with Poisson noise, the MAP-EM OSL algorithm with the thin-plate prior showed the best result in both percent error and stability. Bilinear interpolation was the most effective method for rebinning from the fan-beam to parallel geometry when the accuracy and computation load were considered. Direct fan-beam EM reconstructions were more accurate than the standard EM reconstructions obtained from rebinned parallel data. Conclusion: Direct fan-beam reconstruction algorithms were implemented, which provided significantly improved reconstructions.