• Title/Summary/Keyword: 최근접 데이터 선택

Search Result 28, Processing Time 0.024 seconds

A Study on the Development of Tracking Algorithm for Shipborne Automatic Tracking Aids (선박자동추적장치의 목표물 추적 알고리즘 개발에 관한 연구)

  • 김석재;구자윤;윤수원
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.9 no.2
    • /
    • pp.5-13
    • /
    • 2003
  • Ships of 500 gross tonnage and upwards constructed an or after 1 July 2002 shall have an automatic tracking aids according to SOLAS V/19 but existing ships less than 10,000 gross tonnage constructed before 1 July 2002 have potential collision risks due to the lack of automatic plotting devices like as an ATA. This paper aims to provide a homemade ATA by developing the tracking algorithm for ATA and to prevent collision incidents by distributing ATA system to coasters.

  • PDF

Energy-Efficient Index Scheme in Mobile P2P Environments (모바일 P2P 환경에서의 에너지 효율을 고려한 인덱스 기법)

  • Park, Kwang-Jin
    • Journal of Internet Computing and Services
    • /
    • v.10 no.6
    • /
    • pp.1-6
    • /
    • 2009
  • In this paper, we propose an energy efficient index scheme for Nearest Neighbor queries, specially designed for mobile P2P environments. we propose a PMBR (P2P Minimum Boundary Rectangle) index that provides the client with the ability of selective contacting and tuning from other nodes. The mobile client is able to identify whether or not the neighbor nodes have the desired information by accessing PMBR. Thus, the client immediately switches to change to another node's broadcast channel if the index does not contain the desired information. Furthermore, the client is able to predict the arrival time of the desired data items and only needs to tune into the broadcast channel when the requested data items arrives, by first accessing PMBR. Thus, the mobile client can stay in power save mode most of time, and tune into the broadcast channel only when the requested data items arrive. Experiments are conducted to evaluate the performance of the proposed scheme. Comprehensive experiments illustrate that the proposed scheme is more efficient than the previous techniques in terms of energy consumption.

  • PDF

Variational Bayesian multinomial probit model with Gaussian process classification on mice protein expression level data (가우시안 과정 분류에 대한 변분 베이지안 다항 프로빗 모형: 쥐 단백질 발현 데이터에의 적용)

  • Donghyun Son;Beom Seuk Hwang
    • The Korean Journal of Applied Statistics
    • /
    • v.36 no.2
    • /
    • pp.115-127
    • /
    • 2023
  • Multinomial probit model is a popular model for multiclass classification and choice model. Markov chain Monte Carlo (MCMC) method is widely used for estimating multinomial probit model, but its computational cost is high. However, it is well known that variational Bayesian approximation is more computationally efficient than MCMC, because it uses subsets of samples. In this study, we describe multinomial probit model with Gaussian process classification and how to employ variational Bayesian approximation on the model. This study also compares the results of variational Bayesian multinomial probit model to the results of naive Bayes, K-nearest neighbors and support vector machine for the UCI mice protein expression level data.

Multiresolution 4- 8 Tile Hierarchy Construction for Realtime Visualization of Planetary Scale Geological Information (행성 규모 지리 정보의 실시간 시각화를 위한 다계층 4-8 타일 구조의 구축)

  • Jin, Jong-Wook;Wohn, Kwang-Yun
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.9 no.4
    • /
    • pp.12-21
    • /
    • 2006
  • Recently, Very large and high resolution geological data from aerial or satellite imagery are available. Many researches and applications require to do realtime visualization of interest geological area or entire planet. Important operation of wide-spreaded terrain realtime visualization technique is the appropriate model resolution selection from pre-processed multi-resolution model hierarchy depend upon participant's view. For embodying such realtime rendering system with large geometric data, Preprocessing multi-resolution hierarchy from large scale geological information of interest area is required. In this research, recent Cubic multiresolution 4-8 tile hierarchy is selected for global planetary applications. Based upon the tile hierarchy, It constructs the selective terminal level tile mesh for original geological information area and starts to sample individual generated tiles for terminal level tiles. It completes the hierarchy by constructing intermediate tiles with low pass filtering in bottom-up direction. This research embodies series of efficient cubic 4-8 tile hierarchy construction mechanism with out-of-core storage. The planetary scale Mars' geographical altitude data and image data were selected for the experiment.

  • PDF

K-Nearest Neighbor Course Recommender System using Collaborative Filtering (협동적 필터링을 이용한 K-최근접 이웃 수강 과목 추천 시스템)

  • Sohn, Ki-Rack;Kim, So-Hyun
    • Journal of The Korean Association of Information Education
    • /
    • v.11 no.3
    • /
    • pp.281-288
    • /
    • 2007
  • Collaborative filtering is a method to predict preference items of a user based on the evaluations of items provided by others with similar preferences. Collaborative filtering helps general people make smart decisions in today's information society where information can be easily accumulated and analyzed. We designed, implemented, and evaluated a course recommendation system experimentally. This system can help university students choose courses they prefer to. Firstly, the system needs to collect the course preferences from students and store in a database. Users showing similar preference patterns are considered into similar groups. We use Pearson correlation as a similarity measure. We select K-nearest students to predict the unknown preferences of the student and provide a ranked list of courses based on the course preferences of K-nearest students. We evaluated the accuracy of the recommendation by computing the mean absolute errors of predictions using a survey on the course preferences of students.

  • PDF

Investigating Dynamic Mutation Process of Issues Using Unstructured Text Analysis (부도예측을 위한 KNN 앙상블 모형의 동시 최적화)

  • Min, Sung-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.1
    • /
    • pp.139-157
    • /
    • 2016
  • Bankruptcy involves considerable costs, so it can have significant effects on a country's economy. Thus, bankruptcy prediction is an important issue. Over the past several decades, many researchers have addressed topics associated with bankruptcy prediction. Early research on bankruptcy prediction employed conventional statistical methods such as univariate analysis, discriminant analysis, multiple regression, and logistic regression. Later on, many studies began utilizing artificial intelligence techniques such as inductive learning, neural networks, and case-based reasoning. Currently, ensemble models are being utilized to enhance the accuracy of bankruptcy prediction. Ensemble classification involves combining multiple classifiers to obtain more accurate predictions than those obtained using individual models. Ensemble learning techniques are known to be very useful for improving the generalization ability of the classifier. Base classifiers in the ensemble must be as accurate and diverse as possible in order to enhance the generalization ability of an ensemble model. Commonly used methods for constructing ensemble classifiers include bagging, boosting, and random subspace. The random subspace method selects a random feature subset for each classifier from the original feature space to diversify the base classifiers of an ensemble. Each ensemble member is trained by a randomly chosen feature subspace from the original feature set, and predictions from each ensemble member are combined by an aggregation method. The k-nearest neighbors (KNN) classifier is robust with respect to variations in the dataset but is very sensitive to changes in the feature space. For this reason, KNN is a good classifier for the random subspace method. The KNN random subspace ensemble model has been shown to be very effective for improving an individual KNN model. The k parameter of KNN base classifiers and selected feature subsets for base classifiers play an important role in determining the performance of the KNN ensemble model. However, few studies have focused on optimizing the k parameter and feature subsets of base classifiers in the ensemble. This study proposed a new ensemble method that improves upon the performance KNN ensemble model by optimizing both k parameters and feature subsets of base classifiers. A genetic algorithm was used to optimize the KNN ensemble model and improve the prediction accuracy of the ensemble model. The proposed model was applied to a bankruptcy prediction problem by using a real dataset from Korean companies. The research data included 1800 externally non-audited firms that filed for bankruptcy (900 cases) or non-bankruptcy (900 cases). Initially, the dataset consisted of 134 financial ratios. Prior to the experiments, 75 financial ratios were selected based on an independent sample t-test of each financial ratio as an input variable and bankruptcy or non-bankruptcy as an output variable. Of these, 24 financial ratios were selected by using a logistic regression backward feature selection method. The complete dataset was separated into two parts: training and validation. The training dataset was further divided into two portions: one for the training model and the other to avoid overfitting. The prediction accuracy against this dataset was used to determine the fitness value in order to avoid overfitting. The validation dataset was used to evaluate the effectiveness of the final model. A 10-fold cross-validation was implemented to compare the performances of the proposed model and other models. To evaluate the effectiveness of the proposed model, the classification accuracy of the proposed model was compared with that of other models. The Q-statistic values and average classification accuracies of base classifiers were investigated. The experimental results showed that the proposed model outperformed other models, such as the single model and random subspace ensemble model.

Pre-service mathematics teachers' noticing competency: Focusing on teaching for robust understanding of mathematics (예비 수학교사의 수학적 사고 중심 수업에 관한 노티싱 역량 탐색)

  • Kim, Hee-jeong
    • The Mathematical Education
    • /
    • v.61 no.2
    • /
    • pp.339-357
    • /
    • 2022
  • This study explores pre-service secondary mathematics teachers (PSTs)' noticing competency. 17 PSTs participated in this study as a part of the mathematics teaching method class. Individual PST's essays regarding the question 'what effective mathematics teaching would be?' that they discussed and wrote at the beginning of the course were collected as the first data. PSTs' written analysis of an expert teacher's teaching video, colleague PSTs' demo-teaching video, and own demo-teaching video were also collected and analyzed. Findings showed that most PSTs' noticing level improved as the class progressed and showed a pattern of focusing on each key aspect in terms of the Teaching for Robust Understanding of Mathematics (TRU Math) framework, but their reasoning strategies were somewhat varied. This suggests that the TRU Math framework can support PSTs to improve the competency of 'what to attend' among the noticing components. In addition, the instructional reasoning strategies imply that PSTs' noticing reasoning strategy was mostly related to their interpretation of noticing components, which should be also emphasized in the teacher education program.

A Study on the Effect of Network Centralities on Recommendation Performance (네트워크 중심성 척도가 추천 성능에 미치는 영향에 대한 연구)

  • Lee, Dongwon
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.23-46
    • /
    • 2021
  • Collaborative filtering, which is often used in personalization recommendations, is recognized as a very useful technique to find similar customers and recommend products to them based on their purchase history. However, the traditional collaborative filtering technique has raised the question of having difficulty calculating the similarity for new customers or products due to the method of calculating similaritiesbased on direct connections and common features among customers. For this reason, a hybrid technique was designed to use content-based filtering techniques together. On the one hand, efforts have been made to solve these problems by applying the structural characteristics of social networks. This applies a method of indirectly calculating similarities through their similar customers placed between them. This means creating a customer's network based on purchasing data and calculating the similarity between the two based on the features of the network that indirectly connects the two customers within this network. Such similarity can be used as a measure to predict whether the target customer accepts recommendations. The centrality metrics of networks can be utilized for the calculation of these similarities. Different centrality metrics have important implications in that they may have different effects on recommended performance. In this study, furthermore, the effect of these centrality metrics on the performance of recommendation may vary depending on recommender algorithms. In addition, recommendation techniques using network analysis can be expected to contribute to increasing recommendation performance even if they apply not only to new customers or products but also to entire customers or products. By considering a customer's purchase of an item as a link generated between the customer and the item on the network, the prediction of user acceptance of recommendation is solved as a prediction of whether a new link will be created between them. As the classification models fit the purpose of solving the binary problem of whether the link is engaged or not, decision tree, k-nearest neighbors (KNN), logistic regression, artificial neural network, and support vector machine (SVM) are selected in the research. The data for performance evaluation used order data collected from an online shopping mall over four years and two months. Among them, the previous three years and eight months constitute social networks composed of and the experiment was conducted by organizing the data collected into the social network. The next four months' records were used to train and evaluate recommender models. Experiments with the centrality metrics applied to each model show that the recommendation acceptance rates of the centrality metrics are different for each algorithm at a meaningful level. In this work, we analyzed only four commonly used centrality metrics: degree centrality, betweenness centrality, closeness centrality, and eigenvector centrality. Eigenvector centrality records the lowest performance in all models except support vector machines. Closeness centrality and betweenness centrality show similar performance across all models. Degree centrality ranking moderate across overall models while betweenness centrality always ranking higher than degree centrality. Finally, closeness centrality is characterized by distinct differences in performance according to the model. It ranks first in logistic regression, artificial neural network, and decision tree withnumerically high performance. However, it only records very low rankings in support vector machine and K-neighborhood with low-performance levels. As the experiment results reveal, in a classification model, network centrality metrics over a subnetwork that connects the two nodes can effectively predict the connectivity between two nodes in a social network. Furthermore, each metric has a different performance depending on the classification model type. This result implies that choosing appropriate metrics for each algorithm can lead to achieving higher recommendation performance. In general, betweenness centrality can guarantee a high level of performance in any model. It would be possible to consider the introduction of proximity centrality to obtain higher performance for certain models.