• Title/Summary/Keyword: Confusion Matrix

Search Result 111, Processing Time 0.026 seconds

A Comparison of Classification Techniques in Hyperspectral Image (하이퍼스펙트럴 영상의 분류 기법 비교)

  • 가칠오;김대성;변영기;김용일
    • Proceedings of the Korean Society of Surveying, Geodesy, Photogrammetry, and Cartography Conference
    • /
    • 2004.11a
    • /
    • pp.251-256
    • /
    • 2004
  • The image classification is one of the most important studies in the remote sensing. In general, the MLC(Maximum Likelihood Classification) classification that in consideration of distribution of training information is the most effective way but it produces a bad result when we apply it to actual hyperspectral image with the same classification technique. The purpose of this research is to reveal that which one is the most effective and suitable way of the classification algorithms iii the hyperspectral image classification. To confirm this matter, we apply the MLC classification algorithm which has distribution information and SAM(Spectral Angle Mapper), SFF(Spectral Feature Fitting) algorithm which use average information of the training class to both multispectral image and hyperspectral image. I conclude this result through quantitative and visual analysis using confusion matrix could confirm that SAM and SFF algorithm using of spectral pattern in vector domain is more effective way in the hyperspectral image classification than MLC which considered distribution.

  • PDF

A Cost Effective Reference Data Sampling Algorithm Using Fractal Analysis (프랙탈 분석을 통한 비용효과적인 기준자료추출 알고리즘에 관한 연구)

  • 김창재;이병길;김용일
    • Proceedings of the KSRS Conference
    • /
    • 2000.04a
    • /
    • pp.149-154
    • /
    • 2000
  • 분류기법을 통해 얻어진 원격탐사 자료는 사용되기 이전에 그 정확성에 관한 신뢰도 검증을 해야 한다. 분류 정확도를 평가하기 위해서는 오분류행렬(confusion matrix)을 사용하여 정확도 평가를 하게 되는데, 이때 오분류행렬을 구성하기 위해서는 기준자료(reference data)에 대한 표본추출이 이루어져야 한다. 기준자료의 표본을 추출하는 기법간의 비교 및 표본 크기를 줄이고자 하는 연구는 많이 이루어져 왔으난, 추출된 표본들간의 거리를 줄임으로써 정확도 평가 비용을 감소시키고자 하는 연구는 미미한 실정이다. 따라서, 본 연구에서는 프랙탈 분석을 통하여 기준자료의 표본을 추출하였으며, 이를 바탕으로 기존의 표본추출 기법과 정확도 차이 및 비용효과 측면을 비교 분석하였다. 연구 결과, 프랙탈 분석을 통하여 표본을 추출하는 기법은 그 정확도 추정에 있어 기존적 표본 추출 기법과 큰 차이가 보이지 않았으며, 추출된 화소들이 가까운 거리에 군집해 있어 비용효과측면에서 보다 유리함을 확인하였다.

  • PDF

Equivalent Writing of Loanwords Detection Method based on Korean Alphabet Confusion Matrix (한국어 자모 혼동행렬 기반 유사 외래어 표기 검출 기법)

  • Kwon, Soonho;Kwon, Hyuk-Chul
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2010.04a
    • /
    • pp.433-436
    • /
    • 2010
  • 최근 한국어 문서에는 한국어뿐만 아니라 외래어 표기 등이 혼합되어 사용되고 있다. 외래어 표기는 한 단어에 대해 한 개만 존재하는 것이 아니라 여러 개의 다른 표기로 사용되고 있다. 이러한 표기상 불일치는 하나의 단어가 다른 개념으로 인식되어 정보검색 시스템의 성능 저하의 원인이 된다. 따라서 정보검색 시스템의 성능 향상을 위해 여러 외래어 표기를 같은 개념으로 인식하는 시스템이 필요하다. 본 논문에서는 한국어 자모 혼동행렬을 기반으로 한 유사 외래어 표기 검출 기법을 제안한다. 제안한 기법에 따라 유사 외래어 표기를 검출해줌으로써 정보검색 시스템의 성능을 향상할 수 있다.

Machine-printed Numeral Recognition using Weighted Template Matching with Chain Code Trimming (체인 코드 트리밍과 가중 원형 정합을 이용한 인쇄체 숫자 인식)

  • Jung, Min-Chul
    • Journal of Intelligence and Information Systems
    • /
    • v.13 no.4
    • /
    • pp.35-44
    • /
    • 2007
  • This paper proposes a new method of weighted template matching for machine-printed numeral recognition. The proposed weighted template matching, which emphasizes the feature of a pattern using adaptive Hamming distance on local feature areas, improves the recognition rate while template matching processes an input image as one global feature. Template matching is vulnerable to random noises that generate ragged outlines of a pattern when it is binarized. This paper offers a method of chain code trimming in order to remove ragged outlines. The method corrects specific chain codes within the chain codes of the inner and the outer contour of a pattern. The experiment compares confusion matrices of both the template matching and the proposed weighted template matching with chain code trimming. The result shows that the proposed method improves fairly the recognition rate of the machine-printed numerals.

  • PDF

Shifts of Geographic Distribution of Pinus koraiensis Based on Climate Change Scenarios and GARP Model (GARP 모형과 기후변화 시나리오에 따른 잣나무의 지리적 분포 변화)

  • Chun, Jung Hwa;Lee, Chang Bae;Yoo, So Min
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.17 no.4
    • /
    • pp.348-357
    • /
    • 2015
  • The main purpose of this study is to understand the potential geographic distribution of P. koraiensis, which is known to be one of major economic tree species, based on the RCP (Representative Concentration Pathway) 8.5 scenarios and current geographic distribution from National Forest Inventory(NFI) data using ecological niche modeling. P. koraiensis abundance data extracted from NFI were utilized to estimate current geographic distribution. Also, GARP (Genetic Algorithm for Rule-set Production) model, one of the ecological niche models, was applied to estimate potential geographic distribution and to project future changes. Environmental explanatory variables showing Area Under Curve (AUC) value bigger than 0.6 were selected and constructed into the final model by running the model for each of the 27 variables. The results of the model validation which was performed based on confusion matrix statistics, showed quite high suitability. Currently P. koraiensis is distributed widely from 300m to 1,200m in altitude and from south to north as a result of national greening project in 1970s although major populations are found in elevated and northern area. The results of this study were successful in showing the current distribution of P. koraiensis and projecting their future changes. Future model for P. koraiensis suggest large areas predicted under current climate conditions may be contracted by 2090s showing dramatic habitat loss. Considering the increasing status of atmospheric $CO_2$ and air temperature in Korea, P. koraiensis seems to experience the significant decrease of potential distribution range in the future. The final model in this study may be used to identify climate change impacts on distribution of P. koraiensis in Korea, and a deeper understanding of its correlation may be helpful when planning afforestation strategies.

Analysis of Feature Importance of Ship's Berthing Velocity Using Classification Algorithms of Machine Learning (머신러닝 분류 알고리즘을 활용한 선박 접안속도 영향요소의 중요도 분석)

  • Lee, Hyeong-Tak;Lee, Sang-Won;Cho, Jang-Won;Cho, Ik-Soon
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.26 no.2
    • /
    • pp.139-148
    • /
    • 2020
  • The most important factor affecting the berthing energy generated when a ship berths is the berthing velocity. Thus, an accident may occur if the berthing velocity is extremely high. Several ship features influence the determination of the berthing velocity. However, previous studies have mostly focused on the size of the vessel. Therefore, the aim of this study is to analyze various features that influence berthing velocity and determine their respective importance. The data used in the analysis was based on the berthing velocity of a ship on a jetty in Korea. Using the collected data, machine learning classification algorithms were compared and analyzed, such as decision tree, random forest, logistic regression, and perceptron. As an algorithm evaluation method, indexes according to the confusion matrix were used. Consequently, perceptron demonstrated the best performance, and the feature importance was in the following order: DWT, jetty number, and state. Hence, when berthing a ship, the berthing velocity should be determined in consideration of various features, such as the size of the ship, position of the jetty, and loading condition of the cargo.

The Structure of Korean Consonants as Perceived by the Japanese (일본인이 지각하는 한국어 자음의 구조)

  • Bae, Moon-Jung;Kim, Jung-Oh
    • Korean Journal of Cognitive Science
    • /
    • v.19 no.2
    • /
    • pp.163-175
    • /
    • 2008
  • Twelve Japanese students living in South Korea have been examined for their perceptual identification of an initial consonant in Korean syllables with or without a white noise. A confusion matrix was then subject to analyses of additive clustering, individual difference scaling, and probability of information transmission, the results of which were also compared to those of South Koreans. The Japanese in the present experiment confused /다/and/타/ most frequently, followed by /가/ and /카/, /자, 차, 짜/, /타/ and /따/, and so on. The results of additive clustering analysis of the Japanese significantly differed from those of the South Koreans. Individual difference scaling revealed dimensions of sonorant, aspiration and coronal. While South Koreans showed binary values on aspiration and tenseness dimensions, the Japanese did continuous values on such dimensions. An information transmission probability analysis revealed that the Japanese participants could not perceive very well such larynx features as tenseness and aspiration compared to the South Korean participants. The former group, however, perceived very well place of articulation features such as labial and coronal. The present results suggest that an approach dealing with structures of base representations is important in understanding the phonological categories of languages.

  • PDF

A Study on Classifying Sea Ice of the Summer Arctic Ocean Using Sentinel-1 A/B SAR Data and Deep Learning Models (Sentinel-1 A/B 위성 SAR 자료와 딥러닝 모델을 이용한 여름철 북극해 해빙 분류 연구)

  • Jeon, Hyungyun;Kim, Junwoo;Vadivel, Suresh Krishnan Palanisamy;Kim, Duk-jin
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.6_1
    • /
    • pp.999-1009
    • /
    • 2019
  • The importance of high-resolution sea ice maps of the Arctic Ocean is increasing due to the possibility of pioneering North Pole Routes and the necessity of precise climate prediction models. In this study,sea ice classification algorithms for two deep learning models were examined using Sentinel-1 A/B SAR data to generate high-resolution sea ice classification maps. Based on current ice charts, three classes (Open Water, First Year Ice, Multi Year Ice) of training data sets were generated by Arctic sea ice and remote sensing experts. Ten sea ice classification algorithms were generated by combing two deep learning models (i.e. Simple CNN and Resnet50) and five cases of input bands including incident angles and thermal noise corrected HV bands. For the ten algorithms, analyses were performed by comparing classification results with ground truth points. A confusion matrix and Cohen's kappa coefficient were produced for the case that showed best result. Furthermore, the classification result with the Maximum Likelihood Classifier that has been traditionally employed to classify sea ice. In conclusion, the Convolutional Neural Network case, which has two convolution layers and two max pooling layers, with HV and incident angle input bands shows classification accuracy of 96.66%, and Cohen's kappa coefficient of 0.9499. All deep learning cases shows better classification accuracy than the classification result of the Maximum Likelihood Classifier.

Assessment of Landslide Susceptibility in Jecheon Using Deep Learning Based on Exploratory Data Analysis (데이터 탐색을 활용한 딥러닝 기반 제천 지역 산사태 취약성 분석)

  • Sang-A Ahn;Jung-Hyun Lee;Hyuck-Jin Park
    • The Journal of Engineering Geology
    • /
    • v.33 no.4
    • /
    • pp.673-687
    • /
    • 2023
  • Exploratory data analysis is the process of observing and understanding data collected from various sources to identify their distributions and correlations through their structures and characterization. This process can be used to identify correlations among conditioning factors and select the most effective factors for analysis. This can help the assessment of landslide susceptibility, because landslides are usually triggered by multiple factors, and the impacts of these factors vary by region. This study compared two stages of exploratory data analysis to examine the impact of the data exploration procedure on the landslide prediction model's performance with respect to factor selection. Deep-learning-based landslide susceptibility analysis used either a combinations of selected factors or all 23 factors. During the data exploration phase, we used a Pearson correlation coefficient heat map and a histogram of random forest feature importance. We then assessed the accuracy of our deep-learning-based analysis of landslide susceptibility using a confusion matrix. Finally, a landslide susceptibility map was generated using the landslide susceptibility index derived from the proposed analysis. The analysis revealed that using all 23 factors resulted in low accuracy (55.90%), but using the 13 factors selected in one step of exploration improved the accuracy to 81.25%. This was further improved to 92.80% using only the nine conditioning factors selected during both steps of the data exploration. Therefore, exploratory data analysis selected the conditioning factors most suitable for landslide susceptibility analysis and thereby improving the performance of the analysis.

Software Vulnerability Prediction System Using Machine Learning Algorithm (기계학습 알고리즘을 이용한 소프트웨어 취약 여부 예측 시스템)

  • Choi, Minjun;Kim, Juhwan;Yun, Joobeom
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.28 no.3
    • /
    • pp.635-642
    • /
    • 2018
  • In the Era of the Fourth Industrial Revolution, we live in huge amounts of software. However, as software increases, software vulnerabilities are also increasing. Therefore, it is important to detect and remove software vulnerabilities. Currently, many researches have been studied to predict and detect software security problems, but it takes a long time to detect and does not have high prediction accuracy. Therefore, in this paper, we describe a method for efficiently predicting software vulnerabilities using machine learning algorithms. In addition, various machine learning algorithms are compared through experiments. Experimental results show that the k-nearest neighbors prediction model has the highest prediction rate.