• Title/Summary/Keyword: 기계학습 알고리즘

Search Result 781, Processing Time 0.045 seconds

Machine Learning-Based Atmospheric Correction Based on Radiative Transfer Modeling Using Sentinel-2 MSI Data and ItsValidation Focusing on Forest (농림위성을 위한 기계학습을 활용한 복사전달모델기반 대기보정 모사 알고리즘 개발 및 검증: 식생 지역을 위주로)

  • Yoojin Kang;Yejin Kim ;Jungho Im;Joongbin Lim
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.5_3
    • /
    • pp.891-907
    • /
    • 2023
  • Compact Advanced Satellite 500-4 (CAS500-4) is scheduled to be launched to collect high spatial resolution data focusing on vegetation applications. To achieve this goal, accurate surface reflectance retrieval through atmospheric correction is crucial. Therefore, a machine learning-based atmospheric correction algorithm was developed to simulate atmospheric correction from a radiative transfer model using Sentinel-2 data that have similarspectral characteristics as CAS500-4. The algorithm was then evaluated mainly for forest areas. Utilizing the atmospheric correction parameters extracted from Sentinel-2 and GEOKOMPSAT-2A (GK-2A), the atmospheric correction algorithm was developed based on Random Forest and Light Gradient Boosting Machine (LGBM). Between the two machine learning techniques, LGBM performed better when considering both accuracy and efficiency. Except for one station, the results had a correlation coefficient of more than 0.91 and well-reflected temporal variations of the Normalized Difference Vegetation Index (i.e., vegetation phenology). GK-2A provides Aerosol Optical Depth (AOD) and water vapor, which are essential parameters for atmospheric correction, but additional processing should be required in the future to mitigate the problem caused by their many missing values. This study provided the basis for the atmospheric correction of CAS500-4 by developing a machine learning-based atmospheric correction simulation algorithm.

Improving Classification Accuracy in Hierarchical Trees via Greedy Node Expansion

  • Byungjin Lim;Jong Wook Kim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.6
    • /
    • pp.113-120
    • /
    • 2024
  • With the advancement of information and communication technology, we can easily generate various forms of data in our daily lives. To efficiently manage such a large amount of data, systematic classification into categories is essential. For effective search and navigation, data is organized into a tree-like hierarchical structure known as a category tree, which is commonly seen in news websites and Wikipedia. As a result, various techniques have been proposed to classify large volumes of documents into the terminal nodes of category trees. However, document classification methods using category trees face a problem: as the height of the tree increases, the number of terminal nodes multiplies exponentially, which increases the probability of misclassification and ultimately leads to a reduction in classification accuracy. Therefore, in this paper, we propose a new node expansion-based classification algorithm that satisfies the classification accuracy required by the application, while enabling detailed categorization. The proposed method uses a greedy approach to prioritize the expansion of nodes with high classification accuracy, thereby maximizing the overall classification accuracy of the category tree. Experimental results on real data show that the proposed technique provides improved performance over naive methods.

A Study on the i-YOLOX Architecture for Multiple Object Detection and Classification of Household Waste (생활 폐기물 다중 객체 검출과 분류를 위한 i-YOLOX 구조에 관한 연구)

  • Weiguang Wang;Kyung Kwon Jung;Taewon Lee
    • Convergence Security Journal
    • /
    • v.23 no.5
    • /
    • pp.135-142
    • /
    • 2023
  • In addressing the prominent issues of climate change, resource scarcity, and environmental pollution associated with household waste, extensive research has been conducted on intelligent waste classification methods. These efforts range from traditional classification algorithms to machine learning and neural networks. However, challenges persist in effectively classifying waste in diverse environments and conditions due to insufficient datasets, increased complexity in neural network architectures, and performance limitations for real-world applications. Therefore, this paper proposes i-YOLOX as a solution for rapid classification and improved accuracy. The proposed model is evaluated based on network parameters, detection speed, and accuracy. To achieve this, a dataset comprising 10,000 samples of household waste, spanning 17 waste categories, is created. The i-YOLOX architecture is constructed by introducing the Involution channel convolution operator and the Convolution Branch Attention Module (CBAM) into the YOLOX structure. A comparative analysis is conducted with the performance of the existing YOLO architecture. Experimental results demonstrate that i-YOLOX enhances the detection speed and accuracy of waste objects in complex scenes compared to conventional neural networks. This confirms the effectiveness of the proposed i-YOLOX architecture in the detection and classification of multiple household waste objects.

Development of the Artificial Intelligence Literacy Education Program for Preservice Secondary Teachers (예비 중등교사를 위한 인공지능 리터러시 교육 프로그램 개발)

  • Bong Seok Jang
    • Journal of Practical Engineering Education
    • /
    • v.16 no.1_spc
    • /
    • pp.65-70
    • /
    • 2024
  • As the interest in AI education grows, researchers have made efforts to implement AI education programs. However, research targeting pre-service teachers has been limited thus far. Therefore, this study was conducted to develop an AI literacy education program for preservice secondary teachers. The research results revealed that the weekly topics included the definition and applications of AI, analysis of intelligent agents, the importance of data, understanding machine learning, hands-on exercises on prediction and classification, hands-on exercises on clustering and classification, hands-on exercises on unstructured data, understanding deep learning, application of deep learning algorithms, fairness, transparency, accountability, safety, and social integration. Through this research, it is hoped that AI literacy education programs for preservice teachers will be expanded. In the future, it is anticipated that follow-up studies will be conducted to implement relevant education in teacher training institutions and analyze its effectiveness.

유리화 비정형 탄소(vitreous carbon)를 이용하여 제작한 전계방출 소자의 균일성 증진방법

  • 안상혁;이광렬
    • Proceedings of the Korean Vacuum Society Conference
    • /
    • 1999.07a
    • /
    • pp.53-53
    • /
    • 1999
  • 전계방출을 이용한 평판 표시장치는 CRT가 가진 장점을 모두 갖는 동시에 얇고 가벼우며 낮은 전력소모로 완벽한 색을 구현할 수 있는 차세대 표시장치로서 이에 대한 여국가 활발히 이루어지고 있다. 여기에 사용되는 음극물질로서 실리콘이나 몰리 등을 팁모양으로 제작하여 사용해 왔다. 하지만 잔류가스에 의한 역스퍼터링이나 화학적 반응에 의해서 전계방출 성능이 점차 저하되는 등의 해결해야할 많은 문제가 있다. 이러한 문제들을 해결하기 위하여 탄소계 재료로서 다이아몬드, 다이아몬드상 카본 등을 이용하려는 노력이 진행되어 왔다. 이중 유리화 비정형 탄소는 다량의 결함을 가지고 있는 유리질의 고상 탄소 재로로서, 전기전도도가 우수하면서 outgassing이 적고 기계적 강도가 뛰어나며 고온에서도 화학적으로 안정하여 전계방출 소자의 음극재료로서 알맞은 것으로 생각된다. 유리화 비정형 탄소가루를 전기영동법으로 기판에 코팅하여 전계방출 소자를 제작하였다. 전기영동 용액으로 이소프로필알코올에 질산마그네슘과 소량의 증류수, 유리화 비정형 탄소분말을 섞어주었고 기판으로는 몰리(Mo)가 증착된 유리를 사용하였다. 균일한 증착을 위해서 증착후 역전압을 걸어 주는 방법과 증착 후 플라즈마 처리를 하는 등의 여러 가지 방법을 사용했다. 전계방출 전류는 1$\times$10-7Torr이사에서 측정하였다. 1회 제작된 용액으로 반복해서 증착한 횟수에 따라 표면의 거치기, 입자의 분포, 전계방출 측정 결과 등의 차이가 관찰되었다. 발광이미지는 전압에 따라 변화하였고, 균일한 발광을 관찰하기 위해서 오랜 시간동안 aging 과정을 거쳐야 했다. 그리고 구 모양의 양극을 사용해서 위치를 변화시키며 시동 전기장을 관찰하여 위치에 따른 전계방출의 차이를 조사하여 발광의 균일성을 알 수 있었다.on microscopy로 분석하였으며 구조 분석은 X-선 회절분석, X-ray photoelectron spectroscopy 그리고Auger electron spectroscope로 하였다. 증착된 산화바나듐 박막의 전기화학적 특성을 분석하기 위하여 리튬 메탈을 anode로 하고 EC:DMC=1:1, 1M LiPF6 액체 전해질을 사용한 Half-Cell를 구성하여 200회 이상의 정전류 충 방전 시험을 행하였다. Half-Cell test 결과 박막의 결정성과 표면상태에 따라 매우 다른 전지 특성을 나타내었다.도상승율을 갖는 경우가 다른 베이킹 시나리오 모델에 비해 효과적이라 생각되며 초대 필요 공급열량은 200kW 정도로 산출되었다. 실질적인 수치를 얻기 위해 보다 고차원 모델로의 해석이 필요하리라 생각된다. 끝으로 장기적인 관점에서 KSTAR 장치의 베이킹 계획도 살펴본다.습파라미터와 더불어, 본 연구에서 새롭게 제시된 주기분할층의 파라미터들이 모형의 학습성과를 높이기 위해 함께 고려된다. 한편, 이러한 학습과정에서 추가적으로 고려해야 할 파라미터 갯수가 증가함에 따라서, 본 모델의 학습성과가 local minimum에 빠지는 문제점이 발생될 수 있다. 즉, 웨이블릿분석과 인공신경망모형을 모두 전역적으로 최적화시켜야 하는 문제가 발생한다. 본 연구에서는 이 문제를 해결하기 위해서, 최근 local minimum의 가능성을 최소화하여 전역적인 학습성과를 높여 주는 인공지능기법으로서 유전자알고리즘기법을 본 연구이 통합모델에 반영하였다. 이에 대한 실증사례 분석결과는 일일 환율예측문제를 적용하였을 경우, 기존의 방법론보다 더 나운 예측성과를 타나내었다.pective" to workflow architectural discussions. The vocabulary suggested

  • PDF

An Outlier Detection Using Autoencoder for Ocean Observation Data (해양 이상 자료 탐지를 위한 오토인코더 활용 기법 최적화 연구)

  • Kim, Hyeon-Jae;Kim, Dong-Hoon;Lim, Chaewook;Shin, Yongtak;Lee, Sang-Chul;Choi, Youngjin;Woo, Seung-Buhm
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.33 no.6
    • /
    • pp.265-274
    • /
    • 2021
  • Outlier detection research in ocean data has traditionally been performed using statistical and distance-based machine learning algorithms. Recently, AI-based methods have received a lot of attention and so-called supervised learning methods that require classification information for data are mainly used. This supervised learning method requires a lot of time and costs because classification information (label) must be manually designated for all data required for learning. In this study, an autoencoder based on unsupervised learning was applied as an outlier detection to overcome this problem. For the experiment, two experiments were designed: one is univariate learning, in which only SST data was used among the observation data of Deokjeok Island and the other is multivariate learning, in which SST, air temperature, wind direction, wind speed, air pressure, and humidity were used. Period of data is 25 years from 1996 to 2020, and a pre-processing considering the characteristics of ocean data was applied to the data. An outlier detection of actual SST data was tried with a learned univariate and multivariate autoencoder. We tried to detect outliers in real SST data using trained univariate and multivariate autoencoders. To compare model performance, various outlier detection methods were applied to synthetic data with artificially inserted errors. As a result of quantitatively evaluating the performance of these methods, the multivariate/univariate accuracy was about 96%/91%, respectively, indicating that the multivariate autoencoder had better outlier detection performance. Outlier detection using an unsupervised learning-based autoencoder is expected to be used in various ways in that it can reduce subjective classification errors and cost and time required for data labeling.

IPC Multi-label Classification based on Functional Characteristics of Fields in Patent Documents (특허문서 필드의 기능적 특성을 활용한 IPC 다중 레이블 분류)

  • Lim, Sora;Kwon, YongJin
    • Journal of Internet Computing and Services
    • /
    • v.18 no.1
    • /
    • pp.77-88
    • /
    • 2017
  • Recently, with the advent of knowledge based society where information and knowledge make values, patents which are the representative form of intellectual property have become important, and the number of the patents follows growing trends. Thus, it needs to classify the patents depending on the technological topic of the invention appropriately in order to use a vast amount of the patent information effectively. IPC (International Patent Classification) is widely used for this situation. Researches about IPC automatic classification have been studied using data mining and machine learning algorithms to improve current IPC classification task which categorizes patent documents by hand. However, most of the previous researches have focused on applying various existing machine learning methods to the patent documents rather than considering on the characteristics of the data or the structure of patent documents. In this paper, therefore, we propose to use two structural fields, technical field and background, considered as having impacts on the patent classification, where the two field are selected by applying of the characteristics of patent documents and the role of the structural fields. We also construct multi-label classification model to reflect what a patent document could have multiple IPCs. Furthermore, we propose a method to classify patent documents at the IPC subclass level comprised of 630 categories so that we investigate the possibility of applying the IPC multi-label classification model into the real field. The effect of structural fields of patent documents are examined using 564,793 registered patents in Korea, and 87.2% precision is obtained in the case of using title, abstract, claims, technical field and background. From this sequence, we verify that the technical field and background have an important role in improving the precision of IPC multi-label classification in IPC subclass level.

Analysis of groundwater withdrawal impact in the middle mountainous area of Pyoseon Watershed in Jeju Island using LSTM (LSTM을 활용한 제주도 표선유역 중산간지역의 지하수 취수영향 분석)

  • Shin, Mun-Ju;Moon, Soo-Hyoung;Moon, Duk-Chul;Koh, Hyuk-Joon;Kang, Kyung Goo
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.267-267
    • /
    • 2021
  • 제주도는 화산섬의 지질특성상 강수의 지표침투성이 높아 지표수의 개발이용여건이 취약한 관계로 용수의 대부분을 지하수에 의존하고 있다. 따라서 지하수의 보전관리는 매우 중요한 사항이며 특히 지하수의 안정적인 이용을 위해서는 지하수 취수가 주변지역 지하수위에 미치는 영향 분석이 반드시 필요하다. 본 연구는 딥러닝 알고리즘인 Long Short-Term Memory(LSTM)를 활용하여 제주도 남동쪽 표선유역 중산간지역에 위치한 2개 지하수위 관측정을 대상으로 지하수 취수영향을 분석하였다. 입력자료로써 인근 2개 강우관측소의 일단위 강수량자료와 인근 6개 취수정의 지하수 취수량자료 및 연구대상 관측정의 지하수위 자료(2001. 2. 11. ~ 2019. 10. 31.)를 사용하였다. 지하수위 변동특성을 최대한 반영하기 위해 LSTM의 예측일수를 1일로 설정하였다. 보정 및 검증 기간을 사용하여 매개변수의 과적합을 방지하였으며, 테스트 기간을 사용하여 LSTM의 예측성능을 평가하였다. 평가지수로써 Nash-Sutcliffe Efficiency(NSE)와 평균제곱근오차(RMSE)를 사용하였다. 그리고 지하수 취수가 주변 지하수위 변동에 미치는 영향을 분석하기 위해 취수량을 최대취수량인 2,300 m3/일, 최대취수량의 2/3인 1,533 m3/일 및 0 m3/일로 설정하여 모의하였다. 모의결과, 2개 감시정의 보정, 검증 및 예측기간에 대한 NSE는 최대 0.999, 최소 0.976의 범위를 보였으며, RMSE는 최대 0.494 m, 최소 0.084 m를 보여 LSTM은 우수한 예측성능을 나타내었다. 이것은 LSTM이 지하수위 변동특성을 적절히 학습하였다는 것을 의미하며 따라서 추정된 매개변수를 활용하여 지하수 취수영향을 모의 및 분석하였다. 그 결과, 지하수위 하강량은 최대 0.38 m 였으며 이것은 대상지점에 대한 취수량은 지하수위 하강에 거의 영향을 주지 않는다는 것을 의미한다. 또한 취수량과 지하수위 하강량과의 관계는 한 개 관측정에 대해 선형적인 관계를 보인 반면 나머지 한 개 관측정에 대해서는 비선형적인 관계를 나타내는 것을 확인하였다. 따라서 LSTM 알고리즘을 활용하여 제주도 표선유역 중산간지역의 지하수위 변동특성을 분석할 수 있다.

  • PDF

Fault Detection Technique for PVDF Sensor Based on Support Vector Machine (서포트벡터머신 기반 PVDF 센서의 결함 예측 기법)

  • Seung-Wook Kim;Sang-Min Lee
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.5
    • /
    • pp.785-796
    • /
    • 2023
  • In this study, a methodology for real-time classification and prediction of defects that may appear in PVDF(Polyvinylidene fluoride) sensors, which are widely used for structural integrity monitoring, is proposed. The types of sensor defects appearing according to the sensor attachment environment were classified, and an impact test using an impact hammer was performed to obtain an output signal according to the defect type. In order to cleary identify the difference between the output signal according to the defect types, the time domain statistical features were extracted and a data set was constructed. Among the machine learning based classification algorithms, the learning of the acquired data set and the result were analyzed to select the most suitable algorithm for detecting sensor defect types, and among them, it was confirmed that the highest optimization was performed to show SVM(Support Vector Machine). As a result, sensor defect types were classified with an accuracy of 92.5%, which was up to 13.95% higher than other classification algorithms. It is believed that the sensor defect prediction technique proposed in this study can be used as a base technology to secure the reliability of not only PVDF sensors but also various sensors for real time structural health monitoring.

Target Word Selection Disambiguation using Untagged Text Data in English-Korean Machine Translation (영한 기계 번역에서 미가공 텍스트 데이터를 이용한 대역어 선택 중의성 해소)

  • Kim Yu-Seop;Chang Jeong-Ho
    • The KIPS Transactions:PartB
    • /
    • v.11B no.6
    • /
    • pp.749-758
    • /
    • 2004
  • In this paper, we propose a new method utilizing only raw corpus without additional human effort for disambiguation of target word selection in English-Korean machine translation. We use two data-driven techniques; one is the Latent Semantic Analysis(LSA) and the other the Probabilistic Latent Semantic Analysis(PLSA). These two techniques can represent complex semantic structures in given contexts like text passages. We construct linguistic semantic knowledge by using the two techniques and use the knowledge for target word selection in English-Korean machine translation. For target word selection, we utilize a grammatical relationship stored in a dictionary. We use k- nearest neighbor learning algorithm for the resolution of data sparseness Problem in target word selection and estimate the distance between instances based on these models. In experiments, we use TREC data of AP news for construction of latent semantic space and Wail Street Journal corpus for evaluation of target word selection. Through the Latent Semantic Analysis methods, the accuracy of target word selection has improved over 10% and PLSA has showed better accuracy than LSA method. finally we have showed the relatedness between the accuracy and two important factors ; one is dimensionality of latent space and k value of k-NT learning by using correlation calculation.