• Title/Summary/Keyword: 원 탐지

Search Result 715, Processing Time 0.02 seconds

The Sensitivity Analysis according to Observed Frequency of Daily Composite Insolation based on COMS (관측 빈도에 따른 COMS 기반의 일 평균 일사량 산출의 민감도 분석)

  • Kim, Honghee;Lee, Kyeong-Sang;Seo, Minji;Choi, Sungwon;Sung, Noh-Hun;Lee, Darae;Jin, Donghyun;Kwon, Chaeyoung;Huh, Morang;Han, Kyung-Soo
    • Korean Journal of Remote Sensing
    • /
    • v.32 no.6
    • /
    • pp.733-739
    • /
    • 2016
  • Insolation is an major indicator variable that can serve as an energy source in earth system. It is important to monitor insolation content using remote sensing to evaluate the potential of solar energy. In this study, we performed sensitivity analysis of observed frequency on daily composite insolation over the Korean peninsula. We estimated INS through the channel data of Communication, Ocean and Meteorological Satellite (COMS) and Cloud Mask which have temporal resolution of 1 and 3 hours. We performed Hemispherical Integration by spatial resolution for meaning whole sky. And we performed daily composite insolation. And then we compared the accuracy of estimated COMS insolation data with pyranometer data from 37 points. As a result, there was no great sensitivity in the daily composite INS by observed frequency of satellite that accuracy of the calculated insolation at 1 hour interval was $28.6401W/m^2$ and 3 hours interval was $30.4960W/m^2$. However, there was a great difference in the space distribution of two other INS data by observed frequency of clouds. So, we performed sensitivity analysis with observed frequency of clouds and distinction between the two other INS data. Consequently, there was showed sensitivity up to $19.4392W/m^2$.

Financial Fraud Detection using Text Mining Analysis against Municipal Cybercriminality (지자체 사이버 공간 안전을 위한 금융사기 탐지 텍스트 마이닝 방법)

  • Choi, Sukjae;Lee, Jungwon;Kwon, Ohbyung
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.3
    • /
    • pp.119-138
    • /
    • 2017
  • Recently, SNS has become an important channel for marketing as well as personal communication. However, cybercrime has also evolved with the development of information and communication technology, and illegal advertising is distributed to SNS in large quantity. As a result, personal information is lost and even monetary damages occur more frequently. In this study, we propose a method to analyze which sentences and documents, which have been sent to the SNS, are related to financial fraud. First of all, as a conceptual framework, we developed a matrix of conceptual characteristics of cybercriminality on SNS and emergency management. We also suggested emergency management process which consists of Pre-Cybercriminality (e.g. risk identification) and Post-Cybercriminality steps. Among those we focused on risk identification in this paper. The main process consists of data collection, preprocessing and analysis. First, we selected two words 'daechul(loan)' and 'sachae(private loan)' as seed words and collected data with this word from SNS such as twitter. The collected data are given to the two researchers to decide whether they are related to the cybercriminality, particularly financial fraud, or not. Then we selected some of them as keywords if the vocabularies are related to the nominals and symbols. With the selected keywords, we searched and collected data from web materials such as twitter, news, blog, and more than 820,000 articles collected. The collected articles were refined through preprocessing and made into learning data. The preprocessing process is divided into performing morphological analysis step, removing stop words step, and selecting valid part-of-speech step. In the morphological analysis step, a complex sentence is transformed into some morpheme units to enable mechanical analysis. In the removing stop words step, non-lexical elements such as numbers, punctuation marks, and double spaces are removed from the text. In the step of selecting valid part-of-speech, only two kinds of nouns and symbols are considered. Since nouns could refer to things, the intent of message is expressed better than the other part-of-speech. Moreover, the more illegal the text is, the more frequently symbols are used. The selected data is given 'legal' or 'illegal'. To make the selected data as learning data through the preprocessing process, it is necessary to classify whether each data is legitimate or not. The processed data is then converted into Corpus type and Document-Term Matrix. Finally, the two types of 'legal' and 'illegal' files were mixed and randomly divided into learning data set and test data set. In this study, we set the learning data as 70% and the test data as 30%. SVM was used as the discrimination algorithm. Since SVM requires gamma and cost values as the main parameters, we set gamma as 0.5 and cost as 10, based on the optimal value function. The cost is set higher than general cases. To show the feasibility of the idea proposed in this paper, we compared the proposed method with MLE (Maximum Likelihood Estimation), Term Frequency, and Collective Intelligence method. Overall accuracy and was used as the metric. As a result, the overall accuracy of the proposed method was 92.41% of illegal loan advertisement and 77.75% of illegal visit sales, which is apparently superior to that of the Term Frequency, MLE, etc. Hence, the result suggests that the proposed method is valid and usable practically. In this paper, we propose a framework for crisis management caused by abnormalities of unstructured data sources such as SNS. We hope this study will contribute to the academia by identifying what to consider when applying the SVM-like discrimination algorithm to text analysis. Moreover, the study will also contribute to the practitioners in the field of brand management and opinion mining.

Migration of the Dokdo Cold Eddy in the East Sea (동해 독도 냉수성 소용돌이의 이동 특성)

  • KIM, JAEMIN;CHOI, BYOUNG-JU;LEE, SANG-HO;BYUN, DO-SEONG;KANG, BOONSOON
    • The Sea:JOURNAL OF THE KOREAN SOCIETY OF OCEANOGRAPHY
    • /
    • v.24 no.2
    • /
    • pp.351-373
    • /
    • 2019
  • The cold eddies around the Ulleung Basin in the East Sea were identified from satellite altimeter sea level data using the Winding-Angle method from 1993 to 2015. Among the cold eddies, the Dokdo Cold Eddies (DCEs), which were formed at the first meandering trough of the East Korea Warm Current (EKWC) and were pinched off to the southwest from the eastward flow, were classified and their migration patterns were analyzed. The vertical structures of water temperature, salinity, and flow velocity near the DCE center were also examined using numerical simulation and observation data provided by the Hybrid Coordinate Ocean Model and the National Institute of Fisheries Science, respectively. A total of 112 DCEs were generated for 23 years. Of these, 39 DCEs migrated westward and arrived off the east coast of Korea. The average travel distance was 250.9 km, the average lifespan was 93 days, and the average travel speed was 3.5 cm/s. The other 73 DCEs had moved to the east or had hovered around the generated location until they disappeared. At 50-100 m depth under the DCE, water temperature and salinity (T < $5^{\circ}C$, S < 34.1) were lower than those of ambient water and isotherms made a dome shape. Current faster than 10 cm/s circulates counterclockwise from the surface to 300 m depth at 38 km away from the center of DCE. After the EKWC separates from the coast, it flows eastward and starts to meander near Ulleungdo. The first trough of the meander in the east of Ulleungdo is pushed deep into the southwest and forms a cold eddy (DCE), which is shed from the meander in the south of Ulleungdo. While a DCE moves westward, it circumvents the Ulleung Warm Eddy (UWE) clockwise and follows U shape path toward the east coast of Korea. When the DCE arrives near the coast, the EKWC separates from the coast at the south of DCE and circumvents the DCE. As the DCE near the coast weakens and extinguishes about 30 days later after the arrival, the EKWC flows northward along the coast recovering its original path. The DCE steadily transports heat and salt from the north to the south, which helps to form a cold water region in the southwest of the Ulleung Basin and brings positive vorticity to change the separation latitude and path of the EKWC. Some of the DCEs moving to the west were merged into a coastal cold eddy to form a wide cold water region in the west of Ulleung Basin and to create a elongated anticlockwise circulation, which separated the UWE in the north from the EKWC in the south.

An Artificial Intelligence Approach to Waterbody Detection of the Agricultural Reservoirs in South Korea Using Sentinel-1 SAR Images (Sentinel-1 SAR 영상과 AI 기법을 이용한 국내 중소규모 농업저수지의 수표면적 산출)

  • Choi, Soyeon;Youn, Youjeong;Kang, Jonggu;Park, Ganghyun;Kim, Geunah;Lee, Seulchan;Choi, Minha;Jeong, Hagyu;Lee, Yangwon
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.5_3
    • /
    • pp.925-938
    • /
    • 2022
  • Agricultural reservoirs are an important water resource nationwide and vulnerable to abnormal climate effects such as drought caused by climate change. Therefore, it is required enhanced management for appropriate operation. Although water-level tracking is necessary through continuous monitoring, it is challenging to measure and observe on-site due to practical problems. This study presents an objective comparison between multiple AI models for water-body extraction using radar images that have the advantages of wide coverage, and frequent revisit time. The proposed methods in this study used Sentinel-1 Synthetic Aperture Radar (SAR) images, and unlike common methods of water extraction based on optical images, they are suitable for long-term monitoring because they are less affected by the weather conditions. We built four AI models such as Support Vector Machine (SVM), Random Forest (RF), Artificial Neural Network (ANN), and Automated Machine Learning (AutoML) using drone images, sentinel-1 SAR and DSM data. There are total of 22 reservoirs of less than 1 million tons for the study, including small and medium-sized reservoirs with an effective storage capacity of less than 300,000 tons. 45 images from 22 reservoirs were used for model training and verification, and the results show that the AutoML model was 0.01 to 0.03 better in the water Intersection over Union (IoU) than the other three models, with Accuracy=0.92 and mIoU=0.81 in a test. As the result, AutoML performed as well as the classical machine learning methods and it is expected that the applicability of the water-body extraction technique by AutoML to monitor reservoirs automatically.

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.