• Title/Summary/Keyword: Generation Prediction

Search Result 803, Processing Time 0.036 seconds

The Generation of Westerly Waves by Sobaek Mountains (소백산맥에 의한 서풍 파동 발생)

  • Kim, Jin wook;Youn, Daeok
    • Journal of the Korean earth science society
    • /
    • v.38 no.1
    • /
    • pp.24-34
    • /
    • 2017
  • The westerly waves generation is described in the advanced earth science textbook used at high school as follows: as westerly wind approaches and blows over large mountains, the air flow shows wave motions in downwind side, which can be explained by the conservation of potential vorticity. However, there has been no case study showing the phenomena of the mesoscale westerly waves with observational data in the area of small mountains in Korea. And thus the wind speed and time persistency of westerly winds along with the width and length of mountains have never been studied to explain the generation of the westerly waves. As a first step, we assured the westerly waves generated in the downwind side of Sobaek mountains based on surface station wind data nearby. Furthermore, the critical or minimum wind velocity of the westerly wind over Sobaek mountains to generate the downwind wave were derived and calcuated tobe about $0.6m\;s^{-1}$ for Sobaek mountains, which means that the westerly waves could be generated in most cases of westerly blowing over the mountains. Using surface station data and 4-dimensional assimilation data of RDAPS (Regional Data Assimilation and Prediction System) provided by Korea Meteorological Agency, we also analyzed cases of westerly waves occurrence and life cycle in the downwind side of Sobaek mountains for a year of 2014. The westerly waves occurred in meso-${\beta}$ or -${\gamma}$ scales. The westerly waves generated by the mountains disappeared gradually with wind speed decreasing. The occurrence frequency of the vorticity with meso-${\beta}$ scale got to be higher when the stronger westerly wind blew. When we extended the spatial range of the analysis, phenomena of westerly waves were also observed in the downwind side of Yensan mountains in Northeastern China. Our current work will be a study material to help students understand the atmospheric phenomena perturbed by mountains.

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.

A Study on Knowledge Entity Extraction Method for Individual Stocks Based on Neural Tensor Network (뉴럴 텐서 네트워크 기반 주식 개별종목 지식개체명 추출 방법에 관한 연구)

  • Yang, Yunseok;Lee, Hyun Jun;Oh, Kyong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.25-38
    • /
    • 2019
  • Selecting high-quality information that meets the interests and needs of users among the overflowing contents is becoming more important as the generation continues. In the flood of information, efforts to reflect the intention of the user in the search result better are being tried, rather than recognizing the information request as a simple string. Also, large IT companies such as Google and Microsoft focus on developing knowledge-based technologies including search engines which provide users with satisfaction and convenience. Especially, the finance is one of the fields expected to have the usefulness and potential of text data analysis because it's constantly generating new information, and the earlier the information is, the more valuable it is. Automatic knowledge extraction can be effective in areas where information flow is vast, such as financial sector, and new information continues to emerge. However, there are several practical difficulties faced by automatic knowledge extraction. First, there are difficulties in making corpus from different fields with same algorithm, and it is difficult to extract good quality triple. Second, it becomes more difficult to produce labeled text data by people if the extent and scope of knowledge increases and patterns are constantly updated. Third, performance evaluation is difficult due to the characteristics of unsupervised learning. Finally, problem definition for automatic knowledge extraction is not easy because of ambiguous conceptual characteristics of knowledge. So, in order to overcome limits described above and improve the semantic performance of stock-related information searching, this study attempts to extract the knowledge entity by using neural tensor network and evaluate the performance of them. Different from other references, the purpose of this study is to extract knowledge entity which is related to individual stock items. Various but relatively simple data processing methods are applied in the presented model to solve the problems of previous researches and to enhance the effectiveness of the model. From these processes, this study has the following three significances. First, A practical and simple automatic knowledge extraction method that can be applied. Second, the possibility of performance evaluation is presented through simple problem definition. Finally, the expressiveness of the knowledge increased by generating input data on a sentence basis without complex morphological analysis. The results of the empirical analysis and objective performance evaluation method are also presented. The empirical study to confirm the usefulness of the presented model, experts' reports about individual 30 stocks which are top 30 items based on frequency of publication from May 30, 2017 to May 21, 2018 are used. the total number of reports are 5,600, and 3,074 reports, which accounts about 55% of the total, is designated as a training set, and other 45% of reports are designated as a testing set. Before constructing the model, all reports of a training set are classified by stocks, and their entities are extracted using named entity recognition tool which is the KKMA. for each stocks, top 100 entities based on appearance frequency are selected, and become vectorized using one-hot encoding. After that, by using neural tensor network, the same number of score functions as stocks are trained. Thus, if a new entity from a testing set appears, we can try to calculate the score by putting it into every single score function, and the stock of the function with the highest score is predicted as the related item with the entity. To evaluate presented models, we confirm prediction power and determining whether the score functions are well constructed by calculating hit ratio for all reports of testing set. As a result of the empirical study, the presented model shows 69.3% hit accuracy for testing set which consists of 2,526 reports. this hit ratio is meaningfully high despite of some constraints for conducting research. Looking at the prediction performance of the model for each stocks, only 3 stocks, which are LG ELECTRONICS, KiaMtr, and Mando, show extremely low performance than average. this result maybe due to the interference effect with other similar items and generation of new knowledge. In this paper, we propose a methodology to find out key entities or their combinations which are necessary to search related information in accordance with the user's investment intention. Graph data is generated by using only the named entity recognition tool and applied to the neural tensor network without learning corpus or word vectors for the field. From the empirical test, we confirm the effectiveness of the presented model as described above. However, there also exist some limits and things to complement. Representatively, the phenomenon that the model performance is especially bad for only some stocks shows the need for further researches. Finally, through the empirical study, we confirmed that the learning method presented in this study can be used for the purpose of matching the new text information semantically with the related stocks.

Geostatistical Downscaling of Coarse Scale Remote Sensing Data and Integration with Precise Observation Data for Generation of Fine Scale Thematic Information (고해상도 주제 정보 생성을 위한 저해상도 원격탐사 자료의 지구통계학기반 상세화 및 정밀 관측 자료와의 통합)

  • Park, No-Wook
    • Korean Journal of Remote Sensing
    • /
    • v.29 no.1
    • /
    • pp.69-79
    • /
    • 2013
  • This paper presents a two-stage geostatistical integration approach that aims at downscaling of coarse scale remote sensing data. First, downscaling of the coarse scale sedoncary data is implemented using area-to-point kriging, and this result will be used as trend components on the next integration stage. Then simple kriging with local varying means that integrates sparse precise observation data with the downscaled data is applied to generate thematic information at a finer scale. The presented approach can not only account for the statistical relationships between precise observation and secondary data acquired at the different scales, but also to calibrate the errors in the secondary data through the integration with precise observation data. An experiment for precipitation mapping with weather station data and TRMM (Tropical Rainfall Measuring Mission) data acquired at a coarse scale is carried out to illustrate the applicability of the presented approach. From the experiment, the geostatistical downscaling approach applied in this paper could generate detailed thematic information at various finer target scales that reproduced the original TRMM precipitation values when upscaled. And the integration of the downscaled secondary information with precise observation data showed better prediction capability than that of a conventional univariate kriging algorithm. Thus, it is expected that the presented approach would be effectively used for downscaling of coarse scale data with various data acquired at different scales.

Development of a Method for Calculating the Allowable Storage Capacity of Rivers by Using Drone Images (드론 영상을 이용한 하천의 구간별 허용 저수량 산정 방법 개발)

  • Kim, Han-Gyeol;Kim, Jae-In;Yoon, Sung-Joo;Kim, Taejung
    • Korean Journal of Remote Sensing
    • /
    • v.34 no.2_1
    • /
    • pp.203-211
    • /
    • 2018
  • Dam discharge is carried out for the management of rivers and area around rivers due to rainy season or drought. Dam discharge should be based on an accurate understanding of the flow rate that can be accommodated in the river. Therefore, understanding the allowable storage capacity of river is an important factor in the management of the environment around the river. However, the methods using water level meters and images, which are currently used to determine the allowable flow rate of rivers, show limitations in terms of accuracy and efficiency. In order to solve these problems, this paper proposes a method to automatically calculate the allowable storage capacity of river based on the images taken by drone. In the first step, we create a 3D model of the river by using the drone images. This generation process consists of tiepoint extraction, image orientation, and image matching. In the second step, the allowable storage capacity is calculated by cross section analysis of the river using the generated river 3D model and the road and river layers in the target area. In this step, we determine the maximum water level of the river, extract the cross-sectional profile along the river, and use the 3D model to calculate the allowable storage capacity for the area. To prove our method, we used Bukhan river's data and as a result, the allowable storage volume was automatically extracted. It is expected that the proposed method will be useful for real - time management of rivers and surrounding areas and 3D models using drone.

A Study on Forecasting for Ubiquitous Space with Analysis of Digital Technology Trends (디지털기술의 동향분석을 통한 유비쿼터스 공간의 미래예측에 관한 연구)

  • In, Chi-Ho;Yi, Soo-Hyun
    • Archives of design research
    • /
    • v.19 no.5 s.67
    • /
    • pp.323-334
    • /
    • 2006
  • The continuous development of digital technology has actualized ubiquitous computing, and the environment of human life has come to face changes. Prediction of the future so that human life may be prepared can be very important. A number of predictions of ubiquitous space are presented, ranging from related professional research to mass media, However, proper analyses and understanding are required to understand ubiquitous space and apply it to design. This study seeks to understand ubiquitous space through analysis of the trend in digital technology from a new perspective, to research cases, and predict the future of ubiquitous space. Human, object, and environment have been set as basic factors as the subjects smoothly exchanges various types of information in the physical space, and the trend in digital technology is analyzed. From a human-oriented perspective, the background and development trend of digital technology has been analyzed under the theme of interaction and interlace. From an object-oriented perspective, an analysis was unfolded under the theme of products' evolvement from radios to robots. From an environment-focused perspective, an analysis has been carried out under the theme of situation recognition, intrinsic factors, and integration and connectivity. By applying the analytic results, the types of studies that predict the future of ubiquitous space and generate concepts have been classified and analyzed into three different types of studies for experiment, industry, and public. In this manner, ubiquitous space has been forecasted. This study seeks a systematic analysis of the understanding of the trends in digital technology and employs a case study of ubiquitous space based on systematic analysis. In consideration of all these, this study is expected to contribute to concept generation and development by designers.

  • PDF

A Study on the Use of Criminal Justice Information Big Data in terms of the Structuralization and Categorization (형사사법정보의 빅데이터 활용방안 연구: 구조화 범주화 관점으로)

  • Kim, Mi Ryung;Roh, Yoon Ju;Kim, Seonghun
    • Journal of the Korean Society for information Management
    • /
    • v.36 no.4
    • /
    • pp.253-277
    • /
    • 2019
  • In the era of the 4th Industrial Revolution, the importance of data is intensifying, but there are many cases where it is not easy to use data due to personal information protection. Although criminal justice information is expected to have various useful values such as crime prediction and prevention, scientific investigation of criminal investigations, and rationalization of sentencing, the use of criminal justice information is currently limited as a matter of legal interpretation related to privacy protection and criminal justice information. This study proposed to convert criminal justice information into 'crime data' and use it as big data through the structuralization and categorization of criminal justice information. And when using "crime data," legal issues, value in use, considerations for data generation and use were verified by experts, and future strategic development plans were identified. Finally we found that 'crime data' seems to have solved the privacy problem, but it is necessary to specify in the criminal justice information related law and it is urgent to be organized in a standardized form for analysis to use big data. Future directions are to derive data elements, construct a dictionary thesaurus, define and classify personal sensitive information for data grading, and develop algorithms for shaping unstructured data.

Workflow for Building a Draft Genome Assembly using Public-domain Tools: Toxocara canis as a Case Study (개 회충 게놈 응용 사례에서 공개용 분석 툴을 사용한 드래프트 게놈 어셈블리 생성)

  • Won, JungIm;Kong, JinHwa;Huh, Sun;Yoon, JeeHee
    • KIISE Transactions on Computing Practices
    • /
    • v.20 no.9
    • /
    • pp.513-518
    • /
    • 2014
  • It has become possible for small scale laboratories to interpret large scale genomic DNA, thanks to the reduction of the sequencing cost by the development of next generation sequencing (NGS). De novo assembly is a method which creates a putative original sequence by reconstructing reads without using a reference sequence. There have been various study results on de novo assembly, however, it is still difficult to get the desired results even by using the same assembly procedures and the analysis tools which were suggested in the studies reported. This is mainly because there are no specific guidelines for the assembly procedures or know-hows for the use of such analysis tools. In this study, to resolve these problems, we introduce steps to finding whole genome of an unknown DNA via NGS technology and de novo assembly, while providing the pros and cons of the various analysis tools used in each step. We used 350Mbp of Toxocara canis DNA as an application case for the detailed explanations of each stated step. We also extend our works for prediction of protein-coding genes and their functions from the draft genome sequence by comparing its homology with reference sequences of other nematodes.

The Prediction of Hypoxia Occurrence in Dangdong Bay (당동만의 빈산소 발생 예측)

  • Kang, Hoon;Kwon, Min Sun;You, Sun Jae;Kim, Jong Gu
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.26 no.1
    • /
    • pp.65-74
    • /
    • 2020
  • The purpose of this study was to investigate the physical characteristics of marine environment, and to predict the probability of the occurrence of hypoxia in the Dangdong bay. We predicted hypoxia using the logistic regression model analysis by observing the water temperature, salinity, and dissolved oxygen concentration. The analysis showed that the Brunt-Väisälä frequency which was shallow than the deep bay entrance, was higher inside the bay due to a lesser amount of fresh water inflow from the inner side of the bay, and density stratification was formed. The Richardson number, and Brunt-Väisälä frequency were very high occasionally from June to September; however, after September 2, the stratification had a tendency to decrease. Analysis of dissolved oxygen, water temperature, and salinity data observed in Dangdong bay showed that the dissolved oxygen concentration in the bottom layer was mostly affected by the temperature difference (dt) between the surface layer and bottom layer. Meanwhile, when the depth difference (dz) was set as a fixed variable, the probability of the occurrence of hypoxia varied with respect to the difference in water temperature. The depth difference (dz) was calculated to be 5 m, 10 m, 15 m, 20 m, and the difference in water temperature (dt) was found to be greater than 70 % at 8℃, 7℃, 5℃, and 3℃. This indicated that the larger the difference in depth in the bay, the smaller is the temperature difference required for the generation of hypoxia. In particular, the place in the bay, where the water depth dif erence was approximately 20 m, was found to generate hypoxia.

A Study of Global Ocean Data Assimilation using VAF (VAF 변분법을 이용한 전구 해양자료 동화 연구)

  • Ahn, Joong-Bae;Yoon, Yong-Hoon;Cho, Eek-Hyun;Oh, He-Ram
    • The Sea:JOURNAL OF THE KOREAN SOCIETY OF OCEANOGRAPHY
    • /
    • v.10 no.1
    • /
    • pp.69-78
    • /
    • 2005
  • ARCO and TAO data which supply three dimensional global ocean information are assimilated to the background field from a general circulation model, MOM3. Using a variational Analysis using Filter (VAF), which is a spatial variational filter designed to reduce computational time and space efficiently and economically, observed ARGO and TAO data are assimilated to the OGCM-generated background sea temperature for the generation of initial condition of the model. For the assessment of the assimilation impact, a comparative experiment has been done by integrating the model from different intial conditions: one from ARGO-, TAO-data assimilated initial condition and the other from background state without assimilation. The assimilated analysis field not only depicts major oceanic features more realistically but also reduces several systematic model bias that appear in every current OGCMs experiments. From the 10-month of model integrations with and without assimilated initial conditions, it is found that the major assimilated characteristics in sea temperature appeared in the initial field remain persistently throughout the integration. Such implies that the assimilated characteristics of the reduced sea temperature bias is to last in the integration without rapid restoration to the non-assimilated OGCM integration state by dispersing mass field in the form of internal gravity waves. From our analysis, it is concluded that the data assimilation method adapted in this study to MOM3 is reasonable and applicable with dynamical consistency. The success in generating initial condition with ARGO and TAO data assimilation has significant implication upon the prediction of the long-term climate and weather using ocean-atmosphere coupled model.