• Title/Summary/Keyword: Construction Resource Management

Search Result 496, Processing Time 0.024 seconds

Volume Rendering System of e-Science Electron Microscopy using Grid (Gird를 이용한 e-사이언스 전자현미경 볼륨 랜더링 시스템)

  • Jeong, Won-Gu;Jeong, Jong-Man;Lee, Ho;Choe, Sang-Su;Ahn, Young-heon;Hur, Man-Hoi;Kim, Jay;Kim, Eunsung;Jung, Im Y.;Yeom, Heon Y.;Cho, Kum Won;Kweon, Hee-Seok
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2007.11a
    • /
    • pp.560-564
    • /
    • 2007
  • Korea Basic Science Institute(KBSI) has three general electron microscopes including High Voltage Electron Microscope(HVEM) which is the only one in Korea. Observed images through an electron microscope are what they are tilted by each step and saved, offering the more better circumstances for observers, a reconstruction to 3D could be a essential process. In this process, a warping method decreases distortions maximumly of avoided parts of a camera's focus. All these image treatment processes and 3D reconstruction processes are based on an accompaniment of a highly efficient computer, a number of Grid Node Personal computers share this process in a short time and dispose of it. Grid Node Personal computers' purpose is to make an owner can share different each other and various computing resources efficiently and also Grid Node Personal computers is applying to solve problems like a role scheduling needed for a constructing system, a resource management, a security, a capacity measurement, a condition monitoring and so on. Grid Node Personal computers accomplish roles of a highly efficient computer that general individuals felt hard to use, moreover, a image treatment using the warping method becomes a foundation for reconstructing to more closer shape with an real object of observation. Construction of the electron microscope volume 랜더링 system based on Grid Node Personal computer through the warping process can offer more convenient and speedy experiment circumstances to observers, and makes them meet with experiment outcome that is similar to real shapes and is easy to understand.

  • PDF

Evaluation of Function of Upland Farming for Preventing Flood and Fostering Water Resources (밭농사의 수자원 함양과 홍수조절 기능에 대한 계량화 평가)

  • Hyun, Byung-Keun;Kim, Moo-Sung;Eom, Ki-Cheol;Kang, Ki-Kyung;Yun, Hong-Bae;Seo, Myung-Cheol
    • Korean Journal of Soil Science and Fertilizer
    • /
    • v.36 no.3
    • /
    • pp.163-179
    • /
    • 2003
  • Multifunctionality of agriculture which is not traded on the market now has been an important international issue in that it environmental and public benefits. We carried out to modify and to update the function of upland farming on flood prevention and fostering water resources. Economic values of environmental benefits were evaluated by replacement cost methods. Models to evaluate the function of preventing flood were selected as: (1)precipitation(flood-inducing) - runoff(A), (2) soil depth ${\times}$ soil air phase, (3) precipitation (flood-inducing) - runoff(B), (4) soil depth ${\times}$ effective porosity of soil. Models to estimate the function of fostering water resources were (1) saturated hydraulic conductivity (Ks) ${\times}$ duration of saturation(days) ${\times}$ (1-ratio of water flow directly into river), (2) precipitation ${\times}$ ratio of water fostered by rain resources ${\times}$ (area of upland/total land area), and (3) soil water retention quantity(under standing crop or tree) - SWRQ(in bare soil). Function of preventing flood was $883Mg\;ha^{-1}$ of water per year and 645 million Mg for the whole upland area. Function of fostering water resources was $94.1Mg\;ha^{-1}$ of water per year and 69 million Mg for the whole upland area. The value of flood-preventing function evaluated by replacement cost methods was estimated 1,428 billion won per year as compared to the cost for dam construction. The value of water resource fostering were estimated 8.6 billion won in the price of living water.

Research on the Digital Twin Policy for the Utilization of Administrative Services (행정서비스 활용을 위한 디지털 트윈 정책 연구)

  • Jina Ok;Soonduck Yoo;Hyojin Jung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.3
    • /
    • pp.35-43
    • /
    • 2023
  • The purpose of this study is to research digital twin policies for the use of administrative services. The study was conducted through a mobile survey of 1,000 participants, and the results are as follows. First, in order to utilize digital twin technology, it is necessary to first identify appropriate services that can be applied from the perspective of Gyeonggi Province. Efforts to identify digital twin services that are suitable for Gyeonggi Province's field work should be prioritized, and this should lead to increased efficiency in the work. Second, Gyeonggi Province's digital twin administrative services should prevent duplication with central government projects and establish a model that can be connected and utilized. It should be driven around current issues in Gyeonggi Province and the demands of citizens for administrative services. Third, to develop Gyeonggi Province's digital twin administrative services, a standard model development plan through participation in pilot projects should be considered. Gyeonggi Province should lead the project as the main agency and promote it through a collaborative project agreement. It is suggested that a support system for the overall project be established through the Gyeonggi Province Digital Twin Advisory Committee. Fourth, relevant regulations and systems for the construction, operation, and management of dedicated departments and administrative services should be established. To achieve the realization of digital twins in Gyeonggi Province, a dedicated organization that can perform various roles in project promotion and operation, as well as legal and institutional improvements, is necessary. To designate a dedicated organization, it is necessary to consider expanding and reorganizing existing departments and evaluating the operation of newly established departments. The limitation of this study is that it only surveyed participants from Gyeonggi Province, and it is recommended that future research be conducted nationwide. The expected effect of this study is that it can serve as a foundational resource for applying digital twin services to public work.

Gap-Filling of Sentinel-2 NDVI Using Sentinel-1 Radar Vegetation Indices and AutoML (Sentinel-1 레이더 식생지수와 AutoML을 이용한 Sentinel-2 NDVI 결측화소 복원)

  • Youjeong Youn;Jonggu Kang;Seoyeon Kim;Yemin Jeong;Soyeon Choi;Yungyo Im;Youngmin Seo;Myoungsoo Won;Junghwa Chun;Kyungmin Kim;Keunchang Jang;Joongbin Lim;Yangwon Lee
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.6_1
    • /
    • pp.1341-1352
    • /
    • 2023
  • The normalized difference vegetation index (NDVI) derived from satellite images is a crucial tool to monitor forests and agriculture for broad areas because the periodic acquisition of the data is ensured. However, optical sensor-based vegetation indices(VI) are not accessible in some areas covered by clouds. This paper presented a synthetic aperture radar (SAR) based approach to retrieval of the optical sensor-based NDVI using machine learning. SAR system can observe the land surface day and night in all weather conditions. Radar vegetation indices (RVI) from the Sentinel-1 vertical-vertical (VV) and vertical-horizontal (VH) polarizations, surface elevation, and air temperature are used as the input features for an automated machine learning (AutoML) model to conduct the gap-filling of the Sentinel-2 NDVI. The mean bias error (MAE) was 7.214E-05, and the correlation coefficient (CC) was 0.878, demonstrating the feasibility of the proposed method. This approach can be applied to gap-free nationwide NDVI construction using Sentinel-1 and Sentinel-2 images for environmental monitoring and resource management.

Research Trend Analysis Using Bibliographic Information and Citations of Cloud Computing Articles: Application of Social Network Analysis (클라우드 컴퓨팅 관련 논문의 서지정보 및 인용정보를 활용한 연구 동향 분석: 사회 네트워크 분석의 활용)

  • Kim, Dongsung;Kim, Jongwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.1
    • /
    • pp.195-211
    • /
    • 2014
  • Cloud computing services provide IT resources as services on demand. This is considered a key concept, which will lead a shift from an ownership-based paradigm to a new pay-for-use paradigm, which can reduce the fixed cost for IT resources, and improve flexibility and scalability. As IT services, cloud services have evolved from early similar computing concepts such as network computing, utility computing, server-based computing, and grid computing. So research into cloud computing is highly related to and combined with various relevant computing research areas. To seek promising research issues and topics in cloud computing, it is necessary to understand the research trends in cloud computing more comprehensively. In this study, we collect bibliographic information and citation information for cloud computing related research papers published in major international journals from 1994 to 2012, and analyzes macroscopic trends and network changes to citation relationships among papers and the co-occurrence relationships of key words by utilizing social network analysis measures. Through the analysis, we can identify the relationships and connections among research topics in cloud computing related areas, and highlight new potential research topics. In addition, we visualize dynamic changes of research topics relating to cloud computing using a proposed cloud computing "research trend map." A research trend map visualizes positions of research topics in two-dimensional space. Frequencies of key words (X-axis) and the rates of increase in the degree centrality of key words (Y-axis) are used as the two dimensions of the research trend map. Based on the values of the two dimensions, the two dimensional space of a research map is divided into four areas: maturation, growth, promising, and decline. An area with high keyword frequency, but low rates of increase of degree centrality is defined as a mature technology area; the area where both keyword frequency and the increase rate of degree centrality are high is defined as a growth technology area; the area where the keyword frequency is low, but the rate of increase in the degree centrality is high is defined as a promising technology area; and the area where both keyword frequency and the rate of degree centrality are low is defined as a declining technology area. Based on this method, cloud computing research trend maps make it possible to easily grasp the main research trends in cloud computing, and to explain the evolution of research topics. According to the results of an analysis of citation relationships, research papers on security, distributed processing, and optical networking for cloud computing are on the top based on the page-rank measure. From the analysis of key words in research papers, cloud computing and grid computing showed high centrality in 2009, and key words dealing with main elemental technologies such as data outsourcing, error detection methods, and infrastructure construction showed high centrality in 2010~2011. In 2012, security, virtualization, and resource management showed high centrality. Moreover, it was found that the interest in the technical issues of cloud computing increases gradually. From annual cloud computing research trend maps, it was verified that security is located in the promising area, virtualization has moved from the promising area to the growth area, and grid computing and distributed system has moved to the declining area. The study results indicate that distributed systems and grid computing received a lot of attention as similar computing paradigms in the early stage of cloud computing research. The early stage of cloud computing was a period focused on understanding and investigating cloud computing as an emergent technology, linking to relevant established computing concepts. After the early stage, security and virtualization technologies became main issues in cloud computing, which is reflected in the movement of security and virtualization technologies from the promising area to the growth area in the cloud computing research trend maps. Moreover, this study revealed that current research in cloud computing has rapidly transferred from a focus on technical issues to for a focus on application issues, such as SLAs (Service Level Agreements).

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.