• Title/Summary/Keyword: Capacity of server

Search Result 240, Processing Time 0.035 seconds

Self-Configuration System based on Context Adaptiveness (상황적응기능기반 자가구성 시스템)

  • Lee, Seung-Hwa;Lee, Eun-Seok
    • The KIPS Transactions:PartD
    • /
    • v.12D no.4 s.100
    • /
    • pp.647-656
    • /
    • 2005
  • This paper proposes an adaptive resource self-management system that collects system resources, user information, and usage patterns as context information for utilization in self-configuration. This system ill ease the system maintenance burden on users by automation of large part of configuration tasks such as install, reconfiguration and update, and will also decrease cost and errors. Working from the gathered context information, this system allows users to select appropriate components and install them for user's system context. This also offers a more personalized configuration setting by using user's existing application setting and usage pattern. To avoid the overload on central server to transfer and manage related files, we employ Peer-to-Peer method. h prototype was developed to evaluate the system and a comparison study with the conventional methods of manual configuration and MS-IBM systems was conducted to validate the proposed system in terms of functional capacity, install time and etc.

An Efficient Peer-to-Peer Streaming Scheme Based on a Push-Mesh Structure (푸시-메시 구조 기반의 효율적인 피어투피어 스트리밍 기법)

  • Kim, Jin-Sung;Kim, Dong-Il;Kim, Eun-Sam;Pae, Sung-Il
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.3
    • /
    • pp.81-89
    • /
    • 2010
  • The research on peer-to-peer streaming schemes has largely focused on tree-push and mesh-pull structures. However, the tree-push structure has a defect that the tree restructuring time is long, and the mesh-pull structure has long startup delay and lag time from source servers. In this paper, we propose a new peer-to-peer live streaming scheme based on a push-mesh structure that takes advantages of tree-push and mesh-pull structure simultaneously. This structure basically provides the mesh-pull mechanism for data transmission and utilizes peers with high network upload capacity. It also supports the push mechanism along with paths from a source server, super peers, and selected general peers. By NS-2 simulation experiments, we finally show that our proposed scheme can achieve shorter startup delay than the mesh-pull structure, similar lag time to tree-push structure and best playback continuity among the three schemes.

Improving Fault Tolerance for High-capacity Shared Distributed File Systems using the Rotational Lease Under Network Partitioning (대용량 공유 분산 화일 시스템에서 망 분할 시 순환 리스를 사용한 고장 감내성 향상)

  • Tak, Byung-Chul;Chung, Yon-Dohn;Kim, Myoung-Ho
    • Journal of KIISE:Databases
    • /
    • v.32 no.6
    • /
    • pp.616-627
    • /
    • 2005
  • In the shared storage file system, systems can directly access the shared storage device through specialized data-only subnetwork unlike in the network attached file server system. In this shared-storage architecture, data consistency is maintained by some designated set of lock servers which use control network to send and receive the lock information. Furthermore, lease mechanism is introduced to cope with the control network failure. But when the control network is partitioned, participating systems can no longer make progress after the lease term expires until the network recovers. This paper addresses this limitation and proposes a method that allows partitioned systems to make progress under the partition of control network. The proposed method works in a manner that each participating system is rotationally given a predefined lease term periodically. It is also shown that the proposed mechanism always preserves data consistency.

Accuracy Analysis of FKP for Public Surveying and Cadastral Resurvey (공공측량 및 지적재조사 사업 적용을 위한 FKP 정밀도 분석)

  • Park, Jin Sol;Han, Joong-Hee;Kwon, Jay Hyoun;Shin, Han Sup
    • Spatial Information Research
    • /
    • v.22 no.3
    • /
    • pp.23-24
    • /
    • 2014
  • NGII (National Geographic Information Institute) has been providing VRS (Virtual Reference Station) service so that could determine precise positioning in real time since 2007. However, since the VRS service has to maintain the connected status with VRS server, the number of users who can use VRS service are limited by capacity of VRS server. To solve this problem, NGII has been providing FKP (Virtual Reference Station) service using one way telecommunication from November 1, 2012. Therefore, it is predicted that the usage of FKP service will increase in public surveying and cadastral resurveying in the future. However, the studies with respect to analysis of FKP precision for applying to public surveying and cadastral resurveying is not conducted enough. In this study, to analyse the application possibility of FKP on the public surveying and cadastral resurveying, the two kind analysis were performed. First is the analysis of accuracy according to the configuration of reference station of FKP and VRS. One is consisted of same reference stations, another is consisted of different reference stations. Second is the accuracy anlalysis of horizontal and vertical positioning acquiring VRS and FKP data in various measurement environment based on VRS regulation. Result of first study, Positioning accuracy according to the configuration of the reference stations satisfies related regulation. However, accuracy of FKP in case of different reference stations is worse than in case of same reference stations.. The result of second test shows that the horizontal precision of FKP and VRS in good measurement environment satisfy the allowed precision. However, in some case, horizontal precision of FKP and VRS in poor measurement environment exceed the allowed precision. In addition, the number of exceeding the allowed precision in the FKP is more than the VRS. The vertical precision of the VRS satisfy related work provision. In conclusion, the result of this study shows that the FKP only in open area should be used for public survey and cadastral resurvey. Therefore the additional studies with respect to the improvement of FKP precision should be conducted.

Design of MAHA Supercomputing System for Human Genome Analysis (대용량 유전체 분석을 위한 고성능 컴퓨팅 시스템 MAHA)

  • Kim, Young Woo;Kim, Hong-Yeon;Bae, Seungjo;Kim, Hag-Young;Woo, Young-Choon;Park, Soo-Jun;Choi, Wan
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.2
    • /
    • pp.81-90
    • /
    • 2013
  • During the past decade, many changes and attempts have been tried and are continued developing new technologies in the computing area. The brick wall in computing area, especially power wall, changes computing paradigm from computing hardwares including processor and system architecture to programming environment and application usage. The high performance computing (HPC) area, especially, has been experienced catastrophic changes, and it is now considered as a key to the national competitiveness. In the late 2000's, many leading countries rushed to develop Exascale supercomputing systems, and as a results tens of PetaFLOPS system are prevalent now. In Korea, ICT is well developed and Korea is considered as a one of leading countries in the world, but not for supercomputing area. In this paper, we describe architecture design of MAHA supercomputing system which is aimed to develop 300 TeraFLOPS system for bio-informatics applications like human genome analysis and protein-protein docking. MAHA supercomputing system is consists of four major parts - computing hardware, file system, system software and bio-applications. MAHA supercomputing system is designed to utilize heterogeneous computing accelerators (co-processors like GPGPUs and MICs) to get more performance/$, performance/area, and performance/power. To provide high speed data movement and large capacity, MAHA file system is designed to have asymmetric cluster architecture, and consists of metadata server, data server, and client file system on top of SSD and MAID storage servers. MAHA system softwares are designed to provide user-friendliness and easy-to-use based on integrated system management component - like Bio Workflow management, Integrated Cluster management and Heterogeneous Resource management. MAHA supercomputing system was first installed in Dec., 2011. The theoretical performance of MAHA system was 50 TeraFLOPS and measured performance of 30.3 TeraFLOPS with 32 computing nodes. MAHA system will be upgraded to have 100 TeraFLOPS performance at Jan., 2013.

SSD-based RAID-6 System Architecture for Reliability and Performance Enhancement (신뢰성 향상과 성능개선을 위해 다양한 Erasure 코드를 적용한 SSD 기반 RAID-6 시스템 구조)

  • Song, Jae-Seok;Huh, Joon-Moo;Yang, Yu-Seok;Kim, Deok-Hwan
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.47 no.6
    • /
    • pp.47-56
    • /
    • 2010
  • HDD-based RAIDs have been used in high-capacity storage systems for traditional data server. However, their data reliability are relatively low and they consume lots of power since hard disk drive is weak on shock and its power consumption is high due to frequent spindle motor operation. Therefore, this paper presents new SSD based RAID system architecture using various erasure codes. The proposed methode applys Reed-Solomon, EVENODD, and Liberation coding schemes onto file system level and device driver level, respectively. Besides, it uses data allocation method to minimize the side effect of reducing the lifespan of SSD. Detail experimental results show that Liberation code increase wear-leveling rates of SSD based RAID-6 more than other codes. The SSD based RAID system applying erasure codes at the device driver level shows better performance than that at the file system level. I/O performance of RAID-6 system using SSD is 4.5%~8.5% higher than that of using HDD and the power consumption of the RAID system using SSD is 18%~40% less than that of using HDD.

Incentive Design Considerations for Free-riding Prevention in Cooperative Distributed Systems (협조적 분산시스템 환경에서 무임승차 방지를 위한 인센티브 디자인 고려사항 도출에 관한 연구)

  • Shin, Kyu-Yong;Yoo, Jin-Cheol;Lee, Jong-Deog;Park, Byoung-Chul
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.7
    • /
    • pp.137-148
    • /
    • 2011
  • Different from the traditional client-server model, it is possible for participants in a cooperative distributed system to get quality services regardless of the number of participants in the system since they voluntarily pool or share their resources in order to achieve their common goal. However, some selfish participants try to avoid providing their resources while still enjoying the benefits offered by the system, which is termed free-riding. The results of free-riding in cooperative distributed systems lead to system collapse because the system capacity (per participant) decreases as the number of free-riders increases, widely known as the tragedy of commons. As a consequence, designing an efficient incentive mechanism to prevent free-riding is mandatory for a successful cooperative distributed system. Because of the importance of incentive mechanisms in cooperative distributed system, a myriad of incentives mechanisms have been proposed without a standard for performance evaluation. This paper draws general incentive design considerations which can be used as performance metrics through an extensive survey on this literature, providing future researchers with guidelines for the effective incentive design in cooperative distributed systems.

Study on Development of HDD Integrity Verification System using FirmOS (FirmOS를 이용한 HDD 무결성 검사 시스템 개발에 관한 연구)

  • Yeom, Jae-Hwan;Oh, Se-Jin;Roh, Duk-Gyoo;Jung, Dong-Kyu;Hwang, Ju-Yeon;Oh, Chungsik;Kim, Hyo-Ryoung;Shin, Jae-Sik
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.18 no.2
    • /
    • pp.55-61
    • /
    • 2017
  • In radio astronomy, high-capacity HDDs are being used to save huge amounts of HDDs in order to record the observational data. For VLBI observations, observational speeds increase and huge amounts of observational data must be stored as they expand to broadband. As the HDD is frequently used, the number of failures occurred, and then it takes a lot of time to recover it. In addition, if a failed HDD is continuously used, observational data loss occurs. And it costs a lot of money to buy a new HDD. In this study, we developed the integrity verification system of the Serial ATA HDD using FirmOS. The FirmOS is an OS that has been developed to function exclusively for specific purposes on a system having a general server board and CPU. The developed system performs the process of writing and reading specific patterns of data in a physical area of the SATA HDD based on a FirmOS. In addition, we introduced a method to investigate the integrity of HDD integrity by comparing it with the stored pattern data from the HDD controller. Using the developed system, it was easy to determine whether the disk pack used in VLBI observations has error or not, and it is very useful to improve the observation efficiency. This paper introduces the detail for the design, configuration, testing, etc. of the SATA HDD integrity verification system developed.

  • PDF

Using IoT and Apache Spark Analysis Technique to Monitoring Architecture Model for Fruit Harvest Region (IoT 기반 Apache Spark 분석기법을 이용한 과수 수확 불량 영역 모니터링 아키텍처 모델)

  • Oh, Jung Won;Kim, Hangkon
    • Smart Media Journal
    • /
    • v.6 no.4
    • /
    • pp.58-64
    • /
    • 2017
  • Modern society is characterized by rapid increase in world population, aging of the rural population, decrease of cultivation area due to industrialization. The food problem is becoming an important issue with the farmers and becomes rural. Recently, the researches about the field of the smart farm are actively carried out to increase the profit of the rural area. The existing smart farm researches mainly monitor the cultivation environment of the crops in the greenhouse, another way like in the case of poor quality t is being studied that the system to control cultivation environmental factors is automatically activated to keep the cultivation environment of crops in optimum conditions. The researches focus on the crops cultivated indoors, and there are not many studies applied to the cultivation environment of crops grown outside. In this paper, we propose a method to improve the harvestability of poor areas by monitoring the areas with bad harvests by using big data analysis, by precisely predicting the harvest timing of fruit trees growing in orchards. Factors besides for harvesting include fruit color information and fruit weight information We suggest that a harvest correlation factor data collected in real time. It is analyzed using the Apache Spark engine. The Apache Spark engine has excellent performance in real-time data analysis as well as high capacity batch data analysis. User device receiving service supports PC user and smartphone users. A sensing data receiving device purpose Arduino, because it requires only simple processing to receive a sensed data and transmit it to the server. It regulates a harvest time of fruit which produces a good quality fruit, it is needful to determine a poor harvest area or concentrate a bad area. In this paper, we also present an architectural model to determine the bad areas of fruit harvest using strong data analysis.

A Delay-Bandwidth Normalized Scheduling Model with Service Rate Guarantees (서비스율을 보장하는 지연시간-대역폭 정규화 스케줄링 모델)

  • Lee, Ju-Hyun;Hwang, Ho-Young;Lee, Chang-Gun;Min, Sang-Lyul
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.34 no.10
    • /
    • pp.529-538
    • /
    • 2007
  • Fair Queueing algorithms based on Generalized Processor Sharing (GPS) not only guarantee sessions with service rate and delay, but also provide sessions with instantaneous fair sharing. This fair sharing distributes server capacity to currently backlogged sessions in proportion to their weights without regard to the amount of service that the sessions received in the past. From a long-term perspective, the instantaneous fair sharing leads to a different quality of service in terms of delay and bandwidth to sessions with the same weight depending on their traffic pattern. To minimize such long-term unfairness, we propose a delay-bandwidth normalization model that defines the concept of value of service (VoS) from the aspect of both delay and bandwidth. A model and a packet-by-packet scheduling algorithm are proposed to realize the VoS concept. Performance comparisons between the proposed algorithm and algorithms based on fair queueing and service curve show that the proposed algorithm provides better long-term fairness among sessions and that is more adaptive to dynamic traffic characteristics without compromising its service rate and delay guarantees.