• Title/Summary/Keyword: supercomputer

Search Result 143, Processing Time 0.024 seconds

EDISON Platform to Supporting Education and Integration Research in Computational Science (계산과학 분야의 교육 및 융합연구 지원을 위한 EDISON 플랫폼)

  • Jin, Du-Seok;Jung, Young-Jin;Jung, Hoe-Kyung
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.1
    • /
    • pp.176-182
    • /
    • 2012
  • Recently, a new theoretical and methodological approach for computational science is becoming more and more popular for analyzing and solving scientific problems in various scientific disciplines and applied research. Computational science is a field of study concerned with constructing mathematical models and quantitative analysis techniques and using large computing resources to solve the problems which are difficult to approach in a physical experimentally. In this paper, we present R&D of EDISON open integration platform that allows anyone like professors, researchers, industrial workers, students etc to upload their advanced research result such as simulation SW to use and share based on the cyber infrastructure of supercomputer and network. EDISON platform, which consists of 3 tiers (EDISON application framework, EDISON middleware, and EDISON infra resources) provides Web portal for education and research in 5 areas (CFD, Chemistry, Physics, Structural Dynamics, Computational Design) and user service.

Information and Telecommunications R&D Strategy of United States federal Governments (미국 연방정부의 정보통신연구개발 전략)

  • 이근구;성종진;김장경
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.2 no.4
    • /
    • pp.447-454
    • /
    • 1998
  • The recent rapid evolution of the Internet and the increasing interest on the electronic commerce had been driven by the United States Federal Governments as strategic areas. The real meaning and effect of these technological evolution should be understood to find our right direction in current hard IMF environment wisely by analyzing the continual R&D efforts and policies of the US Federal Governments in the Information and Telecommunication areas. The typical R&D projects were the Supercomputer center program launched at 1985, HPCC program at 1992, NII project at 1993, and CIC R&D program at 1996 which were increasing in scale and budget. This paper will explain the overall key R&D efforts and policies of US Federal Governments in the industrial, science and technologies, and Information and Telecommunications areas as well as the background, concepts, and effect of 1996 Telecommunication Act as a deregulation efforts.

  • PDF

The future of bioinformntics

  • Gribskov, Michael
    • Proceedings of the Korean Society for Bioinformatics Conference
    • /
    • 2003.10a
    • /
    • pp.1-1
    • /
    • 2003
  • It is clear that computers will play a key role in the biology of the future. Even now, it is virtually impossible to keep track of the key proteins, their names and associated gene names, physical constants(e.g. binding constants, reaction constants, etc.), and hewn physical and genetic interactions without computational assistance. In this sense, computers act as an auxiliary brain, allowing one to keep track of thousands of complex molecules and their interactions. With the advent of gene expression array technology, many experiments are simply impossible without this computer assistance. In the future, as we seek to integrate the reductionist description of life provided by genomic sequencing into complex and sophisticated models of living systems, computers will play an increasingly important role in both analyzing data and generating experimentally testable hypotheses. The future of bioinformatics is thus being driven by potent technological and scientific forces. On the technological side, new experimental technologies such as microarrays, protein arrays, high-throughput expression and three-dimensional structure determination prove rapidly increasing amounts of detailed experimental information on a genomic scale. On the computational side, faster computers, ubiquitous computing systems, high-speed networks provide a powerful but rapidly changing environment of potentially immense power. The challenges we face are enormous: How do we create stable data resources when both the science and computational technology change rapidly? How do integrate and synthesize information from many disparate subdisciplines, each with their own vocabulary and viewpoint? How do we 'liberate' the scientific literature so that it can be incorporated into electronic resources? How do we take advantage of advances in computing and networking to build the international infrastructure needed to support a complete understanding of biological systems. The seeds to the solutions of these problems exist, at least partially, today. These solutions emphasize ubiquitous high-speed computation, database interoperation, federation, and integration, and the development of research networks that capture scientific knowledge rather than just the ABCs of genomic sequence. 1 will discuss a number of these solutions, with examples from existing resources, as well as area where solutions do not currently exist with a view to defining what bioinformatics and biology will look like in the future.

  • PDF

Implementation of a GPU Cluster System using Inexpensive Graphics Devices (저가의 그래픽스 장치를 이용한 GPU 클러스터 시스템 구현)

  • Lee, Jong-Min;Lee, Jung-Hwa;Kim, Seong-Woo
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.11
    • /
    • pp.1458-1466
    • /
    • 2011
  • Recently the research on GPGPU has been carried out actively as the performance of GPUs has been increased rapidly. In this paper, we propose the system architecture by benchmarking the existing supercomputer architecture for a cost-effective system using GPUs in low-cost graphics devices and implement a GPU cluster system with eight GPUs. We also make the software development environment that is suitable for the GPU cluster system and use it for the performance evaluation by implementing the n-body problem. According to its result, we found that it is efficient to use multiple GPUs when the problem size is large due to its communication cost. In addition, we could calculate up to eight million celestial bodies by applying the method of calculating block by block to mitigate the problem size constraint due to the limited resource in GPUs.

Implementation of Access Control System Suitable for Meteorological Tasks in Grid Computing Environment (그리드 컴퓨팅 환경에서 기상업무에 적합한 접근 제어 시스템 구현)

  • Na, Seung-kwon;Ju, Jae-han
    • Journal of Advanced Navigation Technology
    • /
    • v.21 no.2
    • /
    • pp.206-211
    • /
    • 2017
  • Recently computing devices by connecting to a network, grid computing, the next generation of digital neural networks that provide maximum service will connect all of the computer such as a PC or server, PDA into one giant network makes the virtual machine. Therefore, we propose the grid computing implementation model to be applied to meteorological business field as follows. First, grid computing will be used for tasks such as the development of numerical models below the mid-scale or test operations, and the final backup of the weather supercomputer. Second, the resources that will constitute grid computing are limited to business PCs and Linux servers operated by the central government considering operational efficiency. Third, the network is restricted to the LAN section, which suggests the implementation of high performance computing.

Design Plan of Secure IoT System based Common Criteria (CC 기반의 안전한 IoT 시스템 설계 방안)

  • Kim, Ju-Hun;Jung, Hyun-Mi;Cho, Han-Jin
    • Journal of the Korea Convergence Society
    • /
    • v.8 no.10
    • /
    • pp.61-66
    • /
    • 2017
  • Recently, IoT technology is rapidly developing with the keyword "Anytime, Anywhere, Convenient". In addition, security problems in IoT systems are exploding and the damage is increasing as well. In this paper, we propose a method to develop IoT system safely by using internationally recognized CC evaluation in ICT by identifying the standardization and security technology development status defining IoT system security requirements. For this purpose, IoT system and service security aspects are analyzed. Based on this, it is possible to design the security functional requirements and to demonstrate the rationale of the security objective through the correspondence relation, and it is possible to design the protection profile for the IoT system. This is a sufficient basis for the development methodology to be presented in this paper because it is used as a means of referring to the set of security requirements of administrators, developers, and users.

PC based Immersive Virtual Environment(PIVE) System by Recognizing Human Motion (인체 동작 인식을 통한 PC 기반의 몰입 형 가상 환경 시스템)

  • Oh Young-Il;Jo Kyoung-Hwan;Lee Ji-Hong
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.43 no.4 s.310
    • /
    • pp.103-112
    • /
    • 2006
  • In this paper, we propose a PC based immersive virtual environment system with expandability and compatibility in contrary to existing immersive virtual environment(IVE) systems which have been implemented by supercomputer or special computing system. The application based on commercial personal compute may have two major advantage: one is variety of resources, the other is user-friendly interface. This system intends to offer easy contact to IVE system, realistic images, and convenience. Also, the system can handle various virtual reality at real-time and make it easier to interface existing complicated haptic device. Geometric techniques are adopted to calculate and visualize the physical phenomenon to speed up the computing time. The proposed implementation method of PC based immersive virtual environment system is implemented to the example in which user move around inside of and interact with virtual office environment wearing data glove, behavior recognition devices, and HMD.

A Rendezvous Node Selection and Routing Algorithm for Mobile Wireless Sensor Network

  • Hu, Yifan;Zheng, Yi;Wu, Xiaoming;Liu, Hailin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.10
    • /
    • pp.4738-4753
    • /
    • 2018
  • Efficient rendezvous node selection and routing algorithm (RNSRA) for wireless sensor networks with mobile sink that visits rendezvous node to gather data from sensor nodes is proposed. In order to plan an optimal moving tour for mobile sink and avoid energy hole problem, we develop the RNSRA to find optimal rendezvous nodes (RN) for the mobile sink to visit. The RNSRA can select the set of RNs to act as store points for the mobile sink, and search for the optimal multi-hop path between source nodes and rendezvous node, so that the rendezvous node could gather information from sensor nodes periodically. Fitness function with several factors is calculated to find suitable RNs from sensor nodes, and the artificial bee colony optimization algorithm (ABC) is used to optimize the selection of optimal multi-hop path, in order to forward data to the nearest RN. Therefore the energy consumption of sensor nodes is minimized and balanced. Our method is validated by extensive simulations and illustrates the novel capability for maintaining the network robustness against sink moving problem, the results show that the RNSRA could reduce energy consumption by 6% and increase network lifetime by 5% as comparing with several existing algorithms.

Design and Implementation of National Supercomputing Service Framework (국가 슈퍼컴퓨팅 서비스 프레임워크의 설계 및 구현)

  • Yu, Jung-Lok;Byun, Hee-Jung;Kim, Han-Gi
    • KIISE Transactions on Computing Practices
    • /
    • v.22 no.12
    • /
    • pp.663-674
    • /
    • 2016
  • Traditional supercomputing services suffer from limited accessibility and low utilization in that users(researchers) may perform computational executions only using terminal-based command line interfaces. To address this problem, in this paper, we provide the design and implementation details of National supercomputing service framework. The proposed framework supports all the fundamental primitive functions such as user management/authentication, heterogeneous computing resource management, HPC (High Performance Computing) job management, etc. so that it enables various 3rd-party applications to be newly built on top of the proposed framework. Our framework also provides Web-based RESTful OpenAPIs and the abstraction interfaces of job schedulers (as well as bundle scheduler plug-ins, for example, LoadLeveler, Open Grid Scheduler, TORQUE) in order to easily integrate the broad spectrum of heterogeneous computing clusters. To show and validate the effectiveness of the proposed framework, we describe the best practice scenario of high energy physics Lattice-QCD as an example application.

Performance Analysis of a NOW According to the Number of Processes and Execution Time (프로세스의 수와 실행시간에 따른 NOW의 성능 분석)

  • 조수현;김영학
    • The Journal of the Korea Contents Association
    • /
    • v.2 no.3
    • /
    • pp.135-145
    • /
    • 2002
  • Recently, instead of a high-cost supercomputer, there haws been widely used a NOW system that consists of low-cost PCs and workstations connected all over the network In a NOW, performance for parallel processing depends on the computation pouter of each computer and communication time. Currently, a lot of methods have been proposed in order to increase the performance of parallel processing. However, the previous results have been studied in the view of balancing work load as the computation pouter of each computer. If a computer has multiple work precesses in a NOW, we can predict a decrease of communication tire needed in message passing, Therefore, in this paper, we analyzes factors of improving the performance in the view of work precesses, and evaluates experimently an effect on total performance as the number of work processes increases. Also, we propose a new broadcasting method to be used in experiment of this paper. This paper uses the LAM/MPI for an experimental evaluation.

  • PDF