• Title/Summary/Keyword: heterogeneous computing resources

Search Result 79, Processing Time 0.031 seconds

A Design of Performance Management System in Heterogeneous Networks (이종 전산망의 통합성능 관리 방법)

  • Hwang, Jun;Kwon, Hyeog-In
    • The Transactions of the Korea Information Processing Society
    • /
    • v.2 no.2
    • /
    • pp.237-246
    • /
    • 1995
  • As the needs for accessing distributed informations and computing resources are increasing the need for network interconnection is growing. There are many of the network management packages including performance management tools; but most of them are not appropriate to be used in heterogeneous interconnected networks. To monitor and to control efficiently the performance of heterogeneous network, first, we have to define all performance parameters in general meaning. We need models and criteria for supporting performance analysis activities of the management system. In this study, we have designed a centralized performance management system based on the OSI management, which can be used in heterogeneous networks.

  • PDF

Multi-Objective Job Scheduling Model Based on NSGA-II for Grid Computing (그리드 컴퓨팅을 위한 NSGA-II 기반 다목적 작업 스케줄링 모델)

  • Kim, Sol-Ji;Kim, Tae-Ho;Lee, Hong-Chul
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.7
    • /
    • pp.13-23
    • /
    • 2011
  • Grid computing is a new generation computing technology which organizes virtual high-performance computing system by connecting and sharing geographically distributed heterogeneous resources, and performing large-scaled computing operations. In order to maximize the performance of grid computing, job scheduling is essential which allocates jobs to resources effectively. Many studies have been performed which minimize total completion times, etc. However, resource costs are also important, and through the minimization of resource costs, the overall performance of grid computing and economic efficiency will be improved. So in this paper, we propose a multi-objective job scheduling model considering both time and cost. This model derives from the optimal scheduling solution using NSGA-II, which is a multi objective genetic algorithm, and guarantees the effectiveness of the proposed model by executing experiments with those of existing scheduling models such as Min-Min and Max-Min models. Through experiments, we prove that the proposed scheduling model minimizes time and cost more efficiently than existing scheduling models.

Global Internet Computing Environment based on Java (자바를 기반으로 한 글로벌 인터넷 컴퓨팅 환경)

  • Kim, Hui-Cheol;Sin, Pil-Seop;Park, Yeong-Jin;Lee, Yong-Du
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.9
    • /
    • pp.2320-2331
    • /
    • 1999
  • Over the Internet, in order to utilize a collection of idle computers as a parallel computing platform, we propose a new scheme called GICE(Global Internet Computing Environment). GICE is motivated to obtain high programmability, efficient support for heterogeneous computing resources, system scalability, and finally high performance. The programming model of GICE is based on a single address space. GICE is featured with a Java based programming environment, a dynamic resource management scheme, and efficient parallel task scheduling and execution mechanisms. Based on a prototype implementation of GICE, we address the concept, feasibility, complexity and performance of Internet computing.

  • PDF

Design of A new Algorithm by Using Standard Deviation Techniques in Multi Edge Computing with IoT Application

  • HASNAIN A. ALMASHHADANI;XIAOHENG DENG;OSAMAH R. AL-HWAIDI;SARMAD T. ABDUL-SAMAD;MOHAMMED M. IBRAHM;SUHAIB N. ABDUL LATIF
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.4
    • /
    • pp.1147-1161
    • /
    • 2023
  • The Internet of Things (IoT) requires a new processing model that will allow scalability in cloud computing while reducing time delay caused by data transmission within a network. Such a model can be achieved by using resources that are closer to the user, i.e., by relying on edge computing (EC). The amount of IoT data also grows with an increase in the number of IoT devices. However, building such a flexible model within a heterogeneous environment is difficult in terms of resources. Moreover, the increasing demand for IoT services necessitates shortening time delay and response time by achieving effective load balancing. IoT devices are expected to generate huge amounts of data within a short amount of time. They will be dynamically deployed, and IoT services will be provided to EC devices or cloud servers to minimize resource costs while meeting the latency and quality of service (QoS) constraints of IoT applications when IoT devices are at the endpoint. EC is an emerging solution to the data processing problem in IoT. In this study, we improve the load balancing process and distribute resources fairly to tasks, which, in turn, will improve QoS in cloud and reduce processing time, and consequently, response time.

A Job Scheduling Scheme based on Analytic Hierarchy Process in Cloud Computing (클라우드 컴퓨팅에서 Analytic hierarchy process를 활용한 작업 스케줄링 기법)

  • Kim, Jeong-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.8
    • /
    • pp.9-15
    • /
    • 2013
  • As the resources of cloud computing are essentially heterogeneous and jobs have various characteristics, resource allocation to jobs is one of important problems. We define this issue as a multi-criteria decision-making problem. This paper proposes a priority-based job scheduling algorithm based on analytic hierarchy process (AHP). On the first step, jobs are classified based on their preferences. On the second step, response time, system utilization, and load becomes decision criteria based on the AHP algorithm. Jobs are allocated to adequate resources through their priorities that are calculated by the AHP algorithm. Through analysis and experiment of the proposed algorithm, we are to confirm that the scheme can schedule jobs as well as utilize its resource efficiently.

Modified Deep Reinforcement Learning Agent for Dynamic Resource Placement in IoT Network Slicing

  • Ros, Seyha;Tam, Prohim;Kim, Seokhoon
    • Journal of Internet Computing and Services
    • /
    • v.23 no.5
    • /
    • pp.17-23
    • /
    • 2022
  • Network slicing is a promising paradigm and significant evolution for adjusting the heterogeneous services based on different requirements by placing dynamic virtual network functions (VNF) forwarding graph (VNFFG) and orchestrating service function chaining (SFC) based on criticalities of Quality of Service (QoS) classes. In system architecture, software-defined networks (SDN), network functions virtualization (NFV), and edge computing are used to provide resourceful data view, configurable virtual resources, and control interfaces for developing the modified deep reinforcement learning agent (MDRL-A). In this paper, task requests, tolerable delays, and required resources are differentiated for input state observations to identify the non-critical/critical classes, since each user equipment can execute different QoS application services. We design intelligent slicing for handing the cross-domain resource with MDRL-A in solving network problems and eliminating resource usage. The agent interacts with controllers and orchestrators to manage the flow rule installation and physical resource allocation in NFV infrastructure (NFVI) with the proposed formulation of completion time and criticality criteria. Simulation is conducted in SDN/NFV environment and capturing the QoS performances between conventional and MDRL-A approaches.

A Grid Service based on OGSA for Process Fault Detection (프로세스 결함 검출을 위한 OGSA 기반 그리드 서비스의 설계 및 구현)

  • Kang, Yun-Hee
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2004.11a
    • /
    • pp.314-317
    • /
    • 2004
  • With the advance of network and software infrastructure, Grid-computing technology on a cluster of heterogeneous computing resources becomes pervasive. Grid computing is required a coordinated use of an assembly of distributed computers, which are linked by WAN. As the number of grid system components increases, the probability of failure in the grid computing is higher than that in a traditional parallel computing. To provide the robustness of grid applications, fault detection is critical and is essential elements in design and implementation. In this paper, a OGSA based process fault-detection services presented to provide high reliability under low network traffic environment.

  • PDF

A Game Theoretic Cross-Layer Design for Resource Allocation in Heterogeneous OFDMA Networks

  • Zarakovitis, Charilaos C.;Nikolaros, Ilias G.;Ni, Qiang
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.1 no.1
    • /
    • pp.50-64
    • /
    • 2012
  • Quality of Service (QoS) and fairness considerations are undoubtedly essential parameters that need to be considered in the design of next generation scheduling algorithms. This work presents a novel game theoretic cross-layer design that offers optimal allocation of wireless resources to heterogeneous services in Orthogonal Frequency Division Multiple Access (OFDMA) networks. The method is based on the Axioms of the Symmetric Nash Bargaining Solution (S-NBS) concept used in cooperative game theory that provides Pareto optimality and symmetrically fair resource distribution. The proposed strategies are determined via convex optimization based on a new solution methodology and by the transformation of the subcarrier indexes by means of time-sharing. Simulation comparisons to relevant schemes in the literature show that the proposed design can be successfully employed to typify ideal resource allocation for next-generation broadband wireless systems by providing enhanced performance in terms of queuing delay, fairness provisions, QoS support, and power consumption, as well as a comparable total throughput.

  • PDF

Energy and Service Level Agreement Aware Resource Allocation Heuristics for Cloud Data Centers

  • Sutha, K.;Nawaz, G.M.Kadhar
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.11
    • /
    • pp.5357-5381
    • /
    • 2018
  • Cloud computing offers a wide range of on-demand resources over the internet. Utility-based resource allocation in cloud data centers significantly increases the number of cloud users. Heavy usage of cloud data center encounters many problems such as sacrificing system performance, increasing operational cost and high-energy consumption. Therefore, the result of the system damages the environment extremely due to heavy carbon (CO2) emission. However, dynamic allocation of energy-efficient resources in cloud data centers overcomes these problems. In this paper, we have proposed Energy and Service Level Agreement (SLA) Aware Resource Allocation Heuristic Algorithms. These algorithms are essential for reducing power consumption and SLA violation without diminishing the performance and Quality-of-Service (QoS) in cloud data centers. Our proposed model is organized as follows: a) SLA violation detection model is used to prevent Virtual Machines (VMs) from overloaded and underloaded host usage; b) for reducing power consumption of VMs, we have introduced Enhanced minPower and maxUtilization (EMPMU) VM migration policy; and c) efficient utilization of cloud resources and VM placement are achieved using SLA-aware Modified Best Fit Decreasing (MBFD) algorithm. We have validated our test results using CloudSim toolkit 3.0.3. Finally, experimental results have shown better resource utilization, reduced energy consumption and SLA violation in heterogeneous dynamic cloud environment.

Optimizing Performance and Energy Efficiency in Cloud Data Centers Through SLA-Aware Consolidation of Virtualized Resources (클라우드 데이터 센터에서 가상화된 자원의 SLA-Aware 조정을 통한 성능 및 에너지 효율의 최적화)

  • Elijorde, Frank I.;Lee, Jaewan
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.1-10
    • /
    • 2014
  • The cloud computing paradigm introduced pay-per-use models in which IT services can be created and scaled on-demand. However, service providers are still concerned about the constraints imposed by their physical infrastructures. In order to keep the required QoS and achieve the goal of upholding the SLA, virtualized resources must be efficiently consolidated to maximize system throughput while keeping energy consumption at a minimum. Using ANN, we propose a predictive SLA-aware approach for consolidating virtualized resources in a cloud environment. To maintain the QoS and to establish an optimal trade-off between performance and energy efficiency, the server's utilization threshold dynamically adapts to the physical machine's resource consumption. Furthermore, resource-intensive VMs are prevented from getting underprovisioned by assigning them to hosts that are both capable and reputable. To verify the performance of our proposed approach, we compare it with non-optimized conventional approaches as well as with other previously proposed techniques in a heterogeneous cloud environment setup.