• 제목/요약/키워드: Task scheduling algorithm

검색결과 207건 처리시간 0.03초

Task Scheduling Algorithm for the Communication, Ocean, and Meteorological Satellite

  • Lee, Soo-Jeon;Jung, Won-Chan;Kim, Jae-Hoon
    • ETRI Journal
    • /
    • 제30권1호
    • /
    • pp.1-12
    • /
    • 2008
  • In this paper, we propose an efficient single-resource task scheduling algorithm for the Communication, Ocean, and Meteorological Satellite. Among general satellite planning functions such as constraint check, priority check, and task scheduling, this paper focuses on the task scheduling algorithm, which resolves conflict among tasks which have an exclusion relation and the same priority. The goal of the proposed task scheduling algorithm is to maximize the number of tasks that can be scheduled. The rationale of the algorithm is that a discarded task can be scheduled instead of a previously selected one depending on the expected benefit acquired by doing so. The evaluation results show that the proposed algorithm enhances the number of tasks that can be scheduled considerably.

  • PDF

Task Scheduling and Resource Management Strategy for Edge Cloud Computing Using Improved Genetic Algorithm

  • Xiuye Yin;Liyong Chen
    • Journal of Information Processing Systems
    • /
    • 제19권4호
    • /
    • pp.450-464
    • /
    • 2023
  • To address the problems of large system overhead and low timeliness when dealing with task scheduling in mobile edge cloud computing, a task scheduling and resource management strategy for edge cloud computing based on an improved genetic algorithm was proposed. First, a user task scheduling system model based on edge cloud computing was constructed using the Shannon theorem, including calculation, communication, and network models. In addition, a multi-objective optimization model, including delay and energy consumption, was constructed to minimize the sum of two weights. Finally, the selection, crossover, and mutation operations of the genetic algorithm were improved using the best reservation selection algorithm and normal distribution crossover operator. Furthermore, an improved legacy algorithm was selected to deal with the multi-objective problem and acquire the optimal solution, that is, the best computing task scheduling scheme. The experimental analysis of the proposed strategy based on the MATLAB simulation platform shows that its energy loss does not exceed 50 J, and the time delay is 23.2 ms, which are better than those of other comparison strategies.

실시간 시스템에서 태스크 이용율을 이용한 스케줄링 가능성 검사 (Schedulability Test using task utilization in Real-Time system)

  • 임경현;서재현;박경우
    • 인터넷정보학회논문지
    • /
    • 제6권2호
    • /
    • pp.25-35
    • /
    • 2005
  • 실시간 스케줄링 알고리즘에서는 비율단조(RM) 스케줄링 알고리즘과 마감시한(EDF) 스케줄링 알고리즘이 가장 일반적으로 사용되고 있다. 이러한 알고리즘에서는 태스크 집합의 전체 이용율 값을 가지고 수행 가능성을 판별하였다. 그러나 임의의 태스크에서 이용율 값이 초과되면 개별 태스크의 한계성을 전혀 예측할 수 없는 문제점이 있었다. 본 논문에서 제안한 알고리즘은 이용율 값이 초과한 태스크를 예측하고, 개별 태스크의 이용율 값을 기반으로 스케줄링 가능성 여부를 판단하는 방법을 제시하였다. 또한, 실시간 시스템에서 스케줄링 가능성 검사의 한계성을 시뮬레이션을 통해 예측하고 결과를 분석하였다.

  • PDF

최소의 스케줄 길이를 유지하는 결함 허용 태스크 스케줄링 알고리즘 (A Fault-tolerant Task Scheduling Algorithm Supporting the Minimum Schedule Length)

  • 민병준
    • 한국정보처리학회논문지
    • /
    • 제7권4호
    • /
    • pp.1201-1210
    • /
    • 2000
  • In order to tolerate faults which may occur during the execution of distributed tasks in high-performance parallel computer systems, tasks are duplicated on different processors. In this paper, by utilizing the task duplication based scheduling algorithm, a new task scheduling algorithm which duplicates each task on more than two different processors with the minimum schedule length is presented, and the number of processors required for the duplication is analyzed with the ratio of communication cost to computation time and the workload of the system. A simulation with various task graphs reveals that the number of processors required for the full-duplex fault-tolerant task scheduling with the obtainable minimum schedule length increases about 30% to 75% when compared with that of the task duplication based scheduling algorithm.

  • PDF

Long-Term Container Allocation via Optimized Task Scheduling Through Deep Learning (OTS-DL) And High-Level Security

  • Muthakshi S;Mahesh K
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제17권4호
    • /
    • pp.1258-1275
    • /
    • 2023
  • Cloud computing is a new technology that has adapted to the traditional way of service providing. Service providers are responsible for managing the allocation of resources. Selecting suitable containers and bandwidth for job scheduling has been a challenging task for the service providers. There are several existing systems that have introduced many algorithms for resource allocation. To overcome these challenges, the proposed system introduces an Optimized Task Scheduling Algorithm with Deep Learning (OTS-DL). When a job is assigned to a Cloud Service Provider (CSP), the containers are allocated automatically. The article segregates the containers as' Long-Term Container (LTC)' and 'Short-Term Container (STC)' for resource allocation. The system leverages an 'Optimized Task Scheduling Algorithm' to maximize the resource utilisation that initially inquires for micro-task and macro-task dependencies. The bottleneck task is chosen and acted upon accordingly. Further, the system initializes a 'Deep Learning' (DL) for implementing all the progressive steps of job scheduling in the cloud. Further, to overcome container attacks and errors, the system formulates a Container Convergence (Fault Tolerance) theory with high-level security. The results demonstrate that the used optimization algorithm is more effective for implementing a complete resource allocation and solving the large-scale optimization problem of resource allocation and security issues.

실시간 시스템에서 성능 향상 기법 (Enhanced Technique for Performance in Real Time Systems)

  • 김명준
    • 한국IT서비스학회지
    • /
    • 제16권3호
    • /
    • pp.103-111
    • /
    • 2017
  • The real time scheduling is a key research area in high performance computing and has been a source of challenging problems. A periodic task is an infinite sequence of task instance where each job of a task comes in a regular period. The RMS (Rate Monotonic Scheduling) algorithm has the advantage of a strong theoretical foundation and holds out the promise of reducing the need for exhaustive testing of the scheduling. Many real-time systems built in the past based their scheduling on the Cyclic Executive Model because it produces predictable schedules which facilitate exhaustive testing. In this work we propose hybrid scheduling method which combines features of both of these scheduling algorithms. The original rate monotonic scheduling algorithm didn't consider the uniform sampling tasks in the real time systems. We have enumerated some issues when the RMS is applied to our hybrid scheduling method. We found the scheduling bound for the hard real-time systems which include the uniform sampling tasks. The suggested hybrid scheduling algorithm turns out to have some advantages from the point of view of the real time system designer, and is particularly useful in the context of large critical systems. Our algorithm can be useful for real time system designer who must guarantee the hard real time tasks.

Multi-factor Evolution for Large-scale Multi-objective Cloud Task Scheduling

  • Tianhao Zhao;Linjie Wu;Di Wu;Jianwei Li;Zhihua Cui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제17권4호
    • /
    • pp.1100-1122
    • /
    • 2023
  • Scheduling user-submitted cloud tasks to the appropriate virtual machine (VM) in cloud computing is critical for cloud providers. However, as the demand for cloud resources from user tasks continues to grow, current evolutionary algorithms (EAs) cannot satisfy the optimal solution of large-scale cloud task scheduling problems. In this paper, we first construct a large- scale multi-objective cloud task problem considering the time and cost functions. Second, a multi-objective optimization algorithm based on multi-factor optimization (MFO) is proposed to solve the established problem. This algorithm solves by decomposing the large-scale optimization problem into multiple optimization subproblems. This reduces the computational burden of the algorithm. Later, the introduction of the MFO strategy provides the algorithm with a parallel evolutionary paradigm for multiple subpopulations of implicit knowledge transfer. Finally, simulation experiments and comparisons are performed on a large-scale task scheduling test set on the CloudSim platform. Experimental results show that our algorithm can obtain the best scheduling solution while maintaining good results of the objective function compared with other optimization algorithms.

위상 정렬과 여유 시간 기반 주기 및 실시간 비주기 태스크 스케줄링 알고리즘 (Periodic and Real-Time Aperiodic Task Scheduling Algorithm based on Topological Sort and Residual Time)

  • 김시완;박홍성
    • 제어로봇시스템학회논문지
    • /
    • 제18권4호
    • /
    • pp.302-307
    • /
    • 2012
  • Real-time systems perform periodic tasks and real-time aperiodic tasks such as alarm processing. Especially the periodic tasks included in control systems such as robots have precedence relationships among them. This paper proposes a new scheduling algorithm based on topological sort and residual time. The precedence relationships among periodic tasks are translated to the priorities of the tasks using topological sort algorithm. During the execution of the system the proposed scheduling algorithm decides on whether or not a newly arrived real-time aperiodic task is accepted based on residual time whenever the aperiodic task such as alarm is arrived. The proposed algorithm is validated using examples.

An On-line Algorithm to Search Minimum Total Error for Imprecise Real-time Tasks with 0/1 Constraint

  • Song Gi-Hyeon
    • 한국멀티미디어학회논문지
    • /
    • 제8권12호
    • /
    • pp.1589-1596
    • /
    • 2005
  • The imprecise real-time system provides flexibility in scheduling time-critical tasks. Most scheduling problems of satisfying both 0/1 constraint and timing constraints, while the total error is minimized, are NP complete when the optional tasks have arbitrary processing times. Liu suggested a reasonable strategy of scheduling tasks with the 0/1 constraint on uniprocessors for minimizing the total error. Song et al suggested a reasonable strategy of scheduling tasks with the 0/1 constraint on multiprocessors for minimizing the total error. But, these algorithms are all off-line algorithms. On the other hand, in the case of on line scheduling, Shih and Liu proposed the NORA algorithm which can find a schedule with the minimum total error for a task system consisting solely of on-line tasks that are ready upon arrival. But, for the task system with 0/1 constraint, it has not been known whether the NORA algorithm can be optimal or not in the sense that it guarantees all mandatory tasks are completed by their deadlines and the total error is minimized. So, this paper suggests an optimal algorithm to search minimum total error for the imprecise on-line real-time task system with 0/1 constraint. Furthermore, the proposed algorithm has the same complexity, O(N log N), as the NORA algorithm, where N is the number of tasks.

  • PDF

이질 시스템에서 통신 시간을 고려한 효율적인 복제 기반 태스크 스케줄링 (Efficient Duplication Based Task Scheduling with Communication Cost in Heterogeneous Systems)

  • 윤완오;백정규;신광식;정진하;최상방
    • 한국통신학회논문지
    • /
    • 제33권3C호
    • /
    • pp.219-233
    • /
    • 2008
  • 스케줄링의 목적은 입력 작업(DAG)에 대한 스케줄 결과 길이를 최소화하는 것이다. 이런 스케줄링 문제는 잘 알려진 '정해진 시간 내에 해결하기 어려운 문제(NP-complete)'이며 최적의 스케줄링 결과 값을 얻기 위해서는 휴리스틱으로 해결해야 한다. 선후 관계의 제약을 갖는 노드들의 스케줄링을 효율적으로 수행하기 위해 부모 노드와 이질 프로세서에 대한 정보를 고려하는 TANH(the Task duplication based scheduling Algorithm for Network of Heterogeneous systems), GDL, BIL, TDS과 같은 많은 알고리즘이 제안되었다. 본 논문은 기존의 TANH 스케줄링에서 나타나는 여러 개의 부모 노드와 이질 프로세서에 대한 다양한 경우를 충분히 고려하지 못한 점을 보안하여 향상된 스케줄링을 수행할 수 있는 DTSC (Duplication based Task Scheduling with Communication Cost in Heterogeneous Systems)알고리즘을 제안하였다. 제안된 알고리즘의 성능은 기존 TANH, GDL 알고리즘과 비교하였으며, 스케줄링의 성능 향상을 보여 주었다.