• Title/Summary/Keyword: Message Scheduling

Search Result 106, Processing Time 0.028 seconds

Performance Evaluation of Semi-Persistent Scheduling in a Narrowband LTE System for Internet of Things (사물인터넷을 위한 협대역 LTE 시스템에서의 준지속적 스케줄링의 성능 평가)

  • Kim, Sunkyung;Cha, Wonjung;So, Jaewoo;Na, Minsoo;Choi, Changsoon
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.9
    • /
    • pp.1001-1009
    • /
    • 2016
  • In LTE networks, the base station transmits control information over the physical downlink control channel (PDCCH) including scheduling grants, which are used to indicate the resources that the user equipment uses to send data to the base station. Because the size of the PDCCH message and the number of the PDCCH transmissions increase in proportion to the number of user equipments, the overhead of the PDCCH may cause serious network congestion problems in the narrowband LTE (NB-LTE) system. This paper proposes the compact PDCCH information bit allocation to reduce the size of the PDCCH message and evaluates the performance of the semi-persistent scheduling (SPS) in the NB-LTE system. The simulation results show that the SPS can significantly reduce the signaling overhead of the PDCCH and therefore increase the system utilization.

Design and Implementation of a Linux-based Message Processor to Minimize the Response-time Delay of Non-real-time Messages in Multi-core Environments (멀티코어 환경에서 비실시간 메시지의 응답시간 지연을 최소화하는 리눅스 기반 메시지 처리기의 설계 및 구현)

  • Wang, Sangho;Park, Younghun;Park, Sungyong;Kim, Seungchun;Kim, Cheolhoe;Kim, Sangjun;Jin, Cheol
    • Journal of KIISE
    • /
    • v.44 no.2
    • /
    • pp.115-123
    • /
    • 2017
  • A message processor is server software that receives non-realtime messages as well as realtime messages from clients that need to be processed within a deadline. With the recent advances of micro-processor technologies and Linux, the message processor is often implemented in Linux-based multi-core servers and it is important to use cores efficiently to maximize the performance of system in multi-core environments. Numerous research efforts on a real-time scheduler for the efficient utilization of the multi-core environments have been conducted. Typically, though, they have been conducted theoretically or via simulation, making a subsequent real-system application difficult. Moreover, many Linux-based real-time schedulers can only be used in a specific Linux version, or the Linux source code needs to be modified. This paper presents the design of a Linux-based message processor for multi-core environments that maps the threads to the cores at user level. The message processor is implemented through a modification of the traditional RM algorithm that consolidates the real-time messages into certain cores using a first-fit-based bin-packing algorithm; this minimizes the response-time delay of the non-real-time messages, while guaranteeing the violation rate of the real-time messages. To compare the performances, the message processor was implemented using the two multi-core-scheduling algorithms GSN-EDF and P-FP, which are provided by the LITMUS framework. The benchmarking results show that the response-time delay of non-real-time messages in the proposed system was improved up to a maximum of 17% to 18%.

Communication Schedule for GEN_BLOCK Redistribution (GEN_BLOCK간 재분산을 위한 통신 스케줄)

  • Yook, Hyun-Gyoo;Park, Myong-Soon
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.27 no.5
    • /
    • pp.450-463
    • /
    • 2000
  • Array redistribution is usually required to enhance algorithm performance in many parallel programs on distributed memory multicomputers. GEN_BLOCK redistribution, which is redistribution between different GEN_BLOCKs, is essential for load balancing. However, prior research on redistribution has been focused on regular redistribution, such as redistribution between different CYCLIC(N)s. GEN_BLOCK redistribution is very different from regular redistribution. Message passing in regular redistribution involves repetitions of basic message passing patterns, while message passing for GEN_BLOCK redistribution shows locality. This paper proves that two optimal condition, reducing the number of communication steps and minimizing redistribution size, are essential in GEN_BLOCK redistribution. Additionally, by adding a relocation phase to list scheduling, we make an optimal scheduling algorithm for GEN_BLOCK redistribution. To evaluate the performance of the algorithm, we have performed experiments on a CRAY T3E. According to the experiments, it was proven that the scheduling algorithm shows better performance and that the conditions are critical in enhancing the communication speed of GEN_BLOCK redistribution.

  • PDF

An Analysis of Effective Throughput in Distributed Wireless Scheduling

  • Radwan, Amr
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.2
    • /
    • pp.155-162
    • /
    • 2016
  • Several distributed scheduling policies have been proposed with the objective of attaining the maximum throughput region or a guaranteed fraction throughput region. These policies consider only the theoretical throughput and do not account the lost in throughput due to the time complexity of implementing an algorithm in practice. Therefore, we propose a novel concept called effective throughput to characterize the actual throughput by taking into account the time complexity. Effective throughput can be viewed as the actual transmitted data without including the control message overhead. Numerical results demonstrate that in practical scheduling, time complexity significantly affects throughput. The performance of throughput degrades when the time complexity is high.

End-to-End Scheduling Method Considering 3-type RT-Data in Distributed Control Systems (분산 제어시스템에서 3가지 형태의 실시간 데이터를 고려하는 양극단 스케줄링 방법)

  • Kim, Hyoung-Yuk;Park, Hong-Seong
    • Proceedings of the KIEE Conference
    • /
    • 2003.11b
    • /
    • pp.311-314
    • /
    • 2003
  • In recent years, distributed control systems(DCS) using fieldbus such as CAN have been applied to process systems but it is very difficult to design the DCS while guaranteeing the given end-to-end constraints such as precedence constraints, time constraints, and periods and priorities of tasks and messages. This paper presents a scheduling method to guarantee the given end-to-end constraints considering aperiodic, periodic and non-real-time message and task simultaneously. The presented scheduling method is the integrated one considering both tasks executed in each node and messages transmitted via the network and is designed to be applied to a general DCS that has multiple loops with several types of constraints, where each loop consists of sensor nodes with multiple sensors, actuator nodes with multiple actuators and controller nodes with multiple tasks.

  • PDF

Improving Network Utilization in FlexRay Using Reallocation of Static Message (정적 메시지 재할당을 이용한 FlexRay 네트워크 사용효율 개선 기법)

  • Seo, Byungseok;Jin, Sungho;Lee, Dongik
    • Transactions of the Korean Society of Automotive Engineers
    • /
    • v.21 no.5
    • /
    • pp.113-120
    • /
    • 2013
  • This paper presents a mathematical model to determine the optimal length of static messages that can achieve more efficient use of a FlexRay network. In order to determine the optimal length of static message, the proposed model evaluates the given set of messages with respect to a network utilization index, which is defined in this work. The efficient use of a FlexRay network is achieved by reallocating any static message whose length is equal or greater than the resulting value to the dynamic segment. The effectiveness of the proposed method is investigated by applying to the SAE benchmark data.

Flow Scheduling in OBS Networks Based on Software-Defined Networking Control Plane

  • Tang, Wan;Chen, Fan;Chen, Min;Liu, Guo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.1
    • /
    • pp.1-17
    • /
    • 2016
  • The separated management and operation of commercial IP/optical multilayer networks makes network operators look for a unified control plane (UCP) to reduce their capital and operational expenditure. Software-defined networking (SDN) provides a central control plane with a programmable mechanism, regarded as a promising UCP for future optical networks. The general control and scheduling mechanism in SDN-based optical burst switching (OBS) networks is insufficient so the controller has to process a large number of messages per second, resulting in low network resource utilization. In view of this, this paper presents the burst-flow scheduling mechanism (BFSM) with a proposed scheduling algorithm considering channel usage. The simulation results show that, compared with the general control and scheduling mechanism, BFSM provides higher resource utilization and controller performance for the SDN-based OBS network in terms of burst loss rate, the number of messages to which the controller responds, and the average latency of the controller to process a message.

An Enhanced Message Priority Mechanism in IEEE 802.11p Based Vehicular Networks

  • Liu, Chang;Chung, Sang-Hwa;Jeong, Han-You;Jung, Ik-Joo
    • Journal of Information Processing Systems
    • /
    • v.11 no.3
    • /
    • pp.465-482
    • /
    • 2015
  • IEEE 802.11p is a standard MAC protocol for wireless access in vehicular environments (WAVEs). If a packet collision happens when a safety message is sent out, IEEE 802.11p chooses a random back-off counter value in a fixed-size contention window. However, depending on the random choice of back-off counter value, it is still possible that less important messages are sent out first while more important messages are delayed longer until sent out. In this paper, we present a new scheme for safety message scheduling, called the enhanced message priority mechanism (EMPM). It consists of the following two components: the benefit-value algorithm, which calculates the priority of the messages depending on the speed, deceleration, and message lifetime; and the back-off counter selection algorithm, which chooses the non-uniform back-off counter value in order to reduce the collision probability and to enhance the throughput of the highly beneficial messages. Numerical results show that the EMPM can significantly improve the throughput and delay of messages with high benefits when compared with existing MAC protocols. Consequently, the EMPM can provide better QoS support for the more important and urgent messages.

DEVS 형식론을 이용한 다중프로세서 운영체제의 모델링 및 성능평가

  • 홍준성
    • Proceedings of the Korea Society for Simulation Conference
    • /
    • 1994.10a
    • /
    • pp.32-32
    • /
    • 1994
  • In this example, a message passing based multicomputer system with general interdonnedtion network is considered. After multicomputer systems are developed with morm-hole routing network, topologies of interconecting network are not major considertion for process management and resource sharing. Tehre is an independeent operating system kernel oneach node. It communicates with other kernels using message passingmechanism. Based on this architecture, the problem is how mech does performance degradation will occur in the case of processor sharing on multicomputer systems. Processor sharing between application programs is veryimprotant decision on system performance. In almost cases, application programs running on massively parallel computer systems are not so much user-interactive. Thus, the main performance index is system throughput. Each application program has various communication patterns. and the sharing of processors causes serious performance degradation in hte worst case such that one processor is shared by two processes and another processes are waiting the messages from those processes. As a result, considering this problem is improtant since it gives the reason whether the system allows processor sharingor not. Input data has many parameters in this simulation . It contains the number of threads per task , communication patterns between threads, data generation and also defects in random inupt data. Many parallel aplication programs has its specific communication patterns, and there are computation and communication phases. Therefore, this phase informatin cannot be obtained random input data. If we get trace data from some real applications. we can simulate the problem more realistic . On the other hand, simualtion results will be waseteful unless sufficient trace data with varisous communication patterns is gathered. In this project , random input data are used for simulation . Only controllable data are the number of threads of each task and mapping strategy. First, each task runs independently. After that , each task shres one and more processors with other tasks. As more processors are shared , there will be performance degradation . Form this degradation rate , we can know the overhead of processor sharing . Process scheduling policy can affects the results of simulation . For process scheduling, priority queue and FIFO queue are implemented to support round-robin scheduling and priority scheduling.

  • PDF

Efficient Message Scheduling Protocol for Variable-sized Messages in a WDM-based Local Network with Nonzero Tuning Time (WDM 기반의 근거리 망에서 변조시간을 고려한 효율적인 가변 길이 메시지 예약 프로토콜)

    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.25 no.8A
    • /
    • pp.1238-1246
    • /
    • 2000
  • In a multi-channel network based on wavelength division multiplexing (WDM), an efficient protocol is needed for transmitter and receiver to be tuned to same wavelength during message transmission time. This paper proposes a message scheduling protocol that can efficiently support variable-sized messages, where tunable transceiver has nonzero tuning time. In this protocol, once a node reserves one data channel, it can persistently use the reserved channel till message transmission is finished. Therefore, overhead due to tuning time of the tunable transceiver can be reduced. Moreover, the protocol does not require any global information. Therefore, it can operate independently of the change of the number of nodes, and any new node can join the network at anytime without requiring network initialization. With this protocol, one can avoid data channel and destination conflicts. The protocol is analyzed with a finite population model and the throughput-delay characteristics are investigated as performance measures.

  • PDF