• Title/Summary/Keyword: 전송 스케줄링

Search Result 369, Processing Time 0.021 seconds

A Study on a Real Time Presentation Method for Playing of a Multimedia mail on Internet (인터넷상의 동영상 메일을 재생하기 위한 실시간 연출 기법 연구)

  • Im, Yeong-Hwan;Lee, Seon-Hye
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.4
    • /
    • pp.877-890
    • /
    • 1999
  • In this paper, a multimedia mail including video, sound, graphic data has been proposed as the next generation mail of the text based mail. In order to develop the multimedia mail, the most outstanding problem is the fact that the multimedia data are too huge to send them to the receiving end directly. The fact of big data may cause many problems in both transferring and storing the data of the multimedia mail. Our main idea is to separate between a control program for the multimedia presentation and multimedia data. Since the size of a control program is as small as a plain text mail, it has no problem to send it attached to the internet mail to the receiver directly. Instead, the big multimedia data themselves may remain on the sender's computer or be sent to a designated server so that the data may be transferred to the receiver only when the receiver activates the play of the multimedia mail. In this scheme, our research focus is paced on the buffer management and the thread scheduling for the real time play of the multimedia mail on internet. Another problem is to provide an easy way of editing a multimedia presentation for an ordinary people having no programming knowledge. For the purposed, VIP(Visual Interface Player) has been used and the results or multimedia mail implemented on LAN has been described.

  • PDF

A Backup Node Based Fault-tolerance Scheme for Coverage Preserving in Wireless Sensor Networks (무선 센서 네트워크에서의 감지범위 보존을 위한 백업 노드 기반 결함 허용 기법)

  • Hahn, Joo-Sun;Ha, Rhan
    • Journal of KIISE:Information Networking
    • /
    • v.36 no.4
    • /
    • pp.339-350
    • /
    • 2009
  • In wireless sensor networks, the limited battery resources of sensor nodes have a direct impact on network lifetime. To reduce unnecessary power consumption, it is often the case that only a minimum number of sensor nodes operate in active mode while the others are kept in sleep mode. In such a case, however, the network service can be easily unreliable if any active node is unable to perform its sensing or communication function because of an unexpected failure. Thus, for achieving reliable sensing, it is important to maintain the sensing level even when some sensor nodes fail. In this paper, we propose a new fault-tolerance scheme, called FCP(Fault-tolerant Coverage Preserving), that gives an efficient way to handle the degradation of the sensing level caused by sensor node failures. In the proposed FCP scheme, a set of backup nodes are pre-designated for each active node to be used to replace the active node in case of its failure. Experimental results show that the FCP scheme provides enhanced performance with reduced overhead in terms of sensing coverage preserving, the number of backup nodes and the amount of control messages. On the average, the percentage of coverage preserving is improved by 87.2% while the additional number of backup nodes and the additional amount of control messages are reduced by 57.6% and 99.5%, respectively, compared with previous fault-tolerance schemes.

A Block Relocation Algorithm for Reducing Network Consumption in Hadoop Cluster (하둡 클러스터의 네트워크 사용량 감소를 위한 블록 재배치 알고리즘)

  • Kim, Jun-Sang;Kim, Chang-Hyeon;Lee, Won-Joo;Jeon, Chang-Ho
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.11
    • /
    • pp.9-15
    • /
    • 2014
  • In this paper, We propose a block reallocation algorithm for reducing network traffic in Hadoop cluster. The scheduler of Hadoop cluster receives a job from users. And the job is divided into multiple tasks assigned to nodes. At this time, the scheduler allocates the task to the node that satisfied data locality. If a task is assigned to the node that does not have the data(block) to be processed, the task is processed after the data transmission from another node. There is difference of workload among nodes because blocks in cluster have different access frequency. Therefore, the proposed algorithm relocates blocks according to the task allocation pattern of Hadoop scheduler. Eventually, workload of nodes are leveled, and the case of the task processing in a node that does not have the block to be processing is reduced. Thus, the network traffic of the cluster is also reduced. We evaluate the proposed block reallocation algorithm by a simulation. The simulation result shows maximum 23.3% reduction of network consumption than default delay scheduling for jobs processing.

A Design of Handoff-aware DiffServ Scheduler in TDD/CDMA Networks (TDD/CDMA망에서 핸드오프를 지원하는 DiffServ 스케줄러 설계)

  • Zang, Seog-Ku;Kim, Young-Han
    • The KIPS Transactions:PartC
    • /
    • v.14C no.6
    • /
    • pp.493-502
    • /
    • 2007
  • In this paper, we propose a handoff-aware DiffServ scheduler which intends to guarantee various QoS requirements of multimedia services for mobile nodes in TDD/CDMA based wireless networks. TDD is widely used duplexing mechanism in wireless communications. Unlike FDD, TDD allows a node to symmetrically communicate with a base station by using a single frequency band, resulting in high utilization of wireless resources. DiffServ is regarded as a relatively simple QoS support mechanism and thus it is easy to be extended. This is because DiffServ is not a per-flow based mechanisms and it does not require any signaling protocol. However, previously proposed DiffServ schedulers for wired networks can not be deployed directly into wireless networks since they do not consider properties of wireless networks. As a solution to the problem, DSS(DiffServ Supporting Scheduler) was proposed. DSS uses uplink channel, which is originally used for a node to require a base station to transmit packets, to support QoS efficiently. However, QoS does not consider handoff so that it can not support QoS for moving nodes from one cell to the other cell. Therefor. the proposed handoff support QoS mechanism is necessary for TCC/CDMA networks. The proposed scheme allows a mobile node to achieve seamless service without QoS degradation even for the handoff duration.

A Sensing Channel Scheduling Scheme for Improving the Cognition Ability in Cognitive Radio Systems (인지 라디오 시스템에서 주파수 상황인지 능력 향상을 위한 감지 채널 스케줄링 기법)

  • Han, Jeong-Ae;Jeon, Wha-Sook
    • Journal of KIISE:Information Networking
    • /
    • v.35 no.2
    • /
    • pp.130-138
    • /
    • 2008
  • The scheme for recognizing the channel availability is one of the most important research issues in cognitive radio systems utilizing unused frequency bands. In this paper, we propose a novel scheme of selecting sensing channel in order to improve the sensing ability of frequency status in cognitive radio ad hoc networks. To fully exploit the sensing ability of each cognitive radio user, we adopt a master for a cluster which is made of several cognitive radio users. By gathering and analyzing the sensing information from cognitive radio users in the cluster, the cooperative sensing is realized. Since the transmission range of a licensed user is limited, it is possible that a master determines different sensing channels to each cognitive radio users based on their location. By making cognitive radio users sense different channels, the proposed scheme can recognize the state of wireless spectrum fast and precisely. Using the simulation, we compare the performance of the proposed scheme with those of two different compared schemes that one makes cognitive radio users recognize the frequency status based on their own sensing results and the other shares frequency status information but does not utilize the location information of licensed user. Simulation results show that the proposed scheme provides available channels as many as possible while detecting the activation of licensed user immediately.

A Data Allocation Method based on Broadcast Disks Using Indices over Multiple Broadcast Channels (다중방송 채널에서 인덱스를 이용한 브로드캐스트 디스크 기반의 데이타 할당 기법)

  • Lee, Won-Taek;Jung, Sung-Won
    • Journal of KIISE:Databases
    • /
    • v.35 no.3
    • /
    • pp.272-285
    • /
    • 2008
  • In this paper, we concentrate on data allocation methods for multiple broadcast channels. When the server broadcasts data, the important issue is to let mobile clients access requested data rapidly. Previous works first sorted data by their access probabilities and allocate the sorted data to the multiple channels by partitioning them into multiple channels. However, they do not reflect the difference of access probabilities among data allocated in the same channel. This paper proposes ZGMD allocation method. ZGMD allocates data item on multiple channels so that the difference of access probability in the same channel is maximized. ZGMD allocates sorted data to each channels and applies Broadcast Disk in each channel. ZGMD requires a proper indexing scheme for the performance improvement. This is because in ZGMD method each channel got allocated both hot and cold data. As a result, the sequential search heuristic does not allow the mobile client to access hot data items quickly. The proposed index scheme is based on using dedicated index channels in order to search the data channel where the requested data is. We show that our method achieve the near-optimal performance in terms of the average access time and significantly outperforms the existing methods.

A Study for Improving Performance of ATM Multicast Switch (ATM 멀티캐스트 스위치의 성능 향상을 위한 연구)

  • 이일영;조양현;오영환
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.12A
    • /
    • pp.1922-1931
    • /
    • 1999
  • A multicast traffic’s feature is the function of providing a point to multipoints cell transmission, which is emerging from the main function of ATM switch. However, when a conventional point-to-point switch executes a multicast function, the excess load is occurred because unicast cell as well as multicast cell passed the copy network. Additionally, due to the excess load, multicast cells collide with other cells in a switch. Thus a deadlock that losses cells raises, extremely diminishes the performance of switch. An input queued switch also has a defect of the HOL (Head of Line) blocking that less lessens the performance of the switch. In the proposed multicast switch, we use shared memory switch to reduce HOL blocking and deadlock. In order to decrease switch’s complexity and cell's processing time, to improve a throughput, we utilize the method that routes a cell on a separated paths by traffic pattern and the scheduling algorithm that processes a maximum 2N cell at once in the control part. Besides, when cells is congested at an output port, a cell loss probability increases. Thus we use the Output Memory (OM) to reduce the cell loss probability. And we make use of the method that stores the assigned memory (UM, MM) with a cell by a traffic pattern and clears the cell of the Output memory after a fixed saving time to improve the memory utilization rate. The performance of the proposed switch is executed and compared with the conventional policy under the burst traffic condition through both the analysis based on Markov chain and simulation.

  • PDF

Time Synchronization Algorithm using the Clock Drift Rate and Reference Signals Between Two Sensor Nodes (클럭 표류율과 기준 신호를 이용한 두 센서 노드간 시간 동기 알고리즘)

  • Kim, Hyoun-Soo;Jeon, Joong-Nam
    • The KIPS Transactions:PartC
    • /
    • v.16C no.1
    • /
    • pp.51-56
    • /
    • 2009
  • Time synchronization algorithm in wireless sensor networks is essential to various applications such as object tracking, data encryption, duplicate detection, and precise TDMA scheduling. This paper describes CDRS that is a time synchronization algorithm using the Clock Drift rate and Reference Signals between two sensor nodes. CDRS is composed of two steps. At first step, the time correction is calculated using offset and the clock drift rate between the two nodes based on the LTS method. Two nodes become a synchronized state and the time variance can be compensated by the clock drift rate. At second step, the synchronization node transmits reference signals periodically. This reference signals are used to calculate the time difference between nodes. When this value exceeds the maximum error tolerance, the first step is performed again for resynchronization. The simulation results on the performance analysis show that the time accuracy of the proposed algorithm is improved, and the energy consumption is reduced 2.5 times compared to the time synchronization algorithm with only LTS, because CDRS reduces the number of message about 50% compared to LTS and reference signals do not use the data space for timestamp.

A Time Slot Assignment Scheme for Sensor Data Compression (센서 데이터의 압축을 위한 시간 슬롯 할당 기법)

  • Yeo, Myung-Ho;Kim, Hak-Sin;Park, Hyoung-Soon;Yoo, Jae-Soo
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.11
    • /
    • pp.846-850
    • /
    • 2009
  • Recently, wireless sensor networks have found their way into a wide variety of applications and systems with vastly varying requirements and characteristics such as environmental monitoring, smart spaces, medical applications, and precision agriculture. The sensor nodes are battery powered. Therefore, the energy is the most precious resource of a wireless sensor network since periodically replacing the battery of the nodes in large scale deployments is infeasible. Energy efficient mechanisms for gathering sensor readings are indispensable to prolong the lifetime of a sensor network as long as possible. There are two energy-efficient approaches to prolong the network lifetime in sensor networks. One is the compression scheme to reduce the size of sensor readings. When the communication conflict is occurred between two sensor nodes, the sender must try to retransmit its reading. The other is the MAC protocol to prevent the communication conflict. In this paper, we propose a novel approaches to reduce the size of the sensor readings in the MAC layer. The proposed scheme compresses sensor readings by allocating the time slots of the TDMA schedule to them dynamically. We also present a mathematical model to predict latency from collecting the sensor readings as the compression ratio is changed. In the simulation result, our proposed scheme reduces the communication cost by about 52% over the existing scheme.