• Title/Summary/Keyword: (Network cache

Search Result 270, Processing Time 0.03 seconds

Implementation and Performance Evaluation of Migration Agent for Seamless Virtual Environment System in Grid Computing Network (그리드 컴퓨팅 네트워크에서 Seamless 가상 환경 시스템 구축을 위한 마이그레이션 에이전트 구현 및 성능 평가)

  • Won, Dong Hyun;An, Dong-Un
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.7 no.11
    • /
    • pp.269-274
    • /
    • 2018
  • MMORPG is a role-playing game that tens of thousands of people access it online at the same time. Users connect to the server through the game client and play with their own characters. If the user moves to a management area of another server beyond the area managed by the server, the user information must be transmitted to the server to be moved. In an actual game, the user is required to synchronize the established and the transferred information. In this paper, we propose a migration agent server in the virtual systems. We implement a seamless virtual server using the grid method to experiment with seamless server architecture for virtual systems. We propose a method to minimize the delay and equalize the load when the user moves to another server region in the virtual environment. Migration Agent acts as a cache server to reduce response time, the response time was reduced by 50% in the case of 70,000 people.

Resource Allocation for Heterogeneous Service in Green Mobile Edge Networks Using Deep Reinforcement Learning

  • Sun, Si-yuan;Zheng, Ying;Zhou, Jun-hua;Weng, Jiu-xing;Wei, Yi-fei;Wang, Xiao-jun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.7
    • /
    • pp.2496-2512
    • /
    • 2021
  • The requirements for powerful computing capability, high capacity, low latency and low energy consumption of emerging services, pose severe challenges to the fifth-generation (5G) network. As a promising paradigm, mobile edge networks can provide services in proximity to users by deploying computing components and cache at the edge, which can effectively decrease service delay. However, the coexistence of heterogeneous services and the sharing of limited resources lead to the competition between various services for multiple resources. This paper considers two typical heterogeneous services: computing services and content delivery services, in order to properly configure resources, it is crucial to develop an effective offloading and caching strategies. Considering the high energy consumption of 5G base stations, this paper considers the hybrid energy supply model of traditional power grid and green energy. Therefore, it is necessary to design a reasonable association mechanism which can allocate more service load to base stations rich in green energy to improve the utilization of green energy. This paper formed the joint optimization problem of computing offloading, caching and resource allocation for heterogeneous services with the objective of minimizing the on-grid power consumption under the constraints of limited resources and QoS guarantee. Since the joint optimization problem is a mixed integer nonlinear programming problem that is impossible to solve, this paper uses deep reinforcement learning method to learn the optimal strategy through a lot of training. Extensive simulation experiments show that compared with other schemes, the proposed scheme can allocate resources to heterogeneous service according to the green energy distribution which can effectively reduce the traditional energy consumption.

Design of an Optimized GPGPU for Data Reuse in DeepLearning Convolution (딥러닝 합성곱에서 데이터 재사용에 최적화된 GPGPU 설계)

  • Nam, Ki-Hun;Lee, Kwang-Yeob;Jung, Jun-Mo
    • Journal of IKEEE
    • /
    • v.25 no.4
    • /
    • pp.664-671
    • /
    • 2021
  • This paper proposes a GPGPU structure that can reduce the number of operations and memory access by effectively applying a data reuse method to a convolutional neural network(CNN). Convolution is a two-dimensional operation using kernel and input data, and the operation is performed by sliding the kernel. In this case, a reuse method using an internal register is proposed instead of loading kernel from a cache memory until the convolution operation is completed. The serial operation method was applied to the convolution to increase the effect of data reuse by using the principle of GPGPU in which instructions are executed by the SIMT method. In this paper, for register-based data reuse, the kernel was fixed at 4×4 and GPGPU was designed considering the warp size and register bank to effectively support it. To verify the performance of the designed GPGPU on the CNN, we implemented it as an FPGA and then ran LeNet and measured the performance on AlexNet by comparison using TensorFlow. As a result of the measurement, 1-iteration learning speed based on AlexNet is 0.468sec and the inference speed is 0.135sec.

A Study on analysis tools in the SWF file URL (SWF 파일의 URL정보 분석도구)

  • Jang, Dong-Hwan;Song, Yu-Jin;Lee, Jae-Yong
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.15 no.5
    • /
    • pp.105-111
    • /
    • 2010
  • SWF(Shock Wave Flash) file is a format file for vector graphics produced by Adobe. It is widely used for a variety of contents such as advertising at websites, widgets, games, education, and videos and it contains various types of data such as sound sources, script, API and images. Many SWF files contain URL information on action script for communication in the network and they can be used as important research data as well as PC users' Web Browser history in terms of forensic investigation. And a decompiler for analyzing SWF files exists by which SWF files can be analysed and URL information can be verified. However, it takes a long time to verify the URL information on action scripts of multiple SWF files by the decompiler. In this paper, analysis of URL information on action scripts and extraction of URL information from multiple SWF files by designing analysis tools for URL information in SWF files is studied.

Fast and Secure Handoff Mechanism for IEEE 802.11i System (IEEE 802.11i 환경에서 신속하고 보안성 있는 핸드오프 메커니즘)

  • Park, Jae-Sung;Lim, Yu-Jin
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.43 no.11 s.353
    • /
    • pp.159-167
    • /
    • 2006
  • IEEE 802.11i standard specifies full authentication and preauthentication for secure handoff in 802.11 wireless LAN (WLAN). However, the full authentication is too slow to provide seamless services for handoff users, and preauthentication may fail in highly populated WLAN where it is highly probable that the cache entry of a preauthenticated user is evicted by other users before handoff. In this paper, we propose a fast and secure handoff scheme by reducing authentication and key management delay in the handoff process. When a user handoffs, security context established between the user and the previous access point (AP) is forwarded from the previous AP to the current AP, and the session key is reused before the handoff session terminates. The freshness of session key is maintained by regenerating session keys after handoff session is terminated. The proposed scheme can achieve considerable reduction in handoff delay with providing the same security level as 802.1X authentication by letting an AP authenticate a handoff user before making an robust security network association (RSNA) with it.

Proxy Caching Scheme Based on the User Access Pattern Analysis for Series Video Data (시리즈 비디오 데이터의 접근 패턴에 기반한 프록시 캐슁 기법)

  • Hong, Hyeon-Ok;Park, Seong-Ho;Chung, Ki-Dong
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.8
    • /
    • pp.1066-1077
    • /
    • 2004
  • Dramatic increase in the number of Internet users want highly qualified service of continuous media contents on the web. To solve these problems, we present two network caching schemes(PPC, PPCwP) which consider the characteristics of continuous media objects and user access pattern in this paper. While there are plenty of reasons to create rich media contents, delivering this high bandwidth contents over the internet presents problems such as server overload, network congestion and client-perceived latency. PPC scheme periodically calculates the popularity of objects based on the playback quantity and determines the optimal size of the initial fraction of a continuous media object to be cached in proportion to the calculated popularity. PPCwP scheme calculates the expected popularity using the series information and prefetches the expected initial fraction of newly created continuous media objects. Under the PPCwP scheme, the initial client-perceived latency and the data transferred from a remote server can be reduced and limited cache storage space can be utilized efficiently. Trace-driven simulation have been performed to evaluate the presented caching schemes using the log-files of iMBC. Through these simulations, PPC and PPCwP outperforms LRU and LFU in terms of BHR and DSR.

  • PDF

A Proxy based QoS Provisioning Mechanism for Streaming Service in Wireless Networks (무선이동통신망에서 스트리밍 서비스를 위한 프락시 기반Qos 보장 방안)

  • Kim Yong-Sul;Hong Jung-Pyo;Kim Hwa-Sung;Yoo Ji-Sang;Kim Dong-Wook
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.7B
    • /
    • pp.608-618
    • /
    • 2006
  • The increasing popularity of multimedia streaming services introduces new challenges in content distribution. Especially, it is important to provide the QoS guarantees as they are increasingly expected to support the multimedia applications. The service providers can improve the performance of multimedia streaming by caching the initial segment (prefix) of the popular streams at proxies near the requesting clients. The proxy can initiate transmission to the client while requesting the remainder of the stream from the server. In this paper, in order to apply the prefix caching service based on IETF's RTSP environment to the wireless networks, we propose the effective RTSP handling scheme that can adapt to the radio situation in wireless network and reduce the cutting phenomenon. Also, we propose the traffic based caching algorithm (TSLRU) to improve the performance of caching proxy. TSLRU classifies the traffic into three types, and improve the performance of caching proxy by reflecting the several elements such as traffic types, recency, frequency, object size when performing the replacement decision. In simulation, TSLRU and RTSP handling scheme performs better than the existing schemes in terms of byte hit rate, hit rate, startup latency, and throughput.

A Dynamic Transaction Routing Algorithm with Primary Copy Authority (주사본 권한을 이용한 동적 트랜잭션 분배 알고리즘)

  • Kim, Ki-Hyung;Cho, Hang-Rae;Nam, Young-Hwan
    • The KIPS Transactions:PartD
    • /
    • v.10D no.7
    • /
    • pp.1067-1076
    • /
    • 2003
  • Database sharing system (DSS) refers to a system for high performance transaction processing. In DSS, the processing nodes are locally coupled via a high speed network and share a common database at the disk level. Each node has a local memory and a separate copy of operating system. To reduce the number of disk accesses, the node caches database pages in its local memory buffer. In this paper, we propose a dynamic transaction routing algorithm to balance the load of each node in the DSS. The proposed algorithm is novel in the sense that it can support node-specific locality of reference by utilizing the primary copy authority assigned to each node; hence, it can achieve better cache hit ratios and thus fewer disk I/Os. Furthermore, the proposed algorithm avoids a specific node being overloaded by considering the current workload of each node. To evaluate the performance of the proposed algorithm, we develop a simulation model of the DSS, and then analyze the simulation results. The results show that the proposed algorithm outperforms the existing algorithms in the transaction processing rate. Especially the proposed algorithm shows better performance when the number of concurrently executed transactions is high and the data page access patterns of the transactions are not equally distributed.

Web Prefetching Scheme for Efficient Internet Bandwidth Usage (효율적인 인터넷 대역폭 사용을 위한 웹 프리페칭 기법)

  • Kim, Suk-Hyang;Hong, Won-Gi
    • Journal of KIISE:Information Networking
    • /
    • v.27 no.3
    • /
    • pp.301-314
    • /
    • 2000
  • As the number of World Wide Web (Web) users grows, Web traffic continues to increase at an exponential rate. Currently, Web traffic is one of the major components of Internet traffic. Also, high bandwodth usage due to Web traffic is observed during peak periods while leaving bandwidth usage idle during off-peak periods. One of the solutions to reduce Web traffic and speed up Web access is through the use of Web caching. Unfortunately, Web caching has limitations for reducing network bandwidth usage during peak periods. In this paper, we focus our attention on the use of a prefetching algorithm for reducing bandwidth during peak periods by using off-peak period bandwidth. We propose a statistical, batch, proxy-side prefetching scheme that improves cache hit rate while only requiring a small amount of storage. Web objects that were accessed many times in previous 24 hours but would be expired in the next 24 hours, are selected and prefetched in our scheme. We present simulation results based on Web proxy and show that this prefetching algorithm can reduce peak time bandwidth using off-peak bandwidth.

  • PDF

A Method for Reducing Path Recovery Overhead of Clustering-based, Cognitive Radio Ad Hoc Routing Protocol (클러스터링 기반 인지 무선 애드혹 라우팅 프로토콜의 경로 복구 오버헤드 감소 기법)

  • Jang, Jin-kyung;Lim, Ji-hun;Kim, Do-Hyung;Ko, Young-Bae;Kim, Joung-Sik;Seo, Myung-hwan
    • Journal of IKEEE
    • /
    • v.23 no.1
    • /
    • pp.280-288
    • /
    • 2019
  • In the CR-enabled MANET, routing paths can be easily destroyed due to node mobility and channel unavailability (due to the emergence of the PU of a channel), resulting in significant overhead to maintain/recover the routing path. In this paper, network caching is actively used for route maintenance, taking into account the properties of the CR. In the proposed scheme, even if a node detects that a path becomes unavailable, it does not generate control messages to establish an alternative path. Instead, the node stores the packets in its local cache and 1) waits for a certain amount of time for the PU to disappear; 2) waits for a little longer while overhearing messages from other flow; 3) after that, the node applies local route recovery process or delay tolerant forwarding strategy. According to the simulation study using the OPNET simulator, it is shown that the proposed scheme successfully reduces the amount of control messages for path recovery and the service latency for the time-sensitive traffic by 13.8% and 45.4%, respectively, compared to the existing scheme. Nevertheless, the delivery ratio of the time-insensitive traffic is improved 14.5% in the proposed scheme.