• Title/Summary/Keyword: Data caching

Search Result 222, Processing Time 0.027 seconds

Many-objective joint optimization for dependency-aware task offloading and service caching in mobile edge computing

  • Xiangyu Shi;Zhixia Zhang;Zhihua Cui;Xingjuan Cai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.5
    • /
    • pp.1238-1259
    • /
    • 2024
  • Previous studies on joint optimization of computation offloading and service caching policies in Mobile Edge Computing (MEC) have often neglected the impact of dependency-aware subtasks, edge server resource constraints, and multiple users on policy formulation. To remedy this deficiency, this paper proposes a many-objective joint optimization dependency-aware task offloading and service caching model (MaJDTOSC). MaJDTOSC considers the impact of dependencies between subtasks on the joint optimization problem of task offloading and service caching in multi-user, resource-constrained MEC scenarios, and takes the task completion time, energy consumption, subtask hit rate, load variability, and storage resource utilization as optimization objectives. Meanwhile, in order to better solve MaJDTOSC, a many-objective evolutionary algorithm TSMSNSGAIII based on a three-stage mating selection strategy is proposed. Simulation results show that TSMSNSGAIII exhibits an excellent and stable performance in solving MaJDTOSC with different number of users setting and can converge faster. Therefore, it is believed that TSMSNSGAIII can provide appropriate sub-task offloading and service caching strategies in multi-user and resource-constrained MEC scenarios, which can greatly improve the system offloading efficiency and enhance the user experience.

Playback Quantity-based Proxy Caching Scheme for Continuous Media Data (재생량에 기반한 연속미디어 데이터 프록시 캐슁 기법)

  • Hong, Hyeon-Ok;Im, Eun-Ji;Jeong, Gi-Dong
    • The KIPS Transactions:PartB
    • /
    • v.9B no.3
    • /
    • pp.303-310
    • /
    • 2002
  • In this paper, we propose a proxy caching scheme that stores a portion of a continuous media object or an entire object on the Internet. The proxy stores the initial fraction of a continuous media object and determines the optimal size of the initial fraction of the continuous media object to be cached based on the object popularity. Under the proposed scheme, the initial latency of most clients and the data transferred from a remote server can be reduced and limited cache storage space can be utilized efficiently. Considering the characteristics of continuous media, we also propose the novel popularity for the continuous media objects based on the amount of the data of each object played by the clients. Finally, we have performed trace-driven simulations to evaluate our caching scheme and the popularity for the continuous media objects. Through these simulations, we have verified that our caching scheme, PPC outperforms other well-known caching schemes in terms of BHR, DSR and replacement and that popularity for the continuous media objects based on the amount of the playback data can enhance the performance of caching scheme.

Optimizing Caching in a Patch Streaming Multimedia-on-Demand System

  • Bulti, Dinkisa Aga;Raimond, Kumudha
    • Journal of Computing Science and Engineering
    • /
    • v.9 no.3
    • /
    • pp.134-141
    • /
    • 2015
  • In on-demand multimedia streaming systems, streaming techniques are usually combined with proxy caching to obtain better performance. The patch streaming technique has no start-up latency inherent to it, but requires extra bandwidth to deliver the media data in patch streams. This paper proposes a proxy caching technique which aims at reducing the bandwidth cost of the patch streaming technique. The proposed approach determines media prefixes with high patching cost and caches the appropriate media prefix at the proxy/local server. Herein the scheme is evaluated using a synthetically generated media access workload and its performance is compared with that of the popularity and prefix-aware interval caching scheme (the prefix part) and with that of patch streaming with no caching. The bandwidth saving, hit ratio and concurrent number of clients are used to compare the performance, and the proposed scheme is found to perform better for different caching capacities of the proxy server.

A Spatio-Temporal Geocasting Protocol Using Regional Caching in Vehicular Ad-Hoc Networks (VANET 망에서 지역적 캐싱을 이용하는 시공간 지오캐스팅 프로토콜)

  • Lee, Hyunkyu;Shin, Yongje;Choi, Hyunsuk;Lee, Euisin;Park, Soochang
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.7 no.6
    • /
    • pp.137-144
    • /
    • 2018
  • Vehicular Ad-hoc Networks (VANETS) have enabled to provide a variety of applications such as accident notification, content usage, etc. These applications have spatio-temporal data which have an interesting region and a lifetime according to their properties. However, geocasting protocols to deliver data to an interesting region can provide data to all vehicles in the region through a single transmission only at the current time, but cannot provide data to vehicles passing through the region during the lifetime of the data. Thus, we propose a spatio-temporal geocasting protocol called STGP using a regional caching scheme to send data to vehicles in an interesting region during a data lifetime in VANETs. For efficient and reliable regional caching, the proposed protocol uses the beacon-based data sharing, the extra caching elimination, and the distance-based caching exchange. Simulation results verify that the proposed protocol achieves more reliable and efficient data delivery compared with the existing protocol.

A Scheme on High-Performance Caching and High-Capacity File Transmission for Cloud Storage Optimization (클라우드 스토리지 최적화를 위한 고속 캐싱 및 대용량 파일 전송 기법)

  • Kim, Tae-Hun;Kim, Jung-Han;Eom, Young-Ik
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.8C
    • /
    • pp.670-679
    • /
    • 2012
  • The recent dissemination of cloud computing makes the amount of data storage to be increased and the cost of storing the data grow rapidly. Accordingly, data and service requests from users also increases the load on the cloud storage. There have been many works that tries to provide low-cost and high-performance schemes on distributed file systems. However, most of them have some weaknesses on performing parallel and random data accesses as well as data accesses of frequent small workloads. Recently, improving the performance of distributed file system based on caching technology is getting much attention. In this paper, we propose a CHPC(Cloud storage High-Performance Caching) framework, providing parallel caching, distributed caching, and proxy caching in distributed file systems. This study compares the proposed framework with existing cloud systems in regard to the reduction of the server's disk I/O, prevention of the server-side bottleneck, deduplication of the page caches in each client, and improvement of overall IOPS. As a results, we show some optimization possibilities on the cloud storage systems based on some evaluations and comparisons with other conventional methods.

Design and Optimality Analysis of Cache Performance Model for Multimedia Streaming Environments (멀티미디어 스트리밍 환경을 위한 캐쉬 성능평가 모델 설계 및 최적성 분석)

  • Hyokyung Bahn;Kyungwoon Cho
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.5
    • /
    • pp.9-13
    • /
    • 2023
  • Multimedia streaming data is very large in size and accessed sequentially, making the LRU(Least Recently Used) algorithm widely used to improve I/O performance in traditional caching environments ineffective. Experimental analysis of this has shown the superiority of interval-based caching over LRU, but the theoretical basis has not been proven. In this paper, we design a cache performance model to analyze the optimality of caching for multimedia streaming environments and design a theoretically optimal caching algorithm based on interval caching. Then, we show that the algorithm we design is an optimal algorithm that minimizes cache misses of streaming data based on the proposed model.

A mobile data caching synchronization strategy based on in-demand replacement priority (수요에 따른 교체 우선 순위 기반 모바일 데이터베이스 캐쉬 동기화 정책)

  • Zhao, Jinhua;Xia, Ying;Lee, Soon-Jo;Bae, Hae-Young
    • Journal of the Korea Society of Computer and Information
    • /
    • v.17 no.2
    • /
    • pp.13-21
    • /
    • 2012
  • Mobile data caching is usually used as an effective way to improve the speed of local transaction processing and reduce server load. In mobile database environment, due to its characters - low bandwidth, excessive latency and intermittent network, caching is especially crucial. A lot of mobile data caching strategies have been proposed to handle these problems over the last few years. However, with smart phone widely application these approaches cannot support vast data requirements efficiently. In this paper, to make full use of cache data, lower wireless transmission quantity and raise transaction success rate, we design a new mobile data caching synchronization strategy based on in-demand and replacement priority. We experimentally verify that our techniques significantly reduce quantity of wireless transmission and improve transaction success rate, especially when mobile client request a large amount of data.

A Push-Caching and a Transmission Scheme of Continuous Media for NOD Service on the Internet (인테넷상에서 NOD 서비스를 위한 연속미디어 전송 및 푸쉬-캐싱 기법)

  • Park, Seong-Ho;Im, Eun-Ji;Choe, Tae-Uk;Jeong, Gi-Dong
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.6
    • /
    • pp.1766-1777
    • /
    • 2000
  • In multimedia new service on the internet, there are problems such as server overload, network congestion and initial latency. To overcome these problems, we propose a proxy push-caching scheme that stores a portion of continuous media stream or entire stream, and a transmission scheme of NOD continuous media, RTP-RR and RTP-nR to exploit push-caching scheme. With the proposed push-caching scheme, NOD server pushes fixed portion of stream to a proxy when new data is generated, and the cached size of each stream changes dynamically according to the caching utility value of each stream. As a result, the initial latency of client side could be reduced and the amount of data transmitted fro ma proxy server to client could be increased. Moreover, we estimate a caching utility value of each stream using correlation between disk space occupied by the stream and the amount of data stream requested by client. And we applied the caching utility value ot replacement policies. The performance of the proxy push-caching and continuous media transmission schemes proposed were compared with other schemes using simulations. In the simulation, these schemes show better results than other schemes in terms of BHR (Byte Hit Rate), initial latency, the number of replacement and packet loss rate.

  • PDF

Big Data Meets Telcos: A Proactive Caching Perspective

  • Bastug, Ejder;Bennis, Mehdi;Zeydan, Engin;Kader, Manhal Abdel;Karatepe, Ilyas Alper;Er, Ahmet Salih;Debbah, Merouane
    • Journal of Communications and Networks
    • /
    • v.17 no.6
    • /
    • pp.549-557
    • /
    • 2015
  • Mobile cellular networks are becoming increasingly complex to manage while classical deployment/optimization techniques and current solutions (i.e., cell densification, acquiring more spectrum, etc.) are cost-ineffective and thus seen as stopgaps. This calls for development of novel approaches that leverage recent advances in storage/memory, context-awareness, edge/cloud computing, and falls into framework of big data. However, the big data by itself is yet another complex phenomena to handle and comes with its notorious 4V: Velocity, voracity, volume, and variety. In this work, we address these issues in optimization of 5G wireless networks via the notion of proactive caching at the base stations. In particular, we investigate the gains of proactive caching in terms of backhaul offloadings and request satisfactions, while tackling the large-amount of available data for content popularity estimation. In order to estimate the content popularity, we first collect users' mobile traffic data from a Turkish telecom operator from several base stations in hours of time interval. Then, an analysis is carried out locally on a big data platformand the gains of proactive caching at the base stations are investigated via numerical simulations. It turns out that several gains are possible depending on the level of available information and storage size. For instance, with 10% of content ratings and 15.4Gbyte of storage size (87%of total catalog size), proactive caching achieves 100% of request satisfaction and offloads 98% of the backhaul when considering 16 base stations.

Edge Caching Strategy with User Mobility in Heterogeneous Cellular Network Environments (이종 셀룰러 네트워크 환경에서 사용자 이동성을 고려한 엣지 캐싱 기법)

  • Choi, Yoonjeong;Lim, Yujin
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.11 no.2
    • /
    • pp.43-50
    • /
    • 2022
  • As the use of mobile data increases, the proportion of video content is increasing steeply. In order to solve problems that arise when mobile users receive data from geographically remote cloud servers, methods of caching data in advance to edge servers geographically close to the users are attracting lots of attention. In this paper, we present a caching policy that stores data on Small Cell Base Station(SBS) to effectively provide content files to mobile users by applying a delayed offloading scheme in a cellular network. The goal of the proposed policy is to minimize the size of data transmitted from Macro Base Station(MBS) because the delayed offloading scheme requires more cost than when downloaded from MBS than from SBS. The caching policy is proposed to determine the size of content file and which content file to be cached to SBS using the probability of mobile users' paths and the popularity of content files, and to replace content files in consideration of the overlapping coverage of SBS. In addition, through performance evaluation, it has been proven that the proposed policy reduces the size of data downloaded from MBS compared to other algorithms.