• Title/Summary/Keyword: Edge computing

Search Result 488, Processing Time 0.042 seconds

Implementation and Performance Aanalysis of Efficient Big Data Processing System Through Dynamic Configuration of Edge Server Computing and Storage Modules (BigCrawler: 엣지 서버 컴퓨팅·스토리지 모듈의 동적 구성을 통한 효율적인 빅데이터 처리 시스템 구현 및 성능 분석)

  • Kim, Yongyeon;Jeon, Jaeho;Kang, Sungjoo
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.16 no.6
    • /
    • pp.259-266
    • /
    • 2021
  • Edge Computing enables real-time big data processing by performing computing close to the physical location of the user or data source. However, in an edge computing environment, various situations that affect big data processing performance may occur depending on temporary service requirements or changes of physical resources in the field. In this paper, we proposed a BigCrawler system that dynamically configures the computing module and storage module according to the big data collection status and computing resource usage status in the edge computing environment. And the feature of big data processing workload according to the arrangement of computing module and storage module were analyzed.

Type of Machine Learning Model for Edge Computing Environment: A Survey (Edge Computing 환경을 위한 기계학습 모델 유형 조사)

  • Kim, Min-Woo;Lee, Tae-Ho;Lee, Byung-Jun;Kim, Kyung-Tae;Youn, Hee-Yong
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2019.07a
    • /
    • pp.111-112
    • /
    • 2019
  • Edge computing 환경에서는 노드끼리 직접 또는 간접적으로 전송되는 많은 수의 데이터가 Computing 노드에 의해 수집된다. Computing 노드에 실시간 적으로 전송되어지는 데이터의 저장 및 처리를 위해 기계학습(Machine learning) 기법이 사용된다. 기존의 기계학습 모델의 학습방법의 경우 Edge computing 노드의 지능화에 다소 맞지 않는 방법이며 노드들 간의 협업 시스템을 기계학습 모델에 구축하는 것 또한 중요개선사항 중 하나이다. 본 논문에서는 Edge computing 환경에서 적용 가능한 기계학습 모델을 조사하였다. 본 조사를 통하여 향후 edge computing 환경에서의 제약사항에 대해 더 구체적이며 다양한 연구방향을 제시할 수 있으며 효율적인 모델 적용을 목표로 한다.

  • PDF

A Survey of Computational Offloading in Cloud/Edge-based Architectures: Strategies, Optimization Models and Challenges

  • Alqarni, Manal M.;Cherif, Asma;Alkayal, Entisar
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.3
    • /
    • pp.952-973
    • /
    • 2021
  • In recent years, mobile devices have become an essential part of daily life. More and more applications are being supported by mobile devices thanks to edge computing, which represents an emergent architecture that provides computing, storage, and networking capabilities for mobile devices. In edge computing, heavy tasks are offloaded to edge nodes to alleviate the computations on the mobile side. However, offloading computational tasks may incur extra energy consumption and delays due to network congestion and server queues. Therefore, it is necessary to optimize offloading decisions to minimize time, energy, and payment costs. In this article, different offloading models are examined to identify the offloading parameters that need to be optimized. The paper investigates and compares several optimization techniques used to optimize offloading decisions, specifically Swarm Intelligence (SI) models, since they are best suited to the distributed aspect of edge computing. Furthermore, based on the literature review, this study concludes that a Cuckoo Search Algorithm (CSA) in an edge-based architecture is a good solution for balancing energy consumption, time, and cost.

A cache placement algorithm based on comprehensive utility in big data multi-access edge computing

  • Liu, Yanpei;Huang, Wei;Han, Li;Wang, Liping
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.11
    • /
    • pp.3892-3912
    • /
    • 2021
  • The recent rapid growth of mobile network traffic places multi-access edge computing in an important position to reduce network load and improve network capacity and service quality. Contrasting with traditional mobile cloud computing, multi-access edge computing includes a base station cooperative cache layer and user cooperative cache layer. Selecting the most appropriate cache content according to actual needs and determining the most appropriate location to optimize the cache performance have emerged as serious issues in multi-access edge computing that must be solved urgently. For this reason, a cache placement algorithm based on comprehensive utility in big data multi-access edge computing (CPBCU) is proposed in this work. Firstly, the cache value generated by cache placement is calculated using the cache capacity, data popularity, and node replacement rate. Secondly, the cache placement problem is then modeled according to the cache value, data object acquisition, and replacement cost. The cache placement model is then transformed into a combinatorial optimization problem and the cache objects are placed on the appropriate data nodes using tabu search algorithm. Finally, to verify the feasibility and effectiveness of the algorithm, a multi-access edge computing experimental environment is built. Experimental results show that CPBCU provides a significant improvement in cache service rate, data response time, and replacement number compared with other cache placement algorithms.

Implementation of SDN Wireless Testbed for Performance Evaluation of Edge Computing (Edge Computing의 성능 평가를 위한 SDN 무선 테스트베드 구축 방안)

  • Lim, Hwan-Hee;Yoo, Seung-Eon;Lee, Byung-Jun;Kim, Kyung-Tae;Youn, Hee-Yong
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2018.07a
    • /
    • pp.133-134
    • /
    • 2018
  • 본 논문에서는 Edge Computing 알고리즘의 성능 분석 및 평가를 위해 SDN 무선 테스트베드를 구현하는 방안을 제안한다. Edge Computing 환경은 수 많은 노드를 연결해 테스트해야 하며, 따라서 무선 테스트베드가 적합하다. Edge Computing 기술이 발전함에 따라 네트워크가 점점 더 복잡해지고 정확하고 확장 가능한 테스트베드 기술에 대한 필요성이 급증하고 있다. 네트워크 연구를 위해 대규모 테스트 베드가 필수적이다. 본 논문에서는 Raspberry-Pi를 이용해 Wireless SDN 테스트베드 구축 방안에 대해 소개하고자 한다.

  • PDF

An Overview of Mobile Edge Computing: Architecture, Technology and Direction

  • Rasheed, Arslan;Chong, Peter Han Joo;Ho, Ivan Wang-Hei;Li, Xue Jun;Liu, William
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.10
    • /
    • pp.4849-4864
    • /
    • 2019
  • Modern applications such as augmented reality, connected vehicles, video streaming and gaming have stringent requirements on latency, bandwidth and computation resources. The explosion in data generation by mobile devices has further exacerbated the situation. Mobile Edge Computing (MEC) is a recent addition to the edge computing paradigm that amalgamates the cloud computing capabilities with cellular communications. The concept of MEC is to relocate the cloud capabilities to the edge of the network for yielding ultra-low latency, high computation, high bandwidth, low burden on the core network, enhanced quality of experience (QoE), and efficient resource utilization. In this paper, we provide a comprehensive overview on different traits of MEC including its use cases, architecture, computation offloading, security, economic aspects, research challenges, and potential future directions.

Mobility-Aware Service Migration (MASM) Algorithms for Multi-Access Edge Computing (멀티 액세스 엣지 컴퓨팅을 위한 Mobility-Aware Service Migration (MASM) 알고리즘)

  • Hamzah, Haziq;Le, Duc-Tai;Kim, Moonseong;Choo, Hyunseung
    • Journal of Internet Computing and Services
    • /
    • v.21 no.4
    • /
    • pp.1-8
    • /
    • 2020
  • In order to reach Ultra-Reliable Low-Latency communication, one of 5G aims, Multi-access Edge Computing paradigm was born. The idea of this paradigm is to bring cloud computing technologies closer to the network edge. User services are hosted in multiple Edge Clouds, deployed at the edge of the network distributedly, to reduce the service latency. For mobile users, migrating their services to the most proper Edge Clouds for maintaining a Quality of Service is a non-convex problem. The service migration problem becomes more complex in high mobility scenarios. The goal of the study is to observe how user mobility affects the selection of Edge Cloud during a fixed mobility path. Mobility-Aware Service Migration (MASM) is proposed to optimize service migration based on two main parameters: routing cost and service migration cost, during a high mobility scenario. The performance of the proposed algorithm is compared with an existing greedy algorithm.

Task Scheduling and Resource Management Strategy for Edge Cloud Computing Using Improved Genetic Algorithm

  • Xiuye Yin;Liyong Chen
    • Journal of Information Processing Systems
    • /
    • v.19 no.4
    • /
    • pp.450-464
    • /
    • 2023
  • To address the problems of large system overhead and low timeliness when dealing with task scheduling in mobile edge cloud computing, a task scheduling and resource management strategy for edge cloud computing based on an improved genetic algorithm was proposed. First, a user task scheduling system model based on edge cloud computing was constructed using the Shannon theorem, including calculation, communication, and network models. In addition, a multi-objective optimization model, including delay and energy consumption, was constructed to minimize the sum of two weights. Finally, the selection, crossover, and mutation operations of the genetic algorithm were improved using the best reservation selection algorithm and normal distribution crossover operator. Furthermore, an improved legacy algorithm was selected to deal with the multi-objective problem and acquire the optimal solution, that is, the best computing task scheduling scheme. The experimental analysis of the proposed strategy based on the MATLAB simulation platform shows that its energy loss does not exceed 50 J, and the time delay is 23.2 ms, which are better than those of other comparison strategies.

A Framework for Real-Time Data Processing in Edge Computing Environment (Edge computing 환경에서 실시간성 데이터 처리를 위한 프레임워크 연구)

  • Kim, Jun Heon
    • Proceedings of The KACE
    • /
    • 2017.08a
    • /
    • pp.61-62
    • /
    • 2017
  • Edge computing 환경은 자원이 제한된 IoT의 디바이스가 증가하면서 데이터가 급등하는 환경을 극복하기 위한 패러다임으로 주목받고 있다. 그러나 자원이 제한된 IoT 디바이스의 특성상 실시간성 데이터의 손실, 지연과 같은 문제가 존재한다. 본 논문에서는 이를 해결하기 위해서 폭발적으로 데이터가 증가하는 Edge computing 환경에 적합한 프레임워크에 대하여 서술한다.

  • PDF

Performance analysis of local exit for distributed deep neural networks over cloud and edge computing

  • Lee, Changsik;Hong, Seungwoo;Hong, Sungback;Kim, Taeyeon
    • ETRI Journal
    • /
    • v.42 no.5
    • /
    • pp.658-668
    • /
    • 2020
  • In edge computing, most procedures, including data collection, data processing, and service provision, are handled at edge nodes and not in the central cloud. This decreases the processing burden on the central cloud, enabling fast responses to end-device service requests in addition to reducing bandwidth consumption. However, edge nodes have restricted computing, storage, and energy resources to support computation-intensive tasks such as processing deep neural network (DNN) inference. In this study, we analyze the effect of models with single and multiple local exits on DNN inference in an edge-computing environment. Our test results show that a single-exit model performs better with respect to the number of local exited samples, inference accuracy, and inference latency than a multi-exit model at all exit points. These results signify that higher accuracy can be achieved with less computation when a single-exit model is adopted. In edge computing infrastructure, it is therefore more efficient to adopt a DNN model with only one or a few exit points to provide a fast and reliable inference service.