• Title/Summary/Keyword: intelligent edge computing

Search Result 52, Processing Time 0.026 seconds

A Study on Intelligent Edge Computing Network Technology for Road Danger Context Aware and Notification

  • Oh, Am-Suk
    • Journal of information and communication convergence engineering
    • /
    • v.18 no.3
    • /
    • pp.183-187
    • /
    • 2020
  • The general Wi-Fi network connection structure is that a number of IoT (Internet of Things) sensor nodes are directly connected to one AP (Access Point) node. In this structure, the range of the network that can be established within the specified specifications such as the range of signal strength (RSSI) to which the AP node can connect and the maximum connection capacity is limited. To overcome these limitations, multiple middleware bridge technologies for dynamic scalability and load balancing were studied. However, these network expansion technologies have difficulties in terms of the rules and conditions of AP nodes installed during the initial network deployment phase In this paper, an intelligent edge computing IoT device is developed for constructing an intelligent autonomous cluster edge computing network and applying it to real-time road danger context aware and notification system through an intelligent risk situation recognition algorithm.

A Performance Comparison of Parallel Programming Models on Edge Devices (엣지 디바이스에서의 병렬 프로그래밍 모델 성능 비교 연구)

  • Dukyun Nam
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.4
    • /
    • pp.165-172
    • /
    • 2023
  • Heterogeneous computing is a technology that utilizes different types of processors to perform parallel processing. It maximizes task processing and energy efficiency by leveraging various computing resources such as CPUs, GPUs, and FPGAs. On the other hand, edge computing has developed with IoT and 5G technologies. It is a distributed computing that utilizes computing resources close to clients, thereby offloading the central server. It has evolved to intelligent edge computing combined with artificial intelligence. Intelligent edge computing enables total data processing, such as context awareness, prediction, control, and simple processing for the data collected on the edge. If heterogeneous computing can be successfully applied in the edge, it is expected to maximize job processing efficiency while minimizing dependence on the central server. In this paper, experiments were conducted to verify the feasibility of various parallel programming models on high-end and low-end edge devices by using benchmark applications. We analyzed the performance of five parallel programming models on the Raspberry Pi 4 and Jetson Orin Nano as low-end and high-end devices, respectively. In the experiment, OpenACC showed the best performance on the low-end edge device and OpenSYCL on the high-end device due to the stability and optimization of system libraries.

Performance Comparison and Optimal Selection of Computing Techniques for Corridor Surveillance (회랑감시를 위한 컴퓨팅 기법의 성능 비교와 최적 선택 연구)

  • Gyeong-rae Jo;Seok-min Hong;Won-hyuck Choi
    • Journal of Advanced Navigation Technology
    • /
    • v.27 no.6
    • /
    • pp.770-775
    • /
    • 2023
  • Recently, as the amount of digital data increases exponentially, the importance of data processing systems is being emphasized. In this situation, the selection and construction of data processing systems are becoming more important. In this study, the performance of cloud computing (CC), edge computing (EC), and UAV-based intelligent edge computing (UEC) was compared as a way to solve this problem. The characteristics, strengths, and weaknesses of each method were analyzed. In particular, this study focused on real-time large-capacity data processing situations such as corridor monitoring. When conducting the experiment, a specific scenario was assumed and a penalty was given to the infrastructure. In this way, it was possible to evaluate performance in real situations more accurately. In addition, the effectiveness and limitations of each computing method were more clearly understood, and through this, the help was provided to enable more effective system selection.

Technologies of Intelligent Edge Computing and Networking (지능형 에지 컴퓨팅 및 네트워킹 기술)

  • Hong, S.W.;Lee, C.S.;Kim, S.C.;Kang, K.S.;Moon, S.;Shim, J.C.;Hong, S.B.;Ryu, H.Y.
    • Electronics and Telecommunications Trends
    • /
    • v.34 no.1
    • /
    • pp.23-35
    • /
    • 2019
  • In the upcoming post-app era, real-time, intelligent and immersive services such as autonomous vehicles, virtual secretaries, virtual reality, and augmented reality are expected to dominate. However, there is a growing demand for new networking and computing infrastructure capabilities because existing physical connection-oriented networks and centralized cloud-based service environments have inherent limitations to effectively accommodate these services. To this end, research on intelligent edge network computing technology is underway to analyze the contextual situation of human and things and to configure the service environment on the network edge so that the application services can be performed optimally. In this article, we describe the technology issues for edge network intelligence and introduce related research trends.

Trends in Ultra Low Power Intelligent Edge Semiconductor Technology (초저전력 엣지 지능형반도체 기술 동향)

  • Oh, K.I.;Kim, S.E.;Bae, Y.H.;Park, S.M.;Lee, J.J.;Kang, S.W.
    • Electronics and Telecommunications Trends
    • /
    • v.33 no.6
    • /
    • pp.24-33
    • /
    • 2018
  • In the age of IoT, in which everything is connected to a network, there have been increases in the amount of data traffic, latency, and the risk of personal privacy breaches that conventional cloud computing technology cannot cope with. The idea of edge computing has emerged as a solution to these issues, and furthermore, the concept of ultra-low power edge intelligent semiconductors in which the IoT device itself performs intelligent decisions and processes data has been established. The key elements of this function are an intelligent semiconductor based on artificial intelligence, connectivity for the efficient connection of neurons and synapses, and a large-scale spiking neural network simulation framework for the performance prediction of a neural network. This paper covers the current trends in ultra-low power edge intelligent semiconductors including issues regarding their technology and application.

An Edge AI Device based Intelligent Transportation System

  • Jeong, Youngwoo;Oh, Hyun Woo;Kim, Soohee;Lee, Seung Eun
    • Journal of information and communication convergence engineering
    • /
    • v.20 no.3
    • /
    • pp.166-173
    • /
    • 2022
  • Recently, studies have been conducted on intelligent transportation systems (ITS) that provide safety and convenience to humans. Systems that compose the ITS adopt architectures that applied the cloud computing which consists of a high-performance general-purpose processor or graphics processing unit. However, an architecture that only used the cloud computing requires a high network bandwidth and consumes much power. Therefore, applying edge computing to ITS is essential for solving these problems. In this paper, we propose an edge artificial intelligence (AI) device based ITS. Edge AI which is applicable to various systems in ITS has been applied to license plate recognition. We implemented edge AI on a field-programmable gate array (FPGA). The accuracy of the edge AI for license plate recognition was 0.94. Finally, we synthesized the edge AI logic with Magnachip/Hynix 180nm CMOS technology and the power consumption measured using the Synopsys's design compiler tool was 482.583mW.

Intelligent Resource Management Schemes for Systems, Services, and Applications of Cloud Computing Based on Artificial Intelligence

  • Lim, JongBeom;Lee, DaeWon;Chung, Kwang-Sik;Yu, HeonChang
    • Journal of Information Processing Systems
    • /
    • v.15 no.5
    • /
    • pp.1192-1200
    • /
    • 2019
  • Recently, artificial intelligence techniques have been widely used in the computer science field, such as the Internet of Things, big data, cloud computing, and mobile computing. In particular, resource management is of utmost importance for maintaining the quality of services, service-level agreements, and the availability of the system. In this paper, we review and analyze various ways to meet the requirements of cloud resource management based on artificial intelligence. We divide cloud resource management techniques based on artificial intelligence into three categories: fog computing systems, edge-cloud systems, and intelligent cloud computing systems. The aim of the paper is to propose an intelligent resource management scheme that manages mobile resources by monitoring devices' statuses and predicting their future stability based on one of the artificial intelligence techniques. We explore how our proposed resource management scheme can be extended to various cloud-based systems.

Distributed Edge Computing for DNA-Based Intelligent Services and Applications: A Review (딥러닝을 사용하는 IoT빅데이터 인프라에 필요한 DNA 기술을 위한 분산 엣지 컴퓨팅기술 리뷰)

  • Alemayehu, Temesgen Seyoum;Cho, We-Duke
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.12
    • /
    • pp.291-306
    • /
    • 2020
  • Nowadays, Data-Network-AI (DNA)-based intelligent services and applications have become a reality to provide a new dimension of services that improve the quality of life and productivity of businesses. Artificial intelligence (AI) can enhance the value of IoT data (data collected by IoT devices). The internet of things (IoT) promotes the learning and intelligence capability of AI. To extract insights from massive volume IoT data in real-time using deep learning, processing capability needs to happen in the IoT end devices where data is generated. However, deep learning requires a significant number of computational resources that may not be available at the IoT end devices. Such problems have been addressed by transporting bulks of data from the IoT end devices to the cloud datacenters for processing. But transferring IoT big data to the cloud incurs prohibitively high transmission delay and privacy issues which are a major concern. Edge computing, where distributed computing nodes are placed close to the IoT end devices, is a viable solution to meet the high computation and low-latency requirements and to preserve the privacy of users. This paper provides a comprehensive review of the current state of leveraging deep learning within edge computing to unleash the potential of IoT big data generated from IoT end devices. We believe that the revision will have a contribution to the development of DNA-based intelligent services and applications. It describes the different distributed training and inference architectures of deep learning models across multiple nodes of the edge computing platform. It also provides the different privacy-preserving approaches of deep learning on the edge computing environment and the various application domains where deep learning on the network edge can be useful. Finally, it discusses open issues and challenges leveraging deep learning within edge computing.

Exploring reward efficacy in traffic management using deep reinforcement learning in intelligent transportation system

  • Paul, Ananya;Mitra, Sulata
    • ETRI Journal
    • /
    • v.44 no.2
    • /
    • pp.194-207
    • /
    • 2022
  • In the last decade, substantial progress has been achieved in intelligent traffic control technologies to overcome consistent difficulties of traffic congestion and its adverse effect on smart cities. Edge computing is one such advanced progress facilitating real-time data transmission among vehicles and roadside units to mitigate congestion. An edge computing-based deep reinforcement learning system is demonstrated in this study that appropriately designs a multiobjective reward function for optimizing different objectives. The system seeks to overcome the challenge of evaluating actions with a simple numerical reward. The selection of reward functions has a significant impact on agents' ability to acquire the ideal behavior for managing multiple traffic signals in a large-scale road network. To ascertain effective reward functions, the agent is trained withusing the proximal policy optimization method in several deep neural network models, including the state-of-the-art transformer network. The system is verified using both hypothetical scenarios and real-world traffic maps. The comprehensive simulation outcomes demonstrate the potency of the suggested reward functions.

Deep Learning based Loss Recovery Mechanism for Video Streaming over Mobile Information-Centric Network

  • Han, Longzhe;Maksymyuk, Taras;Bao, Xuecai;Zhao, Jia;Liu, Yan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.9
    • /
    • pp.4572-4586
    • /
    • 2019
  • Mobile Edge Computing (MEC) and Information-Centric Networking (ICN) are essential network architectures for the future Internet. The advantages of MEC and ICN such as computation and storage capabilities at the edge of the network, in-network caching and named-data communication paradigm can greatly improve the quality of video streaming applications. However, the packet loss in wireless network environments still affects the video streaming performance and the existing loss recovery approaches in ICN does not exploit the capabilities of MEC. This paper proposes a Deep Learning based Loss Recovery Mechanism (DL-LRM) for video streaming over MEC based ICN. Different with existing approaches, the Forward Error Correction (FEC) packets are generated at the edge of the network, which dramatically reduces the workload of core network and backhaul. By monitoring network states, our proposed DL-LRM controls the FEC request rate by deep reinforcement learning algorithm. Considering the characteristics of video streaming and MEC, in this paper we develop content caching detection and fast retransmission algorithm to effectively utilize resources of MEC. Experimental results demonstrate that the DL-LRM is able to adaptively adjust and control the FEC request rate and achieve better video quality than the existing approaches.