• Title/Summary/Keyword: Edge devices

Search Result 438, Processing Time 0.022 seconds

A Secure Subscription-Push Service Scheme Based on Blockchain and Edge Computing for IoT

  • Deng, Yinjuan;Wang, Shangping;Zhang, Qian;Zhang, Duo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.445-466
    • /
    • 2022
  • As everything linking to the internet, people can subscribe to various services from a service provider to facilitate their lives through the Internet of Things (IoT). An obligatory thing for the service provider is that they should push the service data safely and timely to multiple IoT terminal devices regularly after the IoT devices accomplishing the service subscription. In order to control the service message received by the legal devices as while as keep the confidentiality of the data, the public key encryption algorithm is utilized. While the existing public encryption algorithms for push service are too complicated for IoT devices, and almost of the current subscription schemes based on push mode are relying on centralized organization which may suffer from centralized entity corruption or single point of failure. To address these issues, we design a secure subscription-push service scheme based on blockchain and edge computing in this article, which is decentralized with secure architecture for the subscription and push of service. Furthermore, inspired by broadcast encryption and multicast encryption, a new encryption algorithm is designed to manage the permissions of IoT devices together with smart contract, and to protect the confidentiality of push messages, which is suitable for IoT devices. The edge computing nodes, in the new system architecture, maintain the blockchain to ensure the impartiality and traceability of service subscriptions and push messages, meanwhile undertake some calculations for IoT devices with limited computing power. The legalities of subscription services are guaranteed by verifying subscription tags on the smart contract. Lastly, the analysis indicates that the scheme is reliable, and the proposed encryption algorithm is safe and efficient.

Distributed Edge Computing for DNA-Based Intelligent Services and Applications: A Review (딥러닝을 사용하는 IoT빅데이터 인프라에 필요한 DNA 기술을 위한 분산 엣지 컴퓨팅기술 리뷰)

  • Alemayehu, Temesgen Seyoum;Cho, We-Duke
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.12
    • /
    • pp.291-306
    • /
    • 2020
  • Nowadays, Data-Network-AI (DNA)-based intelligent services and applications have become a reality to provide a new dimension of services that improve the quality of life and productivity of businesses. Artificial intelligence (AI) can enhance the value of IoT data (data collected by IoT devices). The internet of things (IoT) promotes the learning and intelligence capability of AI. To extract insights from massive volume IoT data in real-time using deep learning, processing capability needs to happen in the IoT end devices where data is generated. However, deep learning requires a significant number of computational resources that may not be available at the IoT end devices. Such problems have been addressed by transporting bulks of data from the IoT end devices to the cloud datacenters for processing. But transferring IoT big data to the cloud incurs prohibitively high transmission delay and privacy issues which are a major concern. Edge computing, where distributed computing nodes are placed close to the IoT end devices, is a viable solution to meet the high computation and low-latency requirements and to preserve the privacy of users. This paper provides a comprehensive review of the current state of leveraging deep learning within edge computing to unleash the potential of IoT big data generated from IoT end devices. We believe that the revision will have a contribution to the development of DNA-based intelligent services and applications. It describes the different distributed training and inference architectures of deep learning models across multiple nodes of the edge computing platform. It also provides the different privacy-preserving approaches of deep learning on the edge computing environment and the various application domains where deep learning on the network edge can be useful. Finally, it discusses open issues and challenges leveraging deep learning within edge computing.

A Monitoring Scheme Based on Artificial Intelligence in Mobile Edge Cloud Computing Environments (모바일 엣지 클라우드 환경에서 인공지능 기반 모니터링 기법)

  • Lim, JongBeom;Choi, HeeSeok;Yu, HeonChang
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.7 no.2
    • /
    • pp.27-32
    • /
    • 2018
  • One of the crucial issues in mobile edge cloud computing environments is to monitor mobile devices. Due to the inherit properties of mobile devices, they are prone to unstable behavior that leads to failures. In order to satisfy the service level agreement (SLA), the mobile edge cloud administrators should take appropriate measures through a monitoring scheme. In this paper, we propose a monitoring scheme of mobile devices based on artificial intelligence in mobile edge cloud computing environments. The proposed monitoring scheme is able to measure faults of mobile devices based on previous and current monitoring information. To this end, we adapt the hidden markov chain model, one of the artificial intelligence technologies, to monitor mobile devices. We validate our monitoring scheme based on the hidden markov chain model. The proposed monitoring scheme can also be used in general cloud computing environments to monitor virtual machines.

A Four-Layer Robust Storage in Cloud using Privacy Preserving Technique with Reliable Computational Intelligence in Fog-Edge

  • Nirmala, E.;Muthurajkumar, S.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.9
    • /
    • pp.3870-3884
    • /
    • 2020
  • The proposed framework of Four Layer Robust Storage in Cloud (FLRSC) architecture involves host server, local host and edge devices in addition to Virtual Machine Monitoring (VMM). The goal is to protect the privacy of stored data at edge devices. The computational intelligence (CI) part of our algorithm distributes blocks of data to three different layers by partially encoded and forwarded for decoding to the next layer using hash and greed Solomon algorithms. VMM monitoring uses snapshot algorithm to detect intrusion. The proposed system is compared with Tiang Wang method to validate efficiency of data transfer with security. Hence, security is proven against the indexed efficiency. It is an important study to integrate communication between local host software and nearer edge devices through different channels by verifying snapshot using lamport mechanism to ensure integrity and security at software level thereby reducing the latency. It also provides thorough knowledge and understanding about data communication at software level with VMM. The performance evaluation and feasibility study of security in FLRSC against three-layered approach is proven over 232 blocks of data with 98% accuracy. Practical implications and contributions to the growing knowledge base are highlighted along with directions for further research.

Design and Evaluation of an Edge-Fog Cloud-based Hierarchical Data Delivery Scheme for IoT Applications (사물인터넷 응용을 위한 에지-포그 클라우드 기반 계층적 데이터 전달 방법의 설계 및 평가)

  • Bae, Ihn-Han
    • Journal of Internet Computing and Services
    • /
    • v.19 no.1
    • /
    • pp.37-47
    • /
    • 2018
  • The number of capabilities of Internet of Things (IoT) devices will exponentially grow over the next years. These devices may generate a vast amount of time-constrained data. In the context of IoT, data management should act as a layer between the objects and devices generating the data and the applications accessing the data for analysis purposes and services. In addition, most of IoT services will be content-centric rather than host centric to increase the data availability and the efficiency of data delivery. IoT will enable all the communication devices to be interconnected and make the data generated by or associated with devices or objects globally accessible. Also, fog computing keeps data and computation close to end users at the edge of network, and thus provides a new breed of applications and services to end users with low latency, high bandwidth, and geographically distributed. In this paper, we propose Edge-Fog cloud-based Hierarchical Data Delivery ($EFcHD^2$) method that effectively and reliably delivers IoT data to associated with IoT applications with ensuring time sensitivity. The proposed $EFcHD^2$ method stands on basis of fully decentralized hybrid of Edge and Fog compute cloud model, Edge-Fog cloud, and uses information-centric networking and bloom filters. In addition, it stores the replica of IoT data or the pre-processed feature data by edge node in the appropriate locations of Edge-Fog cloud considering the characteristic of IoT data: locality, size, time sensitivity and popularity. Then, the performance of $EFcHD^2$ method is evaluated through an analytical model, and is compared to fog server-based and Content-Centric Networking (CCN)-based data delivery methods.

A Study on Field Ring Design of 600 V Super Junction Power MOSFET (600 V급 Super Junction MOSFET을 위한 Field Ring 설계의 관한 연구)

  • Hong, Young-Sung;Jung, Eun-Sik;Kang, Ey-Goo
    • Journal of the Korean Institute of Electrical and Electronic Material Engineers
    • /
    • v.25 no.4
    • /
    • pp.276-281
    • /
    • 2012
  • Power semiconductor devices are widely used as high voltage applications to inverters and motor drivers, etc. The blocking voltage is one of the most important parameters for power semiconductor devices. Generally most of field effect concentrations shows on the edge of power devices. Can be improve the breakdown characteristic using edge termination technology. In this paper, considering the variables that affect the breakdown voltage and optimization of parameters result for 600 V Super Junction MOSFET Field ring.

Effect of Polishing Grinding Conditions on Alkali-free Glass (무 알칼리 유리의 연마 조건에 따른 영향)

  • Park, Younghee;Hong, Minsung
    • Journal of the Korean Society of Manufacturing Technology Engineers
    • /
    • v.25 no.6
    • /
    • pp.440-444
    • /
    • 2016
  • Owing to the introduction of various IT devices with emphasis on portability and design, the TFT (thin film transistor liquid crystal display) panel applied to IT devices has the same shape as the product, and the portability requirement of IT devices has resulted in a need for panels with higher rigidity. In this study, the effect of grinding conditions such as the feed rate and edge speed of edge grinding on the surface roughness and chipping of the machined surface is investigated using a metal bond wheel. During edge grinding of alkari-free glass, weak mechanical property of glass results in big chipping owing to generation of tensile stress at the end of grining operation. The results of this study show that the grinding characteristics of alkali-free glass are obtained and meet industry requirements.

Lightweight image classifier for CIFAR-10

  • Sharma, Akshay Kumar;Rana, Amrita;Kim, Kyung Ki
    • Journal of Sensor Science and Technology
    • /
    • v.30 no.5
    • /
    • pp.286-289
    • /
    • 2021
  • Image classification is one of the fundamental applications of computer vision. It enables a system to identify an object in an image. Recently, image classification applications have broadened their scope from computer applications to edge devices. The convolutional neural network (CNN) is the main class of deep learning neural networks that are widely used in computer tasks, and it delivers high accuracy. However, CNN algorithms use a large number of parameters and incur high computational costs, which hinder their implementation in edge hardware devices. To address this issue, this paper proposes a lightweight image classifier that provides good accuracy while using fewer parameters. The proposed image classifier diverts the input into three paths and utilizes different scales of receptive fields to extract more feature maps while using fewer parameters at the time of training. This results in the development of a model of small size. This model is tested on the CIFAR-10 dataset and achieves an accuracy of 90% using .26M parameters. This is better than the state-of-the-art models, and it can be implemented on edge devices.

Task Scheduling on Cloudlet in Mobile Cloud Computing with Load Balancing

  • Poonam;Suman Sangwan
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.10
    • /
    • pp.73-80
    • /
    • 2023
  • The recent growth in the use of mobile devices has contributed to increased computing and storage requirements. Cloud computing has been used over the past decade to cater to computational and storage needs over the internet. However, the use of various mobile applications like Augmented Reality (AR), M2M Communications, V2X Communications, and the Internet of Things (IoT) led to the emergence of mobile cloud computing (MCC). All data from mobile devices is offloaded and computed on the cloud, removing all limitations incorporated with mobile devices. However, delays induced by the location of data centers led to the birth of edge computing technologies. In this paper, we discuss one of the edge computing technologies, i.e., cloudlet. Cloudlet brings the cloud close to the end-user leading to reduced delay and response time. An algorithm is proposed for scheduling tasks on cloudlet by considering VM's load. Simulation results indicate that the proposed algorithm provides 12% and 29% improvement over EMACS and QRR while balancing the load.

Game Theory-Based Scheme for Optimizing Energy and Latency in LEO Satellite-Multi-access Edge Computing

  • Ducsun Lim;Dongkyun Lim
    • International journal of advanced smart convergence
    • /
    • v.13 no.2
    • /
    • pp.7-15
    • /
    • 2024
  • 6G network technology represents the next generation of communications, supporting high-speed connectivity, ultra-low latency, and integration with cutting-edge technologies, such as the Internet of Things (IoT), virtual reality, and autonomous vehicles. These advancements promise to drive transformative changes in digital society. However, as technology progresses, the demand for efficient data transmission and energy management between smart devices and network equipment also intensifies. A significant challenge within 6G networks is the optimization of interactions between satellites and smart devices. This study addresses this issue by introducing a new game theory-based technique aimed at minimizing system-wide energy consumption and latency. The proposed technique reduces the processing load on smart devices and optimizes the offloading decision ratio to effectively utilize the resources of Low-Earth Orbit (LEO) satellites. Simulation results demonstrate that the proposed technique achieves a 30% reduction in energy consumption and a 40% improvement in latency compared to existing methods, thereby significantly enhancing performance.