• Title/Summary/Keyword: IoT architectures

Search Result 27, Processing Time 0.023 seconds

A Study on the Efficient Load Balancing Method Considering Real-time Data Entry form in SDN Environment (SDN 환경에서 실시간 데이터 유입형태를 고려한 효율적인 부하분산 기법 연구)

  • Ju-Seong Kim;Tae-Wook Kwon
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.6
    • /
    • pp.1081-1086
    • /
    • 2023
  • The rapid growth and increasing complexity of modern networks have highlighted the limitations of traditional network architectures. The emergence of SDN (Software-Defined Network) in response to these challenges has changed the existing network environment. The SDN separates the control unit and the data unit, and adjusts the network operation using a centralized controller. However, this structure has also recently caused a huge amount of traffic due to the rapid spread of numerous Internet of Things (IoT) devices, which has not only slowed the transmission speed of the network but also made it difficult to ensure quality of service (QoS). Therefore, this paper proposes a method of load distribution by switching the IP and any server (processor) from the existing data processing scheduling technique, RR (Round-Robin), to mapping when a large amount of data flows in from a specific IP, that is, server overload and data loss.

A Development of Fusion Processor Architecture for Efficient Main Memory Access in CPU-GPU Environment (CPU-GPU환경에서 효율적인 메인메모리 접근을 위한 융합 프로세서 구조 개발)

  • Park, Hyun-Moon;Kwon, Jin-San;Hwang, Tae-Ho;Kim, Dong-Sun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.11 no.2
    • /
    • pp.151-158
    • /
    • 2016
  • The HSA resolves an old problem with existing CPU and GPU architectures by allowing both units to directly access each other's memory pools via unified virtual memory. In a physically realized system, however, frequent data exchanges between CPU and GPU for a virtual memory block result bottlenecks and coherence request overheads. In this paper, we propose Fusion Processor Architecture for efficient access of main memory from both CPU and GPU. It consists of Job Manager, Re-mapper, and Pre-fetcher to control, organize, and distribute work loads and working areas for GPU cores. These components help on reducing memory exchanges between the two processors and improving overall efficiency by eliminating faulty page table requests. To verify proposed algorithm architectures, we develop an emulator based on QEMU, and compare several architectures such as CUDA(Compute Unified Device Architecture), OpenMP, OpenCL. As a result, Proposed fusion processor architectures show 198% faster than others by removing unnecessary memory copies and cache-miss overheads.

Machine-to-Machine Communications: Architectures, Standards and Applications

  • Chen, Min;Wan, Jiafu;Li, Fang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.6 no.2
    • /
    • pp.480-497
    • /
    • 2012
  • As a new business concept, machine-to-machine (M2M) communications are born from original telemetry technology with the intrinsic features of automatic data transmissions and measurement from remote sources typically by cable or radio. M2M includes a number of technologies that need to be combined in a compatible manner to enable its deployment over a broad market of consumer electronics. In order to provide better understanding for this emerging concept, the correlations among M2M, wireless sensor networks, cyber-physical systems (CPS), and internet of things are first analyzed in this paper. Then, the basic M2M architecture is introduced and the key elements of the architecture are presented. Furthermore, the progress of global M2M standardization is reviewed, and some representative applications (i.e., smart home, smart grid and health care) are given to show that the M2M technologies are gradually utilized to benefit people's life. Finally, a novel M2M system integrating intelligent road with unmanned vehicle is proposed in the form of CPS, and an example of cyber-transportation systems for improving road safety and efficiency are introduced.

Emerging Technologies for Sustainable Smart City Network Security: Issues, Challenges, and Countermeasures

  • Jo, Jeong Hoon;Sharma, Pradip Kumar;Sicato, Jose Costa Sapalo;Park, Jong Hyuk
    • Journal of Information Processing Systems
    • /
    • v.15 no.4
    • /
    • pp.765-784
    • /
    • 2019
  • The smart city is one of the most promising, prominent, and challenging applications of the Internet of Things (IoT). Smart cities rely on everything connected to each other. This in turn depends heavily on technology. Technology literacy is essential to transform a city into a smart, connected, sustainable, and resilient city where information is not only available but can also be found. The smart city vision combines emerging technologies such as edge computing, blockchain, artificial intelligence, etc. to create a sustainable ecosystem by dramatically reducing latency, bandwidth usage, and power consumption of smart devices running various applications. In this research, we present a comprehensive survey of emerging technologies for a sustainable smart city network. We discuss the requirements and challenges for a sustainable network and the role of heterogeneous integrated technologies in providing smart city solutions. We also discuss different network architectures from a security perspective to create an ecosystem. Finally, we discuss the open issues and challenges of the smart city network and provide suitable recommendations to resolve them.

A Study on CPPS Architecture integrated with Centralized OPC UA Server (중앙 집중식 OPC UA 서버와 통합 된 CPPS 아키텍처에 관한 연구)

  • Jo, Guejong;Jang, Su-Hwan;Jeong, Jongpil
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.3
    • /
    • pp.73-82
    • /
    • 2019
  • In order to build a smart factory, building a CPPS (Cyber Physical Product System) is an important system that must be accompanied. Through the CPPS, it is the reality of smart factories to move physical factories to a digital-based cyber world and to intelligently and autonomously monitor and control them. But The existing CPPS architectures present only an abstract modeling architecture, and the research that applied the OPC UA Framework (Open Platform Communication Unified Architecture), an international standard for data exchange in the smart factory, as the basic system of CPPS It was insufficient. Therefore, it is possible to implement CPPS that can include both cloud and IoT by collecting field data distributed by CPPS architecture applicable to actual factories and concentrating data processing in a centralized In this study, we implemented CPPS architecture through central OPC UA Server based on OPC UA conforming to central processing OPC UA Framework, and how CPPS logical process and data processing process are automatically generated through OPC UA modeling processing We have proposed the CPPS architecture including the model factory and implemented the model factory to study its performance and usability.

Data anomaly detection and Data fusion based on Incremental Principal Component Analysis in Fog Computing

  • Yu, Xue-Yong;Guo, Xin-Hui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.10
    • /
    • pp.3989-4006
    • /
    • 2020
  • The intelligent agriculture monitoring is based on the perception and analysis of environmental data, which enables the monitoring of the production environment and the control of environmental regulation equipment. As the scale of the application continues to expand, a large amount of data will be generated from the perception layer and uploaded to the cloud service, which will bring challenges of insufficient bandwidth and processing capacity. A fog-based offline and real-time hybrid data analysis architecture was proposed in this paper, which combines offline and real-time analysis to enable real-time data processing on resource-constrained IoT devices. Furthermore, we propose a data process-ing algorithm based on the incremental principal component analysis, which can achieve data dimensionality reduction and update of principal components. We also introduce the concept of Squared Prediction Error (SPE) value and realize the abnormal detection of data through the combination of SPE value and data fusion algorithm. To ensure the accuracy and effectiveness of the algorithm, we design a regular-SPE hybrid model update strategy, which enables the principal component to be updated on demand when data anomalies are found. In addition, this strategy can significantly reduce resource consumption growth due to the data analysis architectures. Practical datasets-based simulations have confirmed that the proposed algorithm can perform data fusion and exception processing in real-time on resource-constrained devices; Our model update strategy can reduce the overall system resource consumption while ensuring the accuracy of the algorithm.

An Analysis of Big Video Data with Cloud Computing in Ubiquitous City (클라우드 컴퓨팅을 이용한 유시티 비디오 빅데이터 분석)

  • Lee, Hak Geon;Yun, Chang Ho;Park, Jong Won;Lee, Yong Woo
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.45-52
    • /
    • 2014
  • The Ubiquitous-City (U-City) is a smart or intelligent city to satisfy human beings' desire to enjoy IT services with any device, anytime, anywhere. It is a future city model based on Internet of everything or things (IoE or IoT). It includes a lot of video cameras which are networked together. The networked video cameras support a lot of U-City services as one of the main input data together with sensors. They generate huge amount of video information, real big data for the U-City all the time. It is usually required that the U-City manipulates the big data in real-time. And it is not easy at all. Also, many times, it is required that the accumulated video data are analyzed to detect an event or find a figure among them. It requires a lot of computational power and usually takes a lot of time. Currently we can find researches which try to reduce the processing time of the big video data. Cloud computing can be a good solution to address this matter. There are many cloud computing methodologies which can be used to address the matter. MapReduce is an interesting and attractive methodology for it. It has many advantages and is getting popularity in many areas. Video cameras evolve day by day so that the resolution improves sharply. It leads to the exponential growth of the produced data by the networked video cameras. We are coping with real big data when we have to deal with video image data which are produced by the good quality video cameras. A video surveillance system was not useful until we find the cloud computing. But it is now being widely spread in U-Cities since we find some useful methodologies. Video data are unstructured data thus it is not easy to find a good research result of analyzing the data with MapReduce. This paper presents an analyzing system for the video surveillance system, which is a cloud-computing based video data management system. It is easy to deploy, flexible and reliable. It consists of the video manager, the video monitors, the storage for the video images, the storage client and streaming IN component. The "video monitor" for the video images consists of "video translater" and "protocol manager". The "storage" contains MapReduce analyzer. All components were designed according to the functional requirement of video surveillance system. The "streaming IN" component receives the video data from the networked video cameras and delivers them to the "storage client". It also manages the bottleneck of the network to smooth the data stream. The "storage client" receives the video data from the "streaming IN" component and stores them to the storage. It also helps other components to access the storage. The "video monitor" component transfers the video data by smoothly streaming and manages the protocol. The "video translator" sub-component enables users to manage the resolution, the codec and the frame rate of the video image. The "protocol" sub-component manages the Real Time Streaming Protocol (RTSP) and Real Time Messaging Protocol (RTMP). We use Hadoop Distributed File System(HDFS) for the storage of cloud computing. Hadoop stores the data in HDFS and provides the platform that can process data with simple MapReduce programming model. We suggest our own methodology to analyze the video images using MapReduce in this paper. That is, the workflow of video analysis is presented and detailed explanation is given in this paper. The performance evaluation was experiment and we found that our proposed system worked well. The performance evaluation results are presented in this paper with analysis. With our cluster system, we used compressed $1920{\times}1080(FHD)$ resolution video data, H.264 codec and HDFS as video storage. We measured the processing time according to the number of frame per mapper. Tracing the optimal splitting size of input data and the processing time according to the number of node, we found the linearity of the system performance.