• Title/Summary/Keyword: Distributed Stream Processing

Search Result 55, Processing Time 0.021 seconds

Efficient Locality-Aware Traffic Distribution in Apache Storm (Apache Storm에서 지역성을 고려한 효율적인 트래픽 분배)

  • Son, Siwoon;Lee, Sanghun;Moon, Yang-Sae
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.12
    • /
    • pp.677-683
    • /
    • 2017
  • Apache Storm is a representative real-time distributed processing system, which is able to process data streams quickly over distributed servers. Storm currently provides several stream grouping methods to distribute data traffic to multiple servers. Among them, the shuffle grouping may cause a processing delay problem and the local-or-shuffle grouping used to solve the problem may cause the problem of concentrating the traffic on a specific node. In this paper, we propose the locality-aware grouping to solve the problems that may arise in the existing Storm grouping methods. Experimental results show that the proposed locality-aware grouping is considerably superior to the existing shuffle grouping and the local-or-shuffle grouping. These results show that the new grouping is an excellent approach considering both the locality and load balancing which are limitations of the existing Storm.

Evaluation of Multivariate Stream Data Reduction Techniques (다변량 스트림 데이터 축소 기법 평가)

  • Jung, Hung-Jo;Seo, Sung-Bo;Cheol, Kyung-Joo;Park, Jeong-Seok;Ryu, Keun-Ho
    • The KIPS Transactions:PartD
    • /
    • v.13D no.7 s.110
    • /
    • pp.889-900
    • /
    • 2006
  • Even though sensor networks are different in user requests and data characteristics depending on each application area, the existing researches on stream data transmission problem focus on the performance improvement of their methods rather than considering the original characteristic of stream data. In this paper, we introduce a hierarchical or distributed sensor network architecture and data model, and then evaluate the multivariate data reduction methods suitable for user requirements and data features so as to apply reduction methods alternatively. To assess the relative performance of the proposed multivariate data reduction methods, we used the conventional techniques, such as Wavelet, HCL(Hierarchical Clustering), Sampling and SVD (Singular Value Decomposition) as well as the experimental data sets, such as multivariate time series, synthetic data and robot execution failure data. The experimental results shows that SVD and Sampling method are superior to Wavelet and HCL ia respect to the relative error ratio and execution time. Especially, since relative error ratio of each data reduction method is different according to data characteristic, it shows a good performance using the selective data reduction method for the experimental data set. The findings reported in this paper can serve as a useful guideline for sensor network application design and construction including multivariate stream data.

The Design of Intramedia Synchronization based QoS (QoS 기반의 미디어내 동기화 설계)

  • Park, Hong-Jin;Lee, Jun-Yeon;Kim, Chang-Min
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.8
    • /
    • pp.2042-2049
    • /
    • 1999
  • In contrast to traditional data communication, the multimedia communication must support variable requirements to multimedia application service. Specially, it is a critical issue that the synchronization supports a continuous and temporal relationship media playout in a distributed environment. But the continuously generated media in source can't guarantee the constant interval on playout for the destination due to the inconsistent network transmission delay and transmission error. In this paper, we introduce a intramedia synchronization mechanism for the continuous and synchronous playout of multimedia stream such as audio and video across a communication network. The proposed mechanism is based on the user defined QoS(Quality of Service) and the playout correction algorithm to guarantee the continuous media playout at the receiver.

  • PDF

An Object-Oriented, Client-Server Architecture for a Generalized Multimedia a Processing Model in a Distributed Multimedia System (분산 멀티미디어 시스템을 위한 범용 멀티미디어 처리 모델의 객체지향, 클라이언트 -서버 구조)

  • Kim, Doo-Hyun;Lim, Young-Hwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.3 no.1
    • /
    • pp.9-32
    • /
    • 1996
  • In this paper, we describe a multimedia data processing model that supports wide variety of applications based on multimedia production model. This model supports network-transparent access to stored multimedia data, real-time mult media input devices, and multimedia processing. The model addresses real-time dato a switching and delivery, as well as acquisition, processing, and output. Most translation, compression, and synchronization services are integrated. This model consists of three layers: (1) stream, (2) multimedia presentation, (3) hyperpresentaion. This paper describes the data abstractions associated with each layer. These data abstractions provide a framework for defining the services provided by each layer, and describe the object-oriented mechanisms that provide those services. A sample scenario is presentated to illustrate the use of this model. A server-client architecture and implementation issues, and future directions are also discussed. sed.

  • PDF

A Scheduling Algorithm for Parsing of MPEG Video on the Heterogeneous Distributed Environment (이질적인 분산 환경에서의 MPEG비디오의 파싱을 위한 스케줄링 알고리즘)

  • Nam Yunyoung;Hwang Eenjun
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.12
    • /
    • pp.673-681
    • /
    • 2004
  • As the use of digital videos is getting popular, there is an increasing demand for efficient browsing and retrieval of video. To support such operations, effective video indexing should be incorporated. One of the most fundamental steps in video indexing is to parse video stream into shots and scenes. Generally, it takes long time to parse a video due to the huge amount of computation in a traditional single computing environment. Previous studies had widely used Round Robin scheduling which basically allocates tasks to each slave for a time interval of one quantum. This scheduling is difficult to adapt in a heterogeneous environment. In this paper, we propose two different parallel parsing algorithms which are Size-Adaptive Round Robin and Dynamic Size-Adaptive Round Robin for the heterogeneous distributed computing environments. In order to show their performance, we perform several experiments and show some of the results.

The Method for Real-time Complex Event Detection of Unstructured Big data (비정형 빅데이터의 실시간 복합 이벤트 탐지를 위한 기법)

  • Lee, Jun Heui;Baek, Sung Ha;Lee, Soon Jo;Bae, Hae Young
    • Spatial Information Research
    • /
    • v.20 no.5
    • /
    • pp.99-109
    • /
    • 2012
  • Recently, due to the growth of social media and spread of smart-phone, the amount of data has considerably increased by full use of SNS (Social Network Service). According to it, the Big Data concept is come up and many researchers are seeking solutions to make the best use of big data. To maximize the creative value of the big data held by many companies, it is required to combine them with existing data. The physical and theoretical storage structures of data sources are so different that a system which can integrate and manage them is needed. In order to process big data, MapReduce is developed as a system which has advantages over processing data fast by distributed processing. However, it is difficult to construct and store a system for all key words. Due to the process of storage and search, it is to some extent difficult to do real-time processing. And it makes extra expenses to process complex event without structure of processing different data. In order to solve this problem, the existing Complex Event Processing System is supposed to be used. When it comes to complex event processing system, it gets data from different sources and combines them with each other to make it possible to do complex event processing that is useful for real-time processing specially in stream data. Nevertheless, unstructured data based on text of SNS and internet articles is managed as text type and there is a need to compare strings every time the query processing should be done. And it results in poor performance. Therefore, we try to make it possible to manage unstructured data and do query process fast in complex event processing system. And we extend the data complex function for giving theoretical schema of string. It is completed by changing the string key word into integer type with filtering which uses keyword set. In addition, by using the Complex Event Processing System and processing stream data at real-time of in-memory, we try to reduce the time of reading the query processing after it is stored in the disk.

A Framework and Synchronization Mechanism for Real-time Multimedia Streaming Services based on the Time-triggered Message-triggered Object (실시간 멀티미디어 스트리밍 서비스를 위한 Time-triggered Message-triggered Object 기반의 프레임워크 및 동기화 메커니즘)

  • Jo, Eun-Hwan;Kim, Moon-Hae
    • The KIPS Transactions:PartC
    • /
    • v.13C no.6 s.109
    • /
    • pp.669-676
    • /
    • 2006
  • In this paper, we present a new framework and stream synchronization mechanism to effectively support developing real-time multimedia streaming services by using a real-time object model named TMO (Time-triggered Message-triggered Object). The purpose of the framework is twofold. Firstly, the framework helps developers to design complex distributed real-time multimedia streaming services. Secondly, it supports timely streaming facilities. In order to achieve these goals, our framework is consist of Multimedia Streaming TMO, MMStreaming TMO Support Library and TMO Support Middleware. The time-triggered spontaneous feature of the MMStream TMO and a global-time based synchronization scheme is used as a regulator against the irregular deliveries and processing of media units caused by QoS non-guaranteed systems and communication channels. In conclusion, timely service capability of our framework is expected to contributed to overall enhancement of the real-time multimedia streaming.

An Integrated Synchronization Method for a Hyperpresentation in a distributed Computing Environment (분산 컴퓨팅환경에서 하이퍼 프리젠테이션을 위한 통합 동기화 기법)

  • Lim, Young-Hwan;Kim, Doo-Hyun;Kung, Sang-Hwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.6
    • /
    • pp.1441-1456
    • /
    • 1998
  • The concept of a hyperpresmtation, as an extension of a hypermedia, is the presentation in which time-varying multimedia presentations are dynamically linked together and a hyperlink's context can be changed over time at any time during a continuous presentation. Problems caused by integrating the hyperpresentation into an existing multimedia system which handles a sequential presentation only are, how to describe the hyperprcsentation, how to set up a hyperlink on a continuous media, and how to check the consistency of the synchronized presentations. In this paper. a new synchronization description method for the hyperpresentation and a method for setting a hyper link on a continuous media during" presentation are proposed after havin!; SHrvey of existing methods, The proposed method deals with only the DC value in a stream ut a DCT based compressed data for checking a condition of te link. Finally, the method for checking the consistency of mixed presentations before actual play of the hnlerpresentation is described. Proposed methods are implemented on MuX(Multimedia IO Server) where a sample scenario is tested.

  • PDF

A Study on the Data Collection Methods based Hadoop Distributed Environment (하둡 분산 환경 기반의 데이터 수집 기법 연구)

  • Jin, Go-Whan
    • Journal of the Korea Convergence Society
    • /
    • v.7 no.5
    • /
    • pp.1-6
    • /
    • 2016
  • Many studies have been carried out for the development of big data utilization and analysis technology recently. There is a tendency that government agencies and companies to introduce a Hadoop of a processing platform for analyzing big data is increasing gradually. Increased interest with respect to the processing and analysis of these big data collection technology of data has become a major issue in parallel to it. However, study of the collection technology as compared to the study of data analysis techniques, it is insignificant situation. Therefore, in this paper, to build on the Hadoop cluster is a big data analysis platform, through the Apache sqoop, stylized from relational databases, to collect the data. In addition, to provide a sensor through the Apache flume, a system to collect on the basis of the data file of the Web application, the non-structured data such as log files to stream. The collection of data through these convergence would be able to utilize as a basic material of big data analysis.

Design and Implementation of a Transparent Security Infrastructure using Filter and Transformer (필터와 트랜스포머를 이용한 투명한 보안기반의 설계 및 구현)

  • Kim, Yong-Min;Lee, Do-Heon;Noh, Bong-Nam;Choi, Rak-Man;Ine, So-Ran
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.4
    • /
    • pp.975-983
    • /
    • 1998
  • In these days, information communication systems are based on both open distributed computing technologies and object-oriented techniques like inheritance, encapsulation and object reuse to support various system configuration and application. As information systems are interconnected through unsecure networks, the need for the secure information exchange is more critical than before. In this paper, we have designed and implemented a transparent CORBA-basce Security infrastructure with authentication, security context association, access control and security information management to support a secure applications in distributed object environment. SESAME Ver. 4 was adopted as an external security service to manage user privilege attributes and to distribute keys for data encryption, decryption and integrity. Using filter and transformer with an interface to Object Request Broker, it provides a transparent security service to applications. The filter objects are special classes that allow additional parameters to be inserted into messages before they are sent and removed just after they are received. The transformer objects are special classes that allow direct access to the byte stream of every messages for encryption and decryption before it is sent and just after it is received. This study is to implement the access control interceptor(ACI) and the secure invocation interceptor(SII) of secure ORB defined in CORBA using filter and transformer.

  • PDF