• Title/Summary/Keyword: kafka

Search Result 34, Processing Time 0.025 seconds

Monitoring Tools for Efficient Overload Measurements in Apache Kafka (Apache Kafka에서 효율적인 과부하 측정을 위한 모니터링 도구)

  • Bang, Jiwon;Son, Siwoon;Moon, Yang-Sae;Choi, Mi-Jung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2017.11a
    • /
    • pp.52-54
    • /
    • 2017
  • 실시간으로 빠르게 발생하는 대용량 데이터를 다루기 위해 Apache Storm, Apache Spark 등 실시간 데이터 스트림 처리 기술에 대한 연구가 활발하다. 대부분의 실시간 처리 기술들은 단독으로 사용하기에 어려움이 있으며, 데이터 스트림의 입출력을 위해 메시징 시스템과 함께 사용하는 것이 일반적이다. Apache Kafka는 대표적인 분산 메시징 시스템으로써, 실시간으로 발생하는 대용량의 로그 데이터를 전달하는데 특화된 시스템이다. 현재 Kafka를 위한 다양한 성능 모니터링 도구들이 존재한다. 이러한 모니터링 도구들은 Kafka에서 처리되는 데이터의 양 이외에도 유입 데이터의 크기, 수집 속도, 처리 속도 등 다양한 데이터들을 관찰할 수 있다. 본 논문은 Kafka에서 제공하는 도구와 오픈 소스로 제공되는 여러 개의 도구들을 비교하여, 향후 Kafka의 로드 쉐딩에 대한 연구에 적용할 수 있는 최적의 모니터링 도구를 선별하고자 한다.

A Kafka-based Data Sharing Method for Educational Video Services (교육 동영상 공유 서비스의 카프카 기반 데이터 공유 방안)

  • Lee, Hyeon sup;Kim, Jin-Deog
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.574-576
    • /
    • 2021
  • It is necessary to introduce micro-service techniques when constructing large-scale operating systems or systems that take into account scalability. Kafka is a message queue with the pub/sub model, which has features that are well applied to distributed environments and is also suitable for microservices in that it can utilize various data sources. In this paper, we propose a data sharing method for educational video sharing services using Apache's Kafka. The proposed system builds a Kafka cluster for the educational video sharing service to share various data, and also uses a spark cluster to link with recommendation systems based on similarities in educational videos. We also present a way to share various data sources, such as files, various DBMS, etc.

  • PDF

Establish a security control system through attack packet analysis with Suricata, Elastic Stack, and Kafka (Suricata와 Elastic Stack, Kafka를 이용한 공격 패킷 분석 및 보안관제 시스템 구축)

  • Lee, Da-Eun;Lee, Hye-Rin;Jo, Min-Gyu
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.11a
    • /
    • pp.1144-1147
    • /
    • 2021
  • 코로나19 대유행으로 인해 전 세계가 원격으로 일상을 옮겨가면서 인터넷 트래픽량이 증가하고 보안 위협 또한 높아졌다. 높은 보안성이 요구되는 현 상황에 대응하기 위해 본 논문에서는 Suricata와 Elastic Stack, Kafka를 이용해 보안관제 로그 분석시스템을 구축하였다. 실시간으로 공격을 탐지하고 로그를 수집해 유의미한 데이터를 도출하여 시각화한다. 또한 시각화 한 대시보드를 제공함으로써 사용자는 공격의 위험도를 파악할 수 있고 앞으로의 공격을 대비할 수 있다.

Odd Fellows: Hannah Arendt and Philip Roth

  • Nadel, Ira
    • Journal of English Language & Literature
    • /
    • v.64 no.2
    • /
    • pp.151-170
    • /
    • 2018
  • This paper examines the relationship and ideas of Hannah Arendt and Philip Roth including how they met, their correspondence and intellectual parallels, particularly in their shared criticism of Jewish ideals and culture in Europe and North America. It analyzes similarities in their careers and texts, especially between Eichmann in Jerusalem and Operation Shylock, as well as The Ghost Writer, while measuring their reception as social commentators and writers. Kafka was an important figure for both writers, Arendt's earliest writing engaged with the significance of Kafka in understanding and criticizing twentieth century political and cultural values in Europe. For Roth, Kafka offered a similar critique of moral principles he found corroded in North American Jewish life. Arendt connected with other writers, notably Isak Dinesen, W. H. Auden, Randall Jarrell and William Styron who further linked the two: he knew both Arendt and Roth and cited, incorrectly, a work by Arendt as the source for the key incident in his 1979 novel Sophie's Choice. He claimed it was Eichmann in Jerusalem; it was Arendt's The Origins of Totalitarianism. Arendt's reaction to Roth's fiction, however, remains a mystery: she died in 1975, before Roth began to seriously and consistently engage with Holocaust issues in works like The Ghost Writer (1979) and Operation Shylock (1993). Yet even in death they are joined. Their graves are only steps apart at the Bard College Cemetery in Annandale-on-Hudson, New York.

SaaS application mashup based on High Speed Message Processing

  • Chen, Zhiguo;Kim, Myoungjin;Cui, Yun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.5
    • /
    • pp.1446-1465
    • /
    • 2022
  • Diversified SaaS applications allow users more choices to use, according to their own preferences. However, the diversification of SaaS applications also makes it impossible for users to choose the best one. Furthermore, users can't take advantage of the functionality between SaaS applications. In this paper, we propose a platform that provides an SaaS mashup service, by extracting interoperable service functions from SaaS-based applications that independent vendors deploy and supporting a customized service recommendation function through log data binding in the cloud environment. The proposed SaaS mashup service platform consists of a SaaS aggregation framework and a log data binding framework. Each framework was concreted by using Apache Kafka and rule matrix-based recommendation techniques. We present the theoretical basis of implementing the high-performance message-processing function using Kafka. The SaaS mashup service platform, which provides a new type of mashup service by linking SaaS functions based on the above technology described, allows users to combine the required service functions freely and access the results of a rich service-utilization experience, using the SaaS mashup function. The platform developed through SaaS mashup service technology research will enable various flexible SaaS services, expected to contribute to the development of the smart-contents industry and the open market.

Designing Digital Twin Concept Model for High-Speed Synchronization (고속 동기화를 위한 디지털트윈 개념 모델 설계)

  • Chae-Young Lim;Chae-Eun Yeo;Ho-jin Sung
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.6
    • /
    • pp.245-250
    • /
    • 2023
  • Digital twin technology, which copies information from real space into virtual space, is being used in a variety of fields.Interest in digital twins is increasing, especially in advanced manufacturing fields such as Industry 4.0-based smart manufacturing. Operating a digital twin system generates a large amount of data, and the data generated has different characteristics depending on the technology field, so it is necessary to efficiently manage resources and use an optimized digital twin platform technology. Research on digital twin pipelines has continued, mainly in the advanced manufacturing field, but research on high-speed pipelines suitable for data in the plant field is still lacking. Therefore, in this paper, we propose a pipeline design method that is specialized for digital twin data in the plant field that is rapidly poured through Apache Kafka. The proposed model applies plant information on a Revit basis. and collect plant-specific data through Apache Kafka. Equipped with a lightweight CFD engine, it is possible to create a digital twin model that is more suitable for the plant field than existing digital twin technology for the manufacturing field.

Distributed Moving Objects Management System for a Smart Black Box

  • Lee, Hyunbyung;Song, Seokil
    • International Journal of Contents
    • /
    • v.14 no.1
    • /
    • pp.28-33
    • /
    • 2018
  • In this paper, we design and implement a distributed, moving objects management system for processing locations and sensor data from smart black boxes. The proposed system is designed and implemented based on Apache Kafka, Apache Spark & Spark Streaming, Hbase, HDFS. Apache Kafka is used to collect the data from smart black boxes and queries from users. Received location data from smart black boxes and queries from users becomes input of Apache Spark Streaming. Apache Spark Streaming preprocesses the input data for indexing. Recent location data and indexes are stored in-memory managed by Apache Spark. Old data and indexes are flushed into HBase later. We perform experiments to show the throughput of the index manager. Finally, we describe the implementation detail in Scala function level.

Scalable Big Data Pipeline for Video Stream Analytics Over Commodity Hardware

  • Ayub, Umer;Ahsan, Syed M.;Qureshi, Shavez M.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.4
    • /
    • pp.1146-1165
    • /
    • 2022
  • A huge amount of data in the form of videos and images is being produced owning to advancements in sensor technology. Use of low performance commodity hardware coupled with resource heavy image processing and analyzing approaches to infer and extract actionable insights from this data poses a bottleneck for timely decision making. Current approach of GPU assisted and cloud-based architecture video analysis techniques give significant performance gain, but its usage is constrained by financial considerations and extremely complex architecture level details. In this paper we propose a data pipeline system that uses open-source tools such as Apache Spark, Kafka and OpenCV running over commodity hardware for video stream processing and image processing in a distributed environment. Experimental results show that our proposed approach eliminates the need of GPU based hardware and cloud computing infrastructure to achieve efficient video steam processing for face detection with increased throughput, scalability and better performance.

Log processing using messaging system in SSD Storage Tester (SSD Storage Tester에서 메시징 시스템을 이용한 로그 처리)

  • Nam, Ki-ahn;Kwon, Oh-young
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.8
    • /
    • pp.1531-1539
    • /
    • 2017
  • The existing SSD storage tester processed logs in a 1-N structure between server and client using TCP and network file system. This method causes some problems for example, an increase in CPU usage and difficulty in exception handling, etc. In this paper, we implement a log processing message layer that can deal with asynchronous distributed processing using open source messaging system such as kafka, RabbitMQ and compare this layer with existing log transmission method. A log simulator was implemented to compare the transmission bandwidth and CPU usage. Test results show that the transmission using the message layer has higher performance than the transmission using the message layer, and the CPU usage does not show any significant difference The message layer can be implemented more easily than the conventional method and the efficiency is higher than that of the conventional method.

SDN-Based Collection-path Steering for IoT-Cloud Service Monitoring Data over SmartX-mini Playground (SmartX-mini Playground 상의 IoT-Cloud 서비스에 대한 SDN 기반 모니터링 데이터 수집 경로 설정)

  • Yoon, Heebum;Kim, Seungryong;Kim, JongWon
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.11
    • /
    • pp.1598-1607
    • /
    • 2016
  • Safe transmitting monitoring data is essential for supporting IoT-Cloud services efficiently. In this paper, we find ways to configure data path flexibly in SDN based for IoT-Cloud services utilizing SmartX-mini Playground. To do this, we use ONOS(Open Network Operating System) SDN Controller, ONOS NBI Applications made from us to check flexible and safe data path configuration for IoT-Cloud monitoring data transmitting in real IoT-SDN-Cloud environments.