• Title/Summary/Keyword: Distributed Clustering

Search Result 224, Processing Time 0.022 seconds

A Study of Basic Design Method for High Availability Clustering Framework under Distributed Computing Environment (분산컴퓨팅 환경에서의 고가용성 클러스터링 프레임워크 기본설계 연구)

  • Kim, Jeom Goo;Noh, SiChoon
    • Convergence Security Journal
    • /
    • v.13 no.3
    • /
    • pp.17-23
    • /
    • 2013
  • Clustering is required to configure clustering interdependent structural technology. Clustering handles variable workloads or impede continuity of service to continue operating in the event of a failure. Long as high-availability clustering feature focuses on server operating systems. Active-standby state of two systems when the active server fails, all services are running on the standby server, it takes the service. This function switching or switchover is called failover. Long as high-availability clustering feature focuses on server operating systems. The cluster node that is running on multiple systems and services have to duplicate each other so you can keep track of. In the event of a node failure within a few seconds the second node, the node shall perform the duties broken. Structure for high-availability clustering efficiency should be measured. System performance of infrastructure systems performance, latency, response time, CPU load factor(CPU utilization), CPU processes on the system (system process) channels are represented.

A MapReduce-Based Workflow BIG-Log Clustering Technique (맵리듀스기반 워크플로우 빅-로그 클러스터링 기법)

  • Jin, Min-Hyuck;Kim, Kwanghoon Pio
    • Journal of Internet Computing and Services
    • /
    • v.20 no.1
    • /
    • pp.87-96
    • /
    • 2019
  • In this paper, we propose a MapReduce-supported clustering technique for collecting and classifying distributed workflow enactment event logs as a preprocessing tool. Especially, we would call the distributed workflow enactment event logs as Workflow BIG-Logs, because they are satisfied with as well as well-fitted to the 5V properties of BIG-Data like Volume, Velocity, Variety, Veracity and Value. The clustering technique we develop in this paper is intentionally devised for the preprocessing phase of a specific workflow process mining and analysis algorithm based upon the workflow BIG-Logs. In other words, It uses the Map-Reduce framework as a Workflow BIG-Logs processing platform, it supports the IEEE XES standard data format, and it is eventually dedicated for the preprocessing phase of the ${\rho}$-Algorithm that is a typical workflow process mining algorithm based on the structured information control nets. More precisely, The Workflow BIG-Logs can be classified into two types: of activity-based clustering patterns and performer-based clustering patterns, and we try to implement an activity-based clustering pattern algorithm based upon the Map-Reduce framework. Finally, we try to verify the proposed clustering technique by carrying out an experimental study on the workflow enactment event log dataset released by the BPI Challenges.

A Study on Cluster Head Selection Based on Distance from Sensor to Base Station in Wireless Sensor Network (무선센서 네트워크에서 센서와 기지국과의 거리를 고려한 클러스터 헤드 선택기법)

  • Ko, Sung-Won;Cho, Jeong-Hwan
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.27 no.10
    • /
    • pp.50-58
    • /
    • 2013
  • In Wireless Sensor Network, clustering scheme is used to prolong the lifetime of WSN by efficient usage of energy of sensor. In the distributed clustering protocol just like LEACH, every sensor in a network plays a cluster head role once during each epoch. So the FND is prolonged. But, even though every sensor plays a head role, the energy consumed by each sensor is different because the energy consumed increases according to the distance to the Base Station by the way of multiple increase. In this paper, we propose a mechanism to select a head depending on the distance to Base Station, which extends the timing of FND occurrence by 68% compared to the LEACH and makes network stable.

A study on high availability of the linux clustering web server (리눅스 클러스터링 웹 서버의 고가용성에 대한 연구)

  • 박지현;이상문;홍태화;김학배
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2000.10a
    • /
    • pp.88-88
    • /
    • 2000
  • As more and more critical commercial applications move on the Internet, providing highly available servers becomes increasingly important. One of the advantages of a clustered system is that it has hardware and software redundancy. High availability can be provided by detecting node or daemon failure and reconfiguring the system appropriately so that the workload can be taken over bi the remaining nodes in the cluster. This paper presents how to provide the guaranteeing high availability of clustering web server. The load balancer becomes a single failure point of the whole system. In order to prevent the failure of the load balancer, we setup a backup server using heartbeat, fake, mon, and checkpointing fault-tolerance method. For high availability of file servers in the cluster, we setup coda file system. Coda is a advanced network fault-tolerance distributed file system.

  • PDF

A Heuristic Task Allocation Scheme Based on Clustering (클러스터링을 이용한 경험적 태스크 할당 기법)

  • Kim, Seok-Il;Jeon, Jung-Nam;Kim, Gwan-Yu
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.10
    • /
    • pp.2659-2669
    • /
    • 1999
  • This paper a heuristic, clustering based task allocation scheme applicable to non-directed task graph on a distributed system. This scheme firstly builds a task-machine graph, and then applies a clustering process where in a pair of tasks that are connected to the highest cost edge is merged into a big one or a task is allocated to a machine. During the process, the proposed scheme figure out a machine onto which the task allocation may cause deduction of large communication overhead that has incurred between the task and tasks that are already allocated to the machine while the computation costs is slightly increased in the machine. Simulation for the various task graphs shows that the scheduling using the proposed scheme result far better than ones by using the traditional schemes. A comparison with optimal task scheduling also promises that our scheme derives optimal results more occasionally than the traditional schemes do.

  • PDF

A Dual-layer Energy Efficient Distributed Clustering Algorithm for Wireless Sensor Networks (무선 센서 네트워크를 위한 에너지 효율적인 이중 레이어 분산 클러스터링 기법)

  • Yeo, Myung-Ho;Kim, Yu-Mi;Yoo, Jae-Soo
    • Journal of KIISE:Databases
    • /
    • v.35 no.1
    • /
    • pp.84-95
    • /
    • 2008
  • Wireless sensor networks have recently emerged as a platform for several applications. By deploying wireless sensor nodes and constructing a sensor network, we can remotely obtain information about the behavior, conditions, and positions of objects in a region. Since sensor nodes operate on batteries, energy-efficient mechanisms for gathering sensor data are indispensable to prolong the lifetime of a sensor network as long as possible. In this paper, we propose a novel clustering algorithm that distributes the energy consumption of a cluster head. First, we analyze the energy consumption if cluster heads and divide each cluster into a collection layer and a transmission layer according to their roles. Then, we elect a cluster head for each layer to distribute the energy consumption of single cluster head. In order to show the superiority of our clustering algorithm, we compare it with the existing clustering algorithm in terms of the lifetime of the sensor network. As a result, our experimental results show that the proposed clustering algorithm achieves about $10%{\sim}40%$ performance improvements over the existing clustering algorithms.

Cluster-Based Quantization and Estimation for Distributed Systems

  • Kim, Yoon Hak
    • Journal of information and communication convergence engineering
    • /
    • v.14 no.4
    • /
    • pp.215-221
    • /
    • 2016
  • We consider a design of a combined quantizer and estimator for distributed systems wherein each node quantizes its measurement without any communication among the nodes and transmits it to a fusion node for estimation. Noting that the quantization partitions minimizing the estimation error are not independently encoded at nodes, we focus on the parameter regions created by the partitions and propose a cluster-based quantization algorithm that iteratively finds a given number of clusters of parameter regions with each region being closer to the corresponding codeword than to the other codewords. We introduce a new metric to determine the distance between codewords and parameter regions. We also discuss that the fusion node can perform an efficient estimation by finding the intersection of the clusters sent from the nodes. We demonstrate through experiments that the proposed design achieves a significant performance gain with a low complexity as compared to the previous designs.

Management of Distributed Nodes for Big Data Analysis in Small-and-Medium Sized Hospital (중소병원에서의 빅데이터 분석을 위한 분산 노드 관리 방안)

  • Ryu, Wooseok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2016.05a
    • /
    • pp.376-377
    • /
    • 2016
  • Performance of Hadoop, which is a distributed data processing framework for big data analysis, is affected by several characteristics of each node in distributed cluster such as processing power and network bandwidth. This paper analyzes previous approaches for heterogeneous hadoop clusters, and presents several requirements for distributed node clustering in small-and-medium sized hospitals by considering computing environments of the hospitals.

  • PDF

Selective Encryption Algorithm for 3D Printing Model Based on Clustering and DCT Domain

  • Pham, Giao N.;Kwon, Ki-Ryong;Lee, Eung-Joo;Lee, Suk-Hwan
    • Journal of Computing Science and Engineering
    • /
    • v.11 no.4
    • /
    • pp.152-159
    • /
    • 2017
  • Three-dimensional (3D) printing is applied to many areas of life, but 3D printing models are stolen by pirates and distributed without any permission from the original providers. Moreover, some special models and anti-weapon models in 3D printing must be secured from the unauthorized user. Therefore, 3D printing models must be encrypted before being stored and transmitted to ensure access and to prevent illegal copying. This paper presents a selective encryption algorithm for 3D printing models based on clustering and the frequency domain of discrete cosine transform. All facets are extracted from 3D printing model, divided into groups by the clustering algorithm, and all vertices of facets in each group are transformed to the frequency domain of a discrete cosine transform. The proposed algorithm is based on encrypting the selected coefficients in the frequency domain of discrete cosine transform to generate the encrypted 3D printing model. Experimental results verified that the proposed algorithm is very effective for 3D printing models. The entire 3D printing model is altered after the encryption process. The decrypting error is approximated to be zero. The proposed algorithm provides a better method and more security than previous methods.

Distributed data deduplication technique using similarity based clustering and multi-layer bloom filter (SDS 환경의 유사도 기반 클러스터링 및 다중 계층 블룸필터를 활용한 분산 중복제거 기법)

  • Yoon, Dabin;Kim, Deok-Hwan
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.14 no.5
    • /
    • pp.60-70
    • /
    • 2018
  • A software defined storage (SDS) is being deployed in cloud environment to allow multiple users to virtualize physical servers, but a solution for optimizing space efficiency with limited physical resources is needed. In the conventional data deduplication system, it is difficult to deduplicate redundant data uploaded to distributed storages. In this paper, we propose a distributed deduplication method using similarity-based clustering and multi-layer bloom filter. Rabin hash is applied to determine the degree of similarity between virtual machine servers and cluster similar virtual machines. Therefore, it improves the performance compared to deduplication efficiency for individual storage nodes. In addition, a multi-layer bloom filter incorporated into the deduplication process to shorten processing time by reducing the number of the false positives. Experimental results show that the proposed method improves the deduplication ratio by 9% compared to deduplication method using IP address based clusters without any difference in processing time.