• Title/Summary/Keyword: Scalability Problem

Search Result 273, Processing Time 0.035 seconds

An Application of MapReduce Technique over Peer-to-Peer Network (P2P 네트워크상에서 MapReduce 기법 활용)

  • Ren, Jian-Ji;Lee, Jae-Kee
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.8
    • /
    • pp.586-590
    • /
    • 2009
  • The objective of this paper describes the design of MapReduce over Peer-to-Peer network for dynamic environments applications. MapReduce is a software framework used for Cloud Computing which processing large data sets in a highly-parallel way. Based on the Peer-to-Peer network character which node failures will happen anytime, we focus on using a DHT routing protocol which named Pastry to handle the problem of node failures. Our results are very promising and indicate that the framework could have a wide application in P2P network systems while maintaining good computational efficiency and scalability. We believe that, P2P networks and parallel computing emerge as very hot research and development topics in industry and academia for many years to come.

Hierarchical Location Caching Scheme for Mobile Object Tracking in the Internet of Things

  • Han, Youn-Hee;Lim, Hyun-Kyo;Gil, Joon-Min
    • Journal of Information Processing Systems
    • /
    • v.13 no.5
    • /
    • pp.1410-1429
    • /
    • 2017
  • Mobility arises naturally in the Internet of Things networks, since the location of mobile objects, e.g., mobile agents, mobile software, mobile things, or users with wireless hardware, changes as they move. Tracking their current location is essential to mobile computing. To overcome the scalability problem, hierarchical architectures of location databases have been proposed. When location updates and lookups for mobile objects are localized, these architectures become effective. However, the network signaling costs and the execution number of database operations increase particularly when the scale of the architectures and the numbers of databases becomes large to accommodate a great number of objects. This disadvantage can be alleviated by a location caching scheme which exploits the spatial and temporal locality in location lookup. In this paper, we propose a hierarchical location caching scheme, which acclimates the existing location caching scheme to a hierarchical architecture of location databases. The performance analysis indicates that the adjustment of such thresholds has an impact on cost reduction in the proposed scheme.

Improvement of RRT*-Smart Algorithm for Optimal Path Planning and Application of the Algorithm in 2 & 3-Dimension Environment (최적 경로 계획을 위한 RRT*-Smart 알고리즘의 개선과 2, 3차원 환경에서의 적용)

  • Tak, Hyeong-Tae;Park, Cheon-Geon;Lee, Sang-Chul
    • Journal of the Korean Society for Aviation and Aeronautics
    • /
    • v.27 no.2
    • /
    • pp.1-8
    • /
    • 2019
  • Optimal path planning refers to find the safe route to the destination at a low cost, is a major problem with regard to autonomous navigation. Sampling Based Planning(SBP) approaches, such as Rapidly-exploring Random Tree Star($RRT^*$), are the most influential algorithm in path planning due to their relatively small calculations and scalability to high-dimensional problems. $RRT^*$-Smart introduced path optimization and biased sampling techniques into $RRT^*$ to increase convergent rate. This paper presents an improvement plan that has changed the biased sampling method to increase the initial convergent rate of the $RRT^*$-Smart, which is specified as m$RRT^*$-Smart. With comparison among $RRT^*$, $RRT^*$-Smart and m$RRT^*$-Smart in 2 & 3-D environments, m$RRT^*$-Smart showed similar or increased initial convergent rate than $RRT^*$ and $RRT^*$-Smart.

Dynamic Cluster Management of Hadoop Distributed Filesystem (하둡 분산 파일시스템의 동적 클러스터 관리 기법)

  • Ryu, Wooseok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2016.10a
    • /
    • pp.435-437
    • /
    • 2016
  • Hadoop Distributed File System(HDFS) is a file system for distributed processing of big data by replicating data to distributed data nodes. HDFS cluster shows a great scalability up to thousands of nodes, but it assumes a exclusive node cluster with numerous nodes for the big data processing. Various operational-purpose worker systems used by office are hardly considered as a part of cluster. This paper discusses this problem and proposes a dynamic cluster management technique to increase storage capability and analytic performance of hadoop cluster. The propsed technique can add legacy systems to the cluster and can remove them from the cluster dynamically depending on their availability.

  • PDF

NoSQL-based Sensor Web System for Fine Particles Analysis Services (미세먼지 분석 서비스를 위한 NoSQL 기반 센서 웹 시스템)

  • Kim, Jeong-Joon;Kwak, Kwang-Jin;Park, Jeong-Min
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.2
    • /
    • pp.119-125
    • /
    • 2019
  • Recently, it has become a social problem due to fine particles. There are more people wearing masks, weather alerts and disaster notices. Research and policy are actively underway. Meteorologically, the biggest damage caused by fine particles is the inversion layer phenomenon. In this study, we designed a system to warn fine Particles by analyzing inversion layer and wind direction. This weather information system proposes a system that can efficiently perform scalability and parallel processing by using OGC sensor web enablement system and NoSQL storage for sensor control and data exchange.

An Energy Efficient Multi-hop Cluster-Head Election Strategy for Wireless Sensor Networks

  • Zhao, Liquan;Guo, Shuaichao
    • Journal of Information Processing Systems
    • /
    • v.17 no.1
    • /
    • pp.63-74
    • /
    • 2021
  • According to the double-phase cluster-head election method (DCE), the final cluster heads (CHs) sometimes are located at the edge of cluster. They have a long distance from the base station (BS). Sensor data is directly transmitted to BS by CHs. This makes some nodes consume much energy for transmitting data and die earlier. To address this problem, energy efficient multi-hop cluster-head election strategy (EEMCE) is proposed in this paper. To avoid taking these nodes far from BS as CH, this strategy first introduces the distance from the sensor nodes to the BS into the tentative CH election. Subsequently, in the same cluster, the energy of tentative CH is compared with those of other nodes, and then the node that has more energy than the tentative CH and being nearest the tentative CH are taken as the final CH. Lastly, if the CH is located at the periphery of the network, the multi-hop method will be employed to reduce the energy that is consumed by CHs. The simulation results suggest that the proposed method exhibits higher energy efficiency, longer stability period and better scalability than other protocols.

PESA: Prioritized experience replay for parallel hybrid evolutionary and swarm algorithms - Application to nuclear fuel

  • Radaideh, Majdi I.;Shirvan, Koroush
    • Nuclear Engineering and Technology
    • /
    • v.54 no.10
    • /
    • pp.3864-3877
    • /
    • 2022
  • We propose a new approach called PESA (Prioritized replay Evolutionary and Swarm Algorithms) combining prioritized replay of reinforcement learning with hybrid evolutionary algorithms. PESA hybridizes different evolutionary and swarm algorithms such as particle swarm optimization, evolution strategies, simulated annealing, and differential evolution, with a modular approach to account for other algorithms. PESA hybridizes three algorithms by storing their solutions in a shared replay memory, then applying prioritized replay to redistribute data between the integral algorithms in frequent form based on their fitness and priority values, which significantly enhances sample diversity and algorithm exploration. Additionally, greedy replay is used implicitly to improve PESA exploitation close to the end of evolution. PESA features in balancing exploration and exploitation during search and the parallel computing result in an agnostic excellent performance over a wide range of experiments and problems presented in this work. PESA also shows very good scalability with number of processors in solving an expensive problem of optimizing nuclear fuel in nuclear power plants. PESA's competitive performance and modularity over all experiments allow it to join the family of evolutionary algorithms as a new hybrid algorithm; unleashing the power of parallel computing for expensive optimization.

Enhancement of Semantic Interoper ability in Healthcare Systems Using IFCIoT Architecture

  • Sony P;Siva Shanmugam G;Sureshkumar Nagarajan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.881-902
    • /
    • 2024
  • Fast decision support systems and accurate diagnosis have become significant in the rapidly growing healthcare sector. As the number of disparate medical IoT devices connected to the human body rises, fast and interrelated healthcare data retrieval gets harder and harder. One of the most important requirements for the Healthcare Internet of Things (HIoT) is semantic interoperability. The state-of-the-art HIoT systems have problems with bandwidth and latency. An extension of cloud computing called fog computing not only solves the latency problem but also provides other benefits including resource mobility and on-demand scalability. The recommended approach helps to lower latency and network bandwidth consumption in a system that provides semantic interoperability in healthcare organizations. To evaluate the system's language processing performance, we simulated it in three different contexts. 1. Polysemy resolution system 2. System for hyponymy-hypernymy resolution with polysemy 3. System for resolving polysemy, hypernymy, hyponymy, meronymy, and holonymy. In comparison to the other two systems, the third system has lower latency and network usage. The proposed framework can reduce the computation overhead of heterogeneous healthcare data. The simulation results show that fog computing can reduce delay, network usage, and energy consumption.

Performance Analysis of Frequent Pattern Mining with Multiple Minimum Supports (다중 최소 임계치 기반 빈발 패턴 마이닝의 성능분석)

  • Ryang, Heungmo;Yun, Unil
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.1-8
    • /
    • 2013
  • Data mining techniques are used to find important and meaningful information from huge databases, and pattern mining is one of the significant data mining techniques. Pattern mining is a method of discovering useful patterns from the huge databases. Frequent pattern mining which is one of the pattern mining extracts patterns having higher frequencies than a minimum support threshold from databases, and the patterns are called frequent patterns. Traditional frequent pattern mining is based on a single minimum support threshold for the whole database to perform mining frequent patterns. This single support model implicitly supposes that all of the items in the database have the same nature. In real world applications, however, each item in databases can have relative characteristics, and thus an appropriate pattern mining technique which reflects the characteristics is required. In the framework of frequent pattern mining, where the natures of items are not considered, it needs to set the single minimum support threshold to a too low value for mining patterns containing rare items. It leads to too many patterns including meaningless items though. In contrast, we cannot mine any pattern if a too high threshold is used. This dilemma is called the rare item problem. To solve this problem, the initial researches proposed approximate approaches which split data into several groups according to item frequencies or group related rare items. However, these methods cannot find all of the frequent patterns including rare frequent patterns due to being based on approximate techniques. Hence, pattern mining model with multiple minimum supports is proposed in order to solve the rare item problem. In the model, each item has a corresponding minimum support threshold, called MIS (Minimum Item Support), and it is calculated based on item frequencies in databases. The multiple minimum supports model finds all of the rare frequent patterns without generating meaningless patterns and losing significant patterns by applying the MIS. Meanwhile, candidate patterns are extracted during a process of mining frequent patterns, and the only single minimum support is compared with frequencies of the candidate patterns in the single minimum support model. Therefore, the characteristics of items consist of the candidate patterns are not reflected. In addition, the rare item problem occurs in the model. In order to address this issue in the multiple minimum supports model, the minimum MIS value among all of the values of items in a candidate pattern is used as a minimum support threshold with respect to the candidate pattern for considering its characteristics. For efficiently mining frequent patterns including rare frequent patterns by adopting the above concept, tree based algorithms of the multiple minimum supports model sort items in a tree according to MIS descending order in contrast to those of the single minimum support model, where the items are ordered in frequency descending order. In this paper, we study the characteristics of the frequent pattern mining based on multiple minimum supports and conduct performance evaluation with a general frequent pattern mining algorithm in terms of runtime, memory usage, and scalability. Experimental results show that the multiple minimum supports based algorithm outperforms the single minimum support based one and demands more memory usage for MIS information. Moreover, the compared algorithms have a good scalability in the results.

Hierarchical Mesh Network Design & Implementation (계층적 메쉬 네트워크의 설계 및 구현)

  • Kim, Yong-Hyuck;Kim, Young-Han
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.46 no.6
    • /
    • pp.26-36
    • /
    • 2009
  • ad-hoc network is self-constructed by mobile hosts without any infrastructure. And the hosts collaborate with each other for routing packet by means of multi-hop communication. But it is too hard to deploy ad hoc network at the large area because of the salability problem caused by low network throughput. network throughput decrease is due to contention on the air among the neighbor host and single interface limitation, and broadcast flooding through overall network. In this paper, to solve the ad hoc scalability problem, we propose mesh network based scalable hierarchical ad hoc architecture, and also propose the adaptation methods for inter-working with the host not including ad hoc functions and the legacy infra-network.