• Title/Summary/Keyword: Parallel computing model

Search Result 171, Processing Time 0.024 seconds

Performance Enhancement of Parallel Prime Sieving with Hybrid Programming and Pipeline Scheduling (혼합형 병렬처리 및 파이프라이닝을 활용한 소수 연산 알고리즘)

  • Ryu, Seung-yo;Kim, Dongseung
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.4 no.10
    • /
    • pp.337-342
    • /
    • 2015
  • We develop a new parallelization method for Sieve of Eratosthenes algorithm, which enhances both computation speed and energy efficiency. A pipeline scheduling is included for better load balancing after proper workload partitioning. They run on multicore CPUs with hybrid parallel programming model which uses both message passing and multithreading computation. Experimental results performed on both small scale clusters and a PC with a mobile processor show significant improvement in execution time and energy consumptions.

Knowledge Based Recommender System for Disease Diagnostic and Treatment Using Adaptive Fuzzy-Blocks

  • Navin K.;Mukesh Krishnan M. B.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.2
    • /
    • pp.284-310
    • /
    • 2024
  • Identifying clinical pathways for disease diagnosis and treatment process recommendations are seriously decision-intensive tasks for health care practitioners. It requires them to rely on their expertise and experience to analyze various categories of health parameters from a health record to arrive at a decision in order to provide an accurate diagnosis and treatment recommendations to the end user (patient). Technological adaptation in the area of medical diagnosis using AI is dispensable; using expert systems to assist health care practitioners in decision-making is becoming increasingly popular. Our work architects a novel knowledge-based recommender system model, an expert system that can bring adaptability and transparency in usage, provide in-depth analysis of a patient's medical record, and prescribe diagnostic results and treatment process recommendations to them. The proposed system uses a set of parallel discrete fuzzy rule-based classifier systems, with each of them providing recommended sub-outcomes of discrete medical conditions. A novel knowledge-based combiner unit extracts significant relationships between the sub-outcomes of discrete fuzzy rule-based classifier systems to provide holistic outcomes and solutions for clinical decision support. The work establishes a model to address disease diagnosis and treatment recommendations for primary lung disease issues. In this paper, we provide some samples to demonstrate the usage of the system, and the results from the system show excellent correlation with expert assessments.

Auto Regulated Data Provisioning Scheme with Adaptive Buffer Resilience Control on Federated Clouds

  • Kim, Byungsang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.11
    • /
    • pp.5271-5289
    • /
    • 2016
  • On large-scale data analysis platforms deployed on cloud infrastructures over the Internet, the instability of the data transfer time and the dynamics of the processing rate require a more sophisticated data distribution scheme which maximizes parallel efficiency by achieving the balanced load among participated computing elements and by eliminating the idle time of each computing element. In particular, under the constraints that have the real-time and limited data buffer (in-memory storage) are given, it needs more controllable mechanism to prevent both the overflow and the underflow of the finite buffer. In this paper, we propose an auto regulated data provisioning model based on receiver-driven data pull model. On this model, we provide a synchronized data replenishment mechanism that implicitly avoids the data buffer overflow as well as explicitly regulates the data buffer underflow by adequately adjusting the buffer resilience. To estimate the optimal size of buffer resilience, we exploits an adaptive buffer resilience control scheme that minimizes both data buffer space and idle time of the processing elements based on directly measured sample path analysis. The simulation results show that the proposed scheme provides allowable approximation compared to the numerical results. Also, it is suitably efficient to apply for such a dynamic environment that cannot postulate the stochastic characteristic for the data transfer time, the data processing rate, or even an environment where the fluctuation of the both is presented.

Development of Web Service-based Parallel and Distributed Simulation (웹서비스 기반의 분산 시뮬레이션 프로토타입 개발)

  • Jo, In-Ho;Ju, Jeong-Min;Park, Yang-Seon;Jo, Hyeon-Bo
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2005.05a
    • /
    • pp.1033-1039
    • /
    • 2005
  • Parallel and distributed simulation is concerned with the efficient execution of large-scale discrete event simulation models on multiprocessors and distributed platforms. After the development of WWW, many efforts in the parallel and distributed simulation have been made for modeling, particularly building simulation languages and creating model libraries that can be assembled and executed over WWW. However, web-based parallel and distributed simulation is restricted by heterogeneous computing environments. Recently, the advent of XML and web services technology has made these efforts enter upon a new phase. Especially, the web services as a distributed information technology have demonstrated powerful capabilities for scalable interoperation of heterogeneous systems. This paper aims to develop and evaluate the parallel and distributed simulation using the web services technology. In particular, a prototype multi-pass simulation framework is implemented using Java-based web services technology. It focuses on the efficiency of multi-pass simulation used for optimization through the distribution of simulation replication to several simulation service providers. The development of parallel and distributed simulation using web services will help solve efficiently large-scale problems and also guarantee interoperability among heterogeneous networked systems.

  • PDF

A Study on Distributed Processing of Big Data and User Authentication for Human-friendly Robot Service on Smartphone (인간 친화적 로봇 서비스를 위한 대용량 분산 처리 기술 및 사용자 인증에 관한 연구)

  • Choi, Okkyung;Jung, Wooyeol;Lee, Bong Gyou;Moon, Seungbin
    • Journal of Internet Computing and Services
    • /
    • v.15 no.1
    • /
    • pp.55-61
    • /
    • 2014
  • Various human-friendly robot services have been developed and mobile cloud computing is a real time computing service that allows users to rent IT resources what they want over the internet and has become the new-generation computing paradigm of information society. The enterprises and nations are actively underway of the business process using mobile cloud computing and they are aware of need for implementing mobile cloud computing to their business practice, but it has some week points such as authentication services and distributed processing technologies of big data. Sometimes it is difficult to clarify the objective of cloud computing service. In this study, the vulnerability of authentication services on mobile cloud computing is analyzed and mobile cloud computing model is constructed for efficient and safe business process. We will also be able to study how to process and analyze unstructured data in parallel to this model, so that in the future, providing customized information for individuals may be possible using unstructured data.

Assessment of computational performance for a vector parallel implementation: 3D probabilistic model discrete cracking in concrete

  • Paz, Carmen N.M.;Alves, Jose L.D.;Ebecken, Nelson F.F.
    • Computers and Concrete
    • /
    • v.2 no.5
    • /
    • pp.345-366
    • /
    • 2005
  • This work presents an assessment of the computational performance of a vector-parallel implementation of probabilistic model for concrete cracking in 3D. This paper shows the continuing efforts towards code optimization as reported in earlier works Paz, et al. (2002a,b and 2003). The probabilistic crack approach is based on the direct Monte Carlo method. Cracking is accounted by means of 3D interface elements. This approach considers that all nonlinearities are restricted to interface elements modeling cracks. The heterogeneity governs the overall cracking behavior and related size effects on concrete fracture. Computational kernels in the implementation are the inexact Newton iterative driver to solve the non-linear problem and a preconditioned conjugate gradient (PCG) driver to solve linearized equations, using an element by element (EBE) strategy to compute matrix-vector products. In particular the paper analyzes code behavior using OpenMP directives in parallel vector processors (PVP), such as the CRAY SV1 and CRAY T94. The impact of the memory architecture on code performance, and also some strategies devised to circumvent this issue are addressed by numerical experiment.

An Internet-based computing framework for the simulation of multi-scale response of structural systems

  • Chen, Hung-Ming;Lin, Yu-Chih
    • Structural Engineering and Mechanics
    • /
    • v.37 no.1
    • /
    • pp.17-37
    • /
    • 2011
  • This paper presents a new Internet-based computational framework for the realistic simulation of multi-scale response of structural systems. Two levels of parallel processing are involved in this frame work: multiple local distributed computing environments connected by the Internet to form a cluster-to-cluster distributed computing environment. To utilize such a computing environment for a realistic simulation, the simulation task of a structural system has been separated into a simulation of a simplified global model in association with several detailed component models using various scales. These related multi-scale simulation tasks are distributed amongst clusters and connected to form a multi-level hierarchy. The Internet is used to coordinate geographically distributed simulation tasks. This paper also presents the development of a software framework that can support the multi-level hierarchical simulation approach, in a cluster-to-cluster distributed computing environment. The architectural design of the program also allows the integration of several multi-scale models to be clients and servers under a single platform. Such integration can combine geographically distributed computing resources to produce realistic simulations of structural systems.

Design of an OMNeT++ based Parallel Simulator for a Bio-Inspired System and Its Performance on PC-Clusters (생태계 모방 시스템을 위한 OMNeT++ 기반 병렬 시뮬레이터의 설계 및 PC 클러스터 상에서의 성능 분석)

  • Moon, Joo-Sun;Nang, Jong-Ho
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.34 no.9
    • /
    • pp.416-424
    • /
    • 2007
  • The Bio-Inspired system is a computing model that emulates the objects in ecosystem which are evolving themselves and cooperate each other to perform some tasks. Since it could be used to solved the complex problems that have been very difficult to resolve with previous algorithms, there have been a lot of researches to develop an application based on the Bio-Inspired system. However, since this computing model requires the process of evolving and cooperating with a lot of objects and this process takes a lot of times, it has been very hard to develop an application based on this computing model. This paper presents a parallel simulator for a Bio-Inspired system that is designed and implemented with OMNeT++ on PC clusters, and proves its usefulness by showing its simulation performance for a couple of applications. In the proposed parallel simulator, the functions required in the ERS platform for evolving and cooperating between objects (called Ecogent) are mapped onto the functions of OMNeT++, and they are simulated on PC clusters simultaneously to reduce the total simulation time. The simulation results could be monitored with a GUI In realtime, and they are also recorded into DBMS for systematic analyses afterward. This paper shows the usefulness of the proposed system by analyzing its performances for simulating various applications based on Bio-Inspired system on PC clusters with 4 PCs.

Algorithmic GPGPU Memory Optimization

  • Jang, Byunghyun;Choi, Minsu;Kim, Kyung Ki
    • JSTS:Journal of Semiconductor Technology and Science
    • /
    • v.14 no.4
    • /
    • pp.391-406
    • /
    • 2014
  • The performance of General-Purpose computation on Graphics Processing Units (GPGPU) is heavily dependent on the memory access behavior. This sensitivity is due to a combination of the underlying Massively Parallel Processing (MPP) execution model present on GPUs and the lack of architectural support to handle irregular memory access patterns. Application performance can be significantly improved by applying memory-access-pattern-aware optimizations that can exploit knowledge of the characteristics of each access pattern. In this paper, we present an algorithmic methodology to semi-automatically find the best mapping of memory accesses present in serial loop nest to underlying data-parallel architectures based on a comprehensive static memory access pattern analysis. To that end we present a simple, yet powerful, mathematical model that captures all memory access pattern information present in serial data-parallel loop nests. We then show how this model is used in practice to select the most appropriate memory space for data and to search for an appropriate thread mapping and work group size from a large design space. To evaluate the effectiveness of our methodology, we report on execution speedup using selected benchmark kernels that cover a wide range of memory access patterns commonly found in GPGPU workloads. Our experimental results are reported using the industry standard heterogeneous programming language, OpenCL, targeting the NVIDIA GT200 architecture.

Preliminary Study on the Enhancement of Reconstruction Speed for Emission Computed Tomography Using Parallel Processing (병렬 연산을 이용한 방출 단층 영상의 재구성 속도향상 기초연구)

  • Park, Min-Jae;Lee, Jae-Sung;Kim, Soo-Mee;Kang, Ji-Yeon;Lee, Dong-Soo;Park, Kwang-Suk
    • Nuclear Medicine and Molecular Imaging
    • /
    • v.43 no.5
    • /
    • pp.443-450
    • /
    • 2009
  • Purpose: Conventional image reconstruction uses simplified physical models of projection. However, real physics, for example 3D reconstruction, takes too long time to process all the data in clinic and is unable in a common reconstruction machine because of the large memory for complex physical models. We suggest the realistic distributed memory model of fast-reconstruction using parallel processing on personal computers to enable large-scale technologies. Materials and Methods: The preliminary tests for the possibility on virtual manchines and various performance test on commercial super computer, Tachyon were performed. Expectation maximization algorithm with common 2D projection and realistic 3D line of response were tested. Since the process time was getting slower (max 6 times) after a certain iteration, optimization for compiler was performed to maximize the efficiency of parallelization. Results: Parallel processing of a program on multiple computers was available on Linux with MPICH and NFS. We verified that differences between parallel processed image and single processed image at the same iterations were under the significant digits of floating point number, about 6 bit. Double processors showed good efficiency (1.96 times) of parallel computing. Delay phenomenon was solved by vectorization method using SSE. Conclusion: Through the study, realistic parallel computing system in clinic was established to be able to reconstruct by plenty of memory using the realistic physical models which was impossible to simplify.