• Title/Summary/Keyword: computation-intensive

Search Result 107, Processing Time 0.029 seconds

A Design of Analyzing effects of Distance between a mobile device and Cloudlet (모바일 장치와 구름을 사이에 거리의 효과 분석설계)

  • Eric, Niyonsaba;Jang, Jong-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.19 no.11
    • /
    • pp.2671-2676
    • /
    • 2015
  • Nowadays, Mobile devices are now capable of supporting a wide range of applications. Unfortunately, some of applications demand an ever increasing computational power and mobile devices have limited resources due to their constraints, such as low processing power, limited memory, unpredictable connectivity, and limited battery life. To deal with mobile devices' constraints, researchers envision extending cloud computing services to mobile devices using virtualization techniques to shift the workload from mobile devices to a powerful computational infrastructure. Those techniques consist of migrating resource-intensive computations from a mobile device to the resource-rich cloud, or server (called nearby infrastructure). In this paper, we want to highlight on cloudlet architecture (nearby infrastructure with mobile devices), its functioning and in our future work, analyze effects of distance between cloudlet and mobile devices.

Parallel Computing on Intensity Offset Tracking Using Synthetic Aperture Radar for Retrieval of Glacier Velocity

  • Hong, Sang-Hoon
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.1
    • /
    • pp.29-37
    • /
    • 2019
  • Synthetic Aperture Radar (SAR) observations are powerful tools to monitor surface's displacement very accurately, induced by earthquake, volcano, ground subsidence, glacier movement, etc. Especially, radar interferometry (InSAR) which utilizes phase information related to distance from sensor to target, can generate displacement map in line-of-sight direction with accuracy of a few cm or mm. Due to decorrelation effect, however, degradation of coherence in the InSAR application often prohibit from construction of differential interferogram. Offset tracking method is an alternative approach to make a two-dimensional displacement map using intensity information instead of the phase. However, there is limitation in that the offset tracking requires very intensive computation power and time. In this paper, efficiency of parallel computing has been investigated using high performance computer for estimation of glacier velocity. Two TanDEM-X SAR observations which were acquired on September 15, 2013 and September 26, 2013 over the Narsap Sermia in Southwestern Greenland were collected. Atotal of 56 of 2.4 GHz Intel Xeon processors(28 physical processors with hyperthreading) by operating with linux environment were utilized. The Gamma software was used for application of offset tracking by adjustment of the number of processors for the OpenMP parallel computing. The processing times of the offset tracking at the 256 by 256 pixels of window patch size at single and 56 cores are; 26,344 sec and 2,055 sec, respectively. It is impressive that the processing time could be reduced significantly about thirteen times (12.81) at the 56 cores usage. However, the parallel computing using all the processors prevent other background operations or functions. Except the offset tracking processing, optimum number of processors need to be evaluated for computing efficiency.

Aerodynamic Model Development for Three-dimensional Scramjet Model Based on Two-dimensional CFD Analysis (스크램제트 2차원 모델의 전산해석을 이용한 3차원 비행체의 공력 모델 개발)

  • Han, Song Ee;Shin, Ho Cheol;Park, Soo Hyung
    • Journal of the Korean Society of Propulsion Engineers
    • /
    • v.24 no.5
    • /
    • pp.65-76
    • /
    • 2020
  • On the initial design process of a scramjet vehicle such as the trajectory prediction, it is inevitable to estimate the aerodynamic performance of a three-dimensional effect. Despite the necessity of intensive computing for the three-dimensional model, it is inefficient in predicting a wide range of aerodynamic performance. In this study, an engineering model for aerodynamic performance was developed based on two-dimensional computational fluid analysis and linearized supersonic inviscid flow theory. Correspondingly, the three-dimension aerodynamic performance relations are presented based on the two-dimensional results. And the additional three-dimensional computation was performed to evaluate the adequacy for the extended relations.

Optimization Study of Toom-Cook Algorithm in NIST PQC SABER Utilizing ARM/NEON Processor (ARM/NEON 프로세서를 활용한 NIST PQC SABER에서 Toom-Cook 알고리즘 최적화 구현 연구)

  • Song, JinGyo;Kim, YoungBeom;Seo, Seog Chung
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.31 no.3
    • /
    • pp.463-471
    • /
    • 2021
  • Since 2016, National Institute of Standards and Technology (NIST) has been conducting a post quantum cryptography standardization project in preparation for a quantum computing environment. Three rounds are currently in progress, and most of the candidates (5/7) are lattice-based. Lattice-based post quantum cryptography is evaluated to be applicable even in an embedded environment where resources are limited by providing efficient operation processing and appropriate key length. Among them, SABER KEM provides the efficient modulus and Toom-Cook to process polynomial multiplication with computation-intensive tasks. In this paper, we present the optimized implementation of evaluation and interpolation in Toom-Cook algorithm of SABER utilizing ARM/NEON in ARMv8-A platform. In the evaluation process, we propose an efficient interleaving method of ARM/NEON, and in the interpolation process, we introduce an optimized implementation methodology applicable in various embedded environments. As a result, the proposed implementation achieved 3.5 times faster performance in the evaluation process and 5 times faster in the interpolation process than the previous reference implementation.

Performance Comparison of Task Partitioning Methods in MEC System (MEC 시스템에서 태스크 파티셔닝 기법의 성능 비교)

  • Moon, Sungwon;Lim, Yujin
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.11 no.5
    • /
    • pp.139-146
    • /
    • 2022
  • With the recent development of the Internet of Things (IoT) and the convergence of vehicles and IT technologies, high-performance applications such as autonomous driving are emerging, and multi-access edge computing (MEC) has attracted lots of attentions as next-generation technologies. In order to provide service to these computation-intensive tasks in low latency, many methods have been proposed to partition tasks so that they can be performed through cooperation of multiple MEC servers(MECSs). Conventional methods related to task partitioning have proposed methods for partitioning tasks on vehicles as mobile devices and offloading them to multiple MECSs, and methods for offloading them from vehicles to MECSs and then partitioning and migrating them to other MECSs. In this paper, the performance of task partitioning methods using offloading and migration is compared and analyzed in terms of service delay, blocking rate and energy consumption according to the method of selecting partitioning targets and the number of partitioning. As the number of partitioning increases, the performance of the service delay improves, but the performance of the blocking rate and energy consumption decreases.

Determining chlorine injection intensity in water distribution networks: a comparison of backtracking and water age approaches

  • Flavia D. Frederick;Malvin S. Marlim;Doosun Kang
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2023.05a
    • /
    • pp.170-170
    • /
    • 2023
  • Providing safe and readily available water is vital to maintain public health. One of the most prevalent methods to prevent the spread of waterborne diseases is applying chlorine injection to the treated water before distribution. During the water transmission and distribution, the chlorine will experience a reduction, which can imply potential risks for human health if it falls below the minimum threshold. The ability to determine the appropriate initial intensity of chlorine at the source would be significant to prevent such problems. This study proposes two methods that integrate hydraulic and water quality modeling to determine the suitable intensity of chlorine to be injected into the source water to maintain the minimum chlorine concentration (e.g., 0.2 mg/l) at each demand node. The water quality modeling employs the first-order decay to estimate the rate of chlorine reduction in the water. The first method utilizes a backtracking algorithm to trace the path of water from the demand node to the source during each time step, which helps to accurately determine the travel time through each pipe and node and facilitate the computation of time-dependent chlorine decay in the water delivery process. However, as a backtracking algorithm is computationally intensive, this study also explores an alternative approach using a water age. This approach estimates the elapsed time of water delivery from the source to the demand node and calculate the time-dependent reduction of chlorine in the water. Finally, this study compares the outcomes of two approaches and determines the suitable and effective method for calculating the chlorine intensity at the source to maintain the minimum chlorine level at demand nodes.

  • PDF

Changes in Research Paradigms in Data Intensive Environments

  • Minsoo Park
    • International journal of advanced smart convergence
    • /
    • v.12 no.4
    • /
    • pp.98-103
    • /
    • 2023
  • As technology advanced dramatically in the late 20th century, a new era of science arrived. The emerging era of scientific discovery, variously described as e-Science, cyberscience, and the fourth paradigm, uses technologies required for computation, data curation, analysis, and visualization. The emergence of the fourth research paradigm will have such a huge impact that it will shake the foundations of science, and will also have a huge impact on the role of data-information infrastructure. In the digital age, the roles of data-information professionals are becoming more diverse. As eScience emerges as a sustainable and growing part of research, data-information professionals and centeres are exploring new roles to address the issues that arise from new forms of research. The functions that data-information professionals and centeres can fundamentally provide in the e-Science area are data curation, preservation, access, and metadata. Basically, it involves discovering and using available technical infrastructure and tools, finding relevant data, establishing a data management plan, and developing tools to support research. A further advanced service is archiving and curating relevant data for long-term preservation and integration of datasets and providing curating and data management services as part of a data management plan. Adaptation and change to the new information environment of the 21st century require strong and future-responsive leadership. There is a strong need to effectively respond to future challenges by exploring the role and function of data-information professionals in the future environment. Understanding what types of data-information professionals and skills will be needed in the future is essential to developing the talent that will lead the transformation. The new values and roles of data-information professionals and centers for 21st century researchers in STEAM are discussed.

Design of Omok AI using Genetic Algorithm and Game Trees and Their Parallel Processing on the GPU (유전 알고리즘과 게임 트리를 병합한 오목 인공지능 설계 및 GPU 기반 병렬 처리 기법)

  • Ahn, Il-Jun;Park, In-Kyu
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.37 no.2
    • /
    • pp.66-75
    • /
    • 2010
  • This paper proposes an efficient method for design and implementation of the artificial intelligence (AI) of 'omok' game on the GPU. The proposed AI is designed on a cooperative structure using min-max game tree and genetic algorithm. Since the evaluation function needs intensive computation but is independently performed on a lot of candidates in the solution space, it is computed on the GPU in a massive parallel way. The implementation on NVIDIA CUDA and the experimental results show that it outperforms significantly over the CPU, in which parallel game tree and genetic algorithm on the GPU runs more than 400 times and 300 times faster than on the CPU. In the proposed cooperative AI, selective search using genetic algorithm is performed subsequently after the full search using game tree to search the solution space more efficiently as well as to avoid the thread overflow. Experimental results show that the proposed algorithm enhances the AI significantly and makes it run within the time limit given by the game's rule.

A Design of Integrated Scientific Workflow Execution Environment for A Computational Scientific Application (계산 과학 응용을 위한 과학 워크플로우 통합 수행 환경 설계)

  • Kim, Seo-Young;Yoon, Kyoung-A;Kim, Yoon-Hee
    • Journal of Internet Computing and Services
    • /
    • v.13 no.1
    • /
    • pp.37-44
    • /
    • 2012
  • Numerous scientists who are engaged in compute-intensive researches require more computing facilities than before, while the computing resource and techniques are increasingly becoming more advanced. For this reason, many works for e-Science environment have been actively invested and established around the world, but still the scientists look for an intuitive experimental environment, which is guaranteed the improved environmental facilities without additional configurations or installations. In this paper, we present an integrated scientific workflow execution environment for Scientific applications supporting workflow design with high performance computing infrastructure and accessibility for web browser. This portal supports automated consecutive execution of computation jobs in order of the form defined by workflow design tool and execution service concerning characteristics of each job to batch over distributed grid resources. Workflow editor of the portal presents a high-level frontend and easy-to-use interface with monitoring service, which shows the status of workflow execution in real time so that user can check the intermediate data during experiments. Therefore, the scientists can take advantages of the environment to improve the productivity of study based on HTC.

On Shape Recovery of 3D Object from Multiple Range Images (시점이 다른 다수의 거리 영상으로부터 3차원 물체의 형상 복원)

  • Kim, Jun-Young;Yun, Il-Dong;Lee, Sang-Uk
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.37 no.1
    • /
    • pp.1-15
    • /
    • 2000
  • To reconstruct 3- D shape, It is a common strategy to acquire multiple range Images from different viewpoints and integrate them into a common coordinates In this paper, we particularly focus on the registration and integration processes for combining all range Images into one surface model. For the registration, we propose the 2-step registration algorithm, which consists of 2 steps the rough registration step using all data points and the fine registration step using the high-curved data points For the integration, we propose a new algorithm, referred to as ‘multi-registration’ technique, to alleviate the error accumulation problem, which occurs during applying the pair-wise registration to each range image sequentially, in order to transform them into a common reference frame Intensive experiments are performed on the various real range data In experiments, all range images were registered within 1 minutes on Pentium 150MHz PC The results show that the proposed algorithms registrate and integrate multiple range Images within a tolerable error bound in a reasonable computation time, and the total error between all range Images are equalized with our proposed algorithms.

  • PDF