• 제목/요약/키워드: computation-intensive

검색결과 108건 처리시간 0.025초

A response surface modelling approach for multi-objective optimization of composite plates

  • Kalita, Kanak;Dey, Partha;Joshi, Milan;Haldar, Salil
    • Steel and Composite Structures
    • /
    • 제32권4호
    • /
    • pp.455-466
    • /
    • 2019
  • Despite the rapid advancement in computing resources, many real-life design and optimization problems in structural engineering involve huge computation costs. To counter such challenges, approximate models are often used as surrogates for the highly accurate but time intensive finite element models. In this paper, surrogates for first-order shear deformation based finite element models are built using a polynomial regression approach. Using statistical techniques like Box-Cox transformation and ANOVA, the effectiveness of the surrogates is enhanced. The accuracy of the surrogate models is evaluated using statistical metrics like $R^2$, $R^2{_{adj}}$, $R^2{_{pred}}$ and $Q^2{_{F3}}$. By combining these surrogates with nature-inspired multi-criteria decision-making algorithms, namely multi-objective genetic algorithm (MOGA) and multi-objective particle swarm optimization (MOPSO), the optimal combination of various design variables to simultaneously maximize fundamental frequency and frequency separation is predicted. It is seen that the proposed approach is simple, effective and good at inexpensively producing a host of optimal solutions.

Efficient Resource Slicing Scheme for Optimizing Federated Learning Communications in Software-Defined IoT Networks

  • 담프로힘;맛사;김석훈
    • 인터넷정보학회논문지
    • /
    • 제22권5호
    • /
    • pp.27-33
    • /
    • 2021
  • With the broad adoption of the Internet of Things (IoT) in a variety of scenarios and application services, management and orchestration entities require upgrading the traditional architecture and develop intelligent models with ultra-reliable methods. In a heterogeneous network environment, mission-critical IoT applications are significant to consider. With erroneous priorities and high failure rates, catastrophic losses in terms of human lives, great business assets, and privacy leakage will occur in emergent scenarios. In this paper, an efficient resource slicing scheme for optimizing federated learning in software-defined IoT (SDIoT) is proposed. The decentralized support vector regression (SVR) based controllers predict the IoT slices via packet inspection data during peak hour central congestion to achieve a time-sensitive condition. In off-peak hour intervals, a centralized deep neural networks (DNN) model is used within computation-intensive aspects on fine-grained slicing and remodified decentralized controller outputs. With known slice and prioritization, federated learning communications iteratively process through the adjusted resources by virtual network functions forwarding graph (VNFFG) descriptor set up in software-defined networking (SDN) and network functions virtualization (NFV) enabled architecture. To demonstrate the theoretical approach, Mininet emulator was conducted to evaluate between reference and proposed schemes by capturing the key Quality of Service (QoS) performance metrics.

Public Key Encryption with Equality Test for Heterogeneous Systems in Cloud Computing

  • Elhabob, Rashad;Zhao, Yanan;Sella, Iva;Xiong, Hu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제13권9호
    • /
    • pp.4742-4770
    • /
    • 2019
  • Cloud computing provides a broad range of services like operating systems, hardware, software and resources. Availability of these services encourages data owners to outsource their intensive computations and massive data to the cloud. However, considering the untrusted nature of cloud server, it is essential to encrypt the data before outsourcing it to the cloud. Unfortunately, this leads to a challenge when it comes to providing search functionality for encrypted data located in the cloud. To address this challenge, this paper presents a public key encryption with equality test for heterogeneous systems (PKE-ET-HS). The PKE-ET-HS scheme simulates certificateless public encryption with equality test (CLE-ET) with the identity-based encryption with equality test (IBE-ET). This scheme provides the authorized cloud server the right to actuate the equivalence of two messages having their encryptions performed under heterogeneous systems. Basing on the random oracle model, we construct the security of our proposed scheme under the bilinear Diffie-Hellman (BDH) assumption. Eventually, we evaluate the size of storage, computation complexities, and properties with other related works and illustrations indicate good performance from our scheme.

An Efficient Service Function Chains Orchestration Algorithm for Mobile Edge Computing

  • Wang, Xiulei;Xu, Bo;Jin, Fenglin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제15권12호
    • /
    • pp.4364-4384
    • /
    • 2021
  • The dynamic network state and the mobility of the terminals make the service function chain (SFC) orchestration mechanisms based on static and deterministic assumptions hard to be applied in SDN/NFV mobile edge computing networks. Designing dynamic and online SFC orchestration mechanism can greatly improve the execution efficiency of compute-intensive and resource-hungry applications in mobile edge computing networks. In order to increase the overall profit of service provider and reduce the resource cost, the system running time is divided into a sequence of time slots and a dynamic orchestration scheme based on an improved column generation algorithm is proposed in each slot. Firstly, the SFC dynamic orchestration problem is formulated as an integer linear programming (ILP) model based on layered graph. Then, in order to reduce the computation costs, a column generation model is used to simplify the ILP model. Finally, a two-stage heuristic algorithm based on greedy strategy is proposed. Four metrics are defined and the performance of the proposed algorithm is evaluated based on simulation. The results show that our proposal significantly provides more than 30% reduction of run time and about 12% improvement in service deployment success ratio compared to the Viterbi algorithm based mechanism.

Bitcoin Cryptocurrency: Its Cryptographic Weaknesses and Remedies

  • Anindya Kumar Biswas;Mou Dasgupta
    • Asia pacific journal of information systems
    • /
    • 제30권1호
    • /
    • pp.21-30
    • /
    • 2020
  • Bitcoin (BTC) is a type of cryptocurrency that supports transaction/payment of virtual money between BTC users without the presence of a central authority or any third party like bank. It uses some cryptographic techniques namely public- and private-keys, digital signature and cryptographic-hash functions, and they are used for making secure transactions and maintaining distributed public ledger called blockchain. In BTC system, each transaction signed by sender is broadcasted over the P2P (Peer-to-Peer) Bitcoin network and a set of such transactions collected over a period is hashed together with the previous block/other values to form a block known as candidate block, where the first block known as genesis-block was created independently. Before a candidate block to be the part of existing blockchain (chaining of blocks), a computation-intensive hard problem needs to be solved. A number of miners try to solve it and a winner earns some BTCs as inspiration. The miners have high computing and hardware resources, and they play key roles in BTC for blockchain formation. This paper mainly analyses the underlying cryptographic techniques, identifies some weaknesses and proposes their enhancements. For these, two modifications of BTC are suggested ― (i) All BTC users must use digital certificates for their authentication and (ii) Winning miner must give signature on the compressed data of a block for authentication of public blocks/blockchain.

A hybrid algorithm for the synthesis of computer-generated holograms

  • Nguyen The Anh;An Jun Won;Choe Jae Gwang;Kim Nam
    • 한국광학회:학술대회논문집
    • /
    • 한국광학회 2003년도 하계학술발표회
    • /
    • pp.60-61
    • /
    • 2003
  • A new approach to reduce the computation time of genetic algorithm (GA) for making binary phase holograms is described. Synthesized holograms having diffraction efficiency of 75.8% and uniformity of 5.8% are proven in computer simulation and experimentally demonstrated. Recently, computer-generated holograms (CGHs) having high diffraction efficiency and flexibility of design have been widely developed in many applications such as optical information processing, optical computing, optical interconnection, etc. Among proposed optimization methods, GA has become popular due to its capability of reaching nearly global. However, there exits a drawback to consider when we use the genetic algorithm. It is the large amount of computation time to construct desired holograms. One of the major reasons that the GA' s operation may be time intensive results from the expense of computing the cost function that must Fourier transform the parameters encoded on the hologram into the fitness value. In trying to remedy this drawback, Artificial Neural Network (ANN) has been put forward, allowing CGHs to be created easily and quickly (1), but the quality of reconstructed images is not high enough to use in applications of high preciseness. For that, we are in attempt to find a new approach of combiningthe good properties and performance of both the GA and ANN to make CGHs of high diffraction efficiency in a short time. The optimization of CGH using the genetic algorithm is merely a process of iteration, including selection, crossover, and mutation operators [2]. It is worth noting that the evaluation of the cost function with the aim of selecting better holograms plays an important role in the implementation of the GA. However, this evaluation process wastes much time for Fourier transforming the encoded parameters on the hologram into the value to be solved. Depending on the speed of computer, this process can even last up to ten minutes. It will be more effective if instead of merely generating random holograms in the initial process, a set of approximately desired holograms is employed. By doing so, the initial population will contain less trial holograms equivalent to the reduction of the computation time of GA's. Accordingly, a hybrid algorithm that utilizes a trained neural network to initiate the GA's procedure is proposed. Consequently, the initial population contains less random holograms and is compensated by approximately desired holograms. Figure 1 is the flowchart of the hybrid algorithm in comparison with the classical GA. The procedure of synthesizing a hologram on computer is divided into two steps. First the simulation of holograms based on ANN method [1] to acquire approximately desired holograms is carried. With a teaching data set of 9 characters obtained from the classical GA, the number of layer is 3, the number of hidden node is 100, learning rate is 0.3, and momentum is 0.5, the artificial neural network trained enables us to attain the approximately desired holograms, which are fairly good agreement with what we suggested in the theory. The second step, effect of several parameters on the operation of the hybrid algorithm is investigated. In principle, the operation of the hybrid algorithm and GA are the same except the modification of the initial step. Hence, the verified results in Ref [2] of the parameters such as the probability of crossover and mutation, the tournament size, and the crossover block size are remained unchanged, beside of the reduced population size. The reconstructed image of 76.4% diffraction efficiency and 5.4% uniformity is achieved when the population size is 30, the iteration number is 2000, the probability of crossover is 0.75, and the probability of mutation is 0.001. A comparison between the hybrid algorithm and GA in term of diffraction efficiency and computation time is also evaluated as shown in Fig. 2. With a 66.7% reduction in computation time and a 2% increase in diffraction efficiency compared to the GA method, the hybrid algorithm demonstrates its efficient performance. In the optical experiment, the phase holograms were displayed on a programmable phase modulator (model XGA). Figures 3 are pictures of diffracted patterns of the letter "0" from the holograms generated using the hybrid algorithm. Diffraction efficiency of 75.8% and uniformity of 5.8% are measured. We see that the simulation and experiment results are fairly good agreement with each other. In this paper, Genetic Algorithm and Neural Network have been successfully combined in designing CGHs. This method gives a significant reduction in computation time compared to the GA method while still allowing holograms of high diffraction efficiency and uniformity to be achieved. This work was supported by No.mOl-2001-000-00324-0 (2002)) from the Korea Science & Engineering Foundation.

  • PDF

삼중대각행렬 시스템 풀이의 빠른 GPU 구현 (Fast GPU Implementation for the Solution of Tridiagonal Matrix Systems)

  • 김영희;이성기
    • 한국정보과학회논문지:시스템및이론
    • /
    • 제32권11_12호
    • /
    • pp.692-704
    • /
    • 2005
  • 컴퓨터 하드웨어의 급속한 발전으로 그래픽 프로세서 유닛(Graphics Processor Units : GPUs)은 굉장한 메모리 대역폭과 산술 능역을 보유하게 되어 범용 계산에 많이 활용되고 있으며, 특히 계산 집약적인 물리 기반 시뮬레이션(physics based simulation)의 GPU 구현이 활발하게 연구되고 있다. 물리 기반 시뮬레이션의 기본이 되는 미분방정식 풀이 과정에서 삼중대각행렬(tridiagonal matrix) 시스템은 유한차분(finite-difference) 근사에 의해서 자주 나타나는 선형시스템으로 물리 기반 시뮬레이션 관점에서 삼중대각행렬 시스템의 빠른 풀이는 중요한 연구 분야이다. 본 논문에서는 GPU에서 삼중대각행렬 시스템 풀이를 빠르게 구현할 수 있는 방법을 제안한다. 벡터 프로세서(vector processor) 계산에서 삼중대각행렬 시스템 풀이 방법으로 널리 사용되는 cyclic reduction 또는 odd-even reduction 알고리즘을 GPU에서 구현하였다. 본 논문에서 제안한 방법을 삼중대각행렬 시스템 풀이 방법으로 잘 알려져 있는 Thomas 방법과 GPU를 이용한 선형시스템 풀이에서 좋은 성과를 보이고 있는 conjugate gradient 방법과 비교할 때 상당한 성능 향상을 얻을 수 있었다. 또한, 열전도(heat conduction) 방정식, 이류 확산(advection-diffusion) 방정식, 얕은 물(shallow water) 방정식에 의한 물리 기반 시뮬레이션의 GPU 구현에 본 논문에서 제안한 방법을 사용하여 1024x1024 격자의 계산 영역에서 초당 35프레임 이상의 놀라운 성능을 보여주었다.

CUDA를 활용한 스케일링 필터 및 트랜스코더의 성능향상 (Performance Enhancement of Scaling Filter and Transcoder using CUDA)

  • 한재근;고영섭;서성한;하순회
    • 한국정보과학회논문지:컴퓨팅의 실제 및 레터
    • /
    • 제16권4호
    • /
    • pp.507-511
    • /
    • 2010
  • 본 논문은 GPGPU가속을 이용한 스케일링 필터(scaling filter) 및 트랜스코딩(Transcoding)의 성능 향상 방법을 제안한다. 트랜스코딩 기술은 다양한 요구조건을 지닌 멀티미디어 기기에 적합하게 동영상을 가공하는 기술로, 오늘날 여러 분야에서 활용되는 중요한 기술이다. 그러나 트랜스코딩에는 대량의 연산이 필요하기 때문에 기존 트랜스코더(Transcoder) 사용자들은 오랜 처리시간을 감내 해야만 했는데, 이는 CPU만을 이용한 트랜스코딩이 충분히 효율적이지 못하기 때문이다. 본 연구에서는 고성능의 연산이 가능한 GPGPU기술을 활용하여, 트랜스코더의 스케일링 필터를 GPU 상에서 높은 병렬성을 가지고 동작하도록 개선함으로써 트랜스코더의 전체적인 성능을 향상시켰다. 개선된 트랜스코더는 다양한 크기의 동영상과 여러 종류의 스케일링 필터 옵션들에 대해 잘 동작함이 검증되었으며, 기본 옵션에서 36%, 최대 101%의 성능향상을 보였다.

NTGST 병렬화를 이용한 고해상도 BLU 검사의 고속화 (NTGST-Based Parallel Computer Vision Inspection for High Resolution BLU)

  • 김복만;서경석;최흥문
    • 대한전자공학회논문지SP
    • /
    • 제41권6호
    • /
    • pp.19-24
    • /
    • 2004
  • 본 논문에서는 LCD (liquid crystal display) 생산라인에서 컴퓨터 비전에 의한 BLU (back light unit)의 고해상도 정밀검사를 원활하게 하기 위해 SIMD (single instruction stream and multiple data stream)형 병렬 구조의 다중 프로세서를 이용하여 계산 집약적인 NTGST (noise-tolerant generalized symmetry transform) 검사 알고리즘을 병렬구현 하였다. 먼저 알고리즘 자체의 속도향상을 위해 C 코드의 최적화를 거친 후, 순차형 프로그램을 N개의 데이터를 동시에 처리하는 SIMD형 언어로 변환하고, 검사영상 데이터를 SIMD형 다중프로세서에서 P개의 각 쓰레드에 분할 할당함으로써 O(NP)의 속도향상이 가능하도록 하였다. Dual Pentium Ⅲ 프로세서를 사용하여 실험한 결과, 제안한 병렬시스템은 기존보다 Sp=8 배 이상 고속 처리가 가능하여, 다양한 크기의 BLU에 대한 고해상도 정밀검사장비에도 신축적으로 확장적용 가능함을 확인하였다.

실시간 H.264/AVC 처리를 위한 ASIP설계 (ASIP Design for Real-Time Processing of H.264)

  • 김진수;선우명훈
    • 전자공학회논문지CI
    • /
    • 제44권5호
    • /
    • pp.12-19
    • /
    • 2007
  • 본 논문에서는 ASIP(Application Specific Instruction-set Processor) 기반의 실시간 H.264/AVC 구현 가능한 VSIP(Video Specific Instruction-set Processor) 을 제안한다. 제안한 VSIP은 H.264/AVC의 화면 내 예측, 디블록킹 필터, 정수 변환 등 새로운 기능들을 효율적으로 지원하기 위한 전용의 하드웨어 구조와 명령어를 가지고 있다. 또한 화면 간 예측 및 엔트로피 코딩과 같이 연산량이 많은 부분은 하드웨어 가속기로 만들어 연산 처리 속도 및 효율을 높였다. VSIP은 H.264/AVC에 적합한 하드웨어 구조와 명령어를 통해 기존의 디지털 신호처리 프로세서보다 작은 크기를 가지며, 메모리 접근 횟수를 줄여 전력 소비를 감소시켰다. 제안한 VSIP을 이용하여 실시간 영상 신호처리를 할 수 있으며, 다양한 프로파일과 표준을 지원할 수 있다.