• Title/Summary/Keyword: computation time reduction

Search Result 221, Processing Time 0.029 seconds

Asynchronous Multiplier with Parallel Array Structure (병렬배열구조를 사용한 비동기 곱셈기)

  • Park, Chan-Ho;Choe, Byeong-Su;Lee, Dong-Ik
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.39 no.5
    • /
    • pp.87-94
    • /
    • 2002
  • In this paper an asynchronous away multiplier with a parallel array structure is introduced. This parallel array structure is used to make the computation time faster with a lower Power consumption. Asymmetric parallel away structure is used to minimize the average computation time in an asynchronous multiplier. Simulation shows that this structure reduces the time needed for computation by 55% as compared to conventional booth encoding array structures and that the multiplier with the proposed away structure shows a reduction of 40% in the computational time with a relatively lower power consumption.

A Computational Interactive Approach to Multi-agent Motion Planning

  • Ji, Sang-Hoon;Choi, Jeong-Sik;Lee, Beom-Hee
    • International Journal of Control, Automation, and Systems
    • /
    • v.5 no.3
    • /
    • pp.295-306
    • /
    • 2007
  • It is well known that mathematical solutions for multi-agent planning problems are very difficult to obtain due to the complexity of mutual interactions among multi-agents. Most of the past research results are thus based on the probabilistic completeness. However, the practicality and effectiveness of the solution from the probabilistic completeness is significantly reduced by heavy computational burden. In this paper, we propose a practically applicable solution technique for multi-agent planning problems, which assures a reasonable computation time and a real world application for more than 3 multi-agents, for the case of general shaped paths in agent movement. First, to reduce the computation time, an extended collision map is developed and utilized for detecting potential collisions and obtaining collision-free solutions for multi-agents. Second, a priority for multi-agents is considered for successive and interactive modifications of the agent movements with lower priority. Various solutions using speed reduction and time delay of the relevant agents are investigated and compared in terms of the computation time. A practical implementation is finally provided for three different types of agents to emphasize the effectiveness of the proposed interactive approach to multi-agent planning problems.

Density Adaptive Grid-based k-Nearest Neighbor Regression Model for Large Dataset (대용량 자료에 대한 밀도 적응 격자 기반의 k-NN 회귀 모형)

  • Liu, Yiqi;Uk, Jung
    • Journal of Korean Society for Quality Management
    • /
    • v.49 no.2
    • /
    • pp.201-211
    • /
    • 2021
  • Purpose: This paper proposes a density adaptive grid algorithm for the k-NN regression model to reduce the computation time for large datasets without significant prediction accuracy loss. Methods: The proposed method utilizes the concept of the grid with centroid to reduce the number of reference data points so that the required computation time is much reduced. Since the grid generation process in this paper is based on quantiles of original variables, the proposed method can fully reflect the density information of the original reference data set. Results: Using five real-life datasets, the proposed k-NN regression model is compared with the original k-NN regression model. The results show that the proposed density adaptive grid-based k-NN regression model is superior to the original k-NN regression in terms of data reduction ratio and time efficiency ratio, and provides a similar prediction error if the appropriate number of grids is selected. Conclusion: The proposed density adaptive grid algorithm for the k-NN regression model is a simple and effective model which can help avoid a large loss of prediction accuracy with faster execution speed and fewer memory requirements during the testing phase.

Development of Time Domain Numerical Computation for Predicting Noise Barrier Efficiency (방음벽 성능 예측을 위한 시간영역 수치해석의 개발)

  • 임창우;정철웅;이수갑
    • Proceedings of the Korean Society for Noise and Vibration Engineering Conference
    • /
    • 2001.05a
    • /
    • pp.757-761
    • /
    • 2001
  • In order to study noise barriers of complex shapes and to assess their efficiency, precise prediction model is required. For instance, geometrical approaches cannot deal with complex diffraction effects. So that in this paper, the time domain numerical computation method(Computational Aeroacoustics method) is applied to estimate noise reduction by diffraction and finite impedance condition. The CAA method can be used to calculate exactly the pressure of complex barrier shape with different impedance condition, such as T-shape, cylindrical edge and multi-edge noise barriers.

  • PDF

On a Reduction of Computation Time of FFT Cepstrum (FFT 켑스트럼의 처리시간 단축에 관한 연구)

  • Jo, Wang-Rae;Kim, Jong-Kuk;Bae, Myung-Jin
    • Speech Sciences
    • /
    • v.10 no.2
    • /
    • pp.57-64
    • /
    • 2003
  • The cepstrum coefficients are the most popular feature for speech recognition or speaker recognition. The cepstrum coefficients are also used for speech synthesis and speech coding but has major drawback of long processing time. In this paper, we proposed a new method that can reduce the processing time of FFT cepstrum analysis. We use the normal ordered inputs for FFT function and the bit-reversed inputs for IFFT function. Therefore we can omit the bit-reversing process and reduce the processing time of FFT ceptrum analysis.

  • PDF

A Study of CHMM Reducing Computational Load Using VQ with Multiple Streams (다중 Stream 구조를 가지는 VQ를 이용하여 연산량을 개선한 CHMM에 관한 연구)

  • Bang, Young Gue;Chung, IK Joo
    • Journal of Industrial Technology
    • /
    • v.26 no.B
    • /
    • pp.233-242
    • /
    • 2006
  • Continuous, discrete and semi continuous HMM systems are used for the speech recognition. Discrete systems have the advantage of low run-time computation. However, vector quantization reduces accuracy and this can lead to poor performance. Continuous systems let us get good correctness but they need much calculation so that occasionally they are unable to be used for practice. Although there are semi-continuous systems which apply advantage of continuous and discrete systems, they also require much computation. In this paper, we proposed the way which reduces calculation for continuous systems. The proposed method has the same computational load as discrete systems but can give better recognition accuracy than discrete systems.

  • PDF

A hybrid algorithm for the synthesis of computer-generated holograms

  • Nguyen The Anh;An Jun Won;Choe Jae Gwang;Kim Nam
    • Proceedings of the Optical Society of Korea Conference
    • /
    • 2003.07a
    • /
    • pp.60-61
    • /
    • 2003
  • A new approach to reduce the computation time of genetic algorithm (GA) for making binary phase holograms is described. Synthesized holograms having diffraction efficiency of 75.8% and uniformity of 5.8% are proven in computer simulation and experimentally demonstrated. Recently, computer-generated holograms (CGHs) having high diffraction efficiency and flexibility of design have been widely developed in many applications such as optical information processing, optical computing, optical interconnection, etc. Among proposed optimization methods, GA has become popular due to its capability of reaching nearly global. However, there exits a drawback to consider when we use the genetic algorithm. It is the large amount of computation time to construct desired holograms. One of the major reasons that the GA' s operation may be time intensive results from the expense of computing the cost function that must Fourier transform the parameters encoded on the hologram into the fitness value. In trying to remedy this drawback, Artificial Neural Network (ANN) has been put forward, allowing CGHs to be created easily and quickly (1), but the quality of reconstructed images is not high enough to use in applications of high preciseness. For that, we are in attempt to find a new approach of combiningthe good properties and performance of both the GA and ANN to make CGHs of high diffraction efficiency in a short time. The optimization of CGH using the genetic algorithm is merely a process of iteration, including selection, crossover, and mutation operators [2]. It is worth noting that the evaluation of the cost function with the aim of selecting better holograms plays an important role in the implementation of the GA. However, this evaluation process wastes much time for Fourier transforming the encoded parameters on the hologram into the value to be solved. Depending on the speed of computer, this process can even last up to ten minutes. It will be more effective if instead of merely generating random holograms in the initial process, a set of approximately desired holograms is employed. By doing so, the initial population will contain less trial holograms equivalent to the reduction of the computation time of GA's. Accordingly, a hybrid algorithm that utilizes a trained neural network to initiate the GA's procedure is proposed. Consequently, the initial population contains less random holograms and is compensated by approximately desired holograms. Figure 1 is the flowchart of the hybrid algorithm in comparison with the classical GA. The procedure of synthesizing a hologram on computer is divided into two steps. First the simulation of holograms based on ANN method [1] to acquire approximately desired holograms is carried. With a teaching data set of 9 characters obtained from the classical GA, the number of layer is 3, the number of hidden node is 100, learning rate is 0.3, and momentum is 0.5, the artificial neural network trained enables us to attain the approximately desired holograms, which are fairly good agreement with what we suggested in the theory. The second step, effect of several parameters on the operation of the hybrid algorithm is investigated. In principle, the operation of the hybrid algorithm and GA are the same except the modification of the initial step. Hence, the verified results in Ref [2] of the parameters such as the probability of crossover and mutation, the tournament size, and the crossover block size are remained unchanged, beside of the reduced population size. The reconstructed image of 76.4% diffraction efficiency and 5.4% uniformity is achieved when the population size is 30, the iteration number is 2000, the probability of crossover is 0.75, and the probability of mutation is 0.001. A comparison between the hybrid algorithm and GA in term of diffraction efficiency and computation time is also evaluated as shown in Fig. 2. With a 66.7% reduction in computation time and a 2% increase in diffraction efficiency compared to the GA method, the hybrid algorithm demonstrates its efficient performance. In the optical experiment, the phase holograms were displayed on a programmable phase modulator (model XGA). Figures 3 are pictures of diffracted patterns of the letter "0" from the holograms generated using the hybrid algorithm. Diffraction efficiency of 75.8% and uniformity of 5.8% are measured. We see that the simulation and experiment results are fairly good agreement with each other. In this paper, Genetic Algorithm and Neural Network have been successfully combined in designing CGHs. This method gives a significant reduction in computation time compared to the GA method while still allowing holograms of high diffraction efficiency and uniformity to be achieved. This work was supported by No.mOl-2001-000-00324-0 (2002)) from the Korea Science & Engineering Foundation.

  • PDF

A Study on Fast Block Matching Algorithm for the Motion Vector Estimation (이동벡터 추정을 위한 고속 Block Matching Algorithm에 관한 연구)

  • 이인홍;박래홍
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.25 no.2
    • /
    • pp.211-219
    • /
    • 1988
  • In this paper effective block matching algorithms are proposed to find the motion vector. There are two approaches to the estimation of the motion vector in MCC (motion compensated coding), i.e.pel(pixel element) recursive algorithm and block matching algorithm. The search algorithm in this paper is based on the block matching method. The advantage of this algorithm is the reduction of the computation time. In order to reduce the computation time, three mathods are proposed in this paper. These new algorithms are faster than other methods. Compared with the three step algorithm by Koga et al., the average ratio of the computational savings obtained from the proposed algorithm is about 3-4.

  • PDF

A Study on the Application of Conjugate Gradient Method in Nonlinear Magnetic Field Analysis by FEM. (유한요소법에 의한 비선형 자계 해석에 공액 구배법 적응 연구)

  • 임달호;신흥교
    • The Transactions of the Korean Institute of Electrical Engineers
    • /
    • v.39 no.1
    • /
    • pp.22-28
    • /
    • 1990
  • This paper is a study on the reduction of computation time in case of nonlinear magnetic field analysis by finite element method and Newton-Raphson method. For the purpose, the nonlinear convergence equation is computed by the conjugate gradient method which is known to be applicable to symmetric positive definite matrix equations only. As the results, we can not prove mathematically that the system Jacobian is positive definite, but when we applied this method, the diverging case did not occur. And the computation time is reduced by 25-55% and 15-45% in comparison with the case of direct and successive over-relaxation method, respectively. Therefore, we proved the utility of conjugate gradient method.

  • PDF

Shielding Design of Electron Beam Accelerators Using Supercomputer (슈퍼컴을 이용한 전자빔가속기의 차폐설계)

  • Kang, Won Gu;Kim, In Soo;Kuk, Sung Han;Kim, Jin Kyu;Han, Bum Soo;Jeong, Kwang Young;Kang, Chang Mu
    • Journal of Radiation Industry
    • /
    • v.4 no.1
    • /
    • pp.33-38
    • /
    • 2010
  • The MCNP5 neutron, electron, photon Monte Carlo transport program was installed on the KISTI's SUN Tachyon computer using the parallel programming. Electron beam accelerators were modeled and shielding calculations were performed in order to investigate the reduction of computation time in the supercomputer environment. It was observed that a speedup of 40 to 80 of computation time can be obtained using 64 CPUs compared to an IBM PC.