• Title/Summary/Keyword: Computation time

Search Result 3,169, Processing Time 0.03 seconds

The Computation Reduction Algorithm Independent of the Language for CELP Vocoders (각국 언어 특성에 독립적인 CELP 계열 보코더에서의 계산량 단축 알고리즘)

  • Ju, Sang-Gyu
    • Proceedings of the KAIS Fall Conference
    • /
    • 2010.05a
    • /
    • pp.257-260
    • /
    • 2010
  • In this paper, we propose the computation reduction methods of LSP(Line spectrum pairs) transformation that is mainly used in CELP vocoders. In order to decrease the computational time in real root method the characteristic of four proposed algorithms is as the following. First, scheme to reduce the LSP transformation time uses mel scale. Developed the second scheme is the control of searching order by the distribution characteristic of LSP parameters. Third, scheme to reduce the LSP transformation time uses voice characteristics. Developed the fourth scheme is the control of searching interval and order by the distribution characteristic of LSP parameters. As a result of searching time, computational amount, transformed LSP parameters, SNR, MOS test, waveform of synthesized speech, spectrogram analysis, searching time is reduced about 37.5%, 46.21%, 46.3%, 51.29% in average, computational amount is reduced about 44.76%, 49.44%, 47.03%, 57.40%. But the transformed LSP parameters of the proposed methods were the same as those of real root method.

  • PDF

Sub-pixel Motion Estimation Algorithm with Low Computation Complexity for H.264 Video Compression (H.264 동영상 압축을 위한 낮은 복잡도를 갖는 부 화소 단위에서의 움직임 추정)

  • Lee, Yun-Hwa;Shin, Hyun-Chul
    • Proceedings of the IEEK Conference
    • /
    • 2005.11a
    • /
    • pp.639-642
    • /
    • 2005
  • Motion Estimation(ME) is an important part of video compression, because it requires a large amount of computation. Half-pixel and quarter-pixel motion estimation allows high video compression rates but it also has high computation complexity. In this paper we suggest a new and efficient motion estimation algorithm for half-pixel and quarter-pixel motion estimation using SAD values. In the method, an integer-pixel motion vector is found and then only three neighboring points of the integer-pixel motion vector is evaluated to find the half-pixel motion vector. The quarter-pixel motion vector is also found by using a similar method. Experimental results of our method shows 20% reduction in computation time, when compared with those of a conventional method, while producing same quality motion vectors.

  • PDF

Verifiable Outsourced Ciphertext-Policy Attribute-Based Encryption for Mobile Cloud Computing

  • Zhao, Zhiyuan;Wang, Jianhua
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.6
    • /
    • pp.3254-3272
    • /
    • 2017
  • With the development of wireless access technologies and the popularity of mobile intelligent terminals, cloud computing is expected to expand to mobile environments. Attribute-based encryption, widely applied in cloud computing, incurs massive computational cost during the encryption and decryption phases. The computational cost grows with the complexity of the access policy. This disadvantage becomes more serious for mobile devices because they have limited resources. To address this problem, we present an efficient verifiable outsourced scheme based on the bilinear group of prime order. The scheme is called the verifiable outsourced computation ciphertext-policy attribute-based encryption scheme (VOC-CP-ABE), and it provides a way to outsource intensive computing tasks during encryption and decryption phases to CSP without revealing the private information and leaves only marginal computation to the user. At the same time, the outsourced computation can be verified by two hash functions. Then, the formal security proofs of its (selective) CPA security and verifiability are provided. Finally, we discuss the performance of the proposed scheme with comparisons to several related works.

Automatic Speech Database Verification Method Based on Confidence Measure

  • Kang Jeomja;Jung Hoyoung;Kim Sanghun
    • MALSORI
    • /
    • no.51
    • /
    • pp.71-84
    • /
    • 2004
  • In this paper, we propose the automatic speech database verification method(or called automatic verification) based on confidence measure for a large speech database. This method verifies the consistency between given transcription and speech using the confidence measure. The automatic verification process consists of two stages : the word-level likelihood computation stage and multi-level likelihood ratio computation stage. In the word-level likelihood computation stage, we calculate the word-level likelihood using the viterbi decoding algorithm and make the segment information. In the multi-level likelihood ratio computation stage, we calculate the word-level and the phone-level likelihood ratio based on confidence measure with anti-phone model. By automatic verification, we have achieved about 61% error reduction. And also we can reduce the verification time from 1 month in manual to 1-2 days in automatic.

  • PDF

New systolic arrays for computation of the 1-D and 2-D discrete wavelet transform (1차원 및 2차원 이산 웨이브렛 변환 계산을 위한 새로운 시스톨릭 어레이)

  • 반성범;박래홍
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.34S no.10
    • /
    • pp.132-140
    • /
    • 1997
  • This paper proposes systolic array architectures for compuataion of the 1-D and 2-D discrete wavelet transform (DWT). The proposed systolic array for compuataion of the 1-D DWT consists of L processing element (PE) arrays, where the PE array denotes the systolic array for computation of the one level DWT. The proposed PE array computes only the product terms that are required for further computation and the outputs of low and high frequency filters are computed in alternate clock cycles. Therefore, the proposed architecuter can compute the low and high frequency outputs using a single architecture. The proposed systolic array for computation of the 2-D DWT consists of two systolic array architectures for comutation of the 1-D DWT and memory unit. The required time and hardware cost of the proposed systolic arrays are comparable to those of the conventional architectures. However, the conventional architectures need extra processing units whereas the proposed architectures fo not. The proposed architectures can be applied to subband decomposition by simply changing the filter coefficients.

  • PDF

Three-Dimensional Numerical Computation and Experiment on Periodic Flows under a Background Rotation (배경회전하에서 형성되는 주기적 유동의 3차원 수치해석과 실험)

  • Suh, Yong-Kweon;Park, Jae-Hyun
    • Transactions of the Korean Society of Mechanical Engineers B
    • /
    • v.27 no.5
    • /
    • pp.628-634
    • /
    • 2003
  • We present numerical and experimental results of periodic flows inside a rectangular container under a background rotation. The periodic flows are generated by changing the speed of rotation periodically so that a time-periodic body forces produce the unsteady flows. In numerical computation, a parallel-computation technique with MPI is implemented. Flow visualization and PIV measurement are also performed to obtain velocity fields at the free surface. Through a series of numerical and experimental works, we aim to clarify, if any, the fundamental reasons \ulcornerf discrepancy between the two-dimensional computation and the experimental measurement, which was detected in the previous study for the same flow model. Specifically, we check if the various assumptions prerequisite for the validity of the classical Ekman pumping law are satisfied for periodic flows under a background rotation.

Transient Response Analysis of Frame Structures Using the Finite Element-transfer Stiffness Coefficient Method (FE-TSCM) (유한요소-전달강성계수법을 이용한 골조 구조물의 과도응답해석)

  • 최명수;문덕홍;김성진
    • Transactions of the Korean Society for Noise and Vibration Engineering
    • /
    • v.12 no.9
    • /
    • pp.674-684
    • /
    • 2002
  • In order to decrease remarkably the computation time and storage used in the direct integration method without the loss of accuracy, authors suggest a new transient analysis algorithm. This algorithm is derived from the combination of three techniques, that is, the transfer technique of the transfer stiffness coefficient method, the modeling technique of the finite element method, and the numerical integration technique of the Newmark method. In this paper, the transient analysis algorithm of a frame structure is formulated by the proposed method. The accuracy and computation efficiency of the proposed method are demonstrated through the comparing with the computation results by the direct integration method for three computation models under various excitations.

Speed-up of the Matrix Computation on the Ridge Regression

  • Lee, Woochan;Kim, Moonseong;Park, Jaeyoung
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.10
    • /
    • pp.3482-3497
    • /
    • 2021
  • Artificial intelligence has emerged as the core of the 4th industrial revolution, and large amounts of data processing, such as big data technology and rapid data analysis, are inevitable. The most fundamental and universal data interpretation technique is an analysis of information through regression, which is also the basis of machine learning. Ridge regression is a technique of regression that decreases sensitivity to unique or outlier information. The time-consuming calculation portion of the matrix computation, however, basically includes the introduction of an inverse matrix. As the size of the matrix expands, the matrix solution method becomes a major challenge. In this paper, a new algorithm is introduced to enhance the speed of ridge regression estimator calculation through series expansion and computation recycle without adopting an inverse matrix in the calculation process or other factorization methods. In addition, the performances of the proposed algorithm and the existing algorithm were compared according to the matrix size. Overall, excellent speed-up of the proposed algorithm with good accuracy was demonstrated.

Dynamic Computation Offloading Based on Q-Learning for UAV-Based Mobile Edge Computing

  • Shreya Khisa;Sangman Moh
    • Smart Media Journal
    • /
    • v.12 no.3
    • /
    • pp.68-76
    • /
    • 2023
  • Emerging mobile edge computing (MEC) can be used in battery-constrained Internet of things (IoT). The execution latency of IoT applications can be improved by offloading computation-intensive tasks to an MEC server. Recently, the popularity of unmanned aerial vehicles (UAVs) has increased rapidly, and UAV-based MEC systems are receiving considerable attention. In this paper, we propose a dynamic computation offloading paradigm for UAV-based MEC systems, in which a UAV flies over an urban environment and provides edge services to IoT devices on the ground. Since most IoT devices are energy-constrained, we formulate our problem as a Markov decision process considering the energy level of the battery of each IoT device. We also use model-free Q-learning for time-critical tasks to maximize the system utility. According to our performance study, the proposed scheme can achieve desirable convergence properties and make intelligent offloading decisions.

Improvement of Disparity Map using Loopy Belief Propagation based on Color and Edge (Disparity 보정을 위한 컬러와 윤곽선 기반 루피 신뢰도 전파 기법)

  • Kim, Eun Kyeong;Cho, Hyunhak;Lee, Hansoo;Wibowo, Suryo Adhi;Kim, Sungshin
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.25 no.5
    • /
    • pp.502-508
    • /
    • 2015
  • Stereo images have an advantage of calculating depth(distance) values which can not analyze from 2-D images. However, depth information obtained by stereo images has due to following reasons: it can be obtained by computation process; mismatching occurs when stereo matching is processing in occlusion which has an effect on accuracy of calculating depth information. Also, if global method is used for stereo matching, it needs a lot of computation. Therefore, this paper proposes the method obtaining disparity map which can reduce computation time and has higher accuracy than established method. Edge extraction which is image segmentation based on feature is used for improving accuracy and reducing computation time. Color K-Means method which is image segmentation based on color estimates correlation of objects in an image. And it extracts region of interest for applying Loopy Belief Propagation(LBP). For this, disparity map can be compensated by considering correlation of objects in the image. And it can reduce computation time because of calculating region of interest not all pixels. As a result, disparity map has more accurate and the proposed method reduces computation time.