• Title/Summary/Keyword: computational complexity reduction

Search Result 258, Processing Time 0.023 seconds

Complexity Reduction for Space-Time Block Codes Decoder in MIMO OFDM Systems

  • Pham Van-Su;Le Minh-Tuan;Mai Linh;Yoon Giwan
    • Journal of information and communication convergence engineering
    • /
    • v.3 no.2
    • /
    • pp.72-75
    • /
    • 2005
  • In this work, we first present our study on the decoding schemes for space-time block code as well as our comments on their complexity. Then, we propose a new modified complex sphere decoding scheme, which has lower computational load than that of conventional complex sphere decoders. In the proposed scheme, the boundary for searching is aided with the intersection of approximate polygon of searching circle and disk of constellation. Therefore, the proposed scheme can reduce the searching boundary while can avoid missing searching points. The performance of the proposed scheme is verified by computer simulation that consolidates our scheme.

FPGA Implementation of SC-FDE Timing Synchronization Algorithm

  • Ji, Suyuan;Chen, Chao;Zhang, Yu
    • Journal of Information Processing Systems
    • /
    • v.15 no.4
    • /
    • pp.890-903
    • /
    • 2019
  • The single carrier frequency domain equalization (SC-FDE) technology is an important part of the broadband wireless access communication system, which can effectively combat the frequency selective fading in the wireless channel. In SC-FDE communication system, the accuracy of timing synchronization directly affects the performance of the SC-FDE system. In this paper, on the basis of Schmidl timing synchronization algorithm a timing synchronization algorithm suitable for FPGA (field programmable gate array) implementation is proposed. In the FPGA implementation of the timing synchronization algorithm, the sliding window accumulation, quantization processing and amplitude reduction techniques are adopted to reduce the complexity in the implementation of FPGA. The simulation results show that the algorithm can effectively realize the timing synchronization function under the condition of reducing computational complexity and hardware overhead.

A Fast Intra-Prediction Method in HEVC Using Rate-Distortion Estimation Based on Hadamard Transform

  • Kim, Younhee;Jun, DongSan;Jung, Soon-Heung;Choi, Jin Soo;Kim, Jinwoong
    • ETRI Journal
    • /
    • v.35 no.2
    • /
    • pp.270-280
    • /
    • 2013
  • A fast intra-prediction method is proposed for High Efficiency Video Coding (HEVC) using a fast intra-mode decision and fast coding unit (CU) size decision. HEVC supports very sophisticated intra modes and a recursive quadtree-based CU structure. To provide a high coding efficiency, the mode and CU size are selected in a rate-distortion optimized manner. This causes a high computational complexity in the encoder, and, for practical applications, the complexity should be significantly reduced. In this paper, among the many predefined modes, the intra-prediction mode is chosen without rate-distortion optimization processes, instead using the difference between the minimum and second minimum of the rate-distortion cost estimation based on the Hadamard transform. The experiment results show that the proposed method achieves a 49.04% reduction in the intra-prediction time and a 32.74% reduction in the total encoding time with a nearly similar coding performance to that of HEVC test model 2.1.

PAPR Reduction of an OFDM Signal by use of PTS scheme with MG-PSO Algorithm (MG-PSO 알고리즘을 적용한 PTS 기법에 의한 OFDM 신호의 PAPR 감소)

  • Kim, Wan-Tae;Yoo, Sun-Yong;Cho, Sung-Joon
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.46 no.1
    • /
    • pp.1-9
    • /
    • 2009
  • OFDM(Orthogonal Frequency Division Multiplexing) system is robust to frequency selective fading and narrowband interference in high-speed data communications. However, an OPDM signal consists of a number of independently modulated subcarriers and the superposition of these subcarriers causes a problem that can give a large PARR(Peak-to-Average Power Ratio). PTS(Partial Transmit Sequence) scheme can reduce the PAPR by dividing OFDM signal into subblocks and then multiplying the phase weighting factors to each subblocks, but computational complexity for selecting of phase weighting factors increases exponentially with the number of subblocks. Therefore, in this paper, MG-PSO(Modified Greedy algorithm-Particle Swarm Optimization) algorithm that combines modified greedy algorithm and PSO(Particle Swarm Optimization) algorithm is proposed to use for the phase control method in PTS scheme. This method can solve the computational complexity and guarantee to reduce PAPR. We analyzed the performance of the PAPR reduction when we applied the proposed method to telecommunication systems.

A Study on the PAPR Reduction Using Phase Rotation Method Applying Metaheuristic Algorithm (Metaheuristic 알고리즘을 적용한 위상회전 기법에 의한 PAPR 감소에 관한 연구)

  • Yoo, Sun-Yong;Park, Bee-Ho;Kim, Wan-Tae;Cho, Sung-Joon
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.46 no.5
    • /
    • pp.26-35
    • /
    • 2009
  • OFDM (Orthogonal Frequency Division Multiplexing) system is robust to frequency selective fading and narrowband interference in high-speed data communications. However, an OFDM signal consists of a number of independently modulated subcarriers and the superposition of these subcarriers causes a problem that can give a large PAPR(Peak-to-Average Power Ratio). Phase rotation method can reduce the PAPR without nonlinear distortion by multiplying phase weighting factors. But computational complexity of searching phase weighting factors is increased exponentially with the number of subblocks and considered phase factor. Therefore, a new method, which can reduce computational complexity and detect phase weighting factors efficiently, should be developed. In this paper, a modeling process is introduced, which apply metaheuristic algerian in phase rotation method and optimize in PTS (Particle Swarm Optimization) scheme. Proposed algorithm can solve the computational complexity and guarantee to reduce PAPR We analyzed the efficiency of the PAPR reduction through a simulation when we applied the proposed method to telecommunication systems.

Low Complexity Hybrid Interpolation Algorithm using Weighted Edge Detector (가중치 윤곽선 검출기를 이용한 저 복잡도 하이브리드 보간 알고리듬)

  • Kwon, Hyeok-Jin;Jeon, Gwang-Gil;Jeong, Je-Chang
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.32 no.3C
    • /
    • pp.241-248
    • /
    • 2007
  • In predictive image coding, a LS (Least Squares)-based adaptive predictor is an efficient method to improve image edge predictions. This paper proposes a hybrid interpolation with weighted edge detector. A hybrid approach of switching between bilinear interpolation and EDI (Edge-Directed Interpolation) is proposed in order to reduce the overall computational complexity The objective and subjective quality is also similar to the bilinear interpolation and EDI. Experimental results demonstrate that this hybrid interpolation method that utilizes a weighted edge detector can achieve reduction in complexity with minimal degradation in the interpolation results.

Optimal Path Planning Algorithm for Visiting Multiple Mission Points in Dynamic Environments (동적 변화 환경에서 다중 임무점 방문을 위한 최적 경로 계획 알고리즘)

  • Lee, Hohyeong;Chang, Woohyuk;Jang, Hwanchol
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.47 no.5
    • /
    • pp.379-387
    • /
    • 2019
  • The complexity of path planning for visiting multiple mission points is even larger than that of single pair path planning. Deciding a path for visiting n mission points requires conducting $n^2+n$ times of single pair path planning. We propose Multiple Mission $D^*$ Lite($MMD^*L$) which is an optimal path planning algorithm for visiting multiple mission points in dynamic environments. $MMD^*L$ reduces the complexity by reusing the computational data of preceding single pair path planning. Simulation results show that the complexity reduction is significant while its path optimality is not compromised.

Computational Complexity Reduction of Speech Recognizers Based on the Modified Bucket Box Intersection Algorithm (변형된 BBI 알고리즘에 기반한 음성 인식기의 계산량 감축)

  • Kim, Keun-Yong;Kim, Dong-Hwa
    • MALSORI
    • /
    • no.60
    • /
    • pp.109-123
    • /
    • 2006
  • Since computing the log-likelihood of Gaussian mixture density is a major computational burden for the speech recognizer based on the continuous HMM, several techniques have been proposed to reduce the number of mixtures to be used for recognition. In this paper, we propose a modified Bucket Box Intersection (BBI) algorithm, in which two relative thresholds are employed: one is the relative threshold in the conventional BBI algorithm and the other is used to reduce the number of the Gaussian boxes which are intersected by the hyperplanes at the boxes' edges. The experimental results show that the proposed algorithm reduces the number of Gaussian mixtures by 12.92% during the recognition phase with negligible performance degradation compared to the conventional BBI algorithm.

  • PDF

Comparison of Fall Detection Systems Based on YOLOPose and Long Short-Term Memory

  • Seung Su Jeong;Nam Ho Kim;Yun Seop Yu
    • Journal of information and communication convergence engineering
    • /
    • v.22 no.2
    • /
    • pp.139-144
    • /
    • 2024
  • In this study, four types of fall detection systems - designed with YOLOPose, principal component analysis (PCA), convolutional neural network (CNN), and long short-term memory (LSTM) architectures - were developed and compared in the detection of everyday falls. The experimental dataset encompassed seven types of activities: walking, lying, jumping, jumping in activities of daily living, falling backward, falling forward, and falling sideways. Keypoints extracted from YOLOPose were entered into the following architectures: RAW-LSTM, PCA-LSTM, RAW-PCA-LSTM, and PCA-CNN-LSTM. For the PCA architectures, the reduced input size stemming from a dimensionality reduction enhanced the operational efficiency in terms of computational time and memory at the cost of decreased accuracy. In contrast, the addition of a CNN resulted in higher complexity and lower accuracy. The RAW-LSTM architecture, which did not include either PCA or CNN, had the least number of parameters, which resulted in the best computational time and memory while also achieving the highest accuracy.

Automatic Switching of Clustering Methods based on Fuzzy Inference in Bibliographic Big Data Retrieval System

  • Zolkepli, Maslina;Dong, Fangyan;Hirota, Kaoru
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.14 no.4
    • /
    • pp.256-267
    • /
    • 2014
  • An automatic switch among ensembles of clustering algorithms is proposed as a part of the bibliographic big data retrieval system by utilizing a fuzzy inference engine as a decision support tool to select the fastest performing clustering algorithm between fuzzy C-means (FCM) clustering, Newman-Girvan clustering, and the combination of both. It aims to realize the best clustering performance with the reduction of computational complexity from O($n^3$) to O(n). The automatic switch is developed by using fuzzy logic controller written in Java and accepts 3 inputs from each clustering result, i.e., number of clusters, number of vertices, and time taken to complete the clustering process. The experimental results on PC (Intel Core i5-3210M at 2.50 GHz) demonstrates that the combination of both clustering algorithms is selected as the best performing algorithm in 20 out of 27 cases with the highest percentage of 83.99%, completed in 161 seconds. The self-adapted FCM is selected as the best performing algorithm in 4 cases and the Newman-Girvan is selected in 3 cases.The automatic switch is to be incorporated into the bibliographic big data retrieval system that focuses on visualization of fuzzy relationship using hybrid approach combining FCM and Newman-Girvan algorithm, and is planning to be released to the public through the Internet.