• Title/Summary/Keyword: Fixed Complexity

Search Result 288, Processing Time 0.025 seconds

Improvement of Address Pointer Assignment in DSP Code Generation (DSP용 코드 생성에서 주소 포인터 할당 성능 향상 기법)

  • Lee, Hee-Jin;Lee, Jong-Yeol
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.45 no.1
    • /
    • pp.37-47
    • /
    • 2008
  • Exploitation of address generation units which are typically provided in DSPs plays an important role in DSP code generation since that perform fast address computation in parallel to the central data path. Offset assignment is optimization of memory layout for program variables by taking advantage of the capabilities of address generation units, consists of memory layout generation and address pointer assignment steps. In this paper, we propose an effective address pointer assignment method to minimize the number of address calculation instructions in DSP code generation. The proposed approach reduces the time complexity of a conventional address pointer assignment algorithm with fixed memory layouts by using minimum cost-nodes breaking. In order to contract memory size and processing time, we employ a powerful pruning technique. Moreover our proposed approach improves the initial solution iteratively by changing the memory layout for each iteration because the memory layout affects the result of the address pointer assignment algorithm. We applied the proposed approach to about 3,000 sequences of the OffsetStone benchmarks to demonstrate the effectiveness of the our approach. Experimental results with benchmarks show an average improvement of 25.9% in the address codes over previous works.

Enhancement in BER Performance by Adaptive Sector Antennas in CDMA Wireless Multi-media Communications (CDMA 무선 멀티미디어 통신에서 적응형 섹터 안테나에 의한 오율성능 개선)

  • Lee, Ju-Hyung;Kim, Young-Chul;Oh, Chang-Heon;Cho, Sung-Joon
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.10A
    • /
    • pp.980-989
    • /
    • 2002
  • In this paper, we have proposed the sectorization method of the adaptive sector antenna and the new receiver structure for the CDMA wireless multi-media communication which uses multi-code and single-code. The sectorization and average BER performance of the adaptive sector antenna, which employs the proposed sectorization method, are obtained through simulation. The simulation results have shown that the sizes of sectors are properly controlled according to the distribution of the load and that the BER performance of the adaptive sector antenna is enhanced much more than that of the fixed sector antenna. Also, we have proposed two receivers, which are different from the positions to detect multi-code signals, for the multi-media communication. Each receiver has a simple cancellation scheme and two adaptive sector antennas, which are called the 1st adaptive sector antenna and the 2nd adaptive sector antenna, respectively. In this paper, the average BER performances of the receivers are compared through simulation. As the results of the simulation, we have recognized that all user' BER performances are greatly dependent on the positions to detect the multi-code signals. When we consider both system complexity and terget BERs of signals, we have concluded that the receiver, which detect multi-code signals in the 1st and the 2nd adaptive sector antenna, is appropriate for the CDMA wireless multi-media communication systems.

Visualization of Korean Speech Based on the Distance of Acoustic Features (음성특징의 거리에 기반한 한국어 발음의 시각화)

  • Pok, Gou-Chol
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.13 no.3
    • /
    • pp.197-205
    • /
    • 2020
  • Korean language has the characteristics that the pronunciation of phoneme units such as vowels and consonants are fixed and the pronunciation associated with a notation does not change, so that foreign learners can approach rather easily Korean language. However, when one pronounces words, phrases, or sentences, the pronunciation changes in a manner of a wide variation and complexity at the boundaries of syllables, and the association of notation and pronunciation does not hold any more. Consequently, it is very difficult for foreign learners to study Korean standard pronunciations. Despite these difficulties, it is believed that systematic analysis of pronunciation errors for Korean words is possible according to the advantageous observations that the relationship between Korean notations and pronunciations can be described as a set of firm rules without exceptions unlike other languages including English. In this paper, we propose a visualization framework which shows the differences between standard pronunciations and erratic ones as quantitative measures on the computer screen. Previous researches only show color representation and 3D graphics of speech properties, or an animated view of changing shapes of lips and mouth cavity. Moreover, the features used in the analysis are only point data such as the average of a speech range. In this study, we propose a method which can directly use the time-series data instead of using summary or distorted data. This was realized by using the deep learning-based technique which combines Self-organizing map, variational autoencoder model, and Markov model, and we achieved a superior performance enhancement compared to the method using the point-based data.

Cell ID Detection Schemes Using PSS/SSS for 5G NR System (5G NR 시스템에서 PSS/SSS를 이용한 Cell ID 검출 방법)

  • Ahn, Haesung;Kim, Hyeongseok;Cha, Eunyoung;Kim, Jeongchang
    • Journal of Broadcast Engineering
    • /
    • v.25 no.6
    • /
    • pp.870-881
    • /
    • 2020
  • This paper presents cell ID (cell identity) detection schemes using PSS/SSS (primary synchronization signal/secondary synchronization signal) for 5G NR (new radio) system and evaluates the detection performance. In this paper, we consider two cell ID detection schemes, i.e. two-stage detection and joint detection schemes. The two-stage detection scheme consists of two stages which estimate a channel gain between a transmitter and receiver and detect the PSS and SSS sequences. The joint detection scheme jointly detects the PSS and SSS sequences. In addition, this paper presents coherent and non-coherent combining schemes. The coherent scheme calculates the correlation value for the total length of the given PSS and SSS sequences, and the non-coherent combining scheme calculates the correlation within each group by dividing the total length of the sequence into several groups and then combines them non-coherently. For the detection schemes considered in this paper, the detection error rates of PSS, SSS and overall cell ID are evaluated and compared through computer simulations. The simulation results show that the joint detection scheme outperforms the two-stage detection scheme for both coherent and non-coherent combining schemes, but the two-stage detection scheme can greatly reduce the computational complexity compared to the joint detection scheme. In addition, the non-coherent combining detection scheme shows better performance under the additive white Gaussian noise (AWGN), fixed, and mobile environments.

A Study of the Influencing Factors for Decision Making on Construction Contract Types : Focused on DoD Construction Acquisitions with Firm Fixed Price and Cost Reimbursable in FAR (건설공사 대가지급방식의 의사결정 영향요인에 관한 연구 - 미국 연방조달규정에 따른 미국 국방성의 정액계약과 실비정산계약을 중심으로 -)

  • Son, Young-Hoon;Kim, Kyung-Rai
    • Korean Journal of Construction Engineering and Management
    • /
    • v.25 no.2
    • /
    • pp.23-35
    • /
    • 2024
  • This study analyzed the correlation between each of the 12 influencing factors in FAR 16.04 and the decision-making process for construction contract types, using data from a total of 2,406 DoD Construction Acquisitions spanning from 2008 to 2022. The study considered 12 independent variables, grouped into 4 Characteristics with 3 factors each. Meanwhile, all other contract types were categorized into two types: Firm-Fixed-Price (FFP) and Cost-Reimbursement Contract (CRC), which served as the dependent variables. The findings revealed that FFP contracts significantly dominated in terms of acquisition volume. In line with prevailing beliefs, logistic data analysis and Analytical Hierarchy Process (AHP) analysis of Relative Weights from Experts' Survey demonstrated that independent variables like Uncertainty of the Scope of Work and Complexity found out to be increasing the likelihood of selecting CRC. The number of contractors in the market does indeed influence the possibilities of contract decision-making between CRC and FFP. Meanwhile, the p-values of the top 3 influencing factors on CRC from the AHP analysis-namely, Appropriateness of CAS, Project Urgency, and Cost Analysis-exceeded 0.05 in the binominal regression results, rendering it inconclusive whether they significantly influenced the construction contract type decision, particularly with respect to payment methods. This outcome partly results from the fact that a majority of respondents possessed specific experiences related to the USFK relocation project. Furthermore, influencing factors in construction projects behave differently than common beliefs suggest. As a result, it is imperative to consider the 12 influencing factors categorized into 4 Characteristics areas before establishing acquisition strategies for targeted construction projects.

Transfer Learning using Multiple ConvNet Layers Activation Features with Principal Component Analysis for Image Classification (전이학습 기반 다중 컨볼류션 신경망 레이어의 활성화 특징과 주성분 분석을 이용한 이미지 분류 방법)

  • Byambajav, Batkhuu;Alikhanov, Jumabek;Fang, Yang;Ko, Seunghyun;Jo, Geun Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.205-225
    • /
    • 2018
  • Convolutional Neural Network (ConvNet) is one class of the powerful Deep Neural Network that can analyze and learn hierarchies of visual features. Originally, first neural network (Neocognitron) was introduced in the 80s. At that time, the neural network was not broadly used in both industry and academic field by cause of large-scale dataset shortage and low computational power. However, after a few decades later in 2012, Krizhevsky made a breakthrough on ILSVRC-12 visual recognition competition using Convolutional Neural Network. That breakthrough revived people interest in the neural network. The success of Convolutional Neural Network is achieved with two main factors. First of them is the emergence of advanced hardware (GPUs) for sufficient parallel computation. Second is the availability of large-scale datasets such as ImageNet (ILSVRC) dataset for training. Unfortunately, many new domains are bottlenecked by these factors. For most domains, it is difficult and requires lots of effort to gather large-scale dataset to train a ConvNet. Moreover, even if we have a large-scale dataset, training ConvNet from scratch is required expensive resource and time-consuming. These two obstacles can be solved by using transfer learning. Transfer learning is a method for transferring the knowledge from a source domain to new domain. There are two major Transfer learning cases. First one is ConvNet as fixed feature extractor, and the second one is Fine-tune the ConvNet on a new dataset. In the first case, using pre-trained ConvNet (such as on ImageNet) to compute feed-forward activations of the image into the ConvNet and extract activation features from specific layers. In the second case, replacing and retraining the ConvNet classifier on the new dataset, then fine-tune the weights of the pre-trained network with the backpropagation. In this paper, we focus on using multiple ConvNet layers as a fixed feature extractor only. However, applying features with high dimensional complexity that is directly extracted from multiple ConvNet layers is still a challenging problem. We observe that features extracted from multiple ConvNet layers address the different characteristics of the image which means better representation could be obtained by finding the optimal combination of multiple ConvNet layers. Based on that observation, we propose to employ multiple ConvNet layer representations for transfer learning instead of a single ConvNet layer representation. Overall, our primary pipeline has three steps. Firstly, images from target task are given as input to ConvNet, then that image will be feed-forwarded into pre-trained AlexNet, and the activation features from three fully connected convolutional layers are extracted. Secondly, activation features of three ConvNet layers are concatenated to obtain multiple ConvNet layers representation because it will gain more information about an image. When three fully connected layer features concatenated, the occurring image representation would have 9192 (4096+4096+1000) dimension features. However, features extracted from multiple ConvNet layers are redundant and noisy since they are extracted from the same ConvNet. Thus, a third step, we will use Principal Component Analysis (PCA) to select salient features before the training phase. When salient features are obtained, the classifier can classify image more accurately, and the performance of transfer learning can be improved. To evaluate proposed method, experiments are conducted in three standard datasets (Caltech-256, VOC07, and SUN397) to compare multiple ConvNet layer representations against single ConvNet layer representation by using PCA for feature selection and dimension reduction. Our experiments demonstrated the importance of feature selection for multiple ConvNet layer representation. Moreover, our proposed approach achieved 75.6% accuracy compared to 73.9% accuracy achieved by FC7 layer on the Caltech-256 dataset, 73.1% accuracy compared to 69.2% accuracy achieved by FC8 layer on the VOC07 dataset, 52.2% accuracy compared to 48.7% accuracy achieved by FC7 layer on the SUN397 dataset. We also showed that our proposed approach achieved superior performance, 2.8%, 2.1% and 3.1% accuracy improvement on Caltech-256, VOC07, and SUN397 dataset respectively compare to existing work.

Noise-robust electrocardiogram R-peak detection with adaptive filter and variable threshold (적응형 필터와 가변 임계값을 적용하여 잡음에 강인한 심전도 R-피크 검출)

  • Rahman, MD Saifur;Choi, Chul-Hyung;Kim, Si-Kyung;Park, In-Deok;Kim, Young-Pil
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.18 no.12
    • /
    • pp.126-134
    • /
    • 2017
  • There have been numerous studies on extracting the R-peak from electrocardiogram (ECG) signals. However, most of the detection methods are complicated to implement in a real-time portable electrocardiograph device and have the disadvantage of requiring a large amount of calculations. R-peak detection requires pre-processing and post-processing related to baseline drift and the removal of noise from the commercial power supply for ECG data. An adaptive filter technique is widely used for R-peak detection, but the R-peak value cannot be detected when the input is lower than a threshold value. Moreover, there is a problem in detecting the P-peak and T-peak values due to the derivation of an erroneous threshold value as a result of noise. We propose a robust R-peak detection algorithm with low complexity and simple computation to solve these problems. The proposed scheme removes the baseline drift in ECG signals using an adaptive filter to solve the problems involved in threshold extraction. We also propose a technique to extract the appropriate threshold value automatically using the minimum and maximum values of the filtered ECG signal. To detect the R-peak from the ECG signal, we propose a threshold neighborhood search technique. Through experiments, we confirmed the improvement of the R-peak detection accuracy of the proposed method and achieved a detection speed that is suitable for a mobile system by reducing the amount of calculation. The experimental results show that the heart rate detection accuracy and sensitivity were very high (about 100%).

Discovering Promising Convergence Technologies Using Network Analysis of Maturity and Dependency of Technology (기술 성숙도 및 의존도의 네트워크 분석을 통한 유망 융합 기술 발굴 방법론)

  • Choi, Hochang;Kwahk, Kee-Young;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.101-124
    • /
    • 2018
  • Recently, most of the technologies have been developed in various forms through the advancement of single technology or interaction with other technologies. Particularly, these technologies have the characteristic of the convergence caused by the interaction between two or more techniques. In addition, efforts in responding to technological changes by advance are continuously increasing through forecasting promising convergence technologies that will emerge in the near future. According to this phenomenon, many researchers are attempting to perform various analyses about forecasting promising convergence technologies. A convergence technology has characteristics of various technologies according to the principle of generation. Therefore, forecasting promising convergence technologies is much more difficult than forecasting general technologies with high growth potential. Nevertheless, some achievements have been confirmed in an attempt to forecasting promising technologies using big data analysis and social network analysis. Studies of convergence technology through data analysis are actively conducted with the theme of discovering new convergence technologies and analyzing their trends. According that, information about new convergence technologies is being provided more abundantly than in the past. However, existing methods in analyzing convergence technology have some limitations. Firstly, most studies deal with convergence technology analyze data through predefined technology classifications. The technologies appearing recently tend to have characteristics of convergence and thus consist of technologies from various fields. In other words, the new convergence technologies may not belong to the defined classification. Therefore, the existing method does not properly reflect the dynamic change of the convergence phenomenon. Secondly, in order to forecast the promising convergence technologies, most of the existing analysis method use the general purpose indicators in process. This method does not fully utilize the specificity of convergence phenomenon. The new convergence technology is highly dependent on the existing technology, which is the origin of that technology. Based on that, it can grow into the independent field or disappear rapidly, according to the change of the dependent technology. In the existing analysis, the potential growth of convergence technology is judged through the traditional indicators designed from the general purpose. However, these indicators do not reflect the principle of convergence. In other words, these indicators do not reflect the characteristics of convergence technology, which brings the meaning of new technologies emerge through two or more mature technologies and grown technologies affect the creation of another technology. Thirdly, previous studies do not provide objective methods for evaluating the accuracy of models in forecasting promising convergence technologies. In the studies of convergence technology, the subject of forecasting promising technologies was relatively insufficient due to the complexity of the field. Therefore, it is difficult to find a method to evaluate the accuracy of the model that forecasting promising convergence technologies. In order to activate the field of forecasting promising convergence technology, it is important to establish a method for objectively verifying and evaluating the accuracy of the model proposed by each study. To overcome these limitations, we propose a new method for analysis of convergence technologies. First of all, through topic modeling, we derive a new technology classification in terms of text content. It reflects the dynamic change of the actual technology market, not the existing fixed classification standard. In addition, we identify the influence relationships between technologies through the topic correspondence weights of each document, and structuralize them into a network. In addition, we devise a centrality indicator (PGC, potential growth centrality) to forecast the future growth of technology by utilizing the centrality information of each technology. It reflects the convergence characteristics of each technology, according to technology maturity and interdependence between technologies. Along with this, we propose a method to evaluate the accuracy of forecasting model by measuring the growth rate of promising technology. It is based on the variation of potential growth centrality by period. In this paper, we conduct experiments with 13,477 patent documents dealing with technical contents to evaluate the performance and practical applicability of the proposed method. As a result, it is confirmed that the forecast model based on a centrality indicator of the proposed method has a maximum forecast accuracy of about 2.88 times higher than the accuracy of the forecast model based on the currently used network indicators.