• Title/Summary/Keyword: Scale complexity

Search Result 483, Processing Time 0.024 seconds

The Use of VFG for Measuring the Slice Complexity (슬라이스 복잡도 측정을 위한 VFG의 사용)

  • 문유미;최완규;이성주
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.5 no.1
    • /
    • pp.183-191
    • /
    • 2001
  • We develop the new data s]ice representation, called the data value flow graph(VFG), for modeling the information flow oil data slices. Then we define a slice complexity measure by using the existing flow complexity measure in order to measure the complexity of the information flow on VFG. We show relation of the slice complexity on a slice with the slice complexity on a procedure. We also demonstrate the measurement scale factors through a set of atomic modifications and a concatenation operator on VFG.

  • PDF

An XPDL-Based Workflow Control-Structure and Data-Sequence Analyzer

  • Kim, Kwanghoon Pio
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.3
    • /
    • pp.1702-1721
    • /
    • 2019
  • A workflow process (or business process) management system helps to define, execute, monitor and manage workflow models deployed on a workflow-supported enterprise, and the system is compartmentalized into a modeling subsystem and an enacting subsystem, in general. The modeling subsystem's functionality is to discover and analyze workflow models via a theoretical modeling methodology like ICN, to graphically define them via a graphical representation notation like BPMN, and to systematically deploy those graphically defined models onto the enacting subsystem by transforming into their textual models represented by a standardized workflow process definition language like XPDL. Before deploying those defined workflow models, it is very important to inspect its syntactical correctness as well as its structural properness to minimize the loss of effectiveness and the depreciation of efficiency in managing the corresponding workflow models. In this paper, we are particularly interested in verifying very large-scale and massively parallel workflow models, and so we need a sophisticated analyzer to automatically analyze those specialized and complex styles of workflow models. One of the sophisticated analyzers devised in this paper is able to analyze not only the structural complexity but also the data-sequence complexity, especially. The structural complexity is based upon combinational usages of those control-structure constructs such as subprocesses, exclusive-OR, parallel-AND and iterative-LOOP primitives with preserving matched pairing and proper nesting properties, whereas the data-sequence complexity is based upon combinational usages of those relevant data repositories such as data definition sequences and data use sequences. Through the devised and implemented analyzer in this paper, we are able eventually to achieve the systematic verifications of the syntactical correctness as well as the effective validation of the structural properness on those complicate and large-scale styles of workflow models. As an experimental study, we apply the implemented analyzer to an exemplary large-scale and massively parallel workflow process model, the Large Bank Transaction Workflow Process Model, and show the structural complexity analysis results via a series of operational screens captured from the implemented analyzer.

The Relationship between Syntactic Complexity Indices and Scores on Language Use in the Analytic Rating Scale (통사적 복잡성과 분석적 척도의 언어 사용 점수간의 관계 탐색)

  • Young-Ju Lee
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.5
    • /
    • pp.229-235
    • /
    • 2023
  • This study investigates the relationship between syntactic complexity indices and scores on language use in Jacobs et al.(1981)' analytic rating scale. Syntactic complexity indices obtained from TAASSC program and 440 essays written by EFL students from the ICNALE corpus were analyzed. Specifically, this study explores the relationship between scores on language use and Lu(2011)'s traditional syntactic complexity indices, phrasal complexity indices, and clausal complexity indices, respectively. Results of the stepwise regression analysis showed that phrasal complexity indices turned out to be the best predictor of scores on language use, although the variance in scores on language use was relatively small, compared with the previous study. Implications of the findings of the current study for writing instruction (i.e., syntactic structures at the phrase level) were also discussed.

Fast Detection of Distributed Global Scale Network Attack Symptoms and Patterns in High-speed Backbone Networks

  • Kim, Sun-Ho;Roh, Byeong-Hee
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.2 no.3
    • /
    • pp.135-149
    • /
    • 2008
  • Traditional attack detection schemes based on packets or flows have very high computational complexity. And, network based anomaly detection schemes can reduce the complexity, but they have a limitation to figure out the pattern of the distributed global scale network attack. In this paper, we propose an efficient and fast method for detecting distributed global-scale network attack symptoms in high-speed backbone networks. The proposed method is implemented at the aggregate traffic level. So, our proposed scheme has much lower computational complexity, and is implemented in very high-speed backbone networks. In addition, the proposed method can detect attack patterns, such as attacks in which the target is a certain host or the backbone infrastructure itself, via collaboration of edge routers on the backbone network. The effectiveness of the proposed method are demonstrated via simulation.

Large-Scale Joint Rate and Power Allocation Algorithm Combined with Admission Control in Cognitive Radio Networks

  • Shin, Woo-Jin;Park, Kyoung-Youp;Kim, Dong-In;Kwon, Jang-Woo
    • Journal of Communications and Networks
    • /
    • v.11 no.2
    • /
    • pp.157-165
    • /
    • 2009
  • In this paper, we investigate a dynamic spectrum sharing problem for the centralized uplink cognitive radio networks using orthogonal frequency division multiple access. We formulate a large-scale joint rate and power allocation as an optimization problem under quality of service constraint for secondary users and interference constraint for primary users. We also suggest admission control to nd a feasible solution to the optimization problem. To implement the resource allocation on a large-scale, we introduce a notion of using the conservative factors $\alpha$ and $\beta$ depending on the outage and violation probabilities. Since estimating instantaneous channel gains is costly and requires high complexity, the proposed algorithm pursues a practical and implementation-friendly resource allocation. Simulation results demonstrate that the large-scale joint rate and power allocation incurs a slight loss in system throughput over the instantaneous one, but it achieves lower complexity with less sensitivity to variations in shadowing statistics.

Low-Complexity Massive MIMO Detectors Based on Richardson Method

  • Kang, Byunggi;Yoon, Ji-Hwan;Park, Jongsun
    • ETRI Journal
    • /
    • v.39 no.3
    • /
    • pp.326-335
    • /
    • 2017
  • In the uplink transmission of massive (or large-scale) multi-input multi-output (MIMO) systems, large dimensional signal detection and its hardware design are challenging issues owing to the high computational complexity. In this paper, we propose low-complexity hardware architectures of Richardson iterative method-based massive MIMO detectors. We present two types of massive MIMO detectors, directly mapped (type1) and reformulated (type2) Richardson iterative methods. In the proposed Richardson method (type2), the matrix-by-matrix multiplications are reformulated to matrix-vector multiplications, thus reducing the computational complexity from $O(U^2)$ to O(U). Both massive MIMO detectors are implemented using a 65 nm CMOS process and compared in terms of detection performance under different channel conditions (high-mobility and flat fading channels). The hardware implementation results confirm that the proposed type1 Richardson method-based detector demonstrates up to 50% power savings over the proposed type2 detector under a flat fading channel. The type2 detector indicates a 37% power savings compared to the type1 under a high-mobility channel.

A Novel Adaptive Turbo Receiver for Large-Scale MIMO Communications

  • Chang, Yu-Kuan;Ueng, Fang-Biau;Tsai, Bo-Yi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.7
    • /
    • pp.2998-3017
    • /
    • 2018
  • Massive (large-scale) MIMO (multiple-input multiple-output) is one of the key technologies in next-generation wireless communication systems. This paper proposes a high-performance low-complexity turbo receiver for SC-FDMA (single-carrier frequency-division multiple access) based MMIMO (massive MIMO) systems. Because SC-FDMA technology has the desirable characteristics of OFDMA (orthogonal frequency division multiple access) and the low PAPR (peak-to-average power ratio) of SC transmission schemes, the 3GPP LTE (long-term evolution) has adopted it as the uplink transmission to meet the demand high data rate and low error rate performance. The complexity of computing will be increased greatly in base station with massive MIMO (MMIMO) system. In this paper, a low-complexity adaptive turbo equalization receiver based on normalized minimal symbol-error-rate for MMIMO SC-FDMA system is proposed. The proposed receiver is with low complexity than that of the conventional turbo MMSE (minimum mean square error) equalizer and is also with better bit error rate (BER) performance than that of the conventional adaptive turbo MMSE equalizer. Simulation results confirm the effectiveness of the proposed scheme.

Fast Time-Scale Modification of Speech Using Nonlinear Clipping Methods

  • Jung, Ho-Young;Kim, Hyung-Soon;Lee, Sung-Joo
    • MALSORI
    • /
    • no.59
    • /
    • pp.69-87
    • /
    • 2006
  • Among the conventional time-scale modification (TSM) methods, the synchronized overlap and add (SOLA) method is widely used due to its good performance relative to computational complexity But the SOLA method remains complex due to its synchronization procedure using the normalized cross-correlation function. In this paper, we introduce a computationally efficient SOLA method utilizing 3 level center clipping method, as well as zero-crossing and level-crossing information. The result of subjective preference test indicates that the proposed method can reduce the computational complexity by over 80% compared with the conventional SOLA method without serious degradation of synthesized speech quality.

  • PDF

A Study on the Type and the Facilities in Compositeness of the Domestic Discount Store (국내 대형할인점의 복합화에 따른 유형과 시설에 관한 연구)

  • 문선욱;양정필
    • Korean Institute of Interior Design Journal
    • /
    • no.41
    • /
    • pp.137-145
    • /
    • 2003
  • This research analyzed the space scheme in connection with complexity, one of the new changes in the discount stores, and has a goal of predicting the direction of space scheme in the upcoming complexity era. The research was conducted in the following way. Firstly, this researcher tried to grasp what kinds of changes were required in the overall distribution industry socially and economically. Secondly, the characteristic and situation of discount stores were scrutinized. Thirdly, the domestic stores' complexity status was classified and types of those were elicited. Fourthly, the time-series change and use were analyzed. The result of this analysis reveals that the types of complexity can be divided by location and adjustment to environmental changes. The time-series analysis shows that total operating area, the number of parked cars and the tenant ratio have increased dramatically in 2000 and 2003. And, according to the correlation analysis between factors, the tenant ratio has, a strong correlation with other two factors. Self-complexity takes the basic form of living facilities and complexity with other facilities is combined with other cultural, sales, educational and administrative ones. Mass-complexity is merged with the stadiums, parks or station sites. As you've seen, the concept of complex shopping mall for the realization of one stop shopping and convenience will continue in the days to come. It is desirable that the study on the large-scale shopping spaces will be conducted continually for the preparedness of future life style.

Near-Optimal Low-Complexity Hybrid Precoding for THz Massive MIMO Systems

  • Yuke Sun;Aihua Zhang;Hao Yang;Di Tian;Haowen Xia
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.1042-1058
    • /
    • 2024
  • Terahertz (THz) communication is becoming a key technology for future 6G wireless networks because of its ultra-wide band. However, the implementation of THz communication systems confronts formidable challenges, notably beam splitting effects and high computational complexity associated with them. Our primary objective is to design a hybrid precoder that minimizes the Euclidean distance from the fully digital precoder. The analog precoding part adopts the delay-phase alternating minimization (DP-AltMin) algorithm, which divides the analog precoder into phase shifters and time delayers. This effectively addresses the beam splitting effects within THz communication by incorporating time delays. The traditional digital precoding solution, however, needs matrix inversion in THz massive multiple-input multiple-output (MIMO) communication systems, resulting in significant computational complexity and complicating the design of the analog precoder. To address this issue, we exploit the characteristics of THz massive MIMO communication systems and construct the digital precoder as a product of scale factors and semi-unitary matrices. We utilize Schatten norm and Hölder's inequality to create semi-unitary matrices after initializing the scale factors depending on the power allocation. Finally, the analog precoder and digital precoder are alternately optimized to obtain the ultimate hybrid precoding scheme. Extensive numerical simulations have demonstrated that our proposed algorithm outperforms existing methods in mitigating the beam splitting issue, improving system performance, and exhibiting lower complexity. Furthermore, our approach exhibits a more favorable alignment with practical application requirements, underlying its practicality and efficiency.