• Title/Summary/Keyword: Complexity System

Search Result 3,526, Processing Time 0.031 seconds

A Study on the Propriety of the Medical Insurance Fee Schedule of Surgical Operations - In Regard to the Relative Price System and the Classification of the Price Unit of Insurance Fee Schedule - (수술수가의 적정성에 관한 연구 - 상대가격체계와 항목분류를 중심으로 -)

  • Oh Jin Joo
    • Journal of Korean Public Health Nursing
    • /
    • v.2 no.2
    • /
    • pp.21-44
    • /
    • 1988
  • In Korea, fee-for service reimbursement has been adopted from the begining of medical insurance system in 1977, and the importance of the relative value unit is currently being investigated. The purpose of this study was to find out the level of propriety of the difference in the fees for different surgical services, and the appropriateness of the classification of the insurance fee schedule. For the purpose of this study, specific subjects and the procedural methodology is shown as follows: 1. The propriety of the Relative Price System(RPS). 1) Choice of sample operations. In this study, sample operations were selected and classified by specialists in general surgery, and the number of items they classified were 32. For the same group of operations the Insurance Fee Schedule(IFS) classified the operations into 24 separate items. In order to investigate the propriety of the RPS, one of the purpose of this study, was to examine the 24 items classified by the IFS. 2) Evaluation of the complexity of surgery. The data used in this study was collected The data used in this study was collected from 94 specialists in general surgery by mail survey from November I to 15, 1986. Several independent variables (age, location, number of bed, university hospital, whether the medical institution adopt residents or not) were also investigated for analysis of the characteristics of surgical complexity. 3) Complexity and time calculations. Time data was collected from the records of the Seoul National University' Hospital, and the cost per operation was calculated through cost finding methods. 4) Analysis of the propriety of the Relative Price System of the Insurance Fee Schedule. The Relative Price System of the sample operation was regressed on the cost, time, comlexity relative ,value system (RVS) separately. The coefficient of determination indicates the degree of variation in the RPS of the Insurance Fee Schedule explained by the cost, time, complexity RVS separately. 2. The appropriateness of the classification of the Insurance Fee Schedule. 1) Choice of sample operations. The items which differed between the classification of the specialist and the classification of medical, Insurance Fee Schedule were chosen. 2) Comparisons of cost, time and complexity between the items were done to evaluate which classification was more appropriate. The findings of the study can be summarized as follows: 1. The coefficient of determination of the regression of the RPS on-cost RVS was 0.58, on time RVS was 0.65, and on complexity RVS was 0.72. This means that the RPS of Insurance Fee Schedule is improper with respect to the cost, time, complexity separately. Thus this indicates that RPS must be re-shaped according to the standard element. In this study, the correlation coefficients of cost, time, complexity Relative Value System were very high, and this suggests that RPS could be reshaped I according to anyone standard element. Considering of measurement, time was thought to be the most I appropriate. 2. The classifications of specialist and of the Insurance Fee Schedule were compared with respect to cost, time, and complexity separately. For complexity, ANOVA was done and the others were compared to the different values of different classifications. The result was that the classification of specialist was more reasonable and that the classification of Insurance Fee Schedule grouped inappropriately several into one price unit.

  • PDF

Computational Complexity Analysis of Cascade AOA Estimation Algorithm Based on FMCCA Antenna

  • Kim, Tae-yun;Hwang, Suk-seung
    • Journal of Positioning, Navigation, and Timing
    • /
    • v.11 no.2
    • /
    • pp.91-98
    • /
    • 2022
  • In the next generation wireless communication system, the beamforming technique based on a massive antenna is one of core technologies for transmitting and receiving huge amounts of data, efficiently and accurately. For highly performed and highly reliable beamforming, it is required to accurately estimate the Angle of Arrival (AOA) for the desired signal incident to an antenna. Employing the massive antenna with a large number of elements, although the accuracy of the AOA estimation is enhanced, its computational complexity is dramatically increased so much that real-time communication is difficult. In order to improve this problem, AOA estimation algorithms based on the massive antenna with the low computational complexity have been actively studied. In this paper, we compute and analyze the computational complexity of the cascade AOA estimation algorithm based on the Flexible Massive Concentric Circular Array (FMCCA). In addition, its computational complexity is compared to conventional AOA estimation techniques such as the Multiple Signal Classification (MUSIC) algorithm with the high resolution and the Only Beamspace MUSIC (OBM) algorithm.

A QUALITATIVE METHOD TO ESTIMATE HSI DISPLAY COMPLEXITY

  • Hugo, Jacques;Gertman, David
    • Nuclear Engineering and Technology
    • /
    • v.45 no.2
    • /
    • pp.141-150
    • /
    • 2013
  • There is mounting evidence that complex computer system displays in control rooms contribute to cognitive complexity and, thus, to the probability of human error. Research shows that reaction time increases and response accuracy decreases as the number of elements in the display screen increase. However, in terms of supporting the control room operator, approaches focusing on addressing display complexity solely in terms of information density and its location and patterning, will fall short of delivering a properly designed interface. This paper argues that information complexity and semantic complexity are mandatory components when considering display complexity and that the addition of these concepts assists in understanding and resolving differences between designers and the preferences and performance of operators. This paper concludes that a number of simplified methods, when combined, can be used to estimate the impact that a particular display may have on the operator's ability to perform a function accurately and effectively. We present a mixed qualitative and quantitative approach and a method for complexity estimation.

A Cumulative Incremental Effort Based Software Growth Model Considering System Size and Complexity (시스템 크기와 복잡도를 고려한 누적 노력 기반의 소프트웨어 성장 모델)

  • Park, Jung-Yang;Kim, Seong-Hui;Park, Jae-Heung
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.1
    • /
    • pp.90-95
    • /
    • 1999
  • A software growth model, a mathematical model describing the growth behavior of a software system during the evolution process, enables us to predict the future system size and incremental erfort required to meet the planned system size. This paper first introduces a software growth model defined with respect to the cumulative incremental effort. It was assumed that the incremental growth of a software system is proportional to the incremental effort and function of the system size is suggested as a system complexity and then applied to real data for its validation. such a system complexity additionally provides us with a measure for complexity comparison. since the measure is independent of the system size, it is useful for comparing complexities of software systems of different size.

  • PDF

TOPOLOGICAL COMPLEXITY OF SEMIGROUP ACTIONS

  • Yan, Xinhua;He, Lianfa
    • Journal of the Korean Mathematical Society
    • /
    • v.45 no.1
    • /
    • pp.221-228
    • /
    • 2008
  • In this paper, we study the complexity of semigroup actions using complexity functions of open covers. The main results are as follows: (1) A dynamical system is equicontinuous if and only if any open cover has bounded complexity; (2) Weak-mixing implies scattering; (3) We get a criterion for the scattering property.

Complexity Reduced Blind Subspace Channel Estimation for DS/CDMA DMB Downlink (DS/CDMA DMB 하향 링크에서 복잡도가 감소된 블라인드 부분 공간 채널 추정)

  • Yang Wan-Chul;Lee Byung-Seub
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.15 no.9
    • /
    • pp.863-871
    • /
    • 2004
  • In this paper, we propose a subspace channel estimation technique for DS/CDMA DMB down link system, which can obtain reduction in numerical complexity by using of matched filtering outputs. The complexity reduction is considerable when the channel length is small and the system is moderately loaded. Previously proposed subspace-based blind channel estimation algorithm suffer from high numerical complexity for systems with large spreading gains. Although the proposed algerian suffers a slight performance loss, it becomes negligible for large observation length. Performance is evaluated through simulations and the derivation of the analytical MSE.

The Complexity of the Static Structures of Object-Oriented Systems by Analyzing the Class Diagram of UML (UML 클래스 다이어그램의 분석에 의한 객체지향 시스템의 정적 구조 복잡도 연구)

  • Chung, Hong;Hong, Dong-Kwon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.6
    • /
    • pp.793-799
    • /
    • 2004
  • Many researches and validations for the complexity metrics of the object-oriented systems have been studied. Most of them are aimed for the measurement of the partial aspects of the systems, for example, the coupling between objects, the complexity of inheritance structures, the cohesion of methods, and so on. But the software practitioners want to measure the complexity of overall system, not partial. We studied the complexity of the overall structures of object-oriented systems by analyzing the class diagram of UML. The class diagram is composed of classes and their relations. There are three kinds of relations, association, generalization, and aggregation, which are making the structure of object-oriented systems to be difficult to understand. We proposed a heuristic metric to measure the complexity of object-oriented systems by putting together the three kinds of the relations. This metric will be helpful to the software developers for their designing tasks by evaluating the complexity of the structures of object-oriented system and redesigning tasks of the system.

A Parallel Collaborative Sphere Decoder for a MIMO Communication System

  • Koo, Jihun;Kim, Soo-Yong;Kim, Jaeseok
    • Journal of Communications and Networks
    • /
    • v.16 no.6
    • /
    • pp.620-626
    • /
    • 2014
  • In this paper, we propose a parallel collaborative sphere decoder with a scalable architecture promising quasi-maximum likelyhood performance with a relatively small amount of computational resources. This design offers a hardware-friendly algorithm using a modified node operation through fixing the variable complexity of the critical path caused by the sequential nature of the conventional sphere decoder (SD). It also reduces the computational complexity compared to the fixed-complexity sphere decoder (FSD) algorithm by tree pruning using collaboratively operated node operators. A Monte Carlo simulation shows that our proposed design can be implemented using only half the parallel operators compared to the approach using an ideal fully parallel scheme such as FSD, with only about a 7% increase of the normalized decoding time for MIMO dimensions of $16{\times}16$ with 16-QAM modulation.

Energy-efficiency Optimization Schemes Based on SWIPT in Distributed Antenna Systems

  • Xu, Weiye;Chu, Junya;Yu, Xiangbin;Zhou, Huiyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.2
    • /
    • pp.673-694
    • /
    • 2021
  • In this paper, we intend to study the energy efficiency (EE) optimization for a simultaneous wireless information and power transfer (SWIPT)-based distributed antenna system (DAS). Firstly, a DAS-SWIPT model is formulated, whose goal is to maximize the EE of the system. Next, we propose an optimal resource allocation method by means of the Karush-Kuhn-Tucker condition as well as an ergodic method. Considering the complexity of the ergodic method, a suboptimal scheme with lower complexity is proposed by using an antenna selection scheme. Numerical results illustrate that our suboptimal method is able to achieve satisfactory performance of EE similar to an optimal one while reducing the calculation complexity.

Case study of the large switching software metrics and their fault analysis (대형 교환 소프트웨어의 복잡성과 고장분석 사례 연구)

  • 이재기;남상식;김창봉;이규대
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.10C
    • /
    • pp.887-901
    • /
    • 2002
  • Software management model divided into the software project model and design estimation model, software matrices model, reliability growth model, process improvement model(or process maturity model) etc. Among these software management models, software complexity model make an estimated of the product software. For a practice of software managed, need to guideline of the static analysis of software. Especially, Software complexity model introduced for the estimation of software quantity and program complexity. In case of measurement the software matrices, its need for us to analysis of software quality and products. On the other hand, we known that complexity program include many defects and consuming of source cost. So, we apply to complexity model using of the program complexity, control structure and volume matrices, interface metrics, process complexity metrics method. In this paper, we represent that the analysis of fault data detected during the system test. Also, we analysis of program control structure and interface, volume matrices in various aspect of switching software. Others, their results utilized similar of project and system development.