• Title/Summary/Keyword: statistical-relevance model

Search Result 33, Processing Time 0.023 seconds

Formulating Regional Relevance Index through Covariance Structure Modeling (공분산구조분석을 이용한 자체충족률 모형 검증)

  • 장혜정;김창엽
    • Health Policy and Management
    • /
    • v.11 no.2
    • /
    • pp.123-140
    • /
    • 2001
  • Hypotheses In health services research are becoming increasingly more complex and specific. As a result, health services research studies often include multiple independent, intervening, and dependent variables in a single hypothesis. Nevertheless, the statistical models adopted by health services researchers have failed to keep pace with the increasing complexity and specificity of hypotheses and research designs. This article introduces a statistical model well suited for complex and specific hypotheses tests in health services research studies. The covariance structure modeling(CSM) methodology is especially applied to regional relevance indices(RIs) to assess the impact of health resources and healthcare utilization. Data on secondary statistics and health insurance claims were collected by each catchment area. The model for RI was justified by direct and indirect effects of three latent variables measured by seven observed variables, using ten structural equations. The resulting structural model revealed significant direct effects of the structure of health resources but indirect effects of the quantity on RIs, and explained 82% of correlation matrix of measurement variables. Two variables, the number of beds and the portion of specialists among medical doctors, became to have significant effects on RIs by being analyzed using the CSM methodology, while they were insignificant in the regression model. Recommendations for the CSM methodology on health service research data are provided.

  • PDF

A Study on Semantic Based Indexing and Fuzzy Relevance Model (의미기반 인덱스 추출과 퍼지검색 모델에 관한 연구)

  • Kang, Bo-Yeong;Kim, Dae-Won;Gu, Sang-Ok;Lee, Sang-Jo
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.04b
    • /
    • pp.238-240
    • /
    • 2002
  • If there is an Information Retrieval system which comprehends the semantic content of documents and knows the preference of users. the system can search the information better on the Internet, or improve the IR performance. Therefore we propose the IR model which combines semantic based indexing and fuzzy relevance model. In addition to the statistical approach, we chose the semantic approach in indexing, lexical chains, because we assume it would improve the performance of the index term extraction. Furthermore, we combined the semantic based indexing with the fuzzy model, which finds out the exact relevance of the user preference and index terms. The proposed system works as follows: First, the presented system indexes documents by the efficient index term extraction method using lexical chains. And then, if a user tends to retrieve the information from the indexed document collection, the extended IR model calculates and ranks the relevance of user query. user preference and index terms by some metrics. When we experimented each module, semantic based indexing and extended fuzzy model. it gave noticeable results. The combination of these modules is expected to improve the information retrieval performance.

  • PDF

A Study on Applications of Regression Diagnostic Method to Technometrics, and the Statistical Quality Control

  • Kim, Soon-Kwi
    • Journal of Korean Society for Quality Management
    • /
    • v.21 no.1
    • /
    • pp.55-64
    • /
    • 1993
  • This article is concerned with procedures for detecting one or more outliers or influential observations in a linear regression model. A test procedure, based on recursive residuals is proposed and developed The power of the test procedure to identify one or more outliers is investigated through simulation, and its relevance to the number and configuration of the outlier.

  • PDF

Fault Prediction Using Statistical and Machine Learning Methods for Improving Software Quality

  • Malhotra, Ruchika;Jain, Ankita
    • Journal of Information Processing Systems
    • /
    • v.8 no.2
    • /
    • pp.241-262
    • /
    • 2012
  • An understanding of quality attributes is relevant for the software organization to deliver high software reliability. An empirical assessment of metrics to predict the quality attributes is essential in order to gain insight about the quality of software in the early phases of software development and to ensure corrective actions. In this paper, we predict a model to estimate fault proneness using Object Oriented CK metrics and QMOOD metrics. We apply one statistical method and six machine learning methods to predict the models. The proposed models are validated using dataset collected from Open Source software. The results are analyzed using Area Under the Curve (AUC) obtained from Receiver Operating Characteristics (ROC) analysis. The results show that the model predicted using the random forest and bagging methods outperformed all the other models. Hence, based on these results it is reasonable to claim that quality models have a significant relevance with Object Oriented metrics and that machine learning methods have a comparable performance with statistical methods.

A probabilistic information retrieval model by document ranking using term dependencies (용어간 종속성을 이용한 문서 순위 매기기에 의한 확률적 정보 검색)

  • You, Hyun-Jo;Lee, Jung-Jin
    • The Korean Journal of Applied Statistics
    • /
    • v.32 no.5
    • /
    • pp.763-782
    • /
    • 2019
  • This paper proposes a probabilistic document ranking model incorporating term dependencies. Document ranking is a fundamental information retrieval task. The task is to sort documents in a collection according to the relevance to the user query (Qin et al., Information Retrieval Journal, 13, 346-374, 2010). A probabilistic model is a model for computing the conditional probability of the relevance of each document given query. Most of the widely used models assume the term independence because it is challenging to compute the joint probabilities of multiple terms. Words in natural language texts are obviously highly correlated. In this paper, we assume a multinomial distribution model to calculate the relevance probability of a document by considering the dependency structure of words, and propose an information retrieval model to rank a document by estimating the probability with the maximum entropy method. The results of the ranking simulation experiment in various multinomial situations show better retrieval results than a model that assumes the independence of words. The results of document ranking experiments using real-world datasets LETOR OHSUMED also show better retrieval results.

Statistical Package fo Sample Size and Power Determination (표본의 수와 검정력 분석을 위한 통계팩키지)

  • Lee, Kwan-Jeh
    • Journal of Korean Society for Quality Management
    • /
    • v.28 no.2
    • /
    • pp.17-38
    • /
    • 2000
  • In application, sample size determination is one of the important problems in designing an experiment. A large amount of literature has been published on the problem of determining sample size and power for various statistical models. In practice, however, it is not easy to calculate sample size and/or power because the formula and other results derived from statistical model are scattered in various textbooks and journal articles. This paper describes some previously published theories that have practical relevance for sample size and power determination in various statistical problems, including life-testing problems with censored cases and introduces a statistical package which calculates sample size and power according to the results described. The screens and numerical results made by the package are demonstrated.

  • PDF

Prediction of uplift capacity of suction caisson in clay using extreme learning machine

  • Muduli, Pradyut Kumar;Das, Sarat Kumar;Samui, Pijush;Sahoo, Rupashree
    • Ocean Systems Engineering
    • /
    • v.5 no.1
    • /
    • pp.41-54
    • /
    • 2015
  • This study presents the development of predictive models for uplift capacity of suction caisson in clay using an artificial intelligence technique, extreme learning machine (ELM). Other artificial intelligence models like artificial neural network (ANN), support vector machine (SVM), relevance vector machine (RVM) models are also developed to compare the ELM model with above models and available numerical models in terms of different statistical criteria. A ranking system is presented to evaluate present models in identifying the 'best' model. Sensitivity analyses are made to identify important inputs contributing to the developed models.

Applied linear and nonlinear statistical models for evaluating strength of Geopolymer concrete

  • Prem, Prabhat Ranjan;Thirumalaiselvi, A.;Verma, Mohit
    • Computers and Concrete
    • /
    • v.24 no.1
    • /
    • pp.7-17
    • /
    • 2019
  • The complex phenomenon of the bond formation in geopolymer is not well understood and therefore, difficult to model. This paper present applied statistical models for evaluating the compressive strength of geopolymer. The applied statistical models studied are divided into three different categories - linear regression [least absolute shrinkage and selection operator (LASSO) and elastic net], tree regression [decision and bagging tree] and kernel methods (support vector regression (SVR), kernel ridge regression (KRR), Gaussian process regression (GPR), relevance vector machine (RVM)]. The performance of the methods is compared in terms of error indices, computational effort, convergence and residuals. Based on the present study, kernel based methods (GPR and KRR) are recommended for evaluating compressive strength of Geopolymer concrete.

Industry Stock Returns Prediction Using Neural Networks (신경망을 이용한 산업주가수익율의 예측)

  • Kwon, Young-Sam;Han, In-Goo
    • Asia pacific journal of information systems
    • /
    • v.9 no.3
    • /
    • pp.93-110
    • /
    • 1999
  • The previous studies regarding the stock returns have advocated that industry effects exist over entire industry. As the industry categories are more rigid, the demand for predicting the industry sectors is rapidly increasing. The advances in Artificial Intelligence and Neural Networks suggest the feasibility of a valuable computational model for stock returns prediction. We propose a sector-factor model for predicting the return on industry stock index using neural networks. As a substitute for the traditional models, neural network model may be more accurate and effective alternative when the dynamics between the underlying industry features are not well known or when the industry specific asset pricing equation cannot be solved analytically. To assess the potential value of neural network model, we simulate the resulting network and show that the proposed model can be used successfully for banks and general construction industry. For comparison, we estimate models using traditional statistical method of multiple regression. To illustrate the practical relevance of neural network model, we apply it to the predictions of two industry stock indexes from 1980 to 1995.

  • PDF

Scientific Explanations by Earth Science Teachers in Secondary Schools: Analyses of the Logical Forms and Discursive Features (중등학교 지구과학 교사들의 과학적 설명: 논리적 형식과 담화적 특징 분석)

  • Oh, Phil-Seok
    • Journal of The Korean Association For Science Education
    • /
    • v.27 no.1
    • /
    • pp.37-49
    • /
    • 2007
  • The purpose of this study was to classify the logical forms of scientific explanations provided by teachers in secondary earth science classrooms, to examine the characteristics of the scientific explanations in different forms, and to identify the roles of the teacher and students in discursive practices for scientific explanations. Data came from the earth science teachers who participated in overseas teacher in-service programs in the years 2003 and 2004. A total of 18 video-taped lessons and their verbatim transcriptions were analyzed. The result showed that deductive-nomological explanations occurred most frequently in earth science classrooms and that the deductive-nomological model was well-suited to those problems for which there existed firmly established scientific laws or principles to construct scientific explanations. However, abductive explanations were presented when the classes dealt with retrodictive tasks of earth science. The statistical-probabilistic and statistical-relevance models were also employed in explaining weather proverbs and unusual changes of weather, respectively. Most of the scientific explanations were completed through the teachers' monologic utterances, and students assumed passive roles in discursive practices for developing scientific explanations. Implications for science lessons and science education research were discussed.