• Title/Summary/Keyword: Vector Fit

Search Result 65, Processing Time 0.023 seconds

An approximate fitting for mixture of multivariate skew normal distribution via EM algorithm (EM 알고리즘에 의한 다변량 치우친 정규분포 혼합모형의 근사적 적합)

  • Kim, Seung-Gu
    • The Korean Journal of Applied Statistics
    • /
    • v.29 no.3
    • /
    • pp.513-523
    • /
    • 2016
  • Fitting a mixture of multivariate skew normal distribution (MSNMix) with multiple skewness parameter vectors via EM algorithm often requires a highly expensive computational cost to calculate the moments and probabilities of multivariate truncated normal distribution in E-step. Subsequently, it is common to fit an asymmetric data set with MSNMix with a simple skewness parameter vector since it allows us to compute them in E-step in an univariate manner that guarantees a cheap computational cost. However, the adaptation of a simple skewness parameter is unrealistic in many situations. This paper proposes an approximate estimation for the MSNMix with multiple skewness parameter vectors that also allows us to treat them in an univariate manner. We additionally provide some experiments to show its effectiveness.

Geometrical description based on forward selection & backward elimination methods for regression models (다중회귀모형에서 전진선택과 후진제거의 기하학적 표현)

  • Hong, Chong-Sun;Kim, Moung-Jin
    • Journal of the Korean Data and Information Science Society
    • /
    • v.21 no.5
    • /
    • pp.901-908
    • /
    • 2010
  • A geometrical description method is proposed to represent the process of the forward selection and backward elimination methods among many variable selection methods for multiple regression models. This graphical method shows the process of the forward selection and backward elimination on the first and second quadrants, respectively, of half circle with a unit radius. At each step, the SSR is represented by the norm of vector and the extra SSR or partial determinant coefficient is represented by the angle between two vectors. Some lines are dotted when the partial F test results are statistically significant, so that statistical analysis could be explored. This geometrical description can be obtained the final regression models based on the forward selection and backward elimination methods. And the goodness-of-fit for the model could be explored.

Face Recognition System Based on the Embedded LINUX (임베디드 리눅스 기반의 눈 영역 비교법을 이용한 얼굴인식)

  • Bae, Eun-Dae;Kim, Seok-Min;Nam, Boo-Hee
    • Proceedings of the KIEE Conference
    • /
    • 2006.04a
    • /
    • pp.120-121
    • /
    • 2006
  • In this paper, We have designed a face recognition system based on the embedded Linux. This paper has an aim in embedded system to recognize the face more exactly. At first, the contrast of the face image is adjusted with lightening compensation method, the skin and lip color is founded based on YCbCr values from the compensated image. To take advantage of the method based on feature and appearance, these methods are applied to the eyes which has the most highly recognition rate of all the part of the human face. For eyes detecting, which is the most important component of the face recognition, we calculate the horizontal gradient of the face image and the maximum value. This part of the face is resized for fitting the eye image. The image, which is resized for fit to the eye image stored to be compared, is extracted to be the feature vectors using the continuous wavelet transform and these vectors are decided to be whether the same person or not with PNN, to miminize the error rate, the accuracy is analyzed due to the rotation or movement of the face. Also last part of this paper we represent many cases to prove the algorithm contains the feature vector extraction and accuracy of the comparison method.

  • PDF

Forecasting for a Credit Loan from Households in South Korea

  • Jeong, Dong-Bin
    • The Journal of Industrial Distribution & Business
    • /
    • v.8 no.4
    • /
    • pp.15-21
    • /
    • 2017
  • Purpose - In this work, we examined the causal relationship between credit loans from households (CLH), loan collateralized with housing (LCH) and an interest of certificate of deposit (ICD) among others in South Korea. Furthermore, the optimal forecasts on the underlying model will be obtained and have the potential for applications in the economic field. Research design, data, and methodology - A total of 31 realizations sampled from the 4th quarter in 2008 to the 4th quarter in 2016 was chosen for this research. To achieve the purpose of this study, a regression model with correlated errors was exploited. Furthermore, goodness-of-fit measures was used as tools of optimal model-construction. Results - We found that by applying the regression model with errors component ARMA(1,5) to CLH, the steep and lasting rise can be expected over the next year, with moderate increase of LCH and ICD. Conclusions - Based on 2017-2018 forecasts for CLH, the precipitous and lasting increase can be expected over the next two years, with gradual rise of two major explanatory variables. By affording the assumption that the feedback among variables can exist, we can, in the future, consider more generalized models such as vector autoregressive model and structural equation model, to name a few.

Formal Representation and Query for Digital Contents Data

  • Khamis, Khamis Abdul-Latif;Song, Huazhu;Zhong, Xian
    • Journal of Information Processing Systems
    • /
    • v.16 no.2
    • /
    • pp.261-276
    • /
    • 2020
  • Digital contents services are one of the topics that have been intensively studied in the media industry, where various semantic and ontology techniques are applied. However, query execution for ontology data is still inefficient, lack of sufficient extensible definitions for node relationships, and there is no specific semantic method fit for media data representation. In order to make the machine understand digital contents (DCs) data well, we analyze DCs data, including static data and dynamic data, and use ontology to specify and classify objects and the events of the particular objects. Then the formal representation method is proposed which not only redefines DCs data based on the technology of OWL/RDF, but is also combined with media segmentation methods. At the same time, to speed up the access mechanism of DCs data stored under the persistent database, an ontology-based DCs query solution is proposed, which uses the specified distance vector associated to a surveillance of semantic label (annotation) to detect and track a moving or static object.

Linear Algebra Class Model using Technology(Matlab) - LINEAR SUBSPACES OF $R^n$ - (시각화를 이용한 선형대수학 교수학습모델 - $R^n$의 부분공간 -)

  • Kim, Duk-Sun;Lee, Sang-Gu;Jung, Kyung-Hoon
    • Communications of Mathematical Education
    • /
    • v.21 no.4
    • /
    • pp.621-646
    • /
    • 2007
  • In our new learning environment, we were asked to change our teaching method in our Linear Algebra class. In mathematics class, we could use several math-softwares such as MATHEMATICA, MATLAB, MAPLE, Drive etc.. MATLAB was quite well fit with our Linear Algebra class. In this paper we introduce an efficient way of delivery on important concepts in linear algebra by using well-known MATLAB/ATLAST M-files which we downloded from http://www.umassd.edu/specialprograms/atlast/.

  • PDF

Wind Attribute Time Series Modeling & Forecasting in IRAN

  • Ghorbani, Fahimeh;Raissi, Sadigh;Rafei, Meysam
    • East Asian Journal of Business Economics (EAJBE)
    • /
    • v.3 no.3
    • /
    • pp.14-26
    • /
    • 2015
  • A wind speed forecast is a crucial and sophisticated task in a wind farm for planning turbines and corresponds to an estimate of the expected production of one or more wind turbines in the near future. By production is often meant available power for wind farm considered (with units KW or MW depending on both the wind speed and direction. Such forecasts can also be expressed in terms of energy, by integrating power production over each time interval. In this study, we technically focused on mathematical modeling of wind speed and direction forecast based on locally data set gathered from Aghdasiyeh station in Tehran. The methodology is set on using most common techniques derived from literature review. Hence we applied the most sophisticated forecasting methods to embed seasonality, trend, and irregular pattern for wind speed as an angular variables. Through this research, we carried out the most common techniques such as the Box and Jenkins family, VARMA, the component method, the Weibull function and the Fourier series. Finally, the best fit for each forecasting method validated statistically based on white noise properties and the final comparisons using residual standard errors and mean absolute deviation from real data.

Background Surface Estimation for Reverse Engineering of Reliefs

  • Liu, Shenglan;Martin, Ralph R.;Langbein, Frank C.;Rosin, Paul L.
    • International Journal of CAD/CAM
    • /
    • v.7 no.1
    • /
    • pp.31-40
    • /
    • 2007
  • Reverse engineering of reliefs aims to turn an existing relief superimposed on an underlying surface into a geometric model which may be applied to a different base surface. Steps in this process include segmenting the relief from the background, and describing it as an offset height field relative to the underlying surface. We have previously considered relief segmentation using a geometric snake. Here, we show how to use this initial segmentation to estimate the background surface lying under the relief, which can be used (i) to refine the segmentation and (ii) to express the relief as an offset field. Our approach fits a B-spline surface patch to the measured background data surrounding the relief, while tension terms ensure this background surface smoothly continues underneath the relief where there are no measured background data points to fit. After making an initial estimate of relief offset height everywhere within the patch, we use a support vector machine to refine the segmentation. Tests demonstrate that this approach can accurately model the background surface where it underlies the relief, providing more accurate segmentation, as well as relief height field estimation. In particular, this approach provides significant improvements for relief concavities with narrow mouths and can segment reliefs with small internal holes.

Predicting idiopathic pulmonary fibrosis (IPF) disease in patients using machine approaches

  • Ali, Sikandar;Hussain, Ali;Kim, Hee-Cheol
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.144-146
    • /
    • 2021
  • Idiopathic pulmonary fibrosis (IPF) is one of the most dreadful lung diseases which effects the performance of the lung unpredictably. There is no any authentic natural history discovered yet pertaining to this disease and it has been very difficult for the physicians to diagnosis this disease. With the advent of Artificial intelligent and its related technologies this task has become a little bit easier. The aim of this paper is to develop and to explore the machine learning models for the prediction and diagnosis of this mysterious disease. For our study, we got IPF dataset from Haeundae Paik hospital consisting of 2425 patients. This dataset consists of 502 features. We applied different data preprocessing techniques for data cleaning while making the data fit for the machine learning implementation. After the preprocessing of the data, 18 features were selected for the experiment. In our experiment, we used different machine learning classifiers i.e., Multilayer perceptron (MLP), Support vector machine (SVM), and Random forest (RF). we compared the performance of each classifier. The experimental results showed that MLP outperformed all other compared models with 91.24% accuracy.

  • PDF

Identification of Pb-Zn ore under the condition of low count rate detection of slim hole based on PGNAA technology

  • Haolong Huang;Pingkun Cai;Wenbao Jia;Yan Zhang
    • Nuclear Engineering and Technology
    • /
    • v.55 no.5
    • /
    • pp.1708-1717
    • /
    • 2023
  • The grade analysis of lead-zinc ore is the basis for the optimal development and utilization of deposits. In this study, a method combining Prompt Gamma Neutron Activation Analysis (PGNAA) technology and machine learning is proposed for lead-zinc mine borehole logging, which can identify lead-zinc ores of different grades and gangue in the formation, providing real-time grade information qualitatively and semi-quantitatively. Firstly, Monte Carlo simulation is used to obtain a gamma-ray spectrum data set for training and testing machine learning classification algorithms. These spectra are broadened, normalized and separated into inelastic scattering and capture spectra, and then used to fit different classifier models. When the comprehensive grade boundary of high- and low-grade ores is set to 5%, the evaluation metrics calculated by the 5-fold cross-validation show that the SVM (Support Vector Machine), KNN (K-Nearest Neighbor), GNB (Gaussian Naive Bayes) and RF (Random Forest) models can effectively distinguish lead-zinc ore from gangue. At the same time, the GNB model has achieved the optimal accuracy of 91.45% when identifying high- and low-grade ores, and the F1 score for both types of ores is greater than 0.9.