• Title/Summary/Keyword: Non-clustering

Search Result 398, Processing Time 0.023 seconds

The Association of Smoking Status and Clustering of Obesity and Depression on the Risk of Early-Onset Cardiovascular Disease in Young Adults: A Nationwide Cohort Study

  • Choon-Young Kim;Cheol Min Lee;Seungwoo Lee;Jung Eun Yoo;Heesun Lee;Hyo Eun Park;Kyungdo Han;Su-Yeon Choi
    • Korean Circulation Journal
    • /
    • v.53 no.1
    • /
    • pp.17-30
    • /
    • 2023
  • Background and Objectives: To evaluate the impact of smoking in young adults on the risk of cardiovascular disease (CVD) and the clustering effect of behavioral risk factors such as smoking, obesity, and depression. Methods: A Korean nationwide population-based cohort of a total of 3,280,826 participants aged 20-39 years old who underwent 2 consecutive health examinations were included. They were followed up until the date of CVD (myocardial infarction [MI] or stroke), or December 2018 (median, 6 years). Results: Current smoking, early age of smoking initiation, and smoking intensity were associated with an increased risk of CVD incidence. Even after quitting smoking, the risk of MI was still high in quitters compared with non-smokers. Cigarette smoking, obesity, and depression were independently associated with a 1.3-1.7 times increased risk of CVD, and clustering of 2 or more of these behavioral risk factors was associated with a 2-3 times increased risk of CVD in young adults. Conclusions: In young adults, cigarette smoking was associated with the risk of CVD, and the clustering of 2 or more behavioral risk factors showed an additive risk of CVD.

Alcock-Paczynski Test with the Evolution of Redshift-Space Galaxy Clustering Anisotropy: Understanding the Systematics

  • Park, Hyunbae;Park, Changbom;Tonegawa, Motonari;Zheng, Yi;Sabiu, Cristiano G.;Li, Xiao-dong;Hong, Sungwook E.;Kim, Juhan
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.44 no.1
    • /
    • pp.78.2-78.2
    • /
    • 2019
  • We develop an Alcock-Paczynski (AP) test method that uses the evolution of redshift-space two-point correlation function (2pCF) of galaxies. The method improves the AP test proposed by Li et al. (2015) in that it uses the full two-dimensional shape of the correlation function. Similarly to the original method, the new one uses the 2pCF in redshift space with its amplitude normalized. Cosmological constraints can be obtained by examining the redshift dependence of the normalized 2pCF. This is because the 2pCF should not change apart from the expected small non-linear evolution if galaxy clustering is not distorted by incorrect choice of cosmology used to convert redshift to comoving distance. Our new method decomposes the redshift difference of the 2-dimensional correlation function into the Legendre polynomials whose amplitudes are modelled by radial fitting functions. The shape of the normalized 2pCF suffers from small intrinsic time evolution due to non-linear gravitational evolution and change of type of galaxies between different redshifts. It can be accurately measured by using state of the art cosmological simulations. We use a set of our Multiverse simulations to find that the systematic effects on the shape of the normalized 2pCF are quite insensitive to change of cosmology over \Omega_m=0.21 - 0.31 and w=-0.5 - -1.5. Thanks to this finding, we can now apply our method for the AP test using the non-linear systematics measured from a single simulation of the fiducial cosmological model.

  • PDF

A non-merging data analysis method to localize brain source for gait-related EEG (보행 관련 뇌파의 신호원 추정을 위한 비통합 데이터 분석 방법)

  • Song, Minsu;Jung, Jiuk;Jee, In-Hyeog;Chu, Jun-Uk
    • Journal of IKEEE
    • /
    • v.25 no.4
    • /
    • pp.679-688
    • /
    • 2021
  • Gait is an evaluation index used in various clinical area including brain nervous system diseases. Signal source localizing and time-frequency analysis are mainly used after extracting independent components for Electroencephalogram data as a method of measuring and analyzing brain activation related to gait. Existing treadmill-based walking EEG analysis performs signal preprocessing, independent component analysis(ICA), and source localizing by merging data after the multiple EEG measurements, and extracts representative component clusters through inter-subject clustering. In this study we propose an analysis method, without merging to single dataset, that performs signal preprocessing, ICA, and source localization on each measurements, and inter-subject clustering is conducted for ICs extracted from all subjects. The effect of data merging on the IC clustering and time-frequency analysis was investigated for the proposed method and two conventional methods. As a result, it was confirmed that a more subdivided gait-related brain signal component was derived from the proposed "non-merging" method (4 clusters) despite the small number of subjects, than conventional method (2 clusters).

Prognostic Evaluation of Categorical Platelet-based Indices Using Clustering Methods Based on the Monte Carlo Comparison for Hepatocellular Carcinoma

  • Guo, Pi;Shen, Shun-Li;Zhang, Qin;Zeng, Fang-Fang;Zhang, Wang-Jian;Hu, Xiao-Min;Zhang, Ding-Mei;Peng, Bao-Gang;Hao, Yuan-Tao
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.15 no.14
    • /
    • pp.5721-5727
    • /
    • 2014
  • Objectives: To evaluate the performance of clustering methods used in the prognostic assessment of categorical clinical data for hepatocellular carcinoma (HCC) patients in China, and establish a predictable prognostic nomogram for clinical decisions. Materials and Methods: A total of 332 newly diagnosed HCC patients treated with hepatic resection during 2006-2009 were enrolled. Patients were regularly followed up at outpatient clinics. Clustering methods including the Average linkage, k-modes, fuzzy k-modes, PAM, CLARA, protocluster, and ROCK were compared by Monte Carlo simulation, and the optimal method was applied to investigate the clustering pattern of the indices including platelet count, platelet/lymphocyte ratio (PLR) and serum aspartate aminotransferase activity/platelet count ratio index (APRI). Then the clustering variable, age group, tumor size, number of tumor and vascular invasion were studied in a multivariable Cox regression model. A prognostic nomogram was constructed for clinical decisions. Results: The ROCK was best in both the overlapping and non-overlapping cases performed to assess the prognostic value of platelet-based indices. Patients with categorical platelet-based indices significantly split across two clusters, and those with high values, had a high risk of HCC recurrence (hazard ratio [HR] 1.42, 95% CI 1.09-1.86; p<0.01). Tumor size, number of tumor and blood vessel invasion were also associated with high risk of HCC recurrence (all p< 0.01). The nomogram well predicted HCC patient survival at 3 and 5 years. Conclusions: A cluster of platelet-based indices combined with other clinical covariates could be used for prognosis evaluation in HCC.

Development of a Daily Pattern Clustering Algorithm using Historical Profiles (과거이력자료를 활용한 요일별 패턴분류 알고리즘 개발)

  • Cho, Jun-Han;Kim, Bo-Sung;Kim, Seong-Ho;Kang, Weon-Eui
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.10 no.4
    • /
    • pp.11-23
    • /
    • 2011
  • The objective of this paper is to develop a daily pattern clustering algorithm using historical traffic data that can reliably detect under various traffic flow conditions in urban streets. The developed algorithm in this paper is categorized into two major parts, that is to say a macroscopic and a microscopic points of view. First of all, a macroscopic analysis process deduces a daily peak/non-peak hour and emphasis analysis time zones based on the speed time-series. A microscopic analysis process clusters a daily pattern compared with a similarity between individuals or between individual and group. The name of the developed algorithm in microscopic analysis process is called "Two-step speed clustering (TSC) algorithm". TSC algorithm improves the accuracy of a daily pattern clustering based on the time-series speed variation data. The experiments of the algorithm have been conducted with point detector data, installed at a Ansan city, and verified through comparison with a clustering techniques using SPSS. Our efforts in this study are expected to contribute to developing pattern-based information processing, operations management of daily recurrent congestion, improvement of daily signal optimization based on TOD plans.

Effective Artificial Neural Network Approach for Non-Binary Incidence Matrix-Based Part-Machine Grouping (비이진 연관행렬 기반의 부품-기계 그룹핑을 위한 효과적인 인공신경망 접근법)

  • Won, You-Kyung
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.31 no.4
    • /
    • pp.69-87
    • /
    • 2006
  • This paper proposes an effective approach for the part-machine grouping(PMG) based on the non-binary part-machine incidence matrix in which real manufacturing factors such as the operation sequences with multiple visits to the same machine and production volumes of parts are incorporated and each entry represents actual moves due to different operation sequences. The proposed approach adopts Fuzzy ART neural network to quickly create the Initial part families and their machine cells. A new performance measure to evaluate and compare the goodness of non-binary block diagonal solution is suggested. To enhance the poor solution due to category proliferation inherent to most artificial neural networks, a supplementary procedure reassigning parts and machines is added. To show effectiveness of the proposed approach to large-size PMG problems, a psuedo-replicated clustering procedure is designed. Experimental results with intermediate to large-size data sets show effectiveness of the proposed approach.

A Process for Transforming Non-component Java Programs into EJB Programs (비 컴포넌트 자바 프로그램에서 EJB 프로그램으로의 변환 프로세스)

  • Lee, Sung-Eun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.3
    • /
    • pp.173-186
    • /
    • 2006
  • In this paper, we suggest a process that transforms non-component Java programs into EJB component programs. We approach following methods to increase reusability of existing Java-based programs. We extract proper factors from existing non-component Java programs to construct for component model, and we suggest a transformation technique using extracted factors. Extracted factors are transformed into EJB components. With consideration for reusability of existing programs and EJB's characteristic, we suggest a process that mixes class clustering and method oriented class restructuring.

  • PDF

Music Transcription Using Non-Negative Matrix Factorization (비음수 행렬 분해 (NMF)를 이용한 악보 전사)

  • Park, Sang-Ha;Lee, Seok-Jin;Sung, Koeng-Mo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.29 no.2
    • /
    • pp.102-110
    • /
    • 2010
  • Music transcription is extracting pitch (the height of a musical note) and rhythm (the length of a musical note) information from audio file and making a music score. In this paper, we decomposed a waveform into frequency and rhythm components using Non-Negative Matrix Factorization (NMF) and Non-Negative Sparse coding (NNSC) which are often used for source separation and data clustering. And using the subharmonic summation method, fundamental frequency is calculated from the decomposed frequency components. Therefore, the accurate pitch of each score can be estimated. The proposed method successfully performed music transcription with its results superior to those of the conventional methods which used either NMF or NNSC.

Generic Summarization Using Generic Important of Semantic Features (의미특징의 포괄적 중요도를 이용한 포괄적 문서 요약)

  • Park, Sun;Lee, Jong-Hoon
    • Journal of Advanced Navigation Technology
    • /
    • v.12 no.5
    • /
    • pp.502-508
    • /
    • 2008
  • With the increased use of the internet and the tremendous amount of data it transfers, it is more necessary to summarize documents. We propose a new method using the Non-negative Semantic Variable Matrix (NSVM) and the generic important of semantic features obtained by Non-negative Matrix Factorization (NMF) to extract the sentences for automatic generic summarization. The proposed method use non-negative constraints which is more similar to the human's cognition process. As a result, the proposed method selects more meaningful sentences for summarization than the unsupervised method used the Latent Semantic Analysis (LSA) or clustering methods. The experimental results show that the proposed method archives better performance than other methods.

  • PDF

A New Forest Fire Detection Algorithm using Outlier Detection Method on Regression Analysis between Surface temperature and NDVI

  • Huh, Yong;Byun, Young-Gi;Son, Jeong-Hoon;Yu, Ki-Yun;Kim, Yong-Il
    • Proceedings of the KSRS Conference
    • /
    • v.2
    • /
    • pp.574-577
    • /
    • 2006
  • In this paper, we developed a forest fire detection algorithm which uses a regression function between NDVI and land surface temperature. Previous detection algorithms use the land surface temperature as a main factor to discriminate fire pixels from non-fire pixels. These algorithms assume that the surface temperatures of non-fire pixels are intrinsically analogous and obey Gaussian normal distribution, regardless of land surface types and conditions. And the temperature thresholds for detecting fire pixels are derived from the statistical distribution of non-fire pixels’ temperature using heuristic methods. This assumption makes the temperature distribution of non-fire pixels very diverse and sometimes slightly overlapped with that of fire pixel. So, sometimes there occur omission errors in the cases of small fires. To ease such problem somewhat, we separated non-fire pixels into each land cover type by clustering algorithm and calculated the residuals between the temperature of a pixel under examination whether fire pixel or not and estimated temperature of the pixel using the linear regression between surface temperature and NDVI. As a result, this algorithm could modify the temperature threshold considering land types and conditions and showed improved detection accuracy.

  • PDF