• Title/Summary/Keyword: Data Set Records

Search Result 197, Processing Time 0.036 seconds

Development of BOM System Using Component Based of Urban Transit (도시철도 CBD 기반의 유지보수 BOM 시스템 개발)

  • 이호용;한석윤;박기준;서명원
    • Journal of the Korean Society for Railway
    • /
    • v.7 no.4
    • /
    • pp.406-411
    • /
    • 2004
  • BOM(Bill of Materials) is a listing or description of raw materials, parts, and assemblies that define a product. In order to evaluate the performance of proposed BOM management system, which is very important to maintenance information system of urban transit. We develop component based BOM data and rule-set to design data structure that is mutually independent and integrated efficiently. It divides data whit management interface using component technology. The component based master BOM have advantage in database size and flexibility. Flexibility is measured as the number of updating records in accordance with added new product or engineering change. In database size, component based BOM is the best. we develop master BOM management system in web environment.

Analyzing records of Korean pro-basketball using general linear model (일반선형모형을 적용한 한국남자프로농구 경기기록분석 : 2014-2015 정규리그)

  • Kim, Sae Hyung
    • Journal of the Korean Data and Information Science Society
    • /
    • v.26 no.4
    • /
    • pp.957-970
    • /
    • 2015
  • The purpose of this study was to analyze records of Korean pro-basketball using general linear model (two-way ANOVA and hierarchical multiple regression analysis). Korea Basketball League (KBL) informed the records (2014-2015 season) of this study. The eight variables (TA, 2PA, 3PA, 2P, 3P, Ast, TFB, CH) were selected in content validity. SPSS program was used to analyze general linear model. All alpha level was set at 0.05. Major results were as follow. 3PA had significant interaction effect between victory & defeat variable and home & away variable. Victory teams showed that 3PA was higher in home games than away games, and defeat teams was the other. 2PA, AS, TFB, and CH were selected significant variables affecting victory and defeat. In result of hierarchical regression, Ast had significant moderation effect between 3PA and TS. TFB also had significant moderation effect between AS between 2P. The other construct (Ast between 2PA and TS; TFB between AS between 3P) had no significant moderation effect. In the effect of 2PA, 3PA and Ast to TS, CH also had no significant moderation effect.

Genetic Evaluation of Somatic Cell Counts of Holstein Cattle in Zimbabwe

  • Mangwiro, F.K.;Mhlanga, F.N.;Dzama, K.;Makuza, S.M.
    • Asian-Australasian Journal of Animal Sciences
    • /
    • v.13 no.10
    • /
    • pp.1347-1352
    • /
    • 2000
  • The objectives of the study were to examine non-genetic factors that influence somatic cell counts in dairy cattle and to estimate the genetic parameters of somatic cell counts. A total of 34, 097-test day somatic cell count records were obtained from the Zimbabwe Dairy Services Association (ZDSA). The data were from 5, 615 Holstein daughters of 390 sires and 2, 541 dams tested between May 1994 and December 1998. First lactation cows contributed 22, 147 records to the data set, while 11, 950 records were from second and later parity cows. The model for analysis included fixed effects of month of calving, year of calving, stage of lactation, calving interval and test date. Milk yield and age on test day were fitted in the model as covariates. The additive genetic effects pertaining to cows, sires and dams and the residual error were the random effects. The Average Information Restricted Maximum Likelihood algorithm was used for analysis. The heritability of somatic cell scores was low at $0.027{\pm}0.013$ for parity one cows and $0.087{\pm}0.031$ for parity two and above. Repeatability estimates were $0.22{\pm}0.01$ and $0.30{\pm}0.01$ for the two lactation groups, respectively. Genetic and phenotypic correlations between the somatic cell scores and test day milk production were small and negative. It seems that there is no genetic link between somatic cell counts and milk yield in Holstein cattle in Zimbabwe. The results also seem to indicate that somatic cell count is a trait that is mainly governed by environmental factors.

The Kinetic Analysis of the Approach and Take-off Motion between Performance in Woman's High Jump (여자 높이뛰기에서 경기력 간 도움닫기와 발구름 동작의 운동역학적 분석)

  • Kim, Young-Suk;Ryu, Jae-Kyun;Jang, Jae-Kwan
    • Korean Journal of Applied Biomechanics
    • /
    • v.25 no.1
    • /
    • pp.1-10
    • /
    • 2015
  • Objective : The purpose of this study was to find some kinetic variable's relationships between personal records and low records in female high jump. Methods : Collected data of the subjects(N=8, ages: $25.5{\pm}1.85$, height: $173{\pm}5.83$, mass: $54.75{\pm}6.36$ personal record: $1.71{\pm}0.04$, low record: $1.62{\pm}0.03$) were used for the last three strides and take-off phase. Five video cameras set in 30frames/s were used for recording. After digitizing motion, the Direct Linear Transformation(DLT) technique was employed to obtain 3-D position coordinates. The kinematic and kinetic factors of distance, velocity, angle, impulse, jerk variables were calculated. A paired t-test was applied for the difference of variables between personal records and lower records and for correlation with performances and variables. The significance level was accepted at p<.05. Results : There was no relationship between pattern of stride and performance. However, rate of change of velocity was related with cental of mass height(CMH) at peak point(PP). Knee, hip, backward lean, foot plant, approach and take off angle showed no difference between best record and low record. Vertical impulse momentum also showed no difference between performances. Conclusion : According to a t-test result, there were significant differences in CMH at PP and jerk at touch down between best record and low record.

Improving the Decision-Making Process in the Higher Learning Institutions via Electronic Records Management System Adoption

  • Mukred, Muaadh;Yusof, Zawiyah M.;Mokhtar, Umi Asma';Sadiq, Ali Safaa;Hawash, Burkan;Ahmed, Waleed Abdulkafi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.1
    • /
    • pp.90-113
    • /
    • 2021
  • Electronic Records Management System (ERMS) is a computer program or set of applications that is utilized for keeping up to date records along with their storage. ERMS has been extensively utilized for enhancing the performance of academic institutions. The system assists in the planning and decision-making processes, which in turn enhances the competencies. However, although ERMS is significant in supporting the process of decision-making, the majority of organizations have failed to take an initiative to implement it, taking into account that are some implementing it without an appropriate framework, and thus resulted in the practice which does not meet the accepted standard. Therefore, this study identifies the factors influencing the adoption of ERMS among employees of HLI in Yemen and the role of such adoption in the decision-making process, using the Unified Theory of Acceptance and Use of Technology (UTAUT) along with Technology, Organization and Environment (TOE) as the underpinning theories. The study conducts a cross-sectional survey with a questionnaire as the technique for data collection, distributed to 364 participants in various Yemeni public Higher Learning Institutions (HLI). Using AMOS as a statistical method, the findings revealed there are significant and positive relationships between technology factors (effort expectancy, performance expectancy, IT infrastructure and security), organizational factors (top management support, financial support, training, and policy),environmental factors (competitiveness pressure, facilitating conditions and trust) and behavioral intention to adopt ERMS, which in return has a significant relationship with the process of decision-making in HLI. The study also presents a variety of theoretical and empirical contributions that enrich the body of knowledge in the field of technology adoption and the electronic record's domain.

Deformation-based vulnerability functions for RC bridges

  • Elnashai, A.S.;Borzi, B.;Vlachos, S.
    • Structural Engineering and Mechanics
    • /
    • v.17 no.2
    • /
    • pp.215-244
    • /
    • 2004
  • There is an ever-increasing demand for assessment of earthquake effects on transportation structures, emphasised by the crippling consequences of recent earthquakes hitting developed countries reliant on road transportation. In this work, vulnerability functions for RC bridges are derived analytically using advanced material characterisation, high quality earthquake records and adaptive inelastic dynamic analysis techniques. Four limit states are employed, all based on deformational quantities, in line with recent development of deformation-based seismic assessment. The analytically-derived vulnerability functions are then compared to a data set comprising observational damage data from the Northridge (California 1994) and Hyogo-ken Nanbu (Kobe 1995) earthquakes. The good agreement gives some confidence in the derived formulation that is recommended for use in seismic risk assessment. Furthermore, by varying the dimensions of the prototype bridge used in the study, and the span lengths supported by piers, three more bridges are obtained with different overstrength ratios (ratio of design-to-available base shear). The process of derivation of vulnerability functions is repeated and the ensuing relationships compared. The results point towards the feasibility of deriving scaling factors that may be used to obtain the set of vulnerability functions for a bridge with the knowledge of a 'generic' function and the overstrength ratio. It is demonstrated that this simple procedure gives satisfactory results for the case considered and may be used in the future to facilitate the process of deriving analytical vulnerability functions for classes of bridges once a generic relationship is established.

A Study on Development of Classification Indicators in Transportation Sector Energy Conservation DB (에너지절약 DB 구축을 위한 수송부문 분류지표 설정)

  • Lim, Ki Choo
    • Journal of Energy Engineering
    • /
    • v.25 no.3
    • /
    • pp.149-156
    • /
    • 2016
  • This paper surveyed and analyzed cases of DB development overseas to set the range of DB to be developed for analyzing energy-saving policies in the domestic transportation sector. The foregoing prerequisites were used to establish system for classification in the broad scale under which system for classification in detail indicators that suit one in the broader indicators was set based on analysis of domestic / overseas cases to determine DB development range in the transportation sector required to analysis domestic energy-saving policies. Accordingly, six items subject to the broadest classification were determined, i.e. energy consumption, energy basic unit, emissions of greenhouse gas, economic indicators, transportation volume / transportation records and basic automobile data. Large classification and sub-items determined by surveying expert opinions were set and proposed as DB classification indicators.

Hot Data Verification Method Considering Continuity and Frequency of Write Requests Using Counting Filter

  • Lee, Seung-Woo;Ryu, Kwan-Woo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.6
    • /
    • pp.1-9
    • /
    • 2019
  • Hard disks, which have long been used as secondary storage in computing systems, are increasingly being replaced by solid state drives (SSDs), due to their relatively fast data input / output speeds and small, light weight. SSDs that use NAND flash memory as a storage medium are significantly different from hard disks in terms of physical operation and internal operation. In particular, there is a feature that data overwrite can not be performed, which causes erase operation before writing. In order to solve this problem, a hot data for frequently updating a data for a specific page is distinguished from a cold data for a relatively non-hot data. Hot data identification helps to improve overall performance by identifying and managing hot data separately. Among the various hot data identification methods known so far, there is a technique of recording consecutive write requests by using a Bloom filter and judging the values by hot data. However, the Bloom filter technique has a problem that a new bit array must be generated every time a set of items is changed. In addition, since it is judged based on a continuous write request, it is possible to make a wrong judgment. In this paper, we propose a method using a counting filter for accurate hot data verification. The proposed method examines consecutive write requests. It also records the number of times consecutive write requests occur. The proposed method enables more accurate hot data verification.

A Study on Performing Join Queries over K-anonymous Tables

  • Kim, Dae-Ho;Kim, Jong Wook
    • Journal of the Korea Society of Computer and Information
    • /
    • v.22 no.7
    • /
    • pp.55-62
    • /
    • 2017
  • Recently, there has been an increasing need for the sharing of microdata containing information regarding an individual entity. As microdata usually contains sensitive information on an individual, releasing it directly for public use may violate existing privacy requirements. Thus, to avoid the privacy problems that occur through the release of microdata for public use, extensive studies have been conducted in the area of privacy-preserving data publishing (PPDP). The k-anonymity algorithm, which is the most popular method, guarantees that, for each record, there are at least k-1 other records included in the released data that have the same values for a set of quasi-identifier attributes. Given an original table, the corresponding k-anonymous table is obtained by generalizing each record in the table into an indistinguishable group, called the equivalent class, by replacing the specific values of the quasi-identifier attributes with more general values. However, query processing over the anonymized data is a very challenging task, due to generalized attribute values. In particular, the problem becomes more challenging with an equi-join query (which is the most common type of query in data analysis tasks) over k-anonymous tables, since with the generalized attribute values, it is hard to determine whether two records can be joinable. Thus, to address this challenge, in this paper, we develop a novel scheme that is able to effectively perform an equi-join between k-anonymous tables. The experiment results show that, through the proposed method, significant gains in accuracy over using a naive scheme can be achieved.

Application of Text-Classification Based Machine Learning in Predicting Psychiatric Diagnosis (텍스트 분류 기반 기계학습의 정신과 진단 예측 적용)

  • Pak, Doohyun;Hwang, Mingyu;Lee, Minji;Woo, Sung-Il;Hahn, Sang-Woo;Lee, Yeon Jung;Hwang, Jaeuk
    • Korean Journal of Biological Psychiatry
    • /
    • v.27 no.1
    • /
    • pp.18-26
    • /
    • 2020
  • Objectives The aim was to find effective vectorization and classification models to predict a psychiatric diagnosis from text-based medical records. Methods Electronic medical records (n = 494) of present illness were collected retrospectively in inpatient admission notes with three diagnoses of major depressive disorder, type 1 bipolar disorder, and schizophrenia. Data were split into 400 training data and 94 independent validation data. Data were vectorized by two different models such as term frequency-inverse document frequency (TF-IDF) and Doc2vec. Machine learning models for classification including stochastic gradient descent, logistic regression, support vector classification, and deep learning (DL) were applied to predict three psychiatric diagnoses. Five-fold cross-validation was used to find an effective model. Metrics such as accuracy, precision, recall, and F1-score were measured for comparison between the models. Results Five-fold cross-validation in training data showed DL model with Doc2vec was the most effective model to predict the diagnosis (accuracy = 0.87, F1-score = 0.87). However, these metrics have been reduced in independent test data set with final working DL models (accuracy = 0.79, F1-score = 0.79), while the model of logistic regression and support vector machine with Doc2vec showed slightly better performance (accuracy = 0.80, F1-score = 0.80) than the DL models with Doc2vec and others with TF-IDF. Conclusions The current results suggest that the vectorization may have more impact on the performance of classification than the machine learning model. However, data set had a number of limitations including small sample size, imbalance among the category, and its generalizability. With this regard, the need for research with multi-sites and large samples is suggested to improve the machine learning models.