• Title/Summary/Keyword: Standard Dataset

Search Result 196, Processing Time 0.025 seconds

A Swapping Red-black Tree for Wear-leveling of Non-volatile Memory (비휘발성 메모리의 마모도 평준화를 위한 레드블랙 트리)

  • Jeong, Minseong;Lee, Eunji
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.6
    • /
    • pp.139-144
    • /
    • 2019
  • For recent decades, Non-volatile Memory (NVM) technologies have been drawing a high attention both in industry and academia due to its high density and short latency comparable to that of DRAM. However, NVM devices has write endurance problem and thus the current data structures that have been built around DRAM-specific features including unlimited program cycles is inadequate for NVM, reducing the device lifetime significantly. In this paper, we revisit a red-black tree extensively adopted for data indexing across a wide range of applications, and make it to better fit for NVM. Specifically, we observe that the conventional red-black tree wears out the specific location of memory because of its rebalancing operation to ensure fast access time over a whole dataset. However, this rebalancing operation frequently updates the long-lived nodes, which leads to the skewed wear out across the NVM cells. To resolve this problem, we present a new swapping wear-leveling red-black tree that periodically moves data in the worn-out node into the young node. The performance study with real-world traces demonstrates the proposed red-black tree reduces the standard deviation of the write count across nodes by up to 12.5%.

Estimation of compressive strength of BFS and WTRP blended cement mortars with machine learning models

  • Ozcan, Giyasettin;Kocak, Yilmaz;Gulbandilar, Eyyup
    • Computers and Concrete
    • /
    • v.19 no.3
    • /
    • pp.275-282
    • /
    • 2017
  • The aim of this study is to build Machine Learning models to evaluate the effect of blast furnace slag (BFS) and waste tire rubber powder (WTRP) on the compressive strength of cement mortars. In order to develop these models, 12 different mixes with 288 specimens of the 2, 7, 28, and 90 days compressive strength experimental results of cement mortars containing BFS, WTRP and BFS+WTRP were used in training and testing by Random Forest, Ada Boost, SVM and Bayes classifier machine learning models, which implement standard cement tests. The machine learning models were trained with 288 data that acquired from experimental results. The models had four input parameters that cover the amount of Portland cement, BFS, WTRP and sample ages. Furthermore, it had one output parameter which is compressive strength of cement mortars. Experimental observations from compressive strength tests were compared with predictions of machine learning methods. In order to do predictive experimentation, we exploit R programming language and corresponding packages. During experimentation on the dataset, Random Forest, Ada Boost and SVM models have produced notable good outputs with higher coefficients of determination of R2, RMS and MAPE. Among the machine learning algorithms, Ada Boost presented the best R2, RMS and MAPE values, which are 0.9831, 5.2425 and 0.1105, respectively. As a result, in the model, the testing results indicated that experimental data can be estimated to a notable close extent by the model.

Improving an Ensemble Model by Optimizing Bootstrap Sampling (부트스트랩 샘플링 최적화를 통한 앙상블 모형의 성능 개선)

  • Min, Sung-Hwan
    • Journal of Internet Computing and Services
    • /
    • v.17 no.2
    • /
    • pp.49-57
    • /
    • 2016
  • Ensemble classification involves combining multiple classifiers to obtain more accurate predictions than those obtained using individual models. Ensemble learning techniques are known to be very useful for improving prediction accuracy. Bagging is one of the most popular ensemble learning techniques. Bagging has been known to be successful in increasing the accuracy of prediction of the individual classifiers. Bagging draws bootstrap samples from the training sample, applies the classifier to each bootstrap sample, and then combines the predictions of these classifiers to get the final classification result. Bootstrap samples are simple random samples selected from the original training data, so not all bootstrap samples are equally informative, due to the randomness. In this study, we proposed a new method for improving the performance of the standard bagging ensemble by optimizing bootstrap samples. A genetic algorithm is used to optimize bootstrap samples of the ensemble for improving prediction accuracy of the ensemble model. The proposed model is applied to a bankruptcy prediction problem using a real dataset from Korean companies. The experimental results showed the effectiveness of the proposed model.

lustering of Categorical Data using Rough Entropy (러프 엔트로피를 이용한 범주형 데이터의 클러스터링)

  • Park, Inkyoo
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.13 no.5
    • /
    • pp.183-188
    • /
    • 2013
  • A variety of cluster analysis techniques prerequisite to cluster objects having similar characteristics in data mining. But the clustering of those algorithms have lots of difficulties in dealing with categorical data within the databases. The imprecise handling of uncertainty within categorical data in the clustering process stems from the only algebraic logic of rough set, resulting in the degradation of stability and effectiveness. This paper proposes a information-theoretic rough entropy(RE) by taking into account the dependency of attributes and proposes a technique called min-mean-mean roughness(MMMR) for selecting clustering attribute. We analyze and compare the performance of the proposed technique with K-means, fuzzy techniques and other standard deviation roughness methods based on ZOO dataset. The results verify the better performance of the proposed approach.

The reliability of tablet computers in depicting maxillofacial radiographic landmarks

  • Tadinada, Aditya;Mahdian, Mina;Sheth, Sonam;Chandhoke, Taranpreet K;Gopalakrishna, Aadarsh;Potluri, Anitha;Yadav, Sumit
    • Imaging Science in Dentistry
    • /
    • v.45 no.3
    • /
    • pp.175-180
    • /
    • 2015
  • Purpose: This study was performed to evaluate the reliability of the identification of anatomical landmarks in panoramic and lateral cephalometric radiographs on a standard medical grade picture archiving communication system (PACS) monitor and a tablet computer (iPad 5). Materials and Methods: A total of 1000 radiographs, including 500 panoramic and 500 lateral cephalometric radiographs, were retrieved from the de-identified dataset of the archive of the Section of Oral and Maxillofacial Radiology of the University Of Connecticut School Of Dental Medicine. Major radiographic anatomical landmarks were independently reviewed by two examiners on both displays. The examiners initially reviewed ten panoramic and ten lateral cephalometric radiographs using each imaging system, in order to verify interoperator agreement in landmark identification. The images were scored on a four-point scale reflecting the diagnostic image quality and exposure level of the images. Results: Statistical analysis showed no significant difference between the two displays regarding the visibility and clarity of the landmarks in either the panoramic or cephalometric radiographs. Conclusion: Tablet computers can reliably show anatomical landmarks in panoramic and lateral cephalometric radiographs.

The Study on the Delineation of the Busan Metropolitan Region, Korea (부산광역도시권 설정에 관한 연구)

  • Lee, Hee-Yul;Ju, Mee-Soon
    • Journal of the Economic Geographical Society of Korea
    • /
    • v.10 no.2
    • /
    • pp.167-181
    • /
    • 2007
  • The newly adopted Korea Geodetic Datum (a.k.a. KGD2002) calls for massive reengineering work on geospatial dataset. The main focus of our study is placed on the strategy and system implementations of the required data reengineering with a keen attention to integrated approaches to interoperability, standardization, and database utilization. Our reengineering strategy includes file-to-file, file-to-DB, DB-to-file, and DB-to-DB conversion for the coordinate transformation of KGD2002. In addition to the map formats of existing standards such as DXF and Shapefile, the newly recommended standards such as GML and SVG are also accommodated in our reengineering environment. These four types of standard format may be imported into and exported from spatial database via KGD2002 transformation component. The DB-to-DB conversion, in particular, includes not only intra-database conversion but also inter-database conversion between SDE/Oracle and Oracle Spatial. All these implementations were carried out in multiple computing environments: desktop and the Web. The feasibility test of our system shows that the coordinate differences between Bessel and GRS80 ellipsoid agree with the criteria presented in the existing researches.

  • PDF

Applicable Evaluation of the Latest Land-use Data for Developing a Real-time Atmospheric Field Prediction of RAMS (RAMS의 실시간 기상장 예측 향상을 위한 최신 토지피복도 자료의 적용가능성)

  • Won, Gyeong-Mee;Lee, Hwa-Woon;Yu, Jeong-Ah;Hong, Hyun-Su;Hwang, Man-Sik;Chun, Kwang-Su;Choi, Kwang-Su;Lee, Moon-Soon
    • Journal of Korean Society for Atmospheric Environment
    • /
    • v.24 no.1
    • /
    • pp.1-15
    • /
    • 2008
  • Chemical Accident Response Information System (CARIS) which has been designed for the efficient emergency response of chemical accidents produces the real-time atmospheric fields through the Regional Atmospheric Modeling System, RAMS. The previous studies were emphasized that improving an initial input data had more effective results in developing prediction ability of atmospheric model. In a continuous effort to improve an initial input data, we replaced the land-use dataset using in the RAMS, which is a high resolution USGS digital data constructed in April, 1993, with the latest land-use data of the Korea Ministry of Environment over the South Korea and simulated atmospheric fields for developing a real-time prediction in dispersion of chemicals. The results showed that the new land-use data was written in a standard RAMS format and shown the modified surface characteristics and the landscape heterogeneity resulting from land-use change. In the results of sensitivity experiment we got the improved atmospheric fields and assured that it will give more reliable real-time atmospheric fields to all users of CARIS for the dispersion forecast in associated with hazardous chemical releases as well as general air pollutants.

Design of Echo Classifier Based on Neuro-Fuzzy Algorithm Using Meteorological Radar Data (기상레이더를 이용한 뉴로-퍼지 알고리즘 기반 에코 분류기 설계)

  • Oh, Sung-Kwun;Ko, Jun-Hyun
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.63 no.5
    • /
    • pp.676-682
    • /
    • 2014
  • In this paper, precipitation echo(PRE) and non-precipitaion echo(N-PRE)(including ground echo and clear echo) through weather radar data are identified with the aid of neuro-fuzzy algorithm. The accuracy of the radar information is lowered because meteorological radar data is mixed with the PRE and N-PRE. So this problem is resolved by using RBFNN and judgement module. Structure expression of weather radar data are analyzed in order to classify PRE and N-PRE. Input variables such as Standard deviation of reflectivity(SDZ), Vertical gradient of reflectivity(VGZ), Spin change(SPN), Frequency(FR), cumulation reflectivity during 1 hour(1hDZ), and cumulation reflectivity during 2 hour(2hDZ) are made by using weather radar data and then each characteristic of input variable is analyzed. Input data is built up from the selected input variables among these input variables, which have a critical effect on the classification between PRE and N-PRE. Echo judgment module is developed to do echo classification between PRE and N-PRE by using testing dataset. Polynomial-based radial basis function neural networks(RBFNNs) are used as neuro-fuzzy algorithm, and the proposed neuro-fuzzy echo pattern classifier is designed by combining RBFNN with echo judgement module. Finally, the results of the proposed classifier are compared with both CZ and DZ, as well as QC data, and analyzed from the view point of output performance.

3D Reconstruction of a Single Clothing Image and Its Application to Image-based Virtual Try-On (의상 이미지의 3차원 의상 복원 방법과 가상착용 응용)

  • Ahn, Heejune;Minar, Matiur Rahman
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.25 no.5
    • /
    • pp.1-11
    • /
    • 2020
  • Image-based virtual try-on (VTON) is becoming popular for online apparel shopping, mainly because of not requiring 3D information for try-on clothes and target humans. However, existing 2D algorithms, even when utilizing advanced non-rigid deformation algorithms, cannot handle large spatial transformations for complex target human poses. In this study, we propose a 3D clothing reconstruction method using a 3D human body model. The resulting 3D models of try-on clothes can be more easily deformed when applied to rest posed standard human models. Then, the poses and shapes of 3D clothing models can be transferred to the target human models estimated from 2D images. Finally, the deformed clothing models can be rendered and blended with target human representations. Experimental results with the VITON dataset used in the previous works show that the shapes of reconstructed clothing are significantly more natural, compared to the 2D image-based deformation results when human poses and shapes are estimated accurately.

Evaluation of Delhi Population Based Cancer Registry and Trends of Tobacco Related Cancers

  • Yadav, Rajesh;Garg, Renu;Manoharan, N;Swasticharan, L;Julka, PK;Rath, GK
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.17 no.6
    • /
    • pp.2841-2846
    • /
    • 2016
  • Background: Tobacco use is the single most important preventable risk factor for cancer. Surveillance of tobacco-related cancers (TRC) is critical for monitoring trends and evaluating tobacco control programmes. We analysed the trends of TRC and evaluated the population-based cancer registry (PBCR) in Delhi for simplicity, comparability, validity, timeliness and representativeness. Materials and Methods: We interviewed key informants, observed registry processes and analysed the PBCR dataset for the period 1988-2009 using the 2009 TRC definition of the International Agency for Research on Cancer. We calculated the percentages of morphologically verified cancers, death certificate-only (DCO) cases, missing values of key variables and the time between cancer diagnosis and registration or publication for the year 2009. Results: The number of new cancer cases increased from 5,854 to 15,244 (160%) during 1988-2009. TRC constituted 58% of all cancers among men and 47% among women in 2009. The age-adjusted incidence rates of TRC per 100,000 population increased from 64.2 to 97.3 among men, and from 66.2 to 69.2 among women during 1988-2009. Data on all cancer cases presenting at all major government and private health facilities are actively collected by the PBCR staff using standard paper-based forms. Data abstraction and coding is conducted manually following ICD-10 classifications. Eighty per cent of cases were morphologically verified and 1% were identified by death certificate only. Less than 1% of key variables had missing values. The median time to registration and publishing was 13 and 32 months, respectively. Conclusions: The burden of TRC in Delhi is high and increasing. The Delhi PBCR is well organized and generates high-quality, representative data. However, data could be published earlier if paper-based data are replaced by electronic data abstraction.