• Title/Summary/Keyword: input data

Search Result 8,340, Processing Time 0.038 seconds

Predicting required licensed spectrum for the future considering big data growth

  • Shayea, Ibraheem;Rahman, Tharek Abd.;Azmi, Marwan Hadri;Han, Chua Tien;Arsad, Arsany
    • ETRI Journal
    • /
    • v.41 no.2
    • /
    • pp.224-234
    • /
    • 2019
  • This paper proposes a new spectrum forecasting (SF) model to estimate the spectrum demands for future mobile broadband (MBB) services. The model requires five main input metrics, that is, the current available spectrum, site number growth, mobile data traffic growth, average network utilization, and spectrum efficiency growth. Using the proposed SF model, the future MBB spectrum demand for Malaysia in 2020 is forecasted based on the input market data of four major mobile telecommunication operators represented by A-D, which account for approximately 95% of the local mobile market share. Statistical data to generate the five input metrics were obtained from prominent agencies, such as the Malaysian Communications and Multimedia Commission, OpenSignal, Analysys Mason, GSMA, and Huawei. Our forecasting results indicate that by 2020, Malaysia would require approximately 307 MHz of additional spectrum to fulfill the enormous increase in mobile broadband data demands.

Assessment of Improving SWAT Weather Input Data using Basic Spatial Interpolation Method

  • Felix, Micah Lourdes;Choi, Mikyoung;Zhang, Ning;Jung, Kwansue
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.368-368
    • /
    • 2022
  • The Soil and Water Assessment Tool (SWAT) has been widely used to simulate the long-term hydrological conditions of a catchment. Two output variables, outflow and sediment yield have been widely investigated in the field of water resources management, especially in determining the conditions of ungauged subbasins. The presence of missing data in weather input data can cause poor representation of the climate conditions in a catchment especially for large or mountainous catchments. Therefore, in this study, a custom module was developed and evaluated to determine the efficiency of utilizing basic spatial interpolation methods in the estimation of weather input data. The module has been written in Python language and can be considered as a pre-processing module prior to using the SWAT model. The results of this study suggests that the utilization of the proposed pre-processing module can improve the simulation results for both outflow and sediment yield in a catchment, even in the presence of missing data.

  • PDF

Evaluation of the Input Status of Exposure-related Information of Working Environment Monitoring Database and Special Health Examination Database for the Construction of a National Exposure Surveillance System (국가노출감시체계 구축을 위한 작업환경측정과 특수건강진단 자료의 노출 정보 입력 실태 평가)

  • Choi, Sangjun;Koh, Dong-Hee;Park, Ju-Hyun;Park, Donguk;Kim, Hwan-Cheol;Lim, Dae Sung;Sung, Yeji;Ko, Kyoung Yoon;Lim, Ji Seon;Seo, Hoekyeong
    • Journal of Korean Society of Occupational and Environmental Hygiene
    • /
    • v.32 no.3
    • /
    • pp.231-241
    • /
    • 2022
  • Objectives: The purpose of this study is to evaluate the input status of exposure-related information in the working environment monitoring database (WEMD) and special health examination database (SHED) for the construction of a national exposure surveillance system. Methods: The industrial and process code input status of WEMD and SHED for 21 carcinogens from 2014 to 2016 was compared. Data from workers who performed both work environment monitoring and special health examinations in 2019 and 2020 were extracted and the actual status of input of industrial and process codes was analyzed. We also investigated the cause of input errors through a focus group interview with 12 data input specialists. Results: As a result of analyzing WMED and SHED for 21 carcinogens, the five-digit industrial code matching rate was low at 53.5% and the process code matching rate was 19% or less. Among the data that simultaneously conducted work environment monitoring and special health examination in 2019 and 2020, the process code matching rate was very low at 18.1% and 5.2%, respectively. The main causes of exposure-related data input errors were the difference between the WEMD and SHED process code input systems from 2020, the number of standard process and job codes being too large, and the inefficiency of the standard code search system. Conclusions: In order to use WEMD and SHED as a national surveillance system, it is necessary to simplify the number of standard code input codes and improve the search system efficiency.

Comparative Study on the Accuracy of Surface Air Temperature Prediction based on selection of land use and initial meteorological data (토지이용도와 초기 기상 입력 자료의 선택에 따른 지상 기온 예측 정확도 비교 연구)

  • Hae-Dong Kim;Ha-Young Kim
    • Journal of Environmental Science International
    • /
    • v.33 no.6
    • /
    • pp.435-442
    • /
    • 2024
  • We investigated the accuracy of surface air temperature prediction according to the selection of land-use data and initial meteorological data using the Weather Research and Forecasting model-v4.2.1. A numerical experiment was conducted at the Daegu Dyeing Industrial Complex. We initially used meteorological input data from GFS (Global forecast system)and GDAPS (Global data assimilation and prediction system). High-resolution input data were generated and used as input data for the weather model using the land cover data of the Ministry of Environment and the digital elevation model of the Ministry of Land, Infrastructure, and Transport. The experiment was conducted by classifying the terrestrial and topographic data (land cover data) and meteorological data applied to the model. For simulations using high-resolution terrestrial data(10 m), global data assimilation, and prediction system data(CASE 3), the calculated surface temperature was much closer to the automatic weather station observations than for simulations using low-resolution terrestrial data(900 m) and GFS(CASE 1).

Identification of DEA Determinant Input-Output Variables : an Illustration for Evaluating the Efficiency of Government-Sponsored R&D Projects (DEA 효율성을 결정하는 입력-출력변수 식별 : 정부지원 R&D 과제 효율성 평가를 위한 실례)

  • Park, Sungmin
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.40 no.1
    • /
    • pp.84-99
    • /
    • 2014
  • In this study, determinant input-output variables are identified for calculating Data Envelopment Analysis (DEA) efficiency scores relating to evaluating the efficiency of government-sponsored research and development (R&D) projects. In particular, this study proposes a systematic framework of design and analysis of experiments, called "all possible DEAs", for pinpointing DEA determinant input-output variables. In addition to correlation analyses, two modified measures of time series analysis are developed in order to check the similarities between a DEA complete data structure (CDS) versus the rest of incomplete data structures (IDSs). In this empirical analysis, a few DEA determinant input-output variables are found to be associated with a typical public R&D performance evaluation logic model, especially oriented to a mid- and long-term performance perspective. Among four variables, only two determinants are identified : "R&D manpower" ($x_2$) and "Sales revenue" ($y_1$). However, it should be pointed out that the input variable "R&D funds" ($x_1$) is insignificant for calculating DEA efficiency score even if it is a critical input for measuring efficiency of a government-sonsored R&D project from a practical point of view a priori. In this context, if practitioners' top priority is to see the efficiency between "R&D funds" ($x_1$) and "Sales revenue" ($y_1$), the DEA efficiency score cannot properly meet their expectations. Therefore, meticulous attention is required when using the DEA application for public R&D performance evaluation, considering that discrepancies can occur between practitioners' expectations and DEA efficiency scores.

A New Low Power High Level Synthesis for DSP (DSP를 위한 새로운 저전력 상위 레벨 합성)

  • 한태희;김영숙;인치호;김희석
    • Proceedings of the IEEK Conference
    • /
    • 2002.06b
    • /
    • pp.101-104
    • /
    • 2002
  • This paper propose that is algorithm of power dissipation reduction in the high level synthesis design for DSP(Digital Signal Processor), as the portable terminal system recently demand high power dissipation. This paper obtain effect of power dissipation reduction and switching activity that increase correlation of operands as input data of function unit. The algorithm search loop or repeatedly data to the input operands of function unit. That can be reduce the power dissipation using the new low power high level synthesis algorithm. In this Paper, scheduling operation search same nodes from input DFG(Data Flow Graph) with correlation coefficient of first input node and among nodes. Function units consist a multiplier, an adder and a register. The power estimation method is added switching activity for each bits of nodes. The power estimation have good efficient using proposed algorithm. This paper result obtain more Power reduction of fifty percents after using a new low power algorithm in a function unit as multiplier.

  • PDF

Feature Selection Effect of Classification Tree Using Feature Importance : Case of Credit Card Customer Churn Prediction (특성중요도를 활용한 분류나무의 입력특성 선택효과 : 신용카드 고객이탈 사례)

  • Yoon Hanseong
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.20 no.2
    • /
    • pp.1-10
    • /
    • 2024
  • For the purpose of predicting credit card customer churn accurately through data analysis, a model can be constructed with various machine learning algorithms, including decision tree. And feature importance has been utilized in selecting better input features that can improve performance of data analysis models for several application areas. In this paper, a method of utilizing feature importance calculated from the MDI method and its effects are investigated in the credit card customer churn prediction problem with classification trees. Compared with several random feature selections from case data, a set of input features selected from higher value of feature importance shows higher predictive power. It can be an efficient method for classifying and choosing input features necessary for improving prediction performance. The method organized in this paper can be an alternative to the selection of input features using feature importance in composing and using classification trees, including credit card customer churn prediction.

Smoothed RSSI-Based Distance Estimation Using Deep Neural Network (심층 인공신경망을 활용한 Smoothed RSSI 기반 거리 추정)

  • Hyeok-Don Kwon;Sol-Bee Lee;Jung-Hyok Kwon;Eui-Jik Kim
    • Journal of Internet of Things and Convergence
    • /
    • v.9 no.2
    • /
    • pp.71-76
    • /
    • 2023
  • In this paper, we propose a smoothed received signal strength indicator (RSSI)-based distance estimation using deep neural network (DNN) for accurate distance estimation in an environment where a single receiver is used. The proposed scheme performs a data preprocessing consisting of data splitting, missing value imputation, and smoothing steps to improve distance estimation accuracy, thereby deriving the smoothed RSSI values. The derived smoothed RSSI values are used as input data of the Multi-Input Single-Output (MISO) DNN model, and are finally returned as an estimated distance in the output layer through input layer and hidden layer. To verify the superiority of the proposed scheme, we compared the performance of the proposed scheme with that of the linear regression-based distance estimation scheme. As a result, the proposed scheme showed 29.09% higher distance estimation accuracy than the linear regression-based distance estimation scheme.

Pre-processing Method of Raw Data Based on Ontology for Machine Learning (머신러닝을 위한 온톨로지 기반의 Raw Data 전처리 기법)

  • Hwang, Chi-Gon;Yoon, Chang-Pyo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.5
    • /
    • pp.600-608
    • /
    • 2020
  • Machine learning constructs an objective function from learning data, and predicts the result of the data generated by checking the objective function through test data. In machine learning, input data is subjected to a normalisation process through a preprocessing. In the case of numerical data, normalization is standardized by using the average and standard deviation of the input data. In the case of nominal data, which is non-numerical data, it is converted into a one-hot code form. However, this preprocessing alone cannot solve the problem. For this reason, we propose a method that uses ontology to normalize input data in this paper. The test data for this uses the received signal strength indicator (RSSI) value of the Wi-Fi device collected from the mobile device. These data are solved through ontology because they includes noise and heterogeneous problems.

Comparative Study on Statistical Packages for using Multivariate Q-technique

  • Choi, Yong-Seok;Moon, Hee-jung
    • Communications for Statistical Applications and Methods
    • /
    • v.10 no.2
    • /
    • pp.433-443
    • /
    • 2003
  • In this study, we provide a comparison of multivariate Q-techniques in the up-to-date versions of SAS, SPSS, Minitab and S-plus well known to those who study statistics. We can analyze data through the direct Input method(command) in SAS and use of menu method in SPSS, Minitab and S-plus. The analysis performance method is chosen by the high frequency of use. Widely we compare with each Q-techniques form according to input data, input option, statistical chart and statistical output.