• Title/Summary/Keyword: Statistical estimation

Search Result 2,438, Processing Time 0.028 seconds

Selection Method for Installation of Reduction Facilities to Prevention of Roe Deer(Capreouls pygargus) Road-kill in Jeju Island (제주도 노루 로드킬 방지를 위한 저감시설 대상지 선정방안 연구)

  • Kim, Min-Ji;Jang, Rae-ik;Yoo, Young-jae;Lee, Jun-Won;Song, Eui-Geun;Oh, Hong-Shik;Sung, Hyun-Chan;Kim, Do-kyung;Jeon, Seong-Woo
    • Journal of the Korean Society of Environmental Restoration Technology
    • /
    • v.26 no.5
    • /
    • pp.19-32
    • /
    • 2023
  • The fragmentation of habitats resulting from human activities leads to the isolation of wildlife and it also causes wildlife-vehicle collisions (i.e. Road-kill). In that sense, it is important to predict potential habitats of specific wildlife that causes wildlife-vehicle collisions by considering geographic, environmental and transportation variables. Road-kill, especially by large mammals, threatens human safety as well as financial losses. Therefore, we conducted this study on roe deer (Capreolus pygargus tianschanicus), a large mammal that causes frequently Road-kill in Jeju Island. So, to predict potential wildlife habitats by considering geographic, environmental, and transportation variables for a specific species this study was conducted to identify high-priority restoration sites with both characteristics of potential habitats and road-kill hotspot. we identified high-priority restoration sites that is likely to be potential habitats, and also identified the known location of a Road-kill records. For this purpose, first, we defined the environmental variables and collect the occurrence records of roe deer. After that, the potential habitat map was generated by using Random Forest model. Second, to analyze roadkill hotspots, a kernel density estimation was used to generate a hotspot map. Third, to define high-priority restoration sites, each map was normalized and overlaid. As a result, three northern regions roads and two southern regions roads of Jeju Island were defined as high-priority restoration sites. Regarding Random Forest modeling, in the case of environmental variables, The importace was found to be a lot in the order of distance from the Oreum, elevation, distance from forest edge(outside) and distance from waterbody. The AUC(Area under the curve) value, which means discrimination capacity, was found to be 0.973 and support the statistical accuracy of prediction result. As a result of predicting the habitat of C. pygargus, it was found to be mainly distributed in forests, agricultural lands, and grasslands, indicating that it supported the results of previous studies.

Effect of Areal Mean Rainfall Estimation Technique and Rainfall-Runoff Models on Flood Simulation in Samcheok Osipcheon(Riv.) Basin (면적 강우량 산정 기법과 강우-유출 모형이 삼척오십천 유역의 홍수 모의에 미치는 영향)

  • Lee, Hyeonji;Shin, Youngsub;Kang, Dongho;Kim, Byungsik
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.43 no.6
    • /
    • pp.775-784
    • /
    • 2023
  • In terms of flood management, it is necessary to analyze quantitative rainfall and runoff from a spatial and temporal perspective and to analyze runoff for heavy rainfall events that are concentrated within a short period of time. The simulation and analysis results of rainfall-runoff models vary depending on the type and input data. In particular, rainfall data is an important factor, so calculating areal mean rainfall is very important. In this study, the areal mean rainfall of the Samcheok Osipcheon(Riv.) watersheds located in the mountainous terrain was calculated using the Arithmetic Mean Method, Thiessen's Weighting Method, and the Isohyetal Method, and the rainfall-runoff results were compared by applying the distributional model S-RAT and the lumped model HEC-HMS. The results of the temporal transferability study showed that the combination of the distributional model and the Isohyetal Method had the best statistical performance with MAE of 64.62 m3/s, RMSE of 82.47 m3/s, and R2 and NSE of 0.9383 and 0.8547, respectively. It is considered that this study was properly analyzed because the peak flood volume occurrence time of the observed and simulated flows is within 1 hour. Therefore, the results of this study can be used for frequency analysis in the future, which can be used to improve the accuracy of simulating peak flood volume and peak flood occurrence time in mountainous watersheds with steep slopes.

Estimation of Frost Occurrence using Multi-Input Deep Learning (다중 입력 딥러닝을 이용한 서리 발생 추정)

  • Yongseok Kim;Jina Hur;Eung-Sup Kim;Kyo-Moon Shim;Sera Jo;Min-Gu Kang
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.26 no.1
    • /
    • pp.53-62
    • /
    • 2024
  • In this study, we built a model to estimate frost occurrence in South Korea using single-input deep learning and multi-input deep learning. Meteorological factors used as learning data included minimum temperature, wind speed, relative humidity, cloud cover, and precipitation. As a result of statistical analysis for each factor on days when frost occurred and days when frost did not occur, significant differences were found. When evaluating the frost occurrence models based on single-input deep learning and multi-input deep learning model, the model using both GRU and MLP was highest accuracy at 0.8774 on average. As a result, it was found that frost occurrence model adopting multi-input deep learning improved performance more than using MLP, LSTM, GRU respectively.

Analysis of the cause-specific proportional hazards model with missing covariates (누락된 공변량을 가진 원인별 비례위험모형의 분석)

  • Minjung Lee
    • The Korean Journal of Applied Statistics
    • /
    • v.37 no.2
    • /
    • pp.225-237
    • /
    • 2024
  • In the analysis of competing risks data, some of covariates may not be fully observed for some subjects. In such cases, excluding subjects with missing covariate values from the analysis may result in biased estimates and loss of efficiency. In this paper, we studied multiple imputation and the augmented inverse probability weighting method for regression parameter estimation in the cause-specific proportional hazards model with missing covariates. The performance of estimators obtained from multiple imputation and the augmented inverse probability weighting method is evaluated by simulation studies, which show that those methods perform well. Multiple imputation and the augmented inverse probability weighting method were applied to investigate significant risk factors for the risk of death from breast cancer and from other causes for breast cancer data with missing values for tumor size obtained from the Prostate, Lung, Colorectal, and Ovarian Cancer Screen Trial Study. Under the cause-specific proportional hazards model, the methods show that race, marital status, stage, grade, and tumor size are significant risk factors for breast cancer mortality, and stage has the greatest effect on increasing the risk of breast cancer death. Age at diagnosis and tumor size have significant effects on increasing the risk of other-cause death.

Estimation of Cerchar abrasivity index based on rock strength and petrological characteristics using linear regression and machine learning (선형회귀분석과 머신러닝을 이용한 암석의 강도 및 암석학적 특징 기반 세르샤 마모지수 추정)

  • Ju-Pyo Hong;Yun Seong Kang;Tae Young Ko
    • Journal of Korean Tunnelling and Underground Space Association
    • /
    • v.26 no.1
    • /
    • pp.39-58
    • /
    • 2024
  • Tunnel Boring Machines (TBM) use multiple disc cutters to excavate tunnels through rock. These cutters wear out due to continuous contact and friction with the rock, leading to decreased cutting efficiency and reduced excavation performance. The rock's abrasivity significantly affects cutter wear, with highly abrasive rocks causing more wear and reducing the cutter's lifespan. The Cerchar Abrasivity Index (CAI) is a key indicator for assessing rock abrasivity, essential for predicting disc cutter life and performance. This study aims to develop a new method for effectively estimating CAI using rock strength, petrological characteristics, linear regression, and machine learning. A database including CAI, uniaxial compressive strength, Brazilian tensile strength, and equivalent quartz content was created, with additional derived variables. Variables for multiple linear regression were selected considering statistical significance and multicollinearity, while machine learning model inputs were chosen based on variable importance. Among the machine learning prediction models, the Gradient Boosting model showed the highest predictive performance. Finally, the predictive performance of the multiple linear regression analysis and the Gradient Boosting model derived in this study were compared with the CAI prediction models of previous studies to validate the results of this research.

Empirical Estimation and Diurnal Patterns of Surface PM2.5 Concentration in Seoul Using GOCI AOD (GOCI AOD를 이용한 서울 지역 지상 PM2.5 농도의 경험적 추정 및 일 변동성 분석)

  • Kim, Sang-Min;Yoon, Jongmin;Moon, Kyung-Jung;Kim, Deok-Rae;Koo, Ja-Ho;Choi, Myungje;Kim, Kwang Nyun;Lee, Yun Gon
    • Korean Journal of Remote Sensing
    • /
    • v.34 no.3
    • /
    • pp.451-463
    • /
    • 2018
  • The empirical/statistical models to estimate the ground Particulate Matter ($PM_{2.5}$) concentration from Geostationary Ocean Color Imager (GOCI) Aerosol Optical Depth (AOD) product were developed and analyzed for the period of 2015 in Seoul, South Korea. In the model construction of AOD-$PM_{2.5}$, two vertical correction methods using the planetary boundary layer height and the vertical ratio of aerosol, and humidity correction method using the hygroscopic growth factor were applied to respective models. The vertical correction for AOD and humidity correction for $PM_{2.5}$ concentration played an important role in improving accuracy of overall estimation. The multiple linear regression (MLR) models with additional meteorological factors (wind speed, visibility, and air temperature) affecting AOD and $PM_{2.5}$ relationships were constructed for the whole year and each season. As a result, determination coefficients of MLR models were significantly increased, compared to those of empirical models. In this study, we analyzed the seasonal, monthly and diurnal characteristics of AOD-$PM_{2.5}$model. when the MLR model is seasonally constructed, underestimation tendency in high $PM_{2.5}$ cases for the whole year were improved. The monthly and diurnal patterns of observed $PM_{2.5}$ and estimated $PM_{2.5}$ were similar. The results of this study, which estimates surface $PM_{2.5}$ concentration using geostationary satellite AOD, are expected to be applicable to the future GK-2A and GK-2B.

Intelligent Optimal Route Planning Based on Context Awareness (상황인식 기반 지능형 최적 경로계획)

  • Lee, Hyun-Jung;Chang, Yong-Sik
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.117-137
    • /
    • 2009
  • Recently, intelligent traffic information systems have enabled people to forecast traffic conditions before hitting the road. These convenient systems operate on the basis of data reflecting current road and traffic conditions as well as distance-based data between locations. Thanks to the rapid development of ubiquitous computing, tremendous context data have become readily available making vehicle route planning easier than ever. Previous research in relation to optimization of vehicle route planning merely focused on finding the optimal distance between locations. Contexts reflecting the road and traffic conditions were then not seriously treated as a way to resolve the optimal routing problems based on distance-based route planning, because this kind of information does not have much significant impact on traffic routing until a a complex traffic situation arises. Further, it was also not easy to take into full account the traffic contexts for resolving optimal routing problems because predicting the dynamic traffic situations was regarded a daunting task. However, with rapid increase in traffic complexity the importance of developing contexts reflecting data related to moving costs has emerged. Hence, this research proposes a framework designed to resolve an optimal route planning problem by taking full account of additional moving cost such as road traffic cost and weather cost, among others. Recent technological development particularly in the ubiquitous computing environment has facilitated the collection of such data. This framework is based on the contexts of time, traffic, and environment, which addresses the following issues. First, we clarify and classify the diverse contexts that affect a vehicle's velocity and estimates the optimization of moving cost based on dynamic programming that accounts for the context cost according to the variance of contexts. Second, the velocity reduction rate is applied to find the optimal route (shortest path) using the context data on the current traffic condition. The velocity reduction rate infers to the degree of possible velocity including moving vehicles' considerable road and traffic contexts, indicating the statistical or experimental data. Knowledge generated in this papercan be referenced by several organizations which deal with road and traffic data. Third, in experimentation, we evaluate the effectiveness of the proposed context-based optimal route (shortest path) between locations by comparing it to the previously used distance-based shortest path. A vehicles' optimal route might change due to its diverse velocity caused by unexpected but potential dynamic situations depending on the road condition. This study includes such context variables as 'road congestion', 'work', 'accident', and 'weather' which can alter the traffic condition. The contexts can affect moving vehicle's velocity on the road. Since these context variables except for 'weather' are related to road conditions, relevant data were provided by the Korea Expressway Corporation. The 'weather'-related data were attained from the Korea Meteorological Administration. The aware contexts are classified contexts causing reduction of vehicles' velocity which determines the velocity reduction rate. To find the optimal route (shortest path), we introduced the velocity reduction rate in the context for calculating a vehicle's velocity reflecting composite contexts when one event synchronizes with another. We then proposed a context-based optimal route (shortest path) algorithm based on the dynamic programming. The algorithm is composed of three steps. In the first initialization step, departure and destination locations are given, and the path step is initialized as 0. In the second step, moving costs including composite contexts into account between locations on path are estimated using the velocity reduction rate by context as increasing path steps. In the third step, the optimal route (shortest path) is retrieved through back-tracking. In the provided research model, we designed a framework to account for context awareness, moving cost estimation (taking both composite and single contexts into account), and optimal route (shortest path) algorithm (based on dynamic programming). Through illustrative experimentation using the Wilcoxon signed rank test, we proved that context-based route planning is much more effective than distance-based route planning., In addition, we found that the optimal solution (shortest paths) through the distance-based route planning might not be optimized in real situation because road condition is very dynamic and unpredictable while affecting most vehicles' moving costs. For further study, while more information is needed for a more accurate estimation of moving vehicles' costs, this study still stands viable in the applications to reduce moving costs by effective route planning. For instance, it could be applied to deliverers' decision making to enhance their decision satisfaction when they meet unpredictable dynamic situations in moving vehicles on the road. Overall, we conclude that taking into account the contexts as a part of costs is a meaningful and sensible approach to in resolving the optimal route problem.

The significance of corpus callosal size in the estimation of neurologically abnormal infants (신경학적인 결함이 있었던 영아의 예후 판단에서 뇌량 크기의 중요성)

  • Yu, Seung Taek;Lee, Chang Woo
    • Clinical and Experimental Pediatrics
    • /
    • v.51 no.11
    • /
    • pp.1205-1210
    • /
    • 2008
  • Purpose : The development of the corpus callosum occupies the entire period of cerebral formation. The myelination pattern on magnetic resonance imaging (MRI) is very useful to evaluate neurologic development and to predict neurologic outcome in high risk infants. The thickness of the corpus callosum is believed to depend on the myelination process. It is possible to calculate the length and thickness of the corpus callosum on MRI. Thus, we can quantitatively evaluate the development of the corpus callosum. We investigated the clinical significance of measuring various portions of the corpus callosum in neonate with neurologic disorders such as hypoxic brain damage and seizure disorder. Methods : Forty-two neonates were evaluated by brain MRI. We measured the size of the genu, body, transitional zone, splenium, and length of the corpus callosum. Each measurement was divided by the total length of the corpus callosum to obtain its corrected size. The ratio of corpus callosal length and the anteroposterior diameter of the brain was also measured. Results : There was no statistical significance in the sample size of each part of the corpus callosum. However, the corrected size or the ratio of body of the corpus callosum correlated with periventricular leukomalacia and hypoxic ischemic encephalopathy. Conclusion : The abnormal size of the corpus callosum showed a good correlation with periventricular leukomalacia and hypoxic ischemic encephalopathy in neonates. We can predict clinical neurological problems by estimation of the corpus callosum in the neonatal period.

Developing a Traffic Accident Prediction Model for Freeways (고속도로 본선에서의 교통사고 예측모형 개발)

  • Mun, Sung-Ra;Lee, Young-Ihn;Lee, Soo-Beom
    • Journal of Korean Society of Transportation
    • /
    • v.30 no.2
    • /
    • pp.101-116
    • /
    • 2012
  • Accident prediction models have been utilized to predict accident possibilities in existing or projected freeways and to evaluate programs or policies for improving safety. In this study, a traffic accident prediction model for freeways was developed for the above purposes. When selecting variables for the model, the highest priority was on the ease of both collecting data and applying them into the model. The dependent variable was set as the number of total accidents and the number of accidents including casualties in the unit of IC(or JCT). As a result, two models were developed; the overall accident model and the casualty-related accident model. The error structure adjusted to each model was the negative binomial distribution and the Poisson distribution, respectively. Among the two models, a more appropriate model was selected by statistical estimation. Major nine national freeways were selected and five-year dada of 2003~2007 were utilized. Explanatory variables should take on either a predictable value such as traffic volumes or a fixed value with respect to geometric conditions. As a result of the Maximum Likelihood estimation, significant variables of the overall accident model were found to be the link length between ICs(or JCTs), the daily volumes(AADT), and the ratio of bus volume to the number of curved segments between ICs(or JCTs). For the casualty-related accident model, the link length between ICs(or JCTs), the daily volumes(AADT), and the ratio of bus volumes had a significant impact on the accident. The likelihood ratio test was conducted to verify the spatial and temporal transferability for estimated parameters of each model. It was found that the overall accident model could be transferred only to the road with four or more than six lanes. On the other hand, the casualty-related accident model was transferrable to every road and every time period. In conclusion, the model developed in this study was able to be extended to various applications to establish future plans and evaluate policies.

Development of the Model for Total Quality Management and Cost of Quality using Activity Based Costing in the Hospital (병원의 활동기준원가를 이용한 총체적 질관리 모형 및 질비용 산출 모형 개발)

  • 조우현;전기홍;이해종;박은철;김병조;김보경;이상규
    • Health Policy and Management
    • /
    • v.11 no.2
    • /
    • pp.141-168
    • /
    • 2001
  • Healthcare service organizations can apply the cost of quality(COQ) model as a method to evaluate a service quality improvement project such as Total Quality Management (TQM). COQ model has been used to quantify and evaluate the efficiency and effectiveness of TQM project through estimation between cost and benefit in intervention for a quality Improvement to provide satisfied services for a customer, and to identify a non value added process. For estimating cost of quality, We used activities and activity costs based on Activity Based Costing(ABC) system. These procedures let the researchers know whether the process is value-added by each activity, and identify a process to require improvement in TQM project. Through the series of procedures, health care organizations are service organizations can identify a problem in their quality improvement programs, solve the problem, and improve their quality of care for their costumers with optimized cost. The study subject was a quality improvement program of the department of radiology department in a hospital with n bed sizes in Metropolitan Statistical Area (MSA). The principal source of data for developing the COQ model was total cases of retaking shots for diagnoses during five months period from December of the 1998 to April of the 1999 in the department. First of the procedures, for estimating activity based cost of the department of diagnostic radiology, the researchers analyzed total department health insurance claims to identify activities and activity costs using one year period health insurance claims from September of the 1998 to August of the 1999. COQ model in this study applied Simpson & Multher's COQ(SM's COQ) model, and SM's COQ model divided cost of quality into failure cost with external and internal failure cost, and evaluation/prevention cost. The researchers identified contents for cost of quality, defined activities and activity costs for each content with the SM's COQ model, and finally made the formula for estimating activity costs relating to implementing service quality improvement program. The results from the formula for estimating cost of quality were following: 1. The reasons for retaking shots were largely classified into technique, appliances, patients, quality management, non-appliances, doctors, and unclassified. These classifications by reasons were allocated into each office doing re-taking shots. Therefore, total retaking shots categorized by reasons and offices, the researchers identified internal and external failure costs based on these categories. 2. The researchers have developed cost of quality (COQ) model, identified activities by content for cost of quality, assessed activity driving factors and activity contribution rate, and calculated total cost by each content for cost for quality, except for activity cost. 3. According to estimation of cost of quality for retaking shots in department of diagnostic radiology, the failure cost was ₩35,880, evaluation/preventive cost was ₩72,521, two times as much as failure cost. The proportion between internal failure cost and external failure cost in failure cost is similar. The study cannot identify trends on input cost and quality improving in cost of qualify over the time, because the study employs cross-sectional design. Even with this limitation, results of this study are much meaningful. This study shows possibility to evaluate value on the process of TQM subjects using activities and activity costs by ABC system, and this study can objectively evaluate quality improvement program through quantitative comparing input costs with marginal benefits in quality improvement.

  • PDF