• Title/Summary/Keyword: Grid Data

Search Result 2,310, Processing Time 0.037 seconds

Prediction of Urban Flood Extent by LSTM Model and Logistic Regression (LSTM 모형과 로지스틱 회귀를 통한 도시 침수 범위의 예측)

  • Kim, Hyun Il;Han, Kun Yeun;Lee, Jae Yeong
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.40 no.3
    • /
    • pp.273-283
    • /
    • 2020
  • Because of climate change, the occurrence of localized and heavy rainfall is increasing. It is important to predict floods in urban areas that have suffered inundation in the past. For flood prediction, not only numerical analysis models but also machine learning-based models can be applied. The LSTM (Long Short-Term Memory) neural network used in this study is appropriate for sequence data, but it demands a lot of data. However, rainfall that causes flooding does not appear every year in a single urban basin, meaning it is difficult to collect enough data for deep learning. Therefore, in addition to the rainfall observed in the study area, the observed rainfall in another urban basin was applied in the predictive model. The LSTM neural network was used for predicting the total overflow, and the result of the SWMM (Storm Water Management Model) was applied as target data. The prediction of the inundation map was performed by using logistic regression; the independent variable was the total overflow and the dependent variable was the presence or absence of flooding in each grid. The dependent variable of logistic regression was collected through the simulation results of a two-dimensional flood model. The input data of the two-dimensional flood model were the overflow at each manhole calculated by the SWMM. According to the LSTM neural network parameters, the prediction results of total overflow were compared. Four predictive models were used in this study depending on the parameter of the LSTM. The average RMSE (Root Mean Square Error) for verification and testing was 1.4279 ㎥/s, 1.0079 ㎥/s for the four LSTM models. The minimum RMSE of the verification and testing was calculated as 1.1655 ㎥/s and 0.8797 ㎥/s. It was confirmed that the total overflow can be predicted similarly to the SWMM simulation results. The prediction of inundation extent was performed by linking the logistic regression with the results of the LSTM neural network, and the maximum area fitness was 97.33 % when more than 0.5 m depth was considered. The methodology presented in this study would be helpful in improving urban flood response based on deep learning methodology.

A Study on the Selection of Parameter Values of FUSION Software for Improving Airborne LiDAR DEM Accuracy in Forest Area (산림지역에서의 LiDAR DEM 정확도 향상을 위한 FUSION 패러미터 선정에 관한 연구)

  • Cho, Seungwan;Park, Joowon
    • Journal of Korean Society of Forest Science
    • /
    • v.106 no.3
    • /
    • pp.320-329
    • /
    • 2017
  • This study aims to evaluate whether the accuracy of LiDAR DEM is affected by the changes of the five input levels ('1','3','5','7' and '9') of median parameter ($F_{md}$), mean parameter ($F_{mn}$) of the Filtering Algorithm (FA) in the GroundFilter module and median parameter ($I_{md}$), mean parameter ($I_{mn}$) of the Interpolation Algorithm (IA) in the GridSurfaceCreate module of the FUSION in order to present the combination of parameter levels producing the most accurate LiDAR DEM. The accuracy is measured by the residuals calculated by difference between the field elevation values and their corresponding DEM elevation values. A multi-way ANOVA is used to statistically examine whether there are effects of parameter level changes on the means of the residuals. The Tukey HSD is conducted as a post-hoc test. The results of the multi- way ANOVA test show that the changes in the levels of $F_{md}$, $F_{mn}$, $I_{mn}$ have significant effects on the DEM accuracy with the significant interaction effect between $F_{md}$ and $F_{mn}$. Therefore, the level of $F_{md}$, $F_{mn}$, and the interaction between two variables are considered to be factors affecting the accuracy of LiDAR DEM as well as the level of $I_{mn}$. As the results of the Tukey HSD test on the combination levels of $F_{md}{\ast}F_{mn}$, the mean of residuals of the '$9{\ast}3$' combination provides the highest accuracy while the '$1{\ast}1$' combination provides the lowest one. Regarding $I_{mn}$ levels, the mean of residuals of the both '3' and '1' provides the highest accuracy. This study can contribute to improve the accuracy of the forest attributes as well as the topographic information extracted from the LiDAR data.

Prediction of Seabed Topography Change Due to Construction of Offshore Wind Power Structures in the West-Southern Sea of Korea (서남해에서 해상풍력구조물의 건설에 의한 해저지형의 변화예측)

  • Jeong, Seung Myung;Kwon, Kyung Hwan;Lee, Jong Sup;Park, Il Heum
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.31 no.6
    • /
    • pp.423-433
    • /
    • 2019
  • In order to predict the seabed topography change due to the construction of offshore wind power structures in the west-southern sea of Korea, field observations for tides, tidal currents, suspended sediment concentrations and seabed sediments were carried out at the same time. These data could be used for numerical simulation. In numerical experiments, the empirical constants for the suspended sediment flux were determined by the trial and error method. When a concentration distribution factor was 0.1 and a proportional constant was 0.05 in the suspended sediment equilibrium concentration formulae, the calculated suspended sediment concentrations were reasonably similar with the observed ones. Also, it was appropriate for the open boundary conditions of the suspended sediment when the south-east boundary corner was 11.0 times, the south-west was 0.5 times, the westnorth 1.0 times, the north-west was 1.0 times and the north-east was 1.0 times, respectively, using the time series of the observed suspended sediment concentrations. In this case, the depth change was smooth and not intermittent around the open boundaries. From these calibrations, the annual water depth change before and after construction of the offshore wind power structures was shown under 1 cm. The reason was that the used numerical model for the large scale grid could not reproduce a local scour phenomenon and they showed almost no significant velocity change over ± 2 cm/s because the jacket structures with small size diameter, about 1 m, were a water-permeable. Therefore, it was natural that there was a slight change on seabed topography in the study area.

Evaluation of Importance and Performance for Students and Employees about Sanitary Characteristics for High School Foodservice in Busan (부산지역 고등학교 학생과 급식종사자의 급식위생에 대한 중요도와 수행도 평가)

  • Kim, So-Hee
    • Journal of the Korean Society of Food Science and Nutrition
    • /
    • v.34 no.9
    • /
    • pp.1414-1426
    • /
    • 2005
  • The evaluation of students and employees on the importance and the performance for sanitary characteristics in high school foodservices, was investigated. The questionnaires were administered to 379 students and 141 employees in 13 high school, in Busan and then the data evaluated by 5 scales method of Likert were statistically analyzed. The mean scores of the importance and the performance evaluated by students (4.26/3.24) were significantly (p<0.01) lower than those by employees (4.71/4.51). Both the students and the employees percepted that, among sanitary characteristics, cleanliness of meals was most important. The student assessed the performance of withdrawal of plate waste as lowest scores, however the employees assessed student's hygiene as lowest scores, among sanitary characteristics. The gap score of the student (-1.02) between the importance and performance for sanitary factor was higher than that of the employees (-0.02) in high school foodservice. The importance grid of students and employees revealed that the items of tray cleanliness, dining table cleanliness, restaurant cleanliness, and handwashing before serving were high scores to the students, but low scores to the employees. The performance grid of students and employees revealed that the items of tray cleanliness, dining table cleanliness, restaurant cleanliness, the sanitation of treatment process of Plate waste, cleanliness of utensils of platewaste, not touch utensils before serving, handwashing before serving, handwashing before eating and not touch inside of tray were low scores to both the students and the employees. Therefore, it is suggested that the sanitary operations for dining place, treatment process of plate waste and the student's hygiene might have to increase in high school foodservice.

Application of Flux Average Discharge Equation to Assess the Submarine Fresh Groundwater Discharge in a Coastal Aquifer (연안 대수층의 해저 담지하수 유출량 산정을 위한 유량 평균 유출량 방정식의 적용)

  • Il Hwan Kim;Min-Gyu Kim;Il-Moon Chung;Gyo-Cheol Jeong;Sunwoo Chang
    • The Journal of Engineering Geology
    • /
    • v.33 no.1
    • /
    • pp.105-119
    • /
    • 2023
  • Water supply is decreasing due to climate change, and coastal and island regions are highly dependent on groundwater, reducing the amount of available water. For sustainable water supply in coastal and island regions, it is necessary to accurately diagnose the current condition and efficiently distribute and manage water. For a precise analysis of the groundwater flow in the coastal island region, submarine fresh groundwater discharge was calculated for the Seongsan basin in the eastern part of Jeju Island. Two methods were used to estimate the thickness of the fresh groundwater. One method employed vertical interpolation of measured electrical conductivity in a multi depth monitoring well; the other used theoretical Ghyben-Herzberg ratio. The value using the Ghyben-Herzberg ratio makes it impossible to accurately estimate the changing salt-saltwater interface, and the value analyzed by electrical conductivity can represent the current state of the freshwater-saltwater interface. Observed parameter was distributed on a virtual grid. The average of submarine fresh groundwater discharge fluxes for the virtual grid was determined as the watershed's representative flux. The submarine fresh groundwater discharge and flux distribution by year were also calculated at the basin scale. The method using electrical conductivity estimated the submarine fresh groundwater discharge from 2018 to 2020 to be 6.27 × 106 m3/year; the method using the Ghyben-Herzberg ratio estimated a discharge of 10.87 × 106 m3/year. The results presented in this study can be used as basis data for policies that determine sustainable water supply by using precise water budget analysis in coastal and island areas.

Development of Sentiment Analysis Model for the hot topic detection of online stock forums (온라인 주식 포럼의 핫토픽 탐지를 위한 감성분석 모형의 개발)

  • Hong, Taeho;Lee, Taewon;Li, Jingjing
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.1
    • /
    • pp.187-204
    • /
    • 2016
  • Document classification based on emotional polarity has become a welcomed emerging task owing to the great explosion of data on the Web. In the big data age, there are too many information sources to refer to when making decisions. For example, when considering travel to a city, a person may search reviews from a search engine such as Google or social networking services (SNSs) such as blogs, Twitter, and Facebook. The emotional polarity of positive and negative reviews helps a user decide on whether or not to make a trip. Sentiment analysis of customer reviews has become an important research topic as datamining technology is widely accepted for text mining of the Web. Sentiment analysis has been used to classify documents through machine learning techniques, such as the decision tree, neural networks, and support vector machines (SVMs). is used to determine the attitude, position, and sensibility of people who write articles about various topics that are published on the Web. Regardless of the polarity of customer reviews, emotional reviews are very helpful materials for analyzing the opinions of customers through their reviews. Sentiment analysis helps with understanding what customers really want instantly through the help of automated text mining techniques. Sensitivity analysis utilizes text mining techniques on text on the Web to extract subjective information in the text for text analysis. Sensitivity analysis is utilized to determine the attitudes or positions of the person who wrote the article and presented their opinion about a particular topic. In this study, we developed a model that selects a hot topic from user posts at China's online stock forum by using the k-means algorithm and self-organizing map (SOM). In addition, we developed a detecting model to predict a hot topic by using machine learning techniques such as logit, the decision tree, and SVM. We employed sensitivity analysis to develop our model for the selection and detection of hot topics from China's online stock forum. The sensitivity analysis calculates a sentimental value from a document based on contrast and classification according to the polarity sentimental dictionary (positive or negative). The online stock forum was an attractive site because of its information about stock investment. Users post numerous texts about stock movement by analyzing the market according to government policy announcements, market reports, reports from research institutes on the economy, and even rumors. We divided the online forum's topics into 21 categories to utilize sentiment analysis. One hundred forty-four topics were selected among 21 categories at online forums about stock. The posts were crawled to build a positive and negative text database. We ultimately obtained 21,141 posts on 88 topics by preprocessing the text from March 2013 to February 2015. The interest index was defined to select the hot topics, and the k-means algorithm and SOM presented equivalent results with this data. We developed a decision tree model to detect hot topics with three algorithms: CHAID, CART, and C4.5. The results of CHAID were subpar compared to the others. We also employed SVM to detect the hot topics from negative data. The SVM models were trained with the radial basis function (RBF) kernel function by a grid search to detect the hot topics. The detection of hot topics by using sentiment analysis provides the latest trends and hot topics in the stock forum for investors so that they no longer need to search the vast amounts of information on the Web. Our proposed model is also helpful to rapidly determine customers' signals or attitudes towards government policy and firms' products and services.

Wind Corridor Analysis and Climate Evaluation with Biotop Map and Airborne LiDAR Data (비오톱 지도와 항공라이다 자료를 이용한 바람통로 분석 및 기후평가)

  • Kim, Yeon-Mee;An, Seung-Man;Moon, Soo-Young;Kim, Hyeon-Soo;Jang, Dae-Hee
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.40 no.6
    • /
    • pp.148-160
    • /
    • 2012
  • The main purpose of this paper is to deliver a climate analysis and evaluation method based on GIS by using airborne LiDAR data and Biotop type map and to provide spatial information of climate analysis and evaluation based on Biotop type Map. At first stage, the area, slope, slope length, surface, wind corridor function and width, and obstacle factors were analyzed to obtain cold/fresh air production and wind corridor evaluation. In addition, climate evaluation was derived from those two results in the second stage. Airborne LiDAR data are useful in wind corridor analysis during the study. Correlation analysis results show that ColdAir_GRD grade was highly correlated with Surface_GRD (-0.967461139) and WindCorridor_ GRD was highly correlated with Function_GRD (-0.883883476) and Obstacle_GRD (-0.834057656). Climate Evaluation GRID was highly correlated with WindCorridor_GRD (0.927554516) than ColdAir_GRD (0.855051646). Visual validations of climate analysis and evaluation results were performed by using aerial ortho-photo image, which shows that the climate evaluation results were well related with in-situ condition. At the end, we applied climate analysis and evaluation by using Biotop map and airborne LiDAR data in Gwangmyung-Shiheung City, candidate for the Bogeumjari Housing District. The results show that the aerial percentile of the 1st Grade is 18.5%, 2nd Grade is 18.2%, 3rd Grade is 30.7%, 4th Grade is 25.2%, and 5th Grade is 7.4%. This study process provided both the spatial analysis and evaluation of climate information and statistics on behalf of each Biotop type.

Comparison of Forest Carbon Stocks Estimation Methods Using Forest Type Map and Landsat TM Satellite Imagery (임상도와 Landsat TM 위성영상을 이용한 산림탄소저장량 추정 방법 비교 연구)

  • Kim, Kyoung-Min;Lee, Jung-Bin;Jung, Jaehoon
    • Korean Journal of Remote Sensing
    • /
    • v.31 no.5
    • /
    • pp.449-459
    • /
    • 2015
  • The conventional National Forest Inventory(NFI)-based forest carbon stock estimation method is suitable for national-scale estimation, but is not for regional-scale estimation due to the lack of NFI plots. In this study, for the purpose of regional-scale carbon stock estimation, we created grid-based forest carbon stock maps using spatial ancillary data and two types of up-scaling methods. Chungnam province was chosen to represent the study area and for which the $5^{th}$ NFI (2006~2009) data was collected. The first method (method 1) selects forest type map as ancillary data and uses regression model for forest carbon stock estimation, whereas the second method (method 2) uses satellite imagery and k-Nearest Neighbor(k-NN) algorithm. Additionally, in order to consider uncertainty effects, the final AGB carbon stock maps were generated by performing 200 iterative processes with Monte Carlo simulation. As a result, compared to the NFI-based estimation(21,136,911 tonC), the total carbon stock was over-estimated by method 1(22,948,151 tonC), but was under-estimated by method 2(19,750,315 tonC). In the paired T-test with 186 independent data, the average carbon stock estimation by the NFI-based method was statistically different from method2(p<0.01), but was not different from method1(p>0.01). In particular, by means of Monte Carlo simulation, it was found that the smoothing effect of k-NN algorithm and mis-registration error between NFI plots and satellite image can lead to large uncertainty in carbon stock estimation. Although method 1 was found suitable for carbon stock estimation of forest stands that feature heterogeneous trees in Korea, satellite-based method is still in demand to provide periodic estimates of un-investigated, large forest area. In these respects, future work will focus on spatial and temporal extent of study area and robust carbon stock estimation with various satellite images and estimation methods.

Estimation of Fractional Urban Tree Canopy Cover through Machine Learning Using Optical Satellite Images (기계학습을 이용한 광학 위성 영상 기반의 도시 내 수목 피복률 추정)

  • Sejeong Bae ;Bokyung Son ;Taejun Sung ;Yeonsu Lee ;Jungho Im ;Yoojin Kang
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.5_3
    • /
    • pp.1009-1029
    • /
    • 2023
  • Urban trees play a vital role in urban ecosystems,significantly reducing impervious surfaces and impacting carbon cycling within the city. Although previous research has demonstrated the efficacy of employing artificial intelligence in conjunction with airborne light detection and ranging (LiDAR) data to generate urban tree information, the availability and cost constraints associated with LiDAR data pose limitations. Consequently, this study employed freely accessible, high-resolution multispectral satellite imagery (i.e., Sentinel-2 data) to estimate fractional tree canopy cover (FTC) within the urban confines of Suwon, South Korea, employing machine learning techniques. This study leveraged a median composite image derived from a time series of Sentinel-2 images. In order to account for the diverse land cover found in urban areas, the model incorporated three types of input variables: average (mean) and standard deviation (std) values within a 30-meter grid from 10 m resolution of optical indices from Sentinel-2, and fractional coverage for distinct land cover classes within 30 m grids from the existing level 3 land cover map. Four schemes with different combinations of input variables were compared. Notably, when all three factors (i.e., mean, std, and fractional cover) were used to consider the variation of landcover in urban areas(Scheme 4, S4), the machine learning model exhibited improved performance compared to using only the mean of optical indices (Scheme 1). Of the various models proposed, the random forest (RF) model with S4 demonstrated the most remarkable performance, achieving R2 of 0.8196, and mean absolute error (MAE) of 0.0749, and a root mean squared error (RMSE) of 0.1022. The std variable exhibited the highest impact on model outputs within the heterogeneous land covers based on the variable importance analysis. This trained RF model with S4 was then applied to the entire Suwon region, consistently delivering robust results with an R2 of 0.8702, MAE of 0.0873, and RMSE of 0.1335. The FTC estimation method developed in this study is expected to offer advantages for application in various regions, providing fundamental data for a better understanding of carbon dynamics in urban ecosystems in the future.

Application of Support Vector Regression for Improving the Performance of the Emotion Prediction Model (감정예측모형의 성과개선을 위한 Support Vector Regression 응용)

  • Kim, Seongjin;Ryoo, Eunchung;Jung, Min Kyu;Kim, Jae Kyeong;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.185-202
    • /
    • 2012
  • .Since the value of information has been realized in the information society, the usage and collection of information has become important. A facial expression that contains thousands of information as an artistic painting can be described in thousands of words. Followed by the idea, there has recently been a number of attempts to provide customers and companies with an intelligent service, which enables the perception of human emotions through one's facial expressions. For example, MIT Media Lab, the leading organization in this research area, has developed the human emotion prediction model, and has applied their studies to the commercial business. In the academic area, a number of the conventional methods such as Multiple Regression Analysis (MRA) or Artificial Neural Networks (ANN) have been applied to predict human emotion in prior studies. However, MRA is generally criticized because of its low prediction accuracy. This is inevitable since MRA can only explain the linear relationship between the dependent variables and the independent variable. To mitigate the limitations of MRA, some studies like Jung and Kim (2012) have used ANN as the alternative, and they reported that ANN generated more accurate prediction than the statistical methods like MRA. However, it has also been criticized due to over fitting and the difficulty of the network design (e.g. setting the number of the layers and the number of the nodes in the hidden layers). Under this background, we propose a novel model using Support Vector Regression (SVR) in order to increase the prediction accuracy. SVR is an extensive version of Support Vector Machine (SVM) designated to solve the regression problems. The model produced by SVR only depends on a subset of the training data, because the cost function for building the model ignores any training data that is close (within a threshold ${\varepsilon}$) to the model prediction. Using SVR, we tried to build a model that can measure the level of arousal and valence from the facial features. To validate the usefulness of the proposed model, we collected the data of facial reactions when providing appropriate visual stimulating contents, and extracted the features from the data. Next, the steps of the preprocessing were taken to choose statistically significant variables. In total, 297 cases were used for the experiment. As the comparative models, we also applied MRA and ANN to the same data set. For SVR, we adopted '${\varepsilon}$-insensitive loss function', and 'grid search' technique to find the optimal values of the parameters like C, d, ${\sigma}^2$, and ${\varepsilon}$. In the case of ANN, we adopted a standard three-layer backpropagation network, which has a single hidden layer. The learning rate and momentum rate of ANN were set to 10%, and we used sigmoid function as the transfer function of hidden and output nodes. We performed the experiments repeatedly by varying the number of nodes in the hidden layer to n/2, n, 3n/2, and 2n, where n is the number of the input variables. The stopping condition for ANN was set to 50,000 learning events. And, we used MAE (Mean Absolute Error) as the measure for performance comparison. From the experiment, we found that SVR achieved the highest prediction accuracy for the hold-out data set compared to MRA and ANN. Regardless of the target variables (the level of arousal, or the level of positive / negative valence), SVR showed the best performance for the hold-out data set. ANN also outperformed MRA, however, it showed the considerably lower prediction accuracy than SVR for both target variables. The findings of our research are expected to be useful to the researchers or practitioners who are willing to build the models for recognizing human emotions.