• Title/Summary/Keyword: input variable

Search Result 1,440, Processing Time 0.038 seconds

Assessment of water supply reliability in the Geum River Basin using univariate climate response functions: a case study for changing instreamflow managements (단변량 기후반응함수를 이용한 금강수계 이수안전도 평가: 하천유지유량 관리 변화를 고려한 사례연구)

  • Kim, Daeha;Choi, Si Jung;Jang, Su Hyung;Kang, Dae Hu
    • Journal of Korea Water Resources Association
    • /
    • v.56 no.12
    • /
    • pp.993-1003
    • /
    • 2023
  • Due to the increasing greenhouse gas emissions, the global mean temperature has risen by 1.1℃ compared to pre-industrial levels, and significant changes are expected in functioning of water supply systems. In this study, we assessed impacts of climate change and instreamflow management on water supply reliability in the Geum River basin, Korea. We proposed univariate climate response functions, where mean precipitation and potential evaporation were coupled as an explanatory variable, to assess impacts of climate stress on multiple water supply reliabilities. To this end, natural streamflows were generated in the 19 sub-basins with the conceptual GR6J model. Then, the simulated streamflows were input into the Water Evaluation And Planning (WEAP) model. The dynamic optimization by WEAP allowed us to assess water supply reliability against the 2020 water demand projections. Results showed that when minimizing the water shortage of the entire river basin under the 1991-2020 climate, water supply reliability was lowest in the Bocheongcheon among the sub-basins. In a scenario where the priority of instreamflow maintenance is adjusted to be the same as municipal and industrial water use, water supply reliability in the Bocheongcheon, Chogang, and Nonsancheon sub-basins significantly decreased. The stress tests with 325 sets of climate perturbations showed that water supply reliability in the three sub-basins considerably decreased under all the climate stresses, while the sub-basins connected to large infrastructures did not change significantly. When using the 2021-2050 climate projections with the stress test results, water supply reliability in the Geum River basin was expected to generally improve, but if the priority of instreamflow maintenance is increased, water shortage is expected to worsen in geographically isolated sub-basins. Here, we suggest that the climate response function can be established by a single explanatory variable to assess climate change impacts of many sub-basin's performance simultaneously.

A Time Series Graph based Convolutional Neural Network Model for Effective Input Variable Pattern Learning : Application to the Prediction of Stock Market (효과적인 입력변수 패턴 학습을 위한 시계열 그래프 기반 합성곱 신경망 모형: 주식시장 예측에의 응용)

  • Lee, Mo-Se;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.167-181
    • /
    • 2018
  • Over the past decade, deep learning has been in spotlight among various machine learning algorithms. In particular, CNN(Convolutional Neural Network), which is known as the effective solution for recognizing and classifying images or voices, has been popularly applied to classification and prediction problems. In this study, we investigate the way to apply CNN in business problem solving. Specifically, this study propose to apply CNN to stock market prediction, one of the most challenging tasks in the machine learning research. As mentioned, CNN has strength in interpreting images. Thus, the model proposed in this study adopts CNN as the binary classifier that predicts stock market direction (upward or downward) by using time series graphs as its inputs. That is, our proposal is to build a machine learning algorithm that mimics an experts called 'technical analysts' who examine the graph of past price movement, and predict future financial price movements. Our proposed model named 'CNN-FG(Convolutional Neural Network using Fluctuation Graph)' consists of five steps. In the first step, it divides the dataset into the intervals of 5 days. And then, it creates time series graphs for the divided dataset in step 2. The size of the image in which the graph is drawn is $40(pixels){\times}40(pixels)$, and the graph of each independent variable was drawn using different colors. In step 3, the model converts the images into the matrices. Each image is converted into the combination of three matrices in order to express the value of the color using R(red), G(green), and B(blue) scale. In the next step, it splits the dataset of the graph images into training and validation datasets. We used 80% of the total dataset as the training dataset, and the remaining 20% as the validation dataset. And then, CNN classifiers are trained using the images of training dataset in the final step. Regarding the parameters of CNN-FG, we adopted two convolution filters ($5{\times}5{\times}6$ and $5{\times}5{\times}9$) in the convolution layer. In the pooling layer, $2{\times}2$ max pooling filter was used. The numbers of the nodes in two hidden layers were set to, respectively, 900 and 32, and the number of the nodes in the output layer was set to 2(one is for the prediction of upward trend, and the other one is for downward trend). Activation functions for the convolution layer and the hidden layer were set to ReLU(Rectified Linear Unit), and one for the output layer set to Softmax function. To validate our model - CNN-FG, we applied it to the prediction of KOSPI200 for 2,026 days in eight years (from 2009 to 2016). To match the proportions of the two groups in the independent variable (i.e. tomorrow's stock market movement), we selected 1,950 samples by applying random sampling. Finally, we built the training dataset using 80% of the total dataset (1,560 samples), and the validation dataset using 20% (390 samples). The dependent variables of the experimental dataset included twelve technical indicators popularly been used in the previous studies. They include Stochastic %K, Stochastic %D, Momentum, ROC(rate of change), LW %R(Larry William's %R), A/D oscillator(accumulation/distribution oscillator), OSCP(price oscillator), CCI(commodity channel index), and so on. To confirm the superiority of CNN-FG, we compared its prediction accuracy with the ones of other classification models. Experimental results showed that CNN-FG outperforms LOGIT(logistic regression), ANN(artificial neural network), and SVM(support vector machine) with the statistical significance. These empirical results imply that converting time series business data into graphs and building CNN-based classification models using these graphs can be effective from the perspective of prediction accuracy. Thus, this paper sheds a light on how to apply deep learning techniques to the domain of business problem solving.

The Prediction of DEA based Efficiency Rating for Venture Business Using Multi-class SVM (다분류 SVM을 이용한 DEA기반 벤처기업 효율성등급 예측모형)

  • Park, Ji-Young;Hong, Tae-Ho
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.139-155
    • /
    • 2009
  • For the last few decades, many studies have tried to explore and unveil venture companies' success factors and unique features in order to identify the sources of such companies' competitive advantages over their rivals. Such venture companies have shown tendency to give high returns for investors generally making the best use of information technology. For this reason, many venture companies are keen on attracting avid investors' attention. Investors generally make their investment decisions by carefully examining the evaluation criteria of the alternatives. To them, credit rating information provided by international rating agencies, such as Standard and Poor's, Moody's and Fitch is crucial source as to such pivotal concerns as companies stability, growth, and risk status. But these types of information are generated only for the companies issuing corporate bonds, not venture companies. Therefore, this study proposes a method for evaluating venture businesses by presenting our recent empirical results using financial data of Korean venture companies listed on KOSDAQ in Korea exchange. In addition, this paper used multi-class SVM for the prediction of DEA-based efficiency rating for venture businesses, which was derived from our proposed method. Our approach sheds light on ways to locate efficient companies generating high level of profits. Above all, in determining effective ways to evaluate a venture firm's efficiency, it is important to understand the major contributing factors of such efficiency. Therefore, this paper is constructed on the basis of following two ideas to classify which companies are more efficient venture companies: i) making DEA based multi-class rating for sample companies and ii) developing multi-class SVM-based efficiency prediction model for classifying all companies. First, the Data Envelopment Analysis(DEA) is a non-parametric multiple input-output efficiency technique that measures the relative efficiency of decision making units(DMUs) using a linear programming based model. It is non-parametric because it requires no assumption on the shape or parameters of the underlying production function. DEA has been already widely applied for evaluating the relative efficiency of DMUs. Recently, a number of DEA based studies have evaluated the efficiency of various types of companies, such as internet companies and venture companies. It has been also applied to corporate credit ratings. In this study we utilized DEA for sorting venture companies by efficiency based ratings. The Support Vector Machine(SVM), on the other hand, is a popular technique for solving data classification problems. In this paper, we employed SVM to classify the efficiency ratings in IT venture companies according to the results of DEA. The SVM method was first developed by Vapnik (1995). As one of many machine learning techniques, SVM is based on a statistical theory. Thus far, the method has shown good performances especially in generalizing capacity in classification tasks, resulting in numerous applications in many areas of business, SVM is basically the algorithm that finds the maximum margin hyperplane, which is the maximum separation between classes. According to this method, support vectors are the closest to the maximum margin hyperplane. If it is impossible to classify, we can use the kernel function. In the case of nonlinear class boundaries, we can transform the inputs into a high-dimensional feature space, This is the original input space and is mapped into a high-dimensional dot-product space. Many studies applied SVM to the prediction of bankruptcy, the forecast a financial time series, and the problem of estimating credit rating, In this study we employed SVM for developing data mining-based efficiency prediction model. We used the Gaussian radial function as a kernel function of SVM. In multi-class SVM, we adopted one-against-one approach between binary classification method and two all-together methods, proposed by Weston and Watkins(1999) and Crammer and Singer(2000), respectively. In this research, we used corporate information of 154 companies listed on KOSDAQ market in Korea exchange. We obtained companies' financial information of 2005 from the KIS(Korea Information Service, Inc.). Using this data, we made multi-class rating with DEA efficiency and built multi-class prediction model based data mining. Among three manners of multi-classification, the hit ratio of the Weston and Watkins method is the best in the test data set. In multi classification problems as efficiency ratings of venture business, it is very useful for investors to know the class with errors, one class difference, when it is difficult to find out the accurate class in the actual market. So we presented accuracy results within 1-class errors, and the Weston and Watkins method showed 85.7% accuracy in our test samples. We conclude that the DEA based multi-class approach in venture business generates more information than the binary classification problem, notwithstanding its efficiency level. We believe this model can help investors in decision making as it provides a reliably tool to evaluate venture companies in the financial domain. For the future research, we perceive the need to enhance such areas as the variable selection process, the parameter selection of kernel function, the generalization, and the sample size of multi-class.

Estimation of river discharge using satellite-derived flow signals and artificial neural network model: application to imjin river (Satellite-derived flow 시그널 및 인공신경망 모형을 활용한 임진강 유역 유출량 산정)

  • Li, Li;Kim, Hyunglok;Jun, Kyungsoo;Choi, Minha
    • Journal of Korea Water Resources Association
    • /
    • v.49 no.7
    • /
    • pp.589-597
    • /
    • 2016
  • In this study, we investigated the use of satellite-derived flow (SDF) signals and a data-based model for the estimation of outflow for the river reach where in situ measurements are either completely unavailable or are difficult to access for hydraulic and hydrology analysis such as the upper basin of Imjin River. It has been demonstrated by many studies that the SDF signals can be used as the river width estimates and the correlation between SDF signals and river width is related to the shape of cross sections. To extract the nonlinear relationship between SDF signals and river outflow, Artificial Neural Network (ANN) model with SDF signals as its inputs were applied for the computation of flow discharge at Imjin Bridge located in Imjin River. 15 pixels were considered to extract SDF signals and Partial Mutual Information (PMI) algorithm was applied to identify the most relevant input variables among 150 candidate SDF signals (including 0~10 day lagged observations). The estimated discharges by ANN model were compared with the measured ones at Imjin Bridge gauging station and correlation coefficients of the training and validation were 0.86 and 0.72, respectively. It was found that if the 1 day previous discharge at Imjin bridge is considered as an input variable for ANN model, the correlation coefficients were improved to 0.90 and 0.83, respectively. Based on the results in this study, SDF signals along with some local measured data can play an useful role in river flow estimation and especially in flood forecasting for data-scarce regions as it can simulate the peak discharge and peak time of flood events with satisfactory accuracy.

An Intervention Study on Integration of Family Planning and Maternal/Infant Care Services in Rural Korea (가족계획과 모자보건 통합을 위한 조산원의 투입효과 분석 -서산지역의 개입연구 평가보고-)

  • Bang, Sook;Han, Seung-Hyun;Lee, Chung-Ja;Ahn, Moon-Young;Lee, In-Sook;Kim, Eun-Shil;Kim, Chong-Ho
    • Journal of Preventive Medicine and Public Health
    • /
    • v.20 no.1 s.21
    • /
    • pp.165-203
    • /
    • 1987
  • This project was a service-cum-research effort with a quasi-experimental study design to examine the health benefits of an integrated Family Planning (FP)/Maternal & Child health (MCH) Service approach that provides crucial factors missing in the present on-going programs. The specific objectives were: 1) To test the effectiveness of trained nurse/midwives (MW) assigned as change agents in the Health Sub-Center (HSC) to bring about the changes in the eight FP/MCH indicators, namely; (i)FP/MCH contacts between field workers and their clients (ii) the use of effective FP methods, (iii) the inter-birth interval and/or open interval, (iv) prenatal care by medically qualified personnel, (v) medically supervised deliveries, (vi) the rate of induced abortion, (vii) maternal and infant morbidity, and (viii) preinatal & infant mortality. 2) To measure the integrative linkage (contacts) between MW & HSC workers and between HSC and clients. 3) To examine the organizational or administrative factors influencing integrative linkage between health workers. Study design; The above objectives called for quasi-experimental design setting up a study and control area with and without a midwife. An active intervention program (FP/MCH minimum 'package' program) was conducted for a 2 year period from June 1982-July 1984 in Seosan County and 'before and after' surveys were conducted to measure the change. Service input; This study was undertaken by the Soonchunhyang University in collaboration with WHO. After a baseline survery in 1981, trained nurses/midwives were introduced into two health sub-centers in a rural setting (Seosan county) for a 2 year period from 1982 to 1984. A major service input was the establishment of midwifery services in the existing health delivery system with emphasis on nurse/midwife's role as the link between health workers (nurse aids) and village health workers, and the referral of risk patients to the private physician (OBGY specialist). An evaluation survey was made in August 1984 to assess the effectiveness of this alternative integrated approach in the study areas in comparison with the control area which had normal government services. Method of evaluation; a. In this study, the primary objective was first to examine to what extent the FP/MCH package program brought about changes in the pre-determined eight indicators (outcome and impact measures) and the following relationship was first analyzed; b. Nevertheless, this project did not automatically accept the assumption that if two or more activities were integrated, the results would automatically be better than a non-integrated or categorical program. There is a need to assess the 'integration process' itself within the package program. The process of integration was measured in terms of interactive linkages, or the quantity & quality of contacts between workers & clients and among workers. Intergrative linkages were hypothesized to be influenced by organizational factors at the HSC clinic level including HSC goals, sltrurture, authority, leadership style, resources, and personal characteristics of HSC staff. The extent or degree of integration, as measured by the intensity of integrative linkages, was in turn presumed to influence programme performance. Thus as indicated diagrammatically below, organizational factors constituted the independent variables, integration as the intervening variable and programme performance with respect to family planning and health services as the dependent variable: Concerning organizational factors, however, due to the limited number of HSCs (2 in the study area and 3 in the control area), they were studied by participatory observation of an anthropologist who was independent of the project. In this observation, we examined whether the assumed integration process actually occurred or not. If not, what were the constraints in producing an effective integration process. Summary of Findings; A) Program effects and impact 1. Effects on FP use: During this 2 year action period, FP acceptance increased from 58% in 1981 to 78% in 1984 in both the study and control areas. This increase in both areas was mainly due to the new family planning campaign driven by the Government for the same study period. Therefore, there was no increment of FP acceptance rate due to additional input of MW to the on-going FP program. But in the study area, quality aspects of FP were somewhat improved, having a better continuation rate of IUDs & pills and more use of effective Contraceptive methods in comparison with the control area. 2. Effects of use of MCH services: Between the study and control areas, however, there was a significant difference in maternal and child health care. For example, the coverage of prenatal care was increased from 53% for 1981 birth cohort to 75% for 1984 birth cohort in the study area. In the control area, the same increased from 41% (1981) to 65% (1984). It is noteworthy that almost two thirds of the recent birth cohort received prenatal care even in the control area, indicating that there is a growing demand of MCH care as the size of family norm becomes smaller 3. There has been a substantive increase in delivery care by medical professions in the study area, with an annual increase rate of 10% due to midwives input in the study areas. The project had about two times greater effect on postnatal care (68% vs. 33%) at delivery care(45.2% vs. 26.1%). 4. The study area had better reproductive efficiency (wanted pregancies with FP practice & healthy live births survived by one year old) than the control area, especially among women under 30 (14.1% vs. 9.6%). The proportion of women who preferred the 1st trimester for their first prenatal care rose significantly in the study area as compared to the control area (24% vs 13%). B) Effects on Interactive Linkage 1. This project made a contribution in making several useful steps in the direction of service integration, namely; i) The health workers have become familiar with procedures on how to work together with each other (especially with a midwife) in carrying out their work in FP/MCH and, ii) The health workers have gotten a feeling of the usefulness of family health records (statistical integration) in identifying targets in their own work and their usefulness in caring for family health. 2. On the other hand, because of a lack of required organizational factors, complete linkage was not obtained as the project intended. i) In regards to the government health worker's activities in terms of home visiting there was not much difference between the study & control areas though the MW did more home visiting than Government health workers. ii) In assessing the service performance of MW & health workers, the midwives balanced their workload between 40% FP, 40% MCH & 20% other activities (mainly immunization). However, $85{\sim}90%$ of the services provided by the health workers were other than FP/MCH, mainly for immunizations such as the encephalitis campaign. In the control area, a similar pattern was observed. Over 75% of their service was other than FP/MCH. Therefore, the pattern shows the health workers are a long way from becoming multipurpose workers even though the government is pushing in this direction. 3. Villagers were much more likely to visit the health sub-center clinic in the study area than in the control area (58% vs.31%) and for more combined care (45% vs.23%). C) Organization factors (admistrative integrative issues) 1. When MW (new workers with higher qualification) were introduced to HSC, it was noted that there were conflicts between the existing HSC workers (Nurse aids with less qualification than MW) and the MW for the beginning period of the project. The cause of the conflict was studied by an anthropologist and it was pointed out that these functional integration problems stemmed from the structural inadequacies of the health subcenter organization as indicated below; i) There is still no general consensus about the objectives and goals of the project between the project staff and the existing health workers. ii) There is no formal linkage between the responsibility of each member's job in the health sub-center. iii) There is still little chance for midwives to play a catalytic role or to establish communicative networks between workers in order to link various knowledge and skills to provide better FP/MCH services in the health sub-center. 2. Based on the above findings the project recommended to the County Chief (who has power to control the administrative staff and the technical staff in his county) the following ; i) In order to solve the conflicts between the individual roles and functions in performing health care activities, there must be goals agreed upon by both. ii) The health sub·center must function as an autonomous organization to undertake the integration health project. In order to do that, it is necessary to support administrative considerations, and to establish a communication system for supervision and to control of the health sub-centers. iii) The administrative organization, tentatively, must be organized to bind the health worker's midwive's and director's jobs by an organic relationship in order to achieve the integrative system under the leadership of health sub-center director. After submitting this observation report, there has been better understanding from frequent meetings & communication between HW/MW in FP/MCH work as the program developed. Lessons learned from the Seosan Project (on issues of FP/MCH integration in Korea); 1) A majority or about 80% of the couples are now practicing FP. As indicated by the study, there is a growing demand from clients for the health system to provide more MCH services than FP in order to maintain the achieved small size of family through FP practice. It is fortunate to see that the government is now formulating a MCH policy for the year 2,000 and revising MCH laws and regulations to emphasize more MCH care for achieving a small size family through family planning practice. 2) Goal consensus in FP/MCH shouBd be made among the health workers It administrators, especially to emphasize the need of care of 'wanted' child. But there is a long way to go to realize the 'real' integration of FP into MCH in Korea, unless there is a structural integration FP/MCH because a categorical FP is still first priority to reduce the rate of population growth for economic reasons but not yet for health/welfare reasons in practice. 3) There should be more financial allocation: (i) a midwife should be made available to help to promote the MCH program and coordinate services, (in) there should be a health sub·center director who can provide leadership training for managing the integrated program. There is a need for 'organizational support', if the decision of integration is made to obtain benefit from both FP & MCH. In other words, costs should be paid equally to both FP/MCH. The integration slogan itself, without the commitment of paying such costs, is powerless to advocate it. 4) Need of management training for middle level health personnel is more acute as the Government has already constructed 90 MCH centers attached to the County Health Center but without adequate manpower, facilities, and guidelines for integrating the work of both FP and MCH. 5) The local government still considers these MCH centers only as delivery centers to take care only of those visiting maternity cases. The MCH center should be a center for the managment of all pregnancies occurring in the community and the promotion of FP with a systematic and effective linkage of resources available in the county such as i.e. Village Health Worker, Community Health Practitioner, Health Sub-center Physicians & Health workers, Doctors and Midwives in MCH center, OBGY Specialists in clinics & hospitals as practiced by the Seosan project at primary health care level.

  • PDF

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.

Memory Organization for a Fuzzy Controller.

  • Jee, K.D.S.;Poluzzi, R.;Russo, B.
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1993.06a
    • /
    • pp.1041-1043
    • /
    • 1993
  • Fuzzy logic based Control Theory has gained much interest in the industrial world, thanks to its ability to formalize and solve in a very natural way many problems that are very difficult to quantify at an analytical level. This paper shows a solution for treating membership function inside hardware circuits. The proposed hardware structure optimizes the memoried size by using particular form of the vectorial representation. The process of memorizing fuzzy sets, i.e. their membership function, has always been one of the more problematic issues for the hardware implementation, due to the quite large memory space that is needed. To simplify such an implementation, it is commonly [1,2,8,9,10,11] used to limit the membership functions either to those having triangular or trapezoidal shape, or pre-definite shape. These kinds of functions are able to cover a large spectrum of applications with a limited usage of memory, since they can be memorized by specifying very few parameters ( ight, base, critical points, etc.). This however results in a loss of computational power due to computation on the medium points. A solution to this problem is obtained by discretizing the universe of discourse U, i.e. by fixing a finite number of points and memorizing the value of the membership functions on such points [3,10,14,15]. Such a solution provides a satisfying computational speed, a very high precision of definitions and gives the users the opportunity to choose membership functions of any shape. However, a significant memory waste can as well be registered. It is indeed possible that for each of the given fuzzy sets many elements of the universe of discourse have a membership value equal to zero. It has also been noticed that almost in all cases common points among fuzzy sets, i.e. points with non null membership values are very few. More specifically, in many applications, for each element u of U, there exists at most three fuzzy sets for which the membership value is ot null [3,5,6,7,12,13]. Our proposal is based on such hypotheses. Moreover, we use a technique that even though it does not restrict the shapes of membership functions, it reduces strongly the computational time for the membership values and optimizes the function memorization. In figure 1 it is represented a term set whose characteristics are common for fuzzy controllers and to which we will refer in the following. The above term set has a universe of discourse with 128 elements (so to have a good resolution), 8 fuzzy sets that describe the term set, 32 levels of discretization for the membership values. Clearly, the number of bits necessary for the given specifications are 5 for 32 truth levels, 3 for 8 membership functions and 7 for 128 levels of resolution. The memory depth is given by the dimension of the universe of the discourse (128 in our case) and it will be represented by the memory rows. The length of a world of memory is defined by: Length = nem (dm(m)+dm(fm) Where: fm is the maximum number of non null values in every element of the universe of the discourse, dm(m) is the dimension of the values of the membership function m, dm(fm) is the dimension of the word to represent the index of the highest membership function. In our case then Length=24. The memory dimension is therefore 128*24 bits. If we had chosen to memorize all values of the membership functions we would have needed to memorize on each memory row the membership value of each element. Fuzzy sets word dimension is 8*5 bits. Therefore, the dimension of the memory would have been 128*40 bits. Coherently with our hypothesis, in fig. 1 each element of universe of the discourse has a non null membership value on at most three fuzzy sets. Focusing on the elements 32,64,96 of the universe of discourse, they will be memorized as follows: The computation of the rule weights is done by comparing those bits that represent the index of the membership function, with the word of the program memor . The output bus of the Program Memory (μCOD), is given as input a comparator (Combinatory Net). If the index is equal to the bus value then one of the non null weight derives from the rule and it is produced as output, otherwise the output is zero (fig. 2). It is clear, that the memory dimension of the antecedent is in this way reduced since only non null values are memorized. Moreover, the time performance of the system is equivalent to the performance of a system using vectorial memorization of all weights. The dimensioning of the word is influenced by some parameters of the input variable. The most important parameter is the maximum number membership functions (nfm) having a non null value in each element of the universe of discourse. From our study in the field of fuzzy system, we see that typically nfm 3 and there are at most 16 membership function. At any rate, such a value can be increased up to the physical dimensional limit of the antecedent memory. A less important role n the optimization process of the word dimension is played by the number of membership functions defined for each linguistic term. The table below shows the request word dimension as a function of such parameters and compares our proposed method with the method of vectorial memorization[10]. Summing up, the characteristics of our method are: Users are not restricted to membership functions with specific shapes. The number of the fuzzy sets and the resolution of the vertical axis have a very small influence in increasing memory space. Weight computations are done by combinatorial network and therefore the time performance of the system is equivalent to the one of the vectorial method. The number of non null membership values on any element of the universe of discourse is limited. Such a constraint is usually non very restrictive since many controllers obtain a good precision with only three non null weights. The method here briefly described has been adopted by our group in the design of an optimized version of the coprocessor described in [10].

  • PDF

A Joint Application of DRASTIC and Numerical Groundwater Flow Model for The Assessment of Groundwater Vulnerability of Buyeo-Eup Area (DRASTIC 모델 및 지하수 수치모사 연계 적용에 의한 부여읍 일대의 지하수 오염 취약성 평가)

  • Lee, Hyun-Ju;Park, Eun-Gyu;Kim, Kang-Joo;Park, Ki-Hoon
    • Journal of Soil and Groundwater Environment
    • /
    • v.13 no.1
    • /
    • pp.77-91
    • /
    • 2008
  • In this study, we developed a technique of applying DRASTIC, which is the most widely used tool for estimation of groundwater vulnerability to the aqueous phase contaminant infiltrated from the surface, and a groundwater flow model jointly to assess groundwater contamination potential. The developed technique is then applied to Buyeo-eup area in Buyeo-gun, Chungcheongnam-do, Korea. The input thematic data of a depth to water required in DRASTIC model is known to be the most sensitive to the output while only a few observations at a few time schedules are generally available. To overcome this practical shortcoming, both steady-state and transient groundwater level distributions are simulated using a finite difference numerical model, MODFLOW. In the application for the assessment of groundwater vulnerability, it is found that the vulnerability results from the numerical simulation of a groundwater level is much more practical compared to cokriging methods. Those advantages are, first, the results from the simulation enable a practitioner to see the temporally comprehensive vulnerabilities. The second merit of the technique is that the method considers wide variety of engaging data such as field-observed hydrogeologic parameters as well as geographic relief. The depth to water generated through geostatistical methods in the conventional method is unable to incorporate temporally variable data, that is, the seasonal variation of a recharge rate. As a result, we found that the vulnerability out of both the geostatistical method and the steady-state groundwater flow simulation are in similar patterns. By applying the transient simulation results to DRASTIC model, we also found that the vulnerability shows sharp seasonal variation due to the change of groundwater recharge. The change of the vulnerability is found to be most peculiar during summer with the highest recharge rate and winter with the lowest. Our research indicates that numerical modeling can be a useful tool for temporal as well as spatial interpolation of the depth to water when the number of the observed data is inadequate for the vulnerability assessments through the conventional techniques.

Predicting Regional Soybean Yield using Crop Growth Simulation Model (작물 생육 모델을 이용한 지역단위 콩 수량 예측)

  • Ban, Ho-Young;Choi, Doug-Hwan;Ahn, Joong-Bae;Lee, Byun-Woo
    • Korean Journal of Remote Sensing
    • /
    • v.33 no.5_2
    • /
    • pp.699-708
    • /
    • 2017
  • The present study was to develop an approach for predicting soybean yield using a crop growth simulation model at the regional level where the detailed and site-specific information on cultivation management practices is not easily accessible for model input. CROPGRO-Soybean model included in Decision Support System for Agrotechnology Transfer (DSSAT) was employed for this study, and Illinois which is a major soybean production region of USA was selected as a study region. As a first step to predict soybean yield of Illinois using CROPGRO-Soybean model, genetic coefficients representative for each soybean maturity group (MG I~VI) were estimated through sowing date experiments using domestic and foreign cultivars with diverse maturity in Seoul National University Farm ($37.27^{\circ}N$, $126.99^{\circ}E$) for two years. The model using the representative genetic coefficients simulated the developmental stages of cultivars within each maturity group fairly well. Soybean yields for the grids of $10km{\times}10km$ in Illinois state were simulated from 2,000 to 2,011 with weather data under 18 simulation conditions including the combinations of three maturity groups, three seeding dates and two irrigation regimes. Planting dates and maturity groups were assigned differently to the three sub-regions divided longitudinally. The yearly state yields that were estimated by averaging all the grid yields simulated under non-irrigated and fully-Irrigated conditions showed a big difference from the statistical yields and did not explain the annual trend of yield increase due to the improved cultivation technologies. Using the grain yield data of 9 agricultural districts in Illinois observed and estimated from the simulated grid yield under 18 simulation conditions, a multiple regression model was constructed to estimate soybean yield at agricultural district level. In this model a year variable was also added to reflect the yearly yield trend. This model explained the yearly and district yield variation fairly well with a determination coefficients of $R^2=0.61$ (n = 108). Yearly state yields which were calculated by weighting the model-estimated yearly average agricultural district yield by the cultivation area of each agricultural district showed very close correspondence ($R^2=0.80$) to the yearly statistical state yields. Furthermore, the model predicted state yield fairly well in 2012 in which data were not used for the model construction and severe yield reduction was recorded due to drought.

Analysis of Empirical Multiple Linear Regression Models for the Production of PM2.5 Concentrations (PM2.5농도 산출을 위한 경험적 다중선형 모델 분석)

  • Choo, Gyo-Hwang;Lee, Kyu-Tae;Jeong, Myeong-Jae
    • Journal of the Korean earth science society
    • /
    • v.38 no.4
    • /
    • pp.283-292
    • /
    • 2017
  • In this study, the empirical models were established to estimate the concentrations of surface-level $PM_{2.5}$ over Seoul, Korea from 1 January 2012 to 31 December 2013. We used six different multiple linear regression models with aerosol optical thickness (AOT), ${\AA}ngstr{\ddot{o}}m$ exponents (AE) data from Moderate Resolution Imaging Spectroradiometer (MODIS) aboard Terra and Aqua satellites, meteorological data, and planetary boundary layer depth (PBLD) data. The results showed that $M_6$ was the best empirical model and AOT, AE, relative humidity (RH), wind speed, wind direction, PBLD, and air temperature data were used as input data. Statistical analysis showed that the result between the observed $PM_{2.5}$ and the estimated $PM_{2.5}$ concentrations using $M_6$ model were correlations (R=0.62) and root square mean error ($RMSE=10.70{\mu}gm^{-3}$). In addition, our study show that the relation strongly depends on the seasons due to seasonal observation characteristics of AOT, with a relatively better correlation in spring (R=0.66) and autumntime (R=0.75) than summer and wintertime (R was about 0.38 and 0.56). These results were due to cloud contamination of summertime and the influence of snow/ice surface of wintertime, compared with those of other seasons. Therefore, the empirical multiple linear regression model used in this study showed that the AOT data retrieved from the satellite was important a dominant variable and we will need to use additional weather variables to improve the results of $PM_{2.5}$. Also, the result calculated for $PM_{2.5}$ using empirical multi linear regression model will be useful as a method to enable monitoring of atmospheric environment from satellite and ground meteorological data.