• 제목/요약/키워드: 학습된 패턴

Search Result 1,576, Processing Time 0.034 seconds

A Prediction of N-value Using Artificial Neural Network (인공신경망을 이용한 N치 예측)

  • Kim, Kwang Myung;Park, Hyoung June;Goo, Tae Hun;Kim, Hyung Chan
    • The Journal of Engineering Geology
    • /
    • v.30 no.4
    • /
    • pp.457-468
    • /
    • 2020
  • Problems arising during pile design works for plant construction, civil and architecture work are mostly come from uncertainty of geotechnical characteristics. In particular, obtaining the N-value measured through the Standard Penetration Test (SPT) is the most important data. However, it is difficult to obtain N-value by drilling investigation throughout the all target area. There are many constraints such as licensing, time, cost, equipment access and residential complaints etc. it is impossible to obtain geotechnical characteristics through drilling investigation within a short bidding period in overseas. The geotechnical characteristics at non-drilling investigation points are usually determined by the engineer's empirical judgment, which can leads to errors in pile design and quantity calculation causing construction delay and cost increase. It would be possible to overcome this problem if N-value could be predicted at the non-drilling investigation points using limited minimum drilling investigation data. This study was conducted to predicted the N-value using an Artificial Neural Network (ANN) which one of the Artificial intelligence (AI) method. An Artificial Neural Network treats a limited amount of geotechnical characteristics as a biological logic process, providing more reliable results for input variables. The purpose of this study is to predict N-value at the non-drilling investigation points through patterns which is studied by multi-layer perceptron and error back-propagation algorithms using the minimum geotechnical data. It has been reviewed the reliability of the values that predicted by AI method compared to the measured values, and we were able to confirm the high reliability as a result. To solving geotechnical uncertainty, we will perform sensitivity analysis of input variables to increase learning effect in next steps and it may need some technical update of program. We hope that our study will be helpful to design works in the future.

Estimation of High Resolution Sea Surface Salinity Using Multi Satellite Data and Machine Learning (다종 위성자료와 기계학습을 이용한 고해상도 표층 염분 추정)

  • Sung, Taejun;Sim, Seongmun;Jang, Eunna;Im, Jungho
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.5_2
    • /
    • pp.747-763
    • /
    • 2022
  • Ocean salinity affects ocean circulation on a global scale and low salinity water around coastal areas often has an impact on aquaculture and fisheries. Microwave satellite sensors (e.g., Soil Moisture Active Passive [SMAP]) have provided sea surface salinity (SSS) based on the dielectric characteristics of water associated with SSS and sea surface temperature (SST). In this study, a Light Gradient Boosting Machine (LGBM)-based model for generating high resolution SSS from Geostationary Ocean Color Imager (GOCI) data was proposed, having machine learning-based improved SMAP SSS by Jang et al. (2022) as reference data (SMAP SSS (Jang)). Three schemes with different input variables were tested, and scheme 3 with all variables including Multi-scale Ultra-high Resolution SST yielded the best performance (coefficient of determination = 0.60, root mean square error = 0.91 psu). The proposed LGBM-based GOCI SSS had a similar spatiotemporal pattern with SMAP SSS (Jang), with much higher spatial resolution even in coastal areas, where SMAP SSS (Jang) was not available. In addition, when tested for the great flood occurred in Southern China in August 2020, GOCI SSS well simulated the spatial and temporal change of Changjiang Diluted Water. This research provided a potential that optical satellite data can be used to generate high resolution SSS associated with the improved microwave-based SSS especially in coastal areas.

GEase-K: Linear and Nonlinear Autoencoder-based Recommender System with Side Information (GEase-K: 부가 정보를 활용한 선형 및 비선형 오토인코더 기반의 추천시스템)

  • Taebeom Lee;Seung-hak Lee;Min-jeong Ma;Yoonho Cho
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.3
    • /
    • pp.167-183
    • /
    • 2023
  • In the recent field of recommendation systems, various studies have been conducted to model sparse data effectively. Among these, GLocal-K(Global and Local Kernels for Recommender Systems) is a research endeavor combining global and local kernels to provide personalized recommendations by considering global data patterns and individual user characteristics. However, due to its utilization of kernel tricks, GLocal-K exhibits diminished performance on highly sparse data and struggles to offer recommendations for new users or items due to the absence of side information. In this paper, to address these limitations of GLocal-K, we propose the GEase-K (Global and EASE kernels for Recommender Systems) model, incorporating the EASE(Embarrassingly Shallow Autoencoders for Sparse Data) model and leveraging side information. Initially, we substitute EASE for the local kernel in GLocal-K to enhance recommendation performance on highly sparse data. EASE, functioning as a simple linear operational structure, is an autoencoder that performs highly on extremely sparse data through regularization and learning item similarity. Additionally, we utilize side information to alleviate the cold-start problem. We enhance the understanding of user-item similarities by employing a conditional autoencoder structure during the training process to incorporate side information. In conclusion, GEase-K demonstrates resilience in highly sparse data and cold-start situations by combining linear and nonlinear structures and utilizing side information. Experimental results show that GEase-K outperforms GLocal-K based on the RMSE and MAE metrics on the highly sparse GoodReads and ModCloth datasets. Furthermore, in cold-start experiments divided into four groups using the GoodReads and ModCloth datasets, GEase-K denotes superior performance compared to GLocal-K.

Confucian View of Self-realization and Context of Life: With a focus on Viewpoint of Confucius and Mencius (유교의 자아실현과 삶의 맥락 - 공자와 맹자의 시선을 중심으로 -)

  • Shin, Chang Ho
    • The Journal of Korean Philosophical History
    • /
    • no.29
    • /
    • pp.153-178
    • /
    • 2010
  • The aim of this research was to examine the traditional Confucian view of self-realization in East Asia and the meaning of life implied therein. The researcher closely reviewed the phase of self-realization of both Confucius and Mencius who are central in Confucianism, especially in the primordial Confucianism, and after investigating maturity of personality as well as educational characteristics thereof, the researcher tried to elicit its modern significance. In Analects, Confucius who is the founder of Confucianism mentioned about 'the pleasure of studying and practicing what he has learned'(學而時習 "Hagisiseup" in Korean), since after, his past was then just the process of self-realization that lasted throughout life. That is, the six phases of self-realization, to wit, 'bending on learning(志學, "Jihak")-'standing firm'(而立, "Irip")-'having no doubts'(不惑, "Bulhok")-'knowing the decrees of Heaven'(知天命, "Jicheonmyeong")-'ear being obedient organ for the reception of truth' (耳順, "Isun")-'able to follow what my heart desires without transgressing what is right'(從心, "Jongsim"), are lying hidden and undeveloped during lifetime, and, at the same time, these phases illustrate the state of enlightenment of life in an in-depth manner. By showing the process of living which is being sublimated in respect of quality, and by going through important process of self-innovation up to six times during lifetime, Confucius edifies us the activity of complete self-realization as well as the importance of education and learning. Meanwhile, these are connected to Mencius in a similar pattern, and strong influence of the characteristics of the learning of the mind and heart( 心學, "Simhak") based on his philosophy permeates the self-actualization phase of Mencius. Mencius' self-actualization phase is expressed in terms of six stages, viz., Person of Goodness(善人, "Seonin")-Trustworthy Person(信人, "Sinin")-Person of Beauty(美人, "Miin")-Great Person(大人, "Daein")-Sage(聖人, "Seongin")-Divine Person(神人, "Sinin"), and these six phases of self-actualization process are educational and learning model for people who dream actualization of perfect personality during their lifetime. Confucian and Mencian view of self-realization congruent with self-discipline internally, and it also reveals a stereotype of human externally. These are a process of performing organic ideals in order for cultivating oneself and regulating others(修己治人, pronounced 'sugichiin' in Korean) which has been pursued by Confucianism. Briefly, these self-realization phases are the arts of living that will lay foundation for "Being Born Human, pronounced Saramim' in Korean" and for becoming "Fully Human, 'Sarmadoem'" and finally for "Human Feelingness, 'Saramdaum'

Ensemble of Nested Dichotomies for Activity Recognition Using Accelerometer Data on Smartphone (Ensemble of Nested Dichotomies 기법을 이용한 스마트폰 가속도 센서 데이터 기반의 동작 인지)

  • Ha, Eu Tteum;Kim, Jeongmin;Ryu, Kwang Ryel
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.123-132
    • /
    • 2013
  • As the smartphones are equipped with various sensors such as the accelerometer, GPS, gravity sensor, gyros, ambient light sensor, proximity sensor, and so on, there have been many research works on making use of these sensors to create valuable applications. Human activity recognition is one such application that is motivated by various welfare applications such as the support for the elderly, measurement of calorie consumption, analysis of lifestyles, analysis of exercise patterns, and so on. One of the challenges faced when using the smartphone sensors for activity recognition is that the number of sensors used should be minimized to save the battery power. When the number of sensors used are restricted, it is difficult to realize a highly accurate activity recognizer or a classifier because it is hard to distinguish between subtly different activities relying on only limited information. The difficulty gets especially severe when the number of different activity classes to be distinguished is very large. In this paper, we show that a fairly accurate classifier can be built that can distinguish ten different activities by using only a single sensor data, i.e., the smartphone accelerometer data. The approach that we take to dealing with this ten-class problem is to use the ensemble of nested dichotomy (END) method that transforms a multi-class problem into multiple two-class problems. END builds a committee of binary classifiers in a nested fashion using a binary tree. At the root of the binary tree, the set of all the classes are split into two subsets of classes by using a binary classifier. At a child node of the tree, a subset of classes is again split into two smaller subsets by using another binary classifier. Continuing in this way, we can obtain a binary tree where each leaf node contains a single class. This binary tree can be viewed as a nested dichotomy that can make multi-class predictions. Depending on how a set of classes are split into two subsets at each node, the final tree that we obtain can be different. Since there can be some classes that are correlated, a particular tree may perform better than the others. However, we can hardly identify the best tree without deep domain knowledge. The END method copes with this problem by building multiple dichotomy trees randomly during learning, and then combining the predictions made by each tree during classification. The END method is generally known to perform well even when the base learner is unable to model complex decision boundaries As the base classifier at each node of the dichotomy, we have used another ensemble classifier called the random forest. A random forest is built by repeatedly generating a decision tree each time with a different random subset of features using a bootstrap sample. By combining bagging with random feature subset selection, a random forest enjoys the advantage of having more diverse ensemble members than a simple bagging. As an overall result, our ensemble of nested dichotomy can actually be seen as a committee of committees of decision trees that can deal with a multi-class problem with high accuracy. The ten classes of activities that we distinguish in this paper are 'Sitting', 'Standing', 'Walking', 'Running', 'Walking Uphill', 'Walking Downhill', 'Running Uphill', 'Running Downhill', 'Falling', and 'Hobbling'. The features used for classifying these activities include not only the magnitude of acceleration vector at each time point but also the maximum, the minimum, and the standard deviation of vector magnitude within a time window of the last 2 seconds, etc. For experiments to compare the performance of END with those of other methods, the accelerometer data has been collected at every 0.1 second for 2 minutes for each activity from 5 volunteers. Among these 5,900 ($=5{\times}(60{\times}2-2)/0.1$) data collected for each activity (the data for the first 2 seconds are trashed because they do not have time window data), 4,700 have been used for training and the rest for testing. Although 'Walking Uphill' is often confused with some other similar activities, END has been found to classify all of the ten activities with a fairly high accuracy of 98.4%. On the other hand, the accuracies achieved by a decision tree, a k-nearest neighbor, and a one-versus-rest support vector machine have been observed as 97.6%, 96.5%, and 97.6%, respectively.

Rough Set Analysis for Stock Market Timing (러프집합분석을 이용한 매매시점 결정)

  • Huh, Jin-Nyung;Kim, Kyoung-Jae;Han, In-Goo
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.77-97
    • /
    • 2010
  • Market timing is an investment strategy which is used for obtaining excessive return from financial market. In general, detection of market timing means determining when to buy and sell to get excess return from trading. In many market timing systems, trading rules have been used as an engine to generate signals for trade. On the other hand, some researchers proposed the rough set analysis as a proper tool for market timing because it does not generate a signal for trade when the pattern of the market is uncertain by using the control function. The data for the rough set analysis should be discretized of numeric value because the rough set only accepts categorical data for analysis. Discretization searches for proper "cuts" for numeric data that determine intervals. All values that lie within each interval are transformed into same value. In general, there are four methods for data discretization in rough set analysis including equal frequency scaling, expert's knowledge-based discretization, minimum entropy scaling, and na$\ddot{i}$ve and Boolean reasoning-based discretization. Equal frequency scaling fixes a number of intervals and examines the histogram of each variable, then determines cuts so that approximately the same number of samples fall into each of the intervals. Expert's knowledge-based discretization determines cuts according to knowledge of domain experts through literature review or interview with experts. Minimum entropy scaling implements the algorithm based on recursively partitioning the value set of each variable so that a local measure of entropy is optimized. Na$\ddot{i}$ve and Booleanreasoning-based discretization searches categorical values by using Na$\ddot{i}$ve scaling the data, then finds the optimized dicretization thresholds through Boolean reasoning. Although the rough set analysis is promising for market timing, there is little research on the impact of the various data discretization methods on performance from trading using the rough set analysis. In this study, we compare stock market timing models using rough set analysis with various data discretization methods. The research data used in this study are the KOSPI 200 from May 1996 to October 1998. KOSPI 200 is the underlying index of the KOSPI 200 futures which is the first derivative instrument in the Korean stock market. The KOSPI 200 is a market value weighted index which consists of 200 stocks selected by criteria on liquidity and their status in corresponding industry including manufacturing, construction, communication, electricity and gas, distribution and services, and financing. The total number of samples is 660 trading days. In addition, this study uses popular technical indicators as independent variables. The experimental results show that the most profitable method for the training sample is the na$\ddot{i}$ve and Boolean reasoning but the expert's knowledge-based discretization is the most profitable method for the validation sample. In addition, the expert's knowledge-based discretization produced robust performance for both of training and validation sample. We also compared rough set analysis and decision tree. This study experimented C4.5 for the comparison purpose. The results show that rough set analysis with expert's knowledge-based discretization produced more profitable rules than C4.5.

Importance and Specialization Plan of the Indicators by the Function of the Arboretum (수목원 기능별 지표의 중요도와 특성화방안 - 대구, 경북, 경남 수목원을 대상으로 -)

  • Kim, Yong-Soo;Ha, Sun-Gyone;Park, Chan-Yong
    • Journal of Korean Society of Forest Science
    • /
    • v.98 no.4
    • /
    • pp.370-378
    • /
    • 2009
  • This study tries to provide the basic direction to form the arboretum with the distinct features by providing the basic data to help the differentiated strategy for each arboretum. For this purpose, the users' pattern, importance of the indicator by the function, and the stimulation and specialization importance were examined for Daegu Arboretum, Gyeongbuk Arboretum and Gyeongnam Arboretum in Gyeongsang Province. The result says, looking into the functions of arboretum, the collection function showed the highest importance in the preservation of the endangered crisis species; the display function showed the highest in the use as the nature experiencing spaces through the plant exhibition; the research function showed the highest in the study on Plant Systematics; the education function showed the highest in the protection of the native plants; and the recreational function showed the highest in the healthy recreational space. In the plan for the promotion of the arboretum showed the highest in the public education program operation such as the narration from arboretum and education for plant. Therefore, it is considered to need the system setup such as the education program, material development and specialist training in terms of the arboretum. For the specialization plan for arboretum in this study, it seem desirable to concentrate on the research and education related to the natural resources renewal, for Daegu Arboretum; to concentrate on the resort site for the protection and display of the species and the disabled visitors by utilizing the geographical traits in the mountains, for Gyeongbuk Arboretum; to create the specialization plan mainly for the tree species suitable for the warm weather and for the children.

EFFECT OF THE SOCIAL SKILL TRAINING IN ADHD CHILDREN (주의력 결핍 과잉운동장애 아동에서 사회기술훈련의 효과)

  • Park, Soon-Young;Kwack, Young-Sook;Kim, Mi-Koung
    • Journal of the Korean Academy of Child and Adolescent Psychiatry
    • /
    • v.9 no.2
    • /
    • pp.154-164
    • /
    • 1998
  • Medication is widely accepted as an effective method to reduce the problem of attention deficit, hyperactivity, impulsivity, resistance and violence of ADHD children. However, it does not provide us with the solution on the conflicting routinized behavioral patterns to gain a high level of self-control and acceptable behavior. As a way of replacing medication, this study applies the social skills training program for ADHD children and measures the level of improvement of social skills and the change of the behavioral patterns. The experiment is carried out on 16 children ranged from 6 to 13 years of age for 10 weeks. The patients are divided into three groups:a pure ADHD group, an ADHD group with conduct disorder, an ADHD group with mental retardation and other symptoms. The change of symptoms and the change of social skills are measured by the Child Behavior Checklist(CBCL), the ADD-H Comprehensive Teacher’s Rating Scale(ACTeRS) and the Social Skills Rating Scale(SSRS), and finally Mastson Evaluation of Social Skills for Youth(MESSY). Wilcoxon signed ranks test is used to evaluate the effect of the treatment, and Kruskal-Wallis test is also used to measure the change after the treatment in each of the three groups. In the ADHD group with conduct disorder, the examination of the effect of the treatment shows a significant reduction of violence in the area of behavior(p<.05), and a significant difference of activity and social skills in the area of social competent(p<.001). In the ADHD group with mental retardation and other symptoms, a significant rise of social skills is found in the area of social skills evaluation (p<.05). However, there is no significant difference of effect by the treatment among the three groups. In addition, the current examination shows that the social skills training program does not make a statistically significant contribution to the social skills of the ADHD children. On the other hand, the training helps some children, when it is suitable for the characteristics and accompanying symptoms of the children:it reduces the level of violence in the ADHD group with conduct disorder, and it raises the social skills in the ADHD group with mental retardation. In other words, the social skills training program will reduce the conduct disorder and helps peer relation for ADHD children.

  • PDF

Monitoring Ground-level SO2 Concentrations Based on a Stacking Ensemble Approach Using Satellite Data and Numerical Models (위성 자료와 수치모델 자료를 활용한 스태킹 앙상블 기반 SO2 지상농도 추정)

  • Choi, Hyunyoung;Kang, Yoojin;Im, Jungho;Shin, Minso;Park, Seohui;Kim, Sang-Min
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1053-1066
    • /
    • 2020
  • Sulfur dioxide (SO2) is primarily released through industrial, residential, and transportation activities, and creates secondary air pollutants through chemical reactions in the atmosphere. Long-term exposure to SO2 can result in a negative effect on the human body causing respiratory or cardiovascular disease, which makes the effective and continuous monitoring of SO2 crucial. In South Korea, SO2 monitoring at ground stations has been performed, but this does not provide spatially continuous information of SO2 concentrations. Thus, this research estimated spatially continuous ground-level SO2 concentrations at 1 km resolution over South Korea through the synergistic use of satellite data and numerical models. A stacking ensemble approach, fusing multiple machine learning algorithms at two levels (i.e., base and meta), was adopted for ground-level SO2 estimation using data from January 2015 to April 2019. Random forest and extreme gradient boosting were used as based models and multiple linear regression was adopted for the meta-model. The cross-validation results showed that the meta-model produced the improved performance by 25% compared to the base models, resulting in the correlation coefficient of 0.48 and root-mean-square-error of 0.0032 ppm. In addition, the temporal transferability of the approach was evaluated for one-year data which were not used in the model development. The spatial distribution of ground-level SO2 concentrations based on the proposed model agreed with the general seasonality of SO2 and the temporal patterns of emission sources.

Estimation of Ground-level PM10 and PM2.5 Concentrations Using Boosting-based Machine Learning from Satellite and Numerical Weather Prediction Data (부스팅 기반 기계학습기법을 이용한 지상 미세먼지 농도 산출)

  • Park, Seohui;Kim, Miae;Im, Jungho
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.2
    • /
    • pp.321-335
    • /
    • 2021
  • Particulate matter (PM10 and PM2.5 with a diameter less than 10 and 2.5 ㎛, respectively) can be absorbed by the human body and adversely affect human health. Although most of the PM monitoring are based on ground-based observations, they are limited to point-based measurement sites, which leads to uncertainty in PM estimation for regions without observation sites. It is possible to overcome their spatial limitation by using satellite data. In this study, we developed machine learning-based retrieval algorithm for ground-level PM10 and PM2.5 concentrations using aerosol parameters from Geostationary Ocean Color Imager (GOCI) satellite and various meteorological parameters from a numerical weather prediction model during January to December of 2019. Gradient Boosted Regression Trees (GBRT) and Light Gradient Boosting Machine (LightGBM) were used to estimate PM concentrations. The model performances were examined for two types of feature sets-all input parameters (Feature set 1) and a subset of input parameters without meteorological and land-cover parameters (Feature set 2). Both models showed higher accuracy (about 10 % higher in R2) by using the Feature set 1 than the Feature set 2. The GBRT model using Feature set 1 was chosen as the final model for further analysis(PM10: R2 = 0.82, nRMSE = 34.9 %, PM2.5: R2 = 0.75, nRMSE = 35.6 %). The spatial distribution of the seasonal and annual-averaged PM concentrations was similar with in-situ observations, except for the northeastern part of China with bright surface reflectance. Their spatial distribution and seasonal changes were well matched with in-situ measurements.