• Title/Summary/Keyword: Learning Loss

Search Result 647, Processing Time 0.032 seconds

Application of Support Vector Regression for Improving the Performance of the Emotion Prediction Model (감정예측모형의 성과개선을 위한 Support Vector Regression 응용)

  • Kim, Seongjin;Ryoo, Eunchung;Jung, Min Kyu;Kim, Jae Kyeong;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.185-202
    • /
    • 2012
  • .Since the value of information has been realized in the information society, the usage and collection of information has become important. A facial expression that contains thousands of information as an artistic painting can be described in thousands of words. Followed by the idea, there has recently been a number of attempts to provide customers and companies with an intelligent service, which enables the perception of human emotions through one's facial expressions. For example, MIT Media Lab, the leading organization in this research area, has developed the human emotion prediction model, and has applied their studies to the commercial business. In the academic area, a number of the conventional methods such as Multiple Regression Analysis (MRA) or Artificial Neural Networks (ANN) have been applied to predict human emotion in prior studies. However, MRA is generally criticized because of its low prediction accuracy. This is inevitable since MRA can only explain the linear relationship between the dependent variables and the independent variable. To mitigate the limitations of MRA, some studies like Jung and Kim (2012) have used ANN as the alternative, and they reported that ANN generated more accurate prediction than the statistical methods like MRA. However, it has also been criticized due to over fitting and the difficulty of the network design (e.g. setting the number of the layers and the number of the nodes in the hidden layers). Under this background, we propose a novel model using Support Vector Regression (SVR) in order to increase the prediction accuracy. SVR is an extensive version of Support Vector Machine (SVM) designated to solve the regression problems. The model produced by SVR only depends on a subset of the training data, because the cost function for building the model ignores any training data that is close (within a threshold ${\varepsilon}$) to the model prediction. Using SVR, we tried to build a model that can measure the level of arousal and valence from the facial features. To validate the usefulness of the proposed model, we collected the data of facial reactions when providing appropriate visual stimulating contents, and extracted the features from the data. Next, the steps of the preprocessing were taken to choose statistically significant variables. In total, 297 cases were used for the experiment. As the comparative models, we also applied MRA and ANN to the same data set. For SVR, we adopted '${\varepsilon}$-insensitive loss function', and 'grid search' technique to find the optimal values of the parameters like C, d, ${\sigma}^2$, and ${\varepsilon}$. In the case of ANN, we adopted a standard three-layer backpropagation network, which has a single hidden layer. The learning rate and momentum rate of ANN were set to 10%, and we used sigmoid function as the transfer function of hidden and output nodes. We performed the experiments repeatedly by varying the number of nodes in the hidden layer to n/2, n, 3n/2, and 2n, where n is the number of the input variables. The stopping condition for ANN was set to 50,000 learning events. And, we used MAE (Mean Absolute Error) as the measure for performance comparison. From the experiment, we found that SVR achieved the highest prediction accuracy for the hold-out data set compared to MRA and ANN. Regardless of the target variables (the level of arousal, or the level of positive / negative valence), SVR showed the best performance for the hold-out data set. ANN also outperformed MRA, however, it showed the considerably lower prediction accuracy than SVR for both target variables. The findings of our research are expected to be useful to the researchers or practitioners who are willing to build the models for recognizing human emotions.

Kinematical Analysis of Heel-Brake Stop in Inline Skate (인라인 스케이트(Inline Skate) 힐 브레이크(Heel-Brake) 정지에 관한 운동학적 분석)

  • Han, Jae-Hee;Lim, Yong-Kyu
    • Korean Journal of Applied Biomechanics
    • /
    • v.15 no.2
    • /
    • pp.11-20
    • /
    • 2005
  • This study has a purpose on contributing to apprehend safe and right way to stop to the inline skate beginners and to the instructors who teaches line skating on the basis for the result of the kinematical analysis on Heel brake stop movement of the inline skate, focusing on the displacement on COG, angle displacement of ankle joint, angle displacement of knee joint, angle displacement of hip joint, using a 3D image method by DLT. To achieve this goal, we analysed the kinematical factor of the 3 well-trained inline skating instructors and obtained the following results. 1. During the movement of heel-brake stop, when strong power was given to a stable and balanced stop and the lower limbs, if the physical centroid is lowered the stability increases, and if it is placed high from the base surface, as the stability decreases compared to the case of low physical centroid, we should make a stop by placing a physical centroid in the base surface and lowering the hight of physical centroid. 2. To make a stable and balanced stop and to provide a strong power to the lower limbs, it is advisable to make a stop by decreasing an angle displacement of ankle joint during a "down" movement. In case of the left ankle joint, in all events and phases the dorsiflexion angle showed a decrease. Nevertheless, in the case of the right ankle joint, the dorsiflexion angle shows an increase after a slight decrease. The dorsiflexion angle displacement of ankle joint can be diminished because of the brake pad of the rear axis frame of the right side inline skate by raising a toe, but cannot be more decreased if certain degree of an angle is made by a brake pad touching a ground surface. To provide a power to a brake pad, it is recommended to place a power by lowering a posture making the dorsiflexion angle of the left ankle joint relatively smaller than that of the right ankle. 3. To make a stable and balanced stop and to add a power to a brake pad, the power must be given to the lower limbs in lowering the hight of physical centroid. For this, it is recommended to make a down movement by decreasing the flexion angle of a knee joint and it is necessary to make a down movement by a regular decrease of the angle displacement of knee joint rather than a swift down movement in every event and phase. 4. The right angle displacement of hip joint is made by lowering vertically the hight of physical centroid as leaning slightly forward. If too narrow angle displacement of hip joint is made by leaning forward too much, the balance is lost during the stop by placing the center in front. To make a stable and balance stop and to place a strong power to the lower limbs, it is recommendable to make a narrow angle by lower the hip joint angle. However, excessive leaning of the upper body to make the angle too narrow, can cause an instable stop and loss of physical centroid. After this study, it is considered to assist the kinematical understanding during the heel brake stop movement of the inline skate, and, to present basic data in learning a method of stable and balanced stop for the inline skating beginners or for the inline skate instructors in the present situation of the complete absence of the study in inline skating.

Effect of Natural Plant Mixtures on Behavioral Profiles and Antioxidants Status in SD Rats (자생식물 혼합 추출물이 SD 흰쥐에서의 행동양상 및 항산화 체계에 미치는 영향)

  • Seo, Bo-Young;Kim, Min-Jung;Kim, Hyun-Su;Park, Hae-Ryong;Lee, Seung-Cheol;Park, Eun-Ju
    • Journal of the Korean Society of Food Science and Nutrition
    • /
    • v.40 no.9
    • /
    • pp.1208-1214
    • /
    • 2011
  • Caffeine, a psychoactive stimulant, has been implicated in the modulation of learning and memory functions due to its action as a non-selective adenosine receptors antagonist. On the contrary, some side effects of caffeine have been reported, such as an increased energy loss and metabolic rate, decrease DNA synthesis in the spleen, and increased oxidative damage to exerted on LDL particles. Therefore, the aim of this study was to develop a safe stimulant from natural plants mixture (Aralia elata, Acori graminei Rhizoma, Chrysanthemum, Dandleion, Guarana, Shepherd's purse) that can be used as a substitute for caffeine. Thirty SD rats were divided into three groups; control group, caffeine group (15.0 mg/kg, i.p.), and natural plants mixture group (NP, 1 mL/kg, p.o.). The effect of NP extract on stimulant activity was evaluated with open-field test (OFT) and plus maze test for measurement of behavioral profiles. Plasma lipid profiles, lipid peroxidation in LDL (conjugated dienes), total antioxidant capacity (TRAP) and DNA damage in white blood, liver, and brain cells were measured. In the OFT, immobility time was increased significantly by acute (once) and chronic (3 weeks) supplementation of NP and showed a similar effect to caffeine treatment. Three weeks of caffeine treatment caused plasma lipid peroxidation and DNA damage in liver cells, whereas there were no changes in the NP group. NP group showed a higher plasma HDL cholesterol concentration compared to the caffeine group. The results indicate that the natural plants mixture had a stimulant effect without inducing oxidative stress.

A Study on The Art of War's strategy and its modern application (손자병법의 전략과 그 현대적 응용에 관한 연구)

  • Song, Yong-ho;Jun, Myung-yong
    • (The)Study of the Eastern Classic
    • /
    • no.73
    • /
    • pp.249-279
    • /
    • 2018
  • This paper analyzes the 'strategy' of Sunzi's art of war and verifies the modern application value of it by combining the 'strategy' of the art of war with modern enterprise management. The army adopts 'war strategy' with the aim of minimizing the loss and sacrifice caused by the war and winning in the shortest time. Enterprise aims to maximize profits at the lowest cost and adopt 'business strategy'. Three factors of art of war's strategic, the 'power', 'adaptation', 'trickery', are similar to the 'internal resources analysis', 'external environment analysis' and 'information management' of the modern enterprise's management. In the process of establishing strategic plan, the art of war emphasizes 'strategy of winning' including 'prophet', 'estimates' and 'maneuvering', in the modern enterprise management, 'prophet' is shown as 'competitor analysis' of the '3C analysis' and 'benchmarking learning'. 'Estimates' is shown as 'SWOT analysis' and '4P's analysis'. 'Maneuvering' is shown as 'market positioning strategy' and 'market preemption strategy'. In the stage of implementing the strategy, 'surprise attack strategy', 'strategy of void and actuality' and 'dividing and integrating strategy' of the art of war are shown as follows in modern enterprises ; 'Surprise attack strategy' is shown as 'differentiation strategy' and 'concentration strategy', 'Strategy of void and actuality' is shown as 'information management' and 'rational market positioning strategy'. 'Dividing and integrating strategy' is shown 'diversification strategy', 'concentration strategy', 'change management', 'basic competition strategy', 'synergy effect' and etc. In terms of strategic results, the 'victory of war' of the art or war is shown as 'competitive advantage' and 'maximization of profits' in modern enterprise management strategy. In a word, although there are different names and expressions between the strategy of Sunzi's art of war and modern enterprise, but their connotation is the same. We can see that the art of war which was written in about B.C.500, has left a high utilization value for modern enterprise in rapid environmental change and intense competition.

Comparison of Convolutional Neural Network (CNN) Models for Lettuce Leaf Width and Length Prediction (상추잎 너비와 길이 예측을 위한 합성곱 신경망 모델 비교)

  • Ji Su Song;Dong Suk Kim;Hyo Sung Kim;Eun Ji Jung;Hyun Jung Hwang;Jaesung Park
    • Journal of Bio-Environment Control
    • /
    • v.32 no.4
    • /
    • pp.434-441
    • /
    • 2023
  • Determining the size or area of a plant's leaves is an important factor in predicting plant growth and improving the productivity of indoor farms. In this study, we developed a convolutional neural network (CNN)-based model to accurately predict the length and width of lettuce leaves using photographs of the leaves. A callback function was applied to overcome data limitations and overfitting problems, and K-fold cross-validation was used to improve the generalization ability of the model. In addition, ImageDataGenerator function was used to increase the diversity of training data through data augmentation. To compare model performance, we evaluated pre-trained models such as VGG16, Resnet152, and NASNetMobile. As a result, NASNetMobile showed the highest performance, especially in width prediction, with an R_squared value of 0.9436, and RMSE of 0.5659. In length prediction, the R_squared value was 0.9537, and RMSE of 0.8713. The optimized model adopted the NASNetMobile architecture, the RMSprop optimization tool, the MSE loss functions, and the ELU activation functions. The training time of the model averaged 73 minutes per Epoch, and it took the model an average of 0.29 seconds to process a single lettuce leaf photo. In this study, we developed a CNN-based model to predict the leaf length and leaf width of plants in indoor farms, which is expected to enable rapid and accurate assessment of plant growth status by simply taking images. It is also expected to contribute to increasing the productivity and resource efficiency of farms by taking appropriate agricultural measures such as adjusting nutrient solution in real time.

Subject-Balanced Intelligent Text Summarization Scheme (주제 균형 지능형 텍스트 요약 기법)

  • Yun, Yeoil;Ko, Eunjung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.141-166
    • /
    • 2019
  • Recently, channels like social media and SNS create enormous amount of data. In all kinds of data, portions of unstructured data which represented as text data has increased geometrically. But there are some difficulties to check all text data, so it is important to access those data rapidly and grasp key points of text. Due to needs of efficient understanding, many studies about text summarization for handling and using tremendous amounts of text data have been proposed. Especially, a lot of summarization methods using machine learning and artificial intelligence algorithms have been proposed lately to generate summary objectively and effectively which called "automatic summarization". However almost text summarization methods proposed up to date construct summary focused on frequency of contents in original documents. Those summaries have a limitation for contain small-weight subjects that mentioned less in original text. If summaries include contents with only major subject, bias occurs and it causes loss of information so that it is hard to ascertain every subject documents have. To avoid those bias, it is possible to summarize in point of balance between topics document have so all subject in document can be ascertained, but still unbalance of distribution between those subjects remains. To retain balance of subjects in summary, it is necessary to consider proportion of every subject documents originally have and also allocate the portion of subjects equally so that even sentences of minor subjects can be included in summary sufficiently. In this study, we propose "subject-balanced" text summarization method that procure balance between all subjects and minimize omission of low-frequency subjects. For subject-balanced summary, we use two concept of summary evaluation metrics "completeness" and "succinctness". Completeness is the feature that summary should include contents of original documents fully and succinctness means summary has minimum duplication with contents in itself. Proposed method has 3-phases for summarization. First phase is constructing subject term dictionaries. Topic modeling is used for calculating topic-term weight which indicates degrees that each terms are related to each topic. From derived weight, it is possible to figure out highly related terms for every topic and subjects of documents can be found from various topic composed similar meaning terms. And then, few terms are selected which represent subject well. In this method, it is called "seed terms". However, those terms are too small to explain each subject enough, so sufficient similar terms with seed terms are needed for well-constructed subject dictionary. Word2Vec is used for word expansion, finds similar terms with seed terms. Word vectors are created after Word2Vec modeling, and from those vectors, similarity between all terms can be derived by using cosine-similarity. Higher cosine similarity between two terms calculated, higher relationship between two terms defined. So terms that have high similarity values with seed terms for each subjects are selected and filtering those expanded terms subject dictionary is finally constructed. Next phase is allocating subjects to every sentences which original documents have. To grasp contents of all sentences first, frequency analysis is conducted with specific terms that subject dictionaries compose. TF-IDF weight of each subjects are calculated after frequency analysis, and it is possible to figure out how much sentences are explaining about each subjects. However, TF-IDF weight has limitation that the weight can be increased infinitely, so by normalizing TF-IDF weights for every subject sentences have, all values are changed to 0 to 1 values. Then allocating subject for every sentences with maximum TF-IDF weight between all subjects, sentence group are constructed for each subjects finally. Last phase is summary generation parts. Sen2Vec is used to figure out similarity between subject-sentences, and similarity matrix can be formed. By repetitive sentences selecting, it is possible to generate summary that include contents of original documents fully and minimize duplication in summary itself. For evaluation of proposed method, 50,000 reviews of TripAdvisor are used for constructing subject dictionaries and 23,087 reviews are used for generating summary. Also comparison between proposed method summary and frequency-based summary is performed and as a result, it is verified that summary from proposed method can retain balance of all subject more which documents originally have.

Estimation of GARCH Models and Performance Analysis of Volatility Trading System using Support Vector Regression (Support Vector Regression을 이용한 GARCH 모형의 추정과 투자전략의 성과분석)

  • Kim, Sun Woong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.107-122
    • /
    • 2017
  • Volatility in the stock market returns is a measure of investment risk. It plays a central role in portfolio optimization, asset pricing and risk management as well as most theoretical financial models. Engle(1982) presented a pioneering paper on the stock market volatility that explains the time-variant characteristics embedded in the stock market return volatility. His model, Autoregressive Conditional Heteroscedasticity (ARCH), was generalized by Bollerslev(1986) as GARCH models. Empirical studies have shown that GARCH models describes well the fat-tailed return distributions and volatility clustering phenomenon appearing in stock prices. The parameters of the GARCH models are generally estimated by the maximum likelihood estimation (MLE) based on the standard normal density. But, since 1987 Black Monday, the stock market prices have become very complex and shown a lot of noisy terms. Recent studies start to apply artificial intelligent approach in estimating the GARCH parameters as a substitute for the MLE. The paper presents SVR-based GARCH process and compares with MLE-based GARCH process to estimate the parameters of GARCH models which are known to well forecast stock market volatility. Kernel functions used in SVR estimation process are linear, polynomial and radial. We analyzed the suggested models with KOSPI 200 Index. This index is constituted by 200 blue chip stocks listed in the Korea Exchange. We sampled KOSPI 200 daily closing values from 2010 to 2015. Sample observations are 1487 days. We used 1187 days to train the suggested GARCH models and the remaining 300 days were used as testing data. First, symmetric and asymmetric GARCH models are estimated by MLE. We forecasted KOSPI 200 Index return volatility and the statistical metric MSE shows better results for the asymmetric GARCH models such as E-GARCH or GJR-GARCH. This is consistent with the documented non-normal return distribution characteristics with fat-tail and leptokurtosis. Compared with MLE estimation process, SVR-based GARCH models outperform the MLE methodology in KOSPI 200 Index return volatility forecasting. Polynomial kernel function shows exceptionally lower forecasting accuracy. We suggested Intelligent Volatility Trading System (IVTS) that utilizes the forecasted volatility results. IVTS entry rules are as follows. If forecasted tomorrow volatility will increase then buy volatility today. If forecasted tomorrow volatility will decrease then sell volatility today. If forecasted volatility direction does not change we hold the existing buy or sell positions. IVTS is assumed to buy and sell historical volatility values. This is somewhat unreal because we cannot trade historical volatility values themselves. But our simulation results are meaningful since the Korea Exchange introduced volatility futures contract that traders can trade since November 2014. The trading systems with SVR-based GARCH models show higher returns than MLE-based GARCH in the testing period. And trading profitable percentages of MLE-based GARCH IVTS models range from 47.5% to 50.0%, trading profitable percentages of SVR-based GARCH IVTS models range from 51.8% to 59.7%. MLE-based symmetric S-GARCH shows +150.2% return and SVR-based symmetric S-GARCH shows +526.4% return. MLE-based asymmetric E-GARCH shows -72% return and SVR-based asymmetric E-GARCH shows +245.6% return. MLE-based asymmetric GJR-GARCH shows -98.7% return and SVR-based asymmetric GJR-GARCH shows +126.3% return. Linear kernel function shows higher trading returns than radial kernel function. Best performance of SVR-based IVTS is +526.4% and that of MLE-based IVTS is +150.2%. SVR-based GARCH IVTS shows higher trading frequency. This study has some limitations. Our models are solely based on SVR. Other artificial intelligence models are needed to search for better performance. We do not consider costs incurred in the trading process including brokerage commissions and slippage costs. IVTS trading performance is unreal since we use historical volatility values as trading objects. The exact forecasting of stock market volatility is essential in the real trading as well as asset pricing models. Further studies on other machine learning-based GARCH models can give better information for the stock market investors.