• Title/Summary/Keyword: learning algorithms

Search Result 2,242, Processing Time 0.027 seconds

Online Multi-Task Learning and Wearable Biosensor-based Detection of Multiple Seniors' Stress in Daily Interaction with the Urban Environment

  • Lee, Gaang;Jebelli, Houtan;Lee, SangHyun
    • International conference on construction engineering and project management
    • /
    • 2020.12a
    • /
    • pp.387-396
    • /
    • 2020
  • Wearable biosensors have the potential to non-invasively and continuously monitor seniors' stress in their daily interaction with the urban environment, thereby enabling to address the stress and ultimately advance their outdoor mobility. However, current wearable biosensor-based stress detection methods have several drawbacks in field application due to their dependence on batch-learning algorithms. First, these methods train a single classifier, which might not account for multiple subjects' different physiological reactivity to stress. Second, they require a great deal of computational power to store and reuse all previous data for updating the signle classifier. To address this issue, we tested the feasibility of online multi-task learning (OMTL) algorithms to identify multiple seniors' stress from electrodermal activity (EDA) collected by a wristband-type biosensor in a daily trip setting. As a result, OMTL algorithms showed the higher test accuracy (75.7%, 76.2%, and 71.2%) than a batch-learning algorithm (64.8%). This finding demonstrates that the OMTL algorithms can strengthen the field applicability of the wearable biosensor-based stress detection, thereby contributing to better understanding the seniors' stress in the urban environment and ultimately advancing their mobility.

  • PDF

Machine Learning Frameworks for Automated Software Testing Tools : A Study

  • Kim, Jungho;Ryu, Joung Woo;Shin, Hyun-Jeong;Song, Jin-Hee
    • International Journal of Contents
    • /
    • v.13 no.1
    • /
    • pp.38-44
    • /
    • 2017
  • Increased use of software and complexity of software functions, as well as shortened software quality evaluation periods, have increased the importance and necessity for automation of software testing. Automating software testing by using machine learning not only minimizes errors in manual testing, but also allows a speedier evaluation. Research on machine learning in automated software testing has so far focused on solving special problems with algorithms, leading to difficulties for the software developers and testers, in applying machine learning to software testing automation. This paper, proposes a new machine learning framework for software testing automation through related studies. To maximize the performance of software testing, we analyzed and categorized the machine learning algorithms applicable to each software test phase, including the diverse data that can be used in the algorithms. We believe that our framework allows software developers or testers to choose a machine learning algorithm suitable for their purpose.

Comparative characteristic of ensemble machine learning and deep learning models for turbidity prediction in a river (딥러닝과 앙상블 머신러닝 모형의 하천 탁도 예측 특성 비교 연구)

  • Park, Jungsu
    • Journal of Korean Society of Water and Wastewater
    • /
    • v.35 no.1
    • /
    • pp.83-91
    • /
    • 2021
  • The increased turbidity in rivers during flood events has various effects on water environmental management, including drinking water supply systems. Thus, prediction of turbid water is essential for water environmental management. Recently, various advanced machine learning algorithms have been increasingly used in water environmental management. Ensemble machine learning algorithms such as random forest (RF) and gradient boosting decision tree (GBDT) are some of the most popular machine learning algorithms used for water environmental management, along with deep learning algorithms such as recurrent neural networks. In this study GBDT, an ensemble machine learning algorithm, and gated recurrent unit (GRU), a recurrent neural networks algorithm, are used for model development to predict turbidity in a river. The observation frequencies of input data used for the model were 2, 4, 8, 24, 48, 120 and 168 h. The root-mean-square error-observations standard deviation ratio (RSR) of GRU and GBDT ranges between 0.182~0.766 and 0.400~0.683, respectively. Both models show similar prediction accuracy with RSR of 0.682 for GRU and 0.683 for GBDT. The GRU shows better prediction accuracy when the observation frequency is relatively short (i.e., 2, 4, and 8 h) where GBDT shows better prediction accuracy when the observation frequency is relatively long (i.e. 48, 120, 160 h). The results suggest that the characteristics of input data should be considered to develop an appropriate model to predict turbidity.

Kernel-based actor-critic approach with applications

  • Chu, Baek-Suk;Jung, Keun-Woo;Park, Joo-Young
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.11 no.4
    • /
    • pp.267-274
    • /
    • 2011
  • Recently, actor-critic methods have drawn significant interests in the area of reinforcement learning, and several algorithms have been studied along the line of the actor-critic strategy. In this paper, we consider a new type of actor-critic algorithms employing the kernel methods, which have recently shown to be very effective tools in the various fields of machine learning, and have performed investigations on combining the actor-critic strategy together with kernel methods. More specifically, this paper studies actor-critic algorithms utilizing the kernel-based least-squares estimation and policy gradient, and in its critic's part, the study uses a sliding-window-based kernel least-squares method, which leads to a fast and efficient value-function-estimation in a nonparametric setting. The applicability of the considered algorithms is illustrated via a robot locomotion problem and a tunnel ventilation control problem.

Deep Learning-Based Artificial Intelligence for Mammography

  • Jung Hyun Yoon;Eun-Kyung Kim
    • Korean Journal of Radiology
    • /
    • v.22 no.8
    • /
    • pp.1225-1239
    • /
    • 2021
  • During the past decade, researchers have investigated the use of computer-aided mammography interpretation. With the application of deep learning technology, artificial intelligence (AI)-based algorithms for mammography have shown promising results in the quantitative assessment of parenchymal density, detection and diagnosis of breast cancer, and prediction of breast cancer risk, enabling more precise patient management. AI-based algorithms may also enhance the efficiency of the interpretation workflow by reducing both the workload and interpretation time. However, more in-depth investigation is required to conclusively prove the effectiveness of AI-based algorithms. This review article discusses how AI algorithms can be applied to mammography interpretation as well as the current challenges in its implementation in real-world practice.

A Feasibility Study on the Improvement of Diagnostic Accuracy for Energy-selective Digital Mammography using Machine Learning (머신러닝을 이용한 에너지 선택적 유방촬영의 진단 정확도 향상에 관한 연구)

  • Eom, Jisoo;Lee, Seungwan;Kim, Burnyoung
    • Journal of radiological science and technology
    • /
    • v.42 no.1
    • /
    • pp.9-17
    • /
    • 2019
  • Although digital mammography is a representative method for breast cancer detection. It has a limitation in detecting and classifying breast tumor due to superimposed structures. Machine learning, which is a part of artificial intelligence fields, is a method for analysing a large amount of data using complex algorithms, recognizing patterns and making prediction. In this study, we proposed a technique to improve the diagnostic accuracy of energy-selective mammography by training data using the machine learning algorithm and using dual-energy measurements. A dual-energy images obtained from a photon-counting detector were used for the input data of machine learning algorithms, and we analyzed the accuracy of predicted tumor thickness for verifying the machine learning algorithms. The results showed that the classification accuracy of tumor thickness was above 95% and was improved with an increase of imput data. Therefore, we expect that the diagnostic accuracy of energy-selective mammography can be improved by using machine learning.

An Effective Anomaly Detection Approach based on Hybrid Unsupervised Learning Technologies in NIDS

  • Kangseok Kim
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.2
    • /
    • pp.494-510
    • /
    • 2024
  • Internet users are exposed to sophisticated cyberattacks that intrusion detection systems have difficulty detecting. Therefore, research is increasing on intrusion detection methods that use artificial intelligence technology for detecting novel cyberattacks. Unsupervised learning-based methods are being researched that learn only from normal data and detect abnormal behaviors by finding patterns. This study developed an anomaly-detection method based on unsupervised machines and deep learning for a network intrusion detection system (NIDS). We present a hybrid anomaly detection approach based on unsupervised learning techniques using the autoencoder (AE), Isolation Forest (IF), and Local Outlier Factor (LOF) algorithms. An oversampling approach that increased the detection rate was also examined. A hybrid approach that combined deep learning algorithms and traditional machine learning algorithms was highly effective in setting the thresholds for anomalies without subjective human judgment. It achieved precision and recall rates respectively of 88.2% and 92.8% when combining two AEs, IF, and LOF while using an oversampling approach to learn more unknown normal data improved the detection accuracy. This approach achieved precision and recall rates respectively of 88.2% and 94.6%, further improving the detection accuracy compared with the hybrid method. Therefore, in NIDS the proposed approach provides high reliability for detecting cyberattacks.

Early Diagnosis of anxiety Disorder Using Artificial Intelligence

  • Choi DongOun;Huan-Meng;Yun-Jeong, Kang
    • International Journal of Advanced Culture Technology
    • /
    • v.12 no.1
    • /
    • pp.242-248
    • /
    • 2024
  • Contemporary societal and environmental transformations coincide with the emergence of novel mental health challenges. anxiety disorder, a chronic and highly debilitating illness, presents with diverse clinical manifestations. Epidemiological investigations indicate a global prevalence of 5%, with an additional 10% exhibiting subclinical symptoms. Notably, 9% of adolescents demonstrate clinical features. Untreated, anxiety disorder exerts profound detrimental effects on individuals, families, and the broader community. Therefore, it is very meaningful to predict anxiety disorder through machine learning algorithm analysis model. The main research content of this paper is the analysis of the prediction model of anxiety disorder by machine learning algorithms. The research purpose of machine learning algorithms is to use computers to simulate human learning activities. It is a method to locate existing knowledge, acquire new knowledge, continuously improve performance, and achieve self-improvement by learning computers. This article analyzes the relevant theories and characteristics of machine learning algorithms and integrates them into anxiety disorder prediction analysis. The final results of the study show that the AUC of the artificial neural network model is the largest, reaching 0.8255, indicating that it is better than the other two models in prediction accuracy. In terms of running time, the time of the three models is less than 1 second, which is within the acceptable range.

Design of Learning Module for ERNIE(ERNIE : Expansible & Reconfigurable Neuro Informatics Engine) (범용 신경망 연산기(ERNIE)를 위한 학습 모듈 설계)

  • Jung Je Kyo;Wee Jae Woo;Dong Sung Soo;Lee Chong Ho
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.53 no.12
    • /
    • pp.804-810
    • /
    • 2004
  • There are two important things for the general purpose neural network processor. The first is a capability to build various structures of neural network, and the second is to be able to support suitable learning method for that neural network. Some way to process various learning algorithms is required for on-chip learning, because the more neural network types are to be handled, the more learning methods need to be built into. In this paper, an improved hardware structure is proposed to compute various kinds of learning algorithms flexibly. The hardware structure is based on the existing modular neural network structure. It doesn't need to add a new circuit or a new program for the learning process. It is shown that rearrangements of the existing processing elements can produce several neural network learning modules. The performance and utilization of this module are analyzed by comparing with other neural network chips.

Controller Learning Method of Self-driving Bicycle Using State-of-the-art Deep Reinforcement Learning Algorithms

  • Choi, Seung-Yoon;Le, Tuyen Pham;Chung, Tae-Choong
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.10
    • /
    • pp.23-31
    • /
    • 2018
  • Recently, there have been many studies on machine learning. Among them, studies on reinforcement learning are actively worked. In this study, we propose a controller to control bicycle using DDPG (Deep Deterministic Policy Gradient) algorithm which is the latest deep reinforcement learning method. In this paper, we redefine the compensation function of bicycle dynamics and neural network to learn agents. When using the proposed method for data learning and control, it is possible to perform the function of not allowing the bicycle to fall over and reach the further given destination unlike the existing method. For the performance evaluation, we have experimented that the proposed algorithm works in various environments such as fixed speed, random, target point, and not determined. Finally, as a result, it is confirmed that the proposed algorithm shows better performance than the conventional neural network algorithms NAF and PPO.