• 제목/요약/키워드: machine learning (ML)

검색결과 280건 처리시간 0.031초

A Survey on Predicting Workloads and Optimising QoS in the Cloud Computing

  • Omar F. Aloufi;Karim Djemame;Faisal Saeed;Fahad Ghabban
    • International Journal of Computer Science & Network Security
    • /
    • 제24권2호
    • /
    • pp.59-66
    • /
    • 2024
  • This paper presents the concept and characteristics of cloud computing, and it addresses how cloud computing delivers quality of service (QoS) to the end-user. Next, it discusses how to schedule one's workload in the infrastructure using technologies that have recently emerged such as Machine Learning (ML). That is followed by an overview of how ML can be used for resource management. This paper then looks at the primary goal of this project, which is to outline the benefits of using ML to schedule upcoming demands to achieve QoS and conserve energy. In this survey, we reviewed the research related to ML methods for predicting workloads in cloud computing. It also provides information on the approaches to elasticity, while another section discusses the methods of prediction used in previous studies and those that used in this field. The paper concludes with a summary of the literature on predicting workloads and optimising QoS in the cloud computing.

머신러닝 기법을 이용한 로터리 킬른 공정의 질소산화물 배출예측에 관한 연구 (A Study on the Prediction of Nitrogen Oxide Emissions in Rotary Kiln Process using Machine Learning)

  • 유제형;박정열;배재권
    • 산업융합연구
    • /
    • 제21권7호
    • /
    • pp.19-27
    • /
    • 2023
  • 이차전지 시장의 확대에 따라 니켈 산화광을 로터리 킬른 및 전기로 공법을 이용하여 생산하는 공정이 전 세계적으로 확대되고 있는 상황이며 지속가능한 ESG 경영 확대에 따라 배출가스 내 질소산화물 등 대기오염물질 관리가 강화되고 있다. 건식니켈제련 공정의 주요 설비 중 하나인 로터리 킬른은 광석의 건조와 예비환원을 위한 설비이며 운전 중 질소산화물이 생성되므로 질소산화물 농도 예측 운전이 필요하다. 본 연구에서는 회귀 예측을 위한 LSTM 모델과 분류 예측을 위한 LightGBM 모델을 적용한 AutoML을 사용하여 모델을 최적화 하였다. LSTM을 적용 시 5분 후 예측 값은 상관계수 0.86, MAE 5.13ppm, 40분 후 예측 값은 상관계수 0.38, MAE 10.84ppm의 결과를 얻었다. 분류 예측을 위한 LightGBM 적용 결과 Test 정확도는 5분 후 0.75에서 40분 후 0.61로 상승하여 실제 조업에 활용할 수 있는 수준까지 상승되었고 AutoML을 통한 모델 최적화 결과 5분 후 예측 값의 정확도는 0.75에서 0.80까지, 40분 후의 예측 정확도는 0.61에서 0.70까지 향상되었다. 본 연구를 통해 로터리 킬른 질소산화물 예측 값을 실제 조업에 적용하여 대기오염물질 배출규제 준수 및 ESG 경영에 기여할 수 있다.

YOLO 기반 개체 검출과 Node.js 서버를 이용한 반려견 행동 분류 시스템 구현 (Implementation of a Classification System for Dog Behaviors using YOLI-based Object Detection and a Node.js Server)

  • 조용화;이혁재;김영훈
    • 융합신호처리학회논문지
    • /
    • 제21권1호
    • /
    • pp.29-37
    • /
    • 2020
  • 본 논문은 실시간 영상 분석을 통해서 반려견에 대한 객체를 추출해 내고, 추출된 이미지로부터 반려견 행동을 분류하는 방법을 구현한다. 반려견 객체 탐지를 위해서 Darknet YOLO를 사용하였으며, 추출된 이미지로부터 행동 패턴 분류는 구글에서 제공하고 있는 Teachable Machine을 이용하였다. 학습된 Teachable Machine은 구글 드라이브에 저장되어 node.js 서버 상에서 ml5.js로 구현하여 사용할 수 있다. 분류된 행동 패턴 결과는 사용자의 스마트 폰 또는 PC로 실시간 전송되며, 언제 어디서든 확인 가능할 수 있게 node.js 서버에서 socket.io 모듈을 사용해서 상호반응 웹 서버를 구현하였다.

A Smartphone-based Virtual Reality Visualization System for Human Activities Classification

  • Lomaliza, Jean-Pierre;Moon, Kwang-Seok;Park, Hanhoon
    • 한국방송∙미디어공학회:학술대회논문집
    • /
    • 한국방송∙미디어공학회 2018년도 하계학술대회
    • /
    • pp.45-46
    • /
    • 2018
  • This paper focuses on human activities monitoring problem using onboard smartphone sensors as data generator. Monitoring such activities can be very important to detect anomalies and prevent disease from patients. Machine learning (ML) algorithms appear to be ideal approaches to use for processing data from smartphone to get sense of how to classify human activities. ML algorithms depend on quality, the quantity and even more important, the properties or features, that can be learnt from data. This paper proposes a mobile virtual reality visualization system that helps to view data representation in a very immersive way so that its quality and discriminative characteristics may be evaluated and improved. The proposed system comes as well with a handy data collecting application that can be accessed directly by the VR visualization part.

  • PDF

Design of Distributed Cloud System for Managing large-scale Genomic Data

  • Seine Jang;Seok-Jae Moon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • 제16권2호
    • /
    • pp.119-126
    • /
    • 2024
  • The volume of genomic data is constantly increasing in various modern industries and research fields. This growth presents new challenges and opportunities in terms of the quantity and diversity of genetic data. In this paper, we propose a distributed cloud system for integrating and managing large-scale gene databases. By introducing a distributed data storage and processing system based on the Hadoop Distributed File System (HDFS), various formats and sizes of genomic data can be efficiently integrated. Furthermore, by leveraging Spark on YARN, efficient management of distributed cloud computing tasks and optimal resource allocation are achieved. This establishes a foundation for the rapid processing and analysis of large-scale genomic data. Additionally, by utilizing BigQuery ML, machine learning models are developed to support genetic search and prediction, enabling researchers to more effectively utilize data. It is expected that this will contribute to driving innovative advancements in genetic research and applications.

Automated Prioritization of Construction Project Requirements using Machine Learning and Fuzzy Logic System

  • Hassan, Fahad ul;Le, Tuyen;Le, Chau;Shrestha, K. Joseph
    • 국제학술발표논문집
    • /
    • The 9th International Conference on Construction Engineering and Project Management
    • /
    • pp.304-311
    • /
    • 2022
  • Construction inspection is a crucial stage that ensures that all contractual requirements of a construction project are verified. The construction inspection capabilities among state highway agencies have been greatly affected due to budget reduction. As a result, efficient inspection practices such as risk-based inspection are required to optimize the use of limited resources without compromising inspection quality. Automated prioritization of textual requirements according to their criticality would be extremely helpful since contractual requirements are typically presented in an unstructured natural language in voluminous text documents. The current study introduces a novel model for predicting the risk level of requirements using machine learning (ML) algorithms. The ML algorithms tested in this study included naïve Bayes, support vector machines, logistic regression, and random forest. The training data includes sequences of requirement texts which were labeled with risk levels (such as very low, low, medium, high, very high) using the fuzzy logic systems. The fuzzy model treats the three risk factors (severity, probability, detectability) as fuzzy input variables, and implements the fuzzy inference rules to determine the labels of requirements. The performance of the model was examined on labeled dataset created by fuzzy inference rules and three different membership functions. The developed requirement risk prediction model yielded a precision, recall, and f-score of 78.18%, 77.75%, and 75.82%, respectively. The proposed model is expected to provide construction inspectors with a means for the automated prioritization of voluminous requirements by their importance, thus help to maximize the effectiveness of inspection activities under resource constraints.

  • PDF

Noncontrast Computed Tomography-Based Radiomics Analysis in Discriminating Early Hematoma Expansion after Spontaneous Intracerebral Hemorrhage

  • Zuhua Song;Dajing Guo;Zhuoyue Tang;Huan Liu;Xin Li;Sha Luo;Xueying Yao;Wenlong Song;Junjie Song;Zhiming Zhou
    • Korean Journal of Radiology
    • /
    • 제22권3호
    • /
    • pp.415-424
    • /
    • 2021
  • Objective: To determine whether noncontrast computed tomography (NCCT) models based on multivariable, radiomics features, and machine learning (ML) algorithms could further improve the discrimination of early hematoma expansion (HE) in patients with spontaneous intracerebral hemorrhage (sICH). Materials and Methods: We retrospectively reviewed 261 patients with sICH who underwent initial NCCT within 6 hours of ictus and follow-up CT within 24 hours after initial NCCT, between April 2011 and March 2019. The clinical characteristics, imaging signs and radiomics features extracted from the initial NCCT images were used to construct models to discriminate early HE. A clinical-radiologic model was constructed using a multivariate logistic regression (LR) analysis. Radiomics models, a radiomics-radiologic model, and a combined model were constructed in the training cohort (n = 182) and independently verified in the validation cohort (n = 79). Receiver operating characteristic analysis and the area under the curve (AUC) were used to evaluate the discriminative power. Results: The AUC of the clinical-radiologic model for discriminating early HE was 0.766. The AUCs of the radiomics model for discriminating early HE built using the LR algorithm in the training and validation cohorts were 0.926 and 0.850, respectively. The AUCs of the radiomics-radiologic model in the training and validation cohorts were 0.946 and 0.867, respectively. The AUCs of the combined model in the training and validation cohorts were 0.960 and 0.867, respectively. Conclusion: NCCT models based on multivariable, radiomics features and ML algorithm could improve the discrimination of early HE. The combined model was the best recommended model to identify sICH patients at risk of early HE.

AutoFe-Sel: A Meta-learning based methodology for Recommending Feature Subset Selection Algorithms

  • Irfan Khan;Xianchao Zhang;Ramesh Kumar Ayyasam;Rahman Ali
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제17권7호
    • /
    • pp.1773-1793
    • /
    • 2023
  • Automated machine learning, often referred to as "AutoML," is the process of automating the time-consuming and iterative procedures that are associated with the building of machine learning models. There have been significant contributions in this area across a number of different stages of accomplishing a data-mining task, including model selection, hyper-parameter optimization, and preprocessing method selection. Among them, preprocessing method selection is a relatively new and fast growing research area. The current work is focused on the recommendation of preprocessing methods, i.e., feature subset selection (FSS) algorithms. One limitation in the existing studies regarding FSS algorithm recommendation is the use of a single learner for meta-modeling, which restricts its capabilities in the metamodeling. Moreover, the meta-modeling in the existing studies is typically based on a single group of data characterization measures (DCMs). Nonetheless, there are a number of complementary DCM groups, and their combination will allow them to leverage their diversity, resulting in improved meta-modeling. This study aims to address these limitations by proposing an architecture for preprocess method selection that uses ensemble learning for meta-modeling, namely AutoFE-Sel. To evaluate the proposed method, we performed an extensive experimental evaluation involving 8 FSS algorithms, 3 groups of DCMs, and 125 datasets. Results show that the proposed method achieves better performance compared to three baseline methods. The proposed architecture can also be easily extended to other preprocessing method selections, e.g., noise-filter selection and imbalance handling method selection.

저수지 CO2 배출량 산정을 위한 기계학습 모델의 적용 (Applications of Machine Learning Models for the Estimation of Reservoir CO2 Emissions)

  • 유지수;정세웅;박형석
    • 한국물환경학회지
    • /
    • 제33권3호
    • /
    • pp.326-333
    • /
    • 2017
  • The lakes and reservoirs have been reported as important sources of carbon emissions to the atmosphere in many countries. Although field experiments and theoretical investigations based on the fundamental gas exchange theory have proposed the quantitative amounts of Net Atmospheric Flux (NAF) in various climate regions, there are still large uncertainties at the global scale estimation. Mechanistic models can be used for understanding and estimating the temporal and spatial variations of the NAFs considering complicated hydrodynamic and biogeochemical processes in a reservoir, but these models require extensive and expensive datasets and model parameters. On the other hand, data driven machine learning (ML) algorithms are likely to be alternative tools to estimate the NAFs in responding to independent environmental variables. The objective of this study was to develop random forest (RF) and multi-layer artificial neural network (ANN) models for the estimation of the daily $CO_2$ NAFs in Daecheong Reservoir located in Geum River of Korea, and compare the models performance against the multiple linear regression (MLR) model that proposed in the previous study (Chung et al., 2016). As a result, the RF and ANN models showed much enhanced performance in the estimation of the high NAF values, while MLR model significantly under estimated them. Across validation with 10-fold random samplings was applied to evaluate the performance of three models, and indicated that the ANN model is best, and followed by RF and MLR models.

A gene expression programming-based model to predict water inflow into tunnels

  • Arsalan Mahmoodzadeh;Hawkar Hashim Ibrahim;Laith R. Flaih;Abed Alanazi;Abdullah Alqahtani;Shtwai Alsubai;Nabil Ben Kahla;Adil Hussein Mohammed
    • Geomechanics and Engineering
    • /
    • 제37권1호
    • /
    • pp.65-72
    • /
    • 2024
  • Water ingress poses a common and intricate geological hazard with profound implications for tunnel construction's speed and safety. The project's success hinges significantly on the precision of estimating water inflow during excavation, a critical factor in early-stage decision-making during conception and design. This article introduces an optimized model employing the gene expression programming (GEP) approach to forecast tunnel water inflow. The GEP model was refined by developing an equation that best aligns with predictive outcomes. The equation's outputs were compared with measured data and assessed against practical scenarios to validate its potential applicability in calculating tunnel water input. The optimized GEP model excelled in forecasting tunnel water inflow, outperforming alternative machine learning algorithms like SVR, GPR, DT, and KNN. This positions the GEP model as a leading choice for accurate and superior predictions. A state-of-the-art machine learning-based graphical user interface (GUI) was innovatively crafted for predicting and visualizing tunnel water inflow. This cutting-edge tool leverages ML algorithms, marking a substantial advancement in tunneling prediction technologies, providing accuracy and accessibility in water inflow projections.