• Title/Summary/Keyword: Bayesian deep learning

Search Result 34, Processing Time 0.026 seconds

The faintest quasar luminosity function at z ~ 5 from Deep Learning and Bayesian Inference

  • Shin, Suhyun;Im, Myungshin
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.46 no.1
    • /
    • pp.31.2-31.2
    • /
    • 2021
  • To estimate the contribution of quasars on keeping the IGM ionized, building a quasar luminosity function (LF) is necessary. Quasar LFs derived from multiple quasar surveys, however, are incompatible, especially for the faint regime, emphasizing the need for deep images. In this study, we construct quasar LF reaching M1450~-21.5 AB magnitude at z ~ 5, which is 1.5 mag deeper than previously reported LFs, using deep images from the Hyper Suprime-Cam Subaru Strategic Program (HSC-SSP). We trained an artificial neural network (ANN) by inserting the colors as inputs to classify the quasars at z ~ 5 from the late-type stars and low-redshift galaxies. The accuracy of ANN is > 99 %. We also adopted the Bayesian information criterion to elaborate on the quasar-like objects. As a result, we recovered 5/5 confirmed quasars and remarkably minimized the contamination rate of high-redshift galaxies by up to six times compared to the selection using color selection alone. The constructed quasar parametric LF shows a flatter faint-end slope α=-127+0.16-0.15 similar to the recent LFs. The number of faint quasars (M1450 < -23.5) is too few to be the main contributor to IGM ionizing photons.

  • PDF

Bayesian Game Theoretic Model for Evasive AI Malware Detection in IoT

  • Jun-Won Ho
    • International journal of advanced smart convergence
    • /
    • v.13 no.3
    • /
    • pp.41-47
    • /
    • 2024
  • In this paper, we deal with a game theoretic problem to explore interactions between evasive Artificial Intelligence (AI) malware and detectors in Internet of Things (IoT). Evasive AI malware is defined as malware having capability of eluding detection by exploiting artificial intelligence such as machine learning and deep leaning. Detectors are defined as IoT devices participating in detection of evasive AI malware in IoT. They can be separated into two groups such that one group of detectors can be armed with detection capability powered by AI, the other group cannot be armed with it. Evasive AI malware can take three strategies of Non-attack, Non-AI attack, AI attack. To cope with these strategies of evasive AI malware, detector can adopt three strategies of Non-defense, Non-AI defense, AI defense. We formulate a Bayesian game theoretic model with these strategies employed by evasive AI malware and detector. We derive pure strategy Bayesian Nash Equilibria in a single stage game from the formulated Bayesian game theoretic model. Our devised work is useful in the sense that it can be used as a basic game theoretic model for developing AI malware detection schemes.

Development of Deterioration Model for Cracks in Asphalt Pavement Using Deep Learning-Based Road Asset Monitoring System (딥러닝 기반의 도로자산 모니터링 시스템을 활용한 아스팔트 도로포장 균열률 파손모델 개발)

  • Park, Jeong-Gwon;Kim, Chang-Hak;Choi, Seung-Hyun;Do, Myung-Sik
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.21 no.5
    • /
    • pp.133-148
    • /
    • 2022
  • In this study, a road pavement crack deterioration model was developed for a pavement road sections of the Sejong-city. Data required for model development were acquired using a deep learning-based road asset monitoring system. Road pavement monitoring was conducted on the same sections in 2021 and 2022. The developed model was analyzed by dividing it into a method for estimating the annual average amount of deterioration and a method based on Bayesian Markov Mixture Hazard model. As a result of the analysis, it was found that an analysis results similar to the crack deterioration model developed based on the data acquired from the Automatic pavement investigation equipmen was derived. The results of this study are expected to be used as basic data by local governments to establish road management plans.

Hyperparameter Search for Facies Classification with Bayesian Optimization (베이지안 최적화를 이용한 암상 분류 모델의 하이퍼 파라미터 탐색)

  • Choi, Yonguk;Yoon, Daeung;Choi, Junhwan;Byun, Joongmoo
    • Geophysics and Geophysical Exploration
    • /
    • v.23 no.3
    • /
    • pp.157-167
    • /
    • 2020
  • With the recent advancement of computer hardware and the contribution of open source libraries to facilitate access to artificial intelligence technology, the use of machine learning (ML) and deep learning (DL) technologies in various fields of exploration geophysics has increased. In addition, ML researchers have developed complex algorithms to improve the inference accuracy of various tasks such as image, video, voice, and natural language processing, and now they are expanding their interests into the field of automatic machine learning (AutoML). AutoML can be divided into three areas: feature engineering, architecture search, and hyperparameter search. Among them, this paper focuses on hyperparamter search with Bayesian optimization, and applies it to the problem of facies classification using seismic data and well logs. The effectiveness of the Bayesian optimization technique has been demonstrated using Vincent field data by comparing with the results of the random search technique.

Comprehensive analysis of deep learning-based target classifiers in small and imbalanced active sonar datasets (소량 및 불균형 능동소나 데이터세트에 대한 딥러닝 기반 표적식별기의 종합적인 분석)

  • Geunhwan Kim;Youngsang Hwang;Sungjin Shin;Juho Kim;Soobok Hwang;Youngmin Choo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.42 no.4
    • /
    • pp.329-344
    • /
    • 2023
  • In this study, we comprehensively analyze the generalization performance of various deep learning-based active sonar target classifiers when applied to small and imbalanced active sonar datasets. To generate the active sonar datasets, we use data from two different oceanic experiments conducted at different times and ocean. Each sample in the active sonar datasets is a time-frequency domain image, which is extracted from audio signal of contact after the detection process. For the comprehensive analysis, we utilize 22 Convolutional Neural Networks (CNN) models. Two datasets are used as train/validation datasets and test datasets, alternatively. To calculate the variance in the output of the target classifiers, the train/validation/test datasets are repeated 10 times. Hyperparameters for training are optimized using Bayesian optimization. The results demonstrate that shallow CNN models show superior robustness and generalization performance compared to most of deep CNN models. The results from this paper can serve as a valuable reference for future research directions in deep learning-based active sonar target classification.

Recommender system using BERT sentiment analysis (BERT 기반 감성분석을 이용한 추천시스템)

  • Park, Ho-yeon;Kim, Kyoung-jae
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.2
    • /
    • pp.1-15
    • /
    • 2021
  • If it is difficult for us to make decisions, we ask for advice from friends or people around us. When we decide to buy products online, we read anonymous reviews and buy them. With the advent of the Data-driven era, IT technology's development is spilling out many data from individuals to objects. Companies or individuals have accumulated, processed, and analyzed such a large amount of data that they can now make decisions or execute directly using data that used to depend on experts. Nowadays, the recommender system plays a vital role in determining the user's preferences to purchase goods and uses a recommender system to induce clicks on web services (Facebook, Amazon, Netflix, Youtube). For example, Youtube's recommender system, which is used by 1 billion people worldwide every month, includes videos that users like, "like" and videos they watched. Recommended system research is deeply linked to practical business. Therefore, many researchers are interested in building better solutions. Recommender systems use the information obtained from their users to generate recommendations because the development of the provided recommender systems requires information on items that are likely to be preferred by the user. We began to trust patterns and rules derived from data rather than empirical intuition through the recommender systems. The capacity and development of data have led machine learning to develop deep learning. However, such recommender systems are not all solutions. Proceeding with the recommender systems, there should be no scarcity in all data and a sufficient amount. Also, it requires detailed information about the individual. The recommender systems work correctly when these conditions operate. The recommender systems become a complex problem for both consumers and sellers when the interaction log is insufficient. Because the seller's perspective needs to make recommendations at a personal level to the consumer and receive appropriate recommendations with reliable data from the consumer's perspective. In this paper, to improve the accuracy problem for "appropriate recommendation" to consumers, the recommender systems are proposed in combination with context-based deep learning. This research is to combine user-based data to create hybrid Recommender Systems. The hybrid approach developed is not a collaborative type of Recommender Systems, but a collaborative extension that integrates user data with deep learning. Customer review data were used for the data set. Consumers buy products in online shopping malls and then evaluate product reviews. Rating reviews are based on reviews from buyers who have already purchased, giving users confidence before purchasing the product. However, the recommendation system mainly uses scores or ratings rather than reviews to suggest items purchased by many users. In fact, consumer reviews include product opinions and user sentiment that will be spent on evaluation. By incorporating these parts into the study, this paper aims to improve the recommendation system. This study is an algorithm used when individuals have difficulty in selecting an item. Consumer reviews and record patterns made it possible to rely on recommendations appropriately. The algorithm implements a recommendation system through collaborative filtering. This study's predictive accuracy is measured by Root Mean Squared Error (RMSE) and Mean Absolute Error (MAE). Netflix is strategically using the referral system in its programs through competitions that reduce RMSE every year, making fair use of predictive accuracy. Research on hybrid recommender systems combining the NLP approach for personalization recommender systems, deep learning base, etc. has been increasing. Among NLP studies, sentiment analysis began to take shape in the mid-2000s as user review data increased. Sentiment analysis is a text classification task based on machine learning. The machine learning-based sentiment analysis has a disadvantage in that it is difficult to identify the review's information expression because it is challenging to consider the text's characteristics. In this study, we propose a deep learning recommender system that utilizes BERT's sentiment analysis by minimizing the disadvantages of machine learning. This study offers a deep learning recommender system that uses BERT's sentiment analysis by reducing the disadvantages of machine learning. The comparison model was performed through a recommender system based on Naive-CF(collaborative filtering), SVD(singular value decomposition)-CF, MF(matrix factorization)-CF, BPR-MF(Bayesian personalized ranking matrix factorization)-CF, LSTM, CNN-LSTM, GRU(Gated Recurrent Units). As a result of the experiment, the recommender system based on BERT was the best.

Genetic Algorithm based hyperparameter tuned CNN for identifying IoT intrusions

  • Alexander. R;Pradeep Mohan Kumar. K
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.3
    • /
    • pp.755-778
    • /
    • 2024
  • In recent years, the number of devices being connected to the internet has grown enormously, as has the intrusive behavior in the network. Thus, it is important for intrusion detection systems to report all intrusive behavior. Using deep learning and machine learning algorithms, intrusion detection systems are able to perform well in identifying attacks. However, the concern with these deep learning algorithms is their inability to identify a suitable network based on traffic volume, which requires manual changing of hyperparameters, which consumes a lot of time and effort. So, to address this, this paper offers a solution using the extended compact genetic algorithm for the automatic tuning of the hyperparameters. The novelty in this work comes in the form of modeling the problem of identifying attacks as a multi-objective optimization problem and the usage of linkage learning for solving the optimization problem. The solution is obtained using the feature map-based Convolutional Neural Network that gets encoded into genes, and using the extended compact genetic algorithm the model is optimized for the detection accuracy and latency. The CIC-IDS-2017 and 2018 datasets are used to verify the hypothesis, and the most recent analysis yielded a substantial F1 score of 99.23%. Response time, CPU, and memory consumption evaluations are done to demonstrate the suitability of this model in a fog environment.

Prediction of skewness and kurtosis of pressure coefficients on a low-rise building by deep learning

  • Youqin Huang;Guanheng Ou;Jiyang Fu;Huifan Wu
    • Wind and Structures
    • /
    • v.36 no.6
    • /
    • pp.393-404
    • /
    • 2023
  • Skewness and kurtosis are important higher-order statistics for simulating non-Gaussian wind pressure series on low-rise buildings, but their predictions are less studied in comparison with those of the low order statistics as mean and rms. The distribution gradients of skewness and kurtosis on roofs are evidently higher than those of mean and rms, which increases their prediction difficulty. The conventional artificial neural networks (ANNs) used for predicting mean and rms show unsatisfactory accuracy in predicting skewness and kurtosis owing to the limited capacity of shallow learning of ANNs. In this work, the deep neural networks (DNNs) model with the ability of deep learning is introduced to predict the skewness and kurtosis on a low-rise building. For obtaining the optimal generalization of the DNNs model, the hyper parameters are automatically determined by Bayesian Optimization (BO). Moreover, for providing a benchmark for future studies on predicting higher order statistics, the data sets for training and testing the DNNs model are extracted from the internationally open NIST-UWO database, and the prediction errors of all taps are comprehensively quantified by various error metrices. The results show that the prediction accuracy in this study is apparently better than that in the literature, since the correlation coefficient between the predicted and experimental results is 0.99 and 0.75 in this paper and the literature respectively. In the untrained cornering wind direction, the distributions of skewness and kurtosis are well captured by DNNs on the whole building including the roof corner with strong non-normality, and the correlation coefficients between the predicted and experimental results are 0.99 and 0.95 for skewness and kurtosis respectively.

Optimize rainfall prediction utilize multivariate time series, seasonal adjustment and Stacked Long short term memory

  • Nguyen, Thi Huong;Kwon, Yoon Jeong;Yoo, Je-Ho;Kwon, Hyun-Han
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.373-373
    • /
    • 2021
  • Rainfall forecasting is an important issue that is applied in many areas, such as agriculture, flood warning, and water resources management. In this context, this study proposed a statistical and machine learning-based forecasting model for monthly rainfall. The Bayesian Gaussian process was chosen to optimize the hyperparameters of the Stacked Long Short-term memory (SLSTM) model. The proposed SLSTM model was applied for predicting monthly precipitation of Seoul station, South Korea. Data were retrieved from the Korea Meteorological Administration (KMA) in the period between 1960 and 2019. Four schemes were examined in this study: (i) prediction with only rainfall; (ii) with deseasonalized rainfall; (iii) with rainfall and minimum temperature; (iv) with deseasonalized rainfall and minimum temperature. The error of predicted rainfall based on the root mean squared error (RMSE), 16-17 mm, is relatively small compared with the average monthly rainfall at Seoul station is 117mm. The results showed scheme (iv) gives the best prediction result. Therefore, this approach is more straightforward than the hydrological and hydraulic models, which request much more input data. The result indicated that a deep learning network could be applied successfully in the hydrology field. Overall, the proposed method is promising, given a good solution for rainfall prediction.

  • PDF

A novel radioactive particle tracking algorithm based on deep rectifier neural network

  • Dam, Roos Sophia de Freitas;dos Santos, Marcelo Carvalho;do Desterro, Filipe Santana Moreira;Salgado, William Luna;Schirru, Roberto;Salgado, Cesar Marques
    • Nuclear Engineering and Technology
    • /
    • v.53 no.7
    • /
    • pp.2334-2340
    • /
    • 2021
  • Radioactive particle tracking (RPT) is a minimally invasive nuclear technique that tracks a radioactive particle inside a volume of interest by means of a mathematical location algorithm. During the past decades, many algorithms have been developed including ones based on artificial intelligence techniques. In this study, RPT technique is applied in a simulated test section that employs a simplified mixer filled with concrete, six scintillator detectors and a137Cs radioactive particle emitting gamma rays of 662 keV. The test section was developed using MCNPX code, which is a mathematical code based on Monte Carlo simulation, and 3516 different radioactive particle positions (x,y,z) were simulated. Novelty of this paper is the use of a location algorithm based on a deep learning model, more specifically a 6-layers deep rectifier neural network (DRNN), in which hyperparameters were defined using a Bayesian optimization method. DRNN is a type of deep feedforward neural network that substitutes the usual sigmoid based activation functions, traditionally used in vanilla Multilayer Perceptron Networks, for rectified activation functions. Results show the great accuracy of the DRNN in a RPT tracking system. Root mean squared error for x, y and coordinates of the radioactive particle is, respectively, 0.03064, 0.02523 and 0.07653.