• Title/Summary/Keyword: matrix learning

Search Result 363, Processing Time 0.032 seconds

Sidewalk Gaseous Pollutants Estimation Through UAV Video-based Model

  • Omar, Wael;Lee, Impyeong
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.1
    • /
    • pp.1-20
    • /
    • 2022
  • As unmanned aerial vehicle (UAV) technology grew in popularity over the years, it was introduced for air quality monitoring. This can easily be used to estimate the sidewalk emission concentration by calculating road traffic emission factors of different vehicle types. These calculations require a simulation of the spread of pollutants from one or more sources given for estimation. For this purpose, a Gaussian plume dispersion model was developed based on the US EPA Motor Vehicle Emissions Simulator (MOVES), which provides an accurate estimate of fuel consumption and pollutant emissions from vehicles under a wide range of user-defined conditions. This paper describes a methodology for estimating emission concentration on the sidewalk emitted by different types of vehicles. This line source considers vehicle parameters, wind speed and direction, and pollutant concentration using a UAV equipped with a monocular camera. All were sampled over an hourly interval. In this article, the YOLOv5 deep learning model is developed, vehicle tracking is used through Deep SORT (Simple Online and Realtime Tracking), vehicle localization using a homography transformation matrix to locate each vehicle and calculate the parameters of speed and acceleration, and ultimately a Gaussian plume dispersion model was developed to estimate the CO, NOx concentrations at a sidewalk point. The results demonstrate that these estimated pollutants values are good to give a fast and reasonable indication for any near road receptor point using a cheap UAV without installing air monitoring stations along the road.

Association Rule Mining and Collaborative Filtering-Based Recommendation for Improving University Graduate Attributes

  • Sheta, Osama E.
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.339-345
    • /
    • 2022
  • Outcome-based education (OBE) is a tried-and-true teaching technique based on a set of predetermined goals. Program Educational Objectives (PEOs), Program Outcomes (POs), and Course Outcomes (COs) are the components of OBE. At the end of each year, the Program Outcomes are evaluated, and faculty members can submit many recommended measures which dependent on the relationship between the program outcomes and its courses outcomes to improve the quality of program and hence the overall educational program. When a vast number of courses are considered, bad actions may be proposed, resulting in unwanted and incorrect decisions. In this paper, a recommender system, using collaborative filtering and association rules algorithms, is proposed for predicting the best relationship between the program outcomes and its courses in order to improve the attributes of the graduates. First, a parallel algorithm is used for Collaborative Filtering on Data Model, which is designed to increase the efficiency of processing big data. Then, a parallel similar learning outcomes discovery method based on matrix correlation is proposed by mining association rules. As a case study, the proposed recommender system is applied to the Computer Information Systems program, College of Computer Sciences and Information Technology, Al-Baha University, Saudi Arabia for helping Program Quality Administration improving the quality of program outcomes. The obtained results revealed that the suggested recommender system provides more actions for boosting Graduate Attributes quality.

New Cellular Neural Networks Template for Image Halftoning based on Bayesian Rough Sets

  • Elsayed Radwan;Basem Y. Alkazemi;Ahmed I. Sharaf
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.4
    • /
    • pp.85-94
    • /
    • 2023
  • Image halftoning is a technique for varying grayscale images into two-tone binary images. Unfortunately, the static representation of an image-half toning, wherever each pixel intensity is combined by its local neighbors only, causes missing subjective problem. Also, the existing noise causes an instability criterion. In this paper an image half-toning is represented as a dynamical system for recognizing the global representation. Also, noise is reduced based on a probabilistic model. Since image half-toning is considered as 2-D matrix with a full connected pass, this structure is recognized by the dynamical system of Cellular Neural Networks (CNNs) which is defined by its template. Bayesian Rough Sets is used in exploiting the ideal CNNs construction that synthesis its dynamic. Also, Bayesian rough sets contribute to enhance the quality of the halftone image by removing noise and discovering the effective parameters in the CNNs template. The novelty of this method lies in finding a probabilistic based technique to discover the term of CNNs template and define new learning rules for CNNs internal work. A numerical experiment is conducted on image half-toning corrupted by Gaussian noise.

Blind Signal Separation Using Eigenvectors as Initial Weights in Delayed Mixtures (지연혼합에서의 초기 값으로 고유벡터를 이용하는 암묵신호분리)

  • Park, Jang-Sik;Son, Kyung-Sik;Park, Keun-Soo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.25 no.1
    • /
    • pp.14-20
    • /
    • 2006
  • In this paper. a novel technique to set up the initial weights in BSS of delayed mixtures is proposed. After analyzing Eigendecomposition for the correlation matrix of mixing data. the initial weights are set from the Eigenvectors ith delay information. The Proposed setting of initial weighting method for conventional FDICA technique improved the separation Performance. The computer simulation shows that the Proposed method achieves the improved SIR and faster convergence speed of learning curve.

GAIN-QoS: A Novel QoS Prediction Model for Edge Computing

  • Jiwon Choi;Jaewook Lee;Duksan Ryu;Suntae Kim;Jongmoon Baik
    • Journal of Web Engineering
    • /
    • v.21 no.1
    • /
    • pp.27-52
    • /
    • 2021
  • With recent increases in the number of network-connected devices, the number of edge computing services that provide similar functions has increased. Therefore, it is important to recommend an optimal edge computing service, based on quality-of-service (QoS). However, in the real world, there is a cold-start problem in QoS data: highly sparse invocation. Therefore, it is difficult to recommend a suitable service to the user. Deep learning techniques were applied to address this problem, or context information was used to extract deep features between users and services. However, edge computing environment has not been considered in previous studies. Our goal is to predict the QoS values in real edge computing environments with improved accuracy. To this end, we propose a GAIN-QoS technique. It clusters services based on their location information, calculates the distance between services and users in each cluster, and brings the QoS values of users within a certain distance. We apply a Generative Adversarial Imputation Nets (GAIN) model and perform QoS prediction based on this reconstructed user service invocation matrix. When the density is low, GAIN-QoS shows superior performance to other techniques. In addition, the distance between the service and user slightly affects performance. Thus, compared to other methods, the proposed method can significantly improve the accuracy of QoS prediction for edge computing, which suffers from cold-start problem.

Separations and Feature Extractions for Image Signals Using Independent Component Analysis Based on Neural Networks of Efficient Learning Rule (효율적인 학습규칙의 신경망 기반 독립성분분석을 이용한 영상신호의 분리 및 특징추출)

  • Cho, Yong-Hyun
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.13 no.2
    • /
    • pp.200-208
    • /
    • 2003
  • This paper proposes a separation and feature extraction of image signals using the independent component analysis(ICA) based on neural networks of efficient learning rule. The proposed learning rule is a hybrid fixed-point(FP) algorithm based on secant method and momentum. Secant method is applied to improve the performance by simplifying the 1st-order derivative computation for optimizing the objective function, which is to minimize the mutual informations of the independent components. The momentum is applied for high-speed convergence by restraining the oscillation in the process of converging to the optimal solution. The proposed algorithm has been applied to the composite images generated by random mixing matrix from the 10 images of $512\times512$-pixel. The simulation results show that the proposed algorithm has better performances of the separation speed and rate than those using the FP algorithm based on Newton and secant method. The proposed algorithm has been also applied to extract the features using a 3 set of 10,000 image patches from the 10 fingerprints of $256\times256$-pixel and the front and the rear paper money of $480\times225$-pixel, respectively, The simulation results show that the proposed algorithm has also better extraction speed than those using the another methods. Especially, the 160 basis vectors(features) of $16\times16$-pixel show the local features which have the characteristics of spatial frequency and oriented edges in the images.

Research on Characterizing Urban Color Analysis based on Tourists-Shared Photos and Machine Learning - Focused on Dali City, China - (관광객 공유한 사진 및 머신 러닝을 활용한 도시 색채 특성 분석 연구 - 중국 대리시를 대상으로 -)

  • Yin, Xiaoyan;Jung, Taeyeol
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.52 no.2
    • /
    • pp.39-50
    • /
    • 2024
  • Color is an essential visual element that has a significant impact on the formation of a city's image and people's perceptions. Quantitative analysis of color in urban environments is a complex process that has been difficult to implement in the past. However, with recent rapid advances in Machine Learning, it has become possible to analyze city colors using photos shared by tourists. This study selected Dali City, a popular tourist destination in China, as a case study. Photos of Dali City shared by tourists were collected, and a method to measure large-scale city colors was explored by combining machine learning techniques. Specifically, the DeepLabv3+ model was first applied to perform a semantic segmentation of tourist sharing photos based on the ADE20k dataset, thereby separating artificial elements in the photos. Next, the K-means clustering algorithm was used to extract colors from the artificial elements in Dali City, and an adjacency matrix was constructed to analyze the correlations between the dominant colors. The research results indicate that the main color of the artificial elements in Dali City has the highest percentage of orange-grey. Furthermore, gray tones are often used in combination with other colors. The results indicated that local ethnic and Buddhist cultures influence the color characteristics of artificial elements in Dali City. This research provides a new method of color analysis, and the results not only help Dali City to shape an urban color image that meets the expectations of tourists but also provide reference materials for future urban color planning in Dali City.

Predicting stock movements based on financial news with systematic group identification (시스템적인 군집 확인과 뉴스를 이용한 주가 예측)

  • Seong, NohYoon;Nam, Kihwan
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.1-17
    • /
    • 2019
  • Because stock price forecasting is an important issue both academically and practically, research in stock price prediction has been actively conducted. The stock price forecasting research is classified into using structured data and using unstructured data. With structured data such as historical stock price and financial statements, past studies usually used technical analysis approach and fundamental analysis. In the big data era, the amount of information has rapidly increased, and the artificial intelligence methodology that can find meaning by quantifying string information, which is an unstructured data that takes up a large amount of information, has developed rapidly. With these developments, many attempts with unstructured data are being made to predict stock prices through online news by applying text mining to stock price forecasts. The stock price prediction methodology adopted in many papers is to forecast stock prices with the news of the target companies to be forecasted. However, according to previous research, not only news of a target company affects its stock price, but news of companies that are related to the company can also affect the stock price. However, finding a highly relevant company is not easy because of the market-wide impact and random signs. Thus, existing studies have found highly relevant companies based primarily on pre-determined international industry classification standards. However, according to recent research, global industry classification standard has different homogeneity within the sectors, and it leads to a limitation that forecasting stock prices by taking them all together without considering only relevant companies can adversely affect predictive performance. To overcome the limitation, we first used random matrix theory with text mining for stock prediction. Wherever the dimension of data is large, the classical limit theorems are no longer suitable, because the statistical efficiency will be reduced. Therefore, a simple correlation analysis in the financial market does not mean the true correlation. To solve the issue, we adopt random matrix theory, which is mainly used in econophysics, to remove market-wide effects and random signals and find a true correlation between companies. With the true correlation, we perform cluster analysis to find relevant companies. Also, based on the clustering analysis, we used multiple kernel learning algorithm, which is an ensemble of support vector machine to incorporate the effects of the target firm and its relevant firms simultaneously. Each kernel was assigned to predict stock prices with features of financial news of the target firm and its relevant firms. The results of this study are as follows. The results of this paper are as follows. (1) Following the existing research flow, we confirmed that it is an effective way to forecast stock prices using news from relevant companies. (2) When looking for a relevant company, looking for it in the wrong way can lower AI prediction performance. (3) The proposed approach with random matrix theory shows better performance than previous studies if cluster analysis is performed based on the true correlation by removing market-wide effects and random signals. The contribution of this study is as follows. First, this study shows that random matrix theory, which is used mainly in economic physics, can be combined with artificial intelligence to produce good methodologies. This suggests that it is important not only to develop AI algorithms but also to adopt physics theory. This extends the existing research that presented the methodology by integrating artificial intelligence with complex system theory through transfer entropy. Second, this study stressed that finding the right companies in the stock market is an important issue. This suggests that it is not only important to study artificial intelligence algorithms, but how to theoretically adjust the input values. Third, we confirmed that firms classified as Global Industrial Classification Standard (GICS) might have low relevance and suggested it is necessary to theoretically define the relevance rather than simply finding it in the GICS.

Label Embedding for Improving Classification Accuracy UsingAutoEncoderwithSkip-Connections (다중 레이블 분류의 정확도 향상을 위한 스킵 연결 오토인코더 기반 레이블 임베딩 방법론)

  • Kim, Museong;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.175-197
    • /
    • 2021
  • Recently, with the development of deep learning technology, research on unstructured data analysis is being actively conducted, and it is showing remarkable results in various fields such as classification, summary, and generation. Among various text analysis fields, text classification is the most widely used technology in academia and industry. Text classification includes binary class classification with one label among two classes, multi-class classification with one label among several classes, and multi-label classification with multiple labels among several classes. In particular, multi-label classification requires a different training method from binary class classification and multi-class classification because of the characteristic of having multiple labels. In addition, since the number of labels to be predicted increases as the number of labels and classes increases, there is a limitation in that performance improvement is difficult due to an increase in prediction difficulty. To overcome these limitations, (i) compressing the initially given high-dimensional label space into a low-dimensional latent label space, (ii) after performing training to predict the compressed label, (iii) restoring the predicted label to the high-dimensional original label space, research on label embedding is being actively conducted. Typical label embedding techniques include Principal Label Space Transformation (PLST), Multi-Label Classification via Boolean Matrix Decomposition (MLC-BMaD), and Bayesian Multi-Label Compressed Sensing (BML-CS). However, since these techniques consider only the linear relationship between labels or compress the labels by random transformation, it is difficult to understand the non-linear relationship between labels, so there is a limitation in that it is not possible to create a latent label space sufficiently containing the information of the original label. Recently, there have been increasing attempts to improve performance by applying deep learning technology to label embedding. Label embedding using an autoencoder, a deep learning model that is effective for data compression and restoration, is representative. However, the traditional autoencoder-based label embedding has a limitation in that a large amount of information loss occurs when compressing a high-dimensional label space having a myriad of classes into a low-dimensional latent label space. This can be found in the gradient loss problem that occurs in the backpropagation process of learning. To solve this problem, skip connection was devised, and by adding the input of the layer to the output to prevent gradient loss during backpropagation, efficient learning is possible even when the layer is deep. Skip connection is mainly used for image feature extraction in convolutional neural networks, but studies using skip connection in autoencoder or label embedding process are still lacking. Therefore, in this study, we propose an autoencoder-based label embedding methodology in which skip connections are added to each of the encoder and decoder to form a low-dimensional latent label space that reflects the information of the high-dimensional label space well. In addition, the proposed methodology was applied to actual paper keywords to derive the high-dimensional keyword label space and the low-dimensional latent label space. Using this, we conducted an experiment to predict the compressed keyword vector existing in the latent label space from the paper abstract and to evaluate the multi-label classification by restoring the predicted keyword vector back to the original label space. As a result, the accuracy, precision, recall, and F1 score used as performance indicators showed far superior performance in multi-label classification based on the proposed methodology compared to traditional multi-label classification methods. This can be seen that the low-dimensional latent label space derived through the proposed methodology well reflected the information of the high-dimensional label space, which ultimately led to the improvement of the performance of the multi-label classification itself. In addition, the utility of the proposed methodology was identified by comparing the performance of the proposed methodology according to the domain characteristics and the number of dimensions of the latent label space.

Improvement of the PFCM(Possibilistic Fuzzy C-Means) Clustering Method (PFCM 클러스터링 기법의 개선)

  • Heo, Gyeong-Yong;Choe, Se-Woon;Woo, Young-Woon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.13 no.1
    • /
    • pp.177-185
    • /
    • 2009
  • Cluster analysis or clustering is a kind of unsupervised learning method in which a set of data points is divided into a given number of homogeneous groups. Fuzzy clustering method, one of the most popular clustering method, allows a point to belong to all the clusters with different degrees, so produces more intuitive and natural clusters than hard clustering method does. Even more some of fuzzy clustering variants have noise-immunity. In this paper, we improved the Possibilistic Fuzzy C-Means (PFCM), which generates a membership matrix as well as a typicality matrix, using Gath-Geva (GG) method. The proposed method has a focus on the boundaries of clusters, which is different from most of the other methods having a focus on the centers of clusters. The generated membership values are suitable for the classification-type applications. As the typicality values generated from the algorithm have a similar distribution with the values of density function of Gaussian distribution, it is useful for Gaussian-type density estimation. Even more GG method can handle the clusters having different numbers of data points, which the other well-known method by Gustafson and Kessel can not. All of these points are obvious in the experimental results.