• 제목/요약/키워드: Deep features

검색결과 1,078건 처리시간 0.024초

Predicting Session Conversion on E-commerce: A Deep Learning-based Multimodal Fusion Approach

  • Minsu Kim;Woosik Shin;SeongBeom Kim;Hee-Woong Kim
    • Asia pacific journal of information systems
    • /
    • 제33권3호
    • /
    • pp.737-767
    • /
    • 2023
  • With the availability of big customer data and advances in machine learning techniques, the prediction of customer behavior at the session-level has attracted considerable attention from marketing practitioners and scholars. This study aims to predict customer purchase conversion at the session-level by employing customer profile, transaction, and clickstream data. For this purpose, we develop a multimodal deep learning fusion model with dynamic and static features (i.e., DS-fusion). Specifically, we base page views within focal visist and recency, frequency, monetary value, and clumpiness (RFMC) for dynamic and static features, respectively, to comprehensively capture customer characteristics for buying behaviors. Our model with deep learning architectures combines these features for conversion prediction. We validate the proposed model using real-world e-commerce data. The experimental results reveal that our model outperforms unimodal classifiers with each feature and the classical machine learning models with dynamic and static features, including random forest and logistic regression. In this regard, this study sheds light on the promise of the machine learning approach with the complementary method for different modalities in predicting customer behaviors.

LSTM Android Malicious Behavior Analysis Based on Feature Weighting

  • Yang, Qing;Wang, Xiaoliang;Zheng, Jing;Ge, Wenqi;Bai, Ming;Jiang, Frank
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제15권6호
    • /
    • pp.2188-2203
    • /
    • 2021
  • With the rapid development of mobile Internet, smart phones have been widely popularized, among which Android platform dominates. Due to it is open source, malware on the Android platform is rampant. In order to improve the efficiency of malware detection, this paper proposes deep learning Android malicious detection system based on behavior features. First of all, the detection system adopts the static analysis method to extract different types of behavior features from Android applications, and extract sensitive behavior features through Term frequency-inverse Document Frequency algorithm for each extracted behavior feature to construct detection features through unified abstract expression. Secondly, Long Short-Term Memory neural network model is established to select and learn from the extracted attributes and the learned attributes are used to detect Android malicious applications, Analysis and further optimization of the application behavior parameters, so as to build a deep learning Android malicious detection method based on feature analysis. We use different types of features to evaluate our method and compare it with various machine learning-based methods. Study shows that it outperforms most existing machine learning based approaches and detects 95.31% of the malware.

A Video Expression Recognition Method Based on Multi-mode Convolution Neural Network and Multiplicative Feature Fusion

  • Ren, Qun
    • Journal of Information Processing Systems
    • /
    • 제17권3호
    • /
    • pp.556-570
    • /
    • 2021
  • The existing video expression recognition methods mainly focus on the spatial feature extraction of video expression images, but tend to ignore the dynamic features of video sequences. To solve this problem, a multi-mode convolution neural network method is proposed to effectively improve the performance of facial expression recognition in video. Firstly, OpenFace 2.0 is used to detect face images in video, and two deep convolution neural networks are used to extract spatiotemporal expression features. Furthermore, spatial convolution neural network is used to extract the spatial information features of each static expression image, and the dynamic information feature is extracted from the optical flow information of multiple expression images based on temporal convolution neural network. Then, the spatiotemporal features learned by the two deep convolution neural networks are fused by multiplication. Finally, the fused features are input into support vector machine to realize the facial expression classification. Experimental results show that the recognition accuracy of the proposed method can reach 64.57% and 60.89%, respectively on RML and Baum-ls datasets. It is better than that of other contrast methods.

텐서플로우를 이용한 주가 예측에서 가격-기반 입력 피쳐의 예측 성능 평가 (Performance Evaluation of Price-based Input Features in Stock Price Prediction using Tensorflow)

  • 송유정;이재원;이종우
    • 정보과학회 컴퓨팅의 실제 논문지
    • /
    • 제23권11호
    • /
    • pp.625-631
    • /
    • 2017
  • 과거부터 현재까지 주식시장에 대한 주가 변동 예측은 풀리지 않는 난제이다. 주가를 과학적으로 예측하기 위해 다양한 시도 및 연구들이 있어왔지만, 아직까지 정확한 미래를 예측하는 것은 불가능하다. 하지만, 주가 예측은 경제, 수학, 물리 그리고 전산학 등 여러 관련 분야에서 오랜 관심의 대상이 되어왔다. 본 논문에서는 최근 각광 받고 있는 딥러닝(Deep-Learning)을 이용하여 주가의 변동패턴을 학습하고 미래를 예측하고자한다. 본 연구에서는 오픈소스 딥러닝 프레임워크인 텐서플로우를 이용하여 총 3가지 학습 모델을 제시하였으며, 각 학습모델은 각기 다른 입력 피쳐들을 받아들여 학습을 진행한다. 입력 피쳐는 이전 연구에서 사용한 단순 가격 데이터를 확장해 입력 피쳐 개수를 증가시켜가며 실험을 하였다. 세 가지 예측 모델의 학습 성능을 측정했으며, 이를 통해 가격-기반 입력 피쳐에 따라 달라지는 예측 모델의 성능 변화 비교 분석하여 가격-기반 입력 피쳐가 주가예측에 미치는 영향을 평가하였다.

깊은 신경망 특징 기반 화자 검증 시스템의 성능 비교 (Performance Comparison of Deep Feature Based Speaker Verification Systems)

  • 김대현;성우경;김홍국
    • 말소리와 음성과학
    • /
    • 제7권4호
    • /
    • pp.9-16
    • /
    • 2015
  • In this paper, several experiments are performed according to deep neural network (DNN) based features for the performance comparison of speaker verification (SV) systems. To this end, input features for a DNN, such as mel-frequency cepstral coefficient (MFCC), linear-frequency cepstral coefficient (LFCC), and perceptual linear prediction (PLP), are first compared in a view of the SV performance. After that, the effect of a DNN training method and a structure of hidden layers of DNNs on the SV performance is investigated depending on the type of features. The performance of an SV system is then evaluated on the basis of I-vector or probabilistic linear discriminant analysis (PLDA) scoring method. It is shown from SV experiments that a tandem feature of DNN bottleneck feature and MFCC feature gives the best performance when DNNs are configured using a rectangular type of hidden layers and trained with a supervised training method.

Tracking by Detection of Multiple Faces using SSD and CNN Features

  • Tai, Do Nhu;Kim, Soo-Hyung;Lee, Guee-Sang;Yang, Hyung-Jeong;Na, In-Seop;Oh, A-Ran
    • 스마트미디어저널
    • /
    • 제7권4호
    • /
    • pp.61-69
    • /
    • 2018
  • Multi-tracking of general objects and specific faces is an important topic in the field of computer vision applicable to many branches of industry such as biometrics, security, etc. The rapid development of deep neural networks has resulted in a dramatic improvement in face recognition and object detection problems, which helps improve the multiple-face tracking techniques exploiting the tracking-by-detection method. Our proposed method uses face detection trained with a head dataset to resolve the face deformation problem in the tracking process. Further, we use robust face features extracted from the deep face recognition network to match the tracklets with tracking faces using Hungarian matching method. We achieved promising results regarding the usage of deep face features and head detection in a face tracking benchmark.

자율 주행을 위한 심층 학습 기반 차선 인식 모델 분석 (Analysis of Deep Learning-Based Lane Detection Models for Autonomous Driving)

  • 이현종;윤의현;하정민;이재구
    • 대한임베디드공학회논문지
    • /
    • 제18권5호
    • /
    • pp.225-231
    • /
    • 2023
  • With the recent surge in the autonomous driving market, the significance of lane detection technology has escalated. Lane detection plays a pivotal role in autonomous driving systems by identifying lanes to ensure safe vehicle operation. Traditional lane detection models rely on engineers manually extracting lane features from predefined environments. However, real-world road conditions present diverse challenges, hampering the engineers' ability to extract adaptable lane features, resulting in limited performance. Consequently, recent research has focused on developing deep learning based lane detection models to extract lane features directly from data. In this paper, we classify lane detection models into four categories: cluster-based, curve-based, information propagation-based, and anchor-based methods. We conduct an extensive analysis of the strengths and weaknesses of each approach, evaluate the model's performance on an embedded board, and assess their practicality and effectiveness. Based on our findings, we propose future research directions and potential enhancements.

Video augmentation technique for human action recognition using genetic algorithm

  • Nida, Nudrat;Yousaf, Muhammad Haroon;Irtaza, Aun;Velastin, Sergio A.
    • ETRI Journal
    • /
    • 제44권2호
    • /
    • pp.327-338
    • /
    • 2022
  • Classification models for human action recognition require robust features and large training sets for good generalization. However, data augmentation methods are employed for imbalanced training sets to achieve higher accuracy. These samples generated using data augmentation only reflect existing samples within the training set, their feature representations are less diverse and hence, contribute to less precise classification. This paper presents new data augmentation and action representation approaches to grow training sets. The proposed approach is based on two fundamental concepts: virtual video generation for augmentation and representation of the action videos through robust features. Virtual videos are generated from the motion history templates of action videos, which are convolved using a convolutional neural network, to generate deep features. Furthermore, by observing an objective function of the genetic algorithm, the spatiotemporal features of different samples are combined, to generate the representations of the virtual videos and then classified through an extreme learning machine classifier on MuHAVi-Uncut, iXMAS, and IAVID-1 datasets.

하지 심부정맥혈전증을 가진 중환자의 특성 및 발생 영향 요인 (Risk Factors and Features of Critically Ill Patients with Deep Vein Thrombosis in Lower Extremities)

  • 김화순;조옥민;조효임;김주연
    • 대한간호학회지
    • /
    • 제42권3호
    • /
    • pp.396-404
    • /
    • 2012
  • Purpose: The purpose of this study was to identify the features, risk scores and risk factors for deep vein thrombosis in critically ill patients who developed deep vein thrombosis in their lower extremities. Methods: The participants in this prospective descriptive study were 175 adult patients who did not receive any prophylactic medication or mechanical therapy during their admission in the intensive care unit. Results: The mean age was 62.24 (${\pm}17.28$) years. Men made up 54.9% of the participating patients. There were significant differences in age, body mass index, and leg swelling between patients who developed deep vein thrombosis and those who did not have deep vein thrombosis. The mean risk score was 6.71(${\pm}2.94$) and they had on average 4.01(${\pm}1.35$) risk factors. In the multiple logistic regression, body mass index (odds ratio=1.14) and leg swelling (odds ratio=6.05) were significant predictors of deep vein thrombosis. Conclusion: Most critically ill patients are in the potentially high risk group for deep vein thrombosis. However, patients who are elderly, obese or have leg edema should be closely assessed and more than one type of active prophylactic intervention should be provided.

Multi-type Image Noise Classification by Using Deep Learning

  • Waqar Ahmed;Zahid Hussain Khand;Sajid Khan;Ghulam Mujtaba;Muhammad Asif Khan;Ahmad Waqas
    • International Journal of Computer Science & Network Security
    • /
    • 제24권7호
    • /
    • pp.143-147
    • /
    • 2024
  • Image noise classification is a classical problem in the field of image processing, machine learning, deep learning and computer vision. In this paper, image noise classification is performed using deep learning. Keras deep learning library of TensorFlow is used for this purpose. 6900 images images are selected from the Kaggle database for the classification purpose. Dataset for labeled noisy images of multiple type was generated with the help of Matlab from a dataset of non-noisy images. Labeled dataset comprised of Salt & Pepper, Gaussian and Sinusoidal noise. Different training and tests sets were partitioned to train and test the model for image classification. In deep neural networks CNN (Convolutional Neural Network) is used due to its in-depth and hidden patterns and features learning in the images to be classified. This deep learning of features and patterns in images make CNN outperform the other classical methods in many classification problems.