• Title/Summary/Keyword: Auto detection

Search Result 355, Processing Time 0.021 seconds

An Adaptive ROI Decision for Real-time Performance in an Autonomous Driving Perception Module (자율주행 인지 모듈의 실시간 성능을 위한 적응형 관심 영역 판단)

  • Lee, Ayoung;Lee, Hojoon;Yi, Kyongsu
    • Journal of Auto-vehicle Safety Association
    • /
    • v.14 no.2
    • /
    • pp.20-25
    • /
    • 2022
  • This paper represents an adaptive Region of Interest (ROI) decision for real-time performance in an autonomous driving perception module. Since the whole automated driving system consists of numerous modules and subdivisions of module occur, it is necessary to consider the characteristics, complexity, and limitations of each module. Furthermore, Light Detection And Ranging (Lidar) sensors require a considerable amount of time. In view of these limitations, division of submodule is inevitable to represent high real-time performance for stable system. This paper proposes ROI to reduce the number of data respect to computation time. ROI is set by a road's design speed and the corresponding ROI is applied differently to each vehicle considering its speed. The simulation model is constructed by ROS, and overall data analysis is conducted by Matlab. The algorithm is validated using real-time driving data in urban environment, and the result shows that ROI provides low computational costs.

Multimodal Biometrics Recognition from Facial Video with Missing Modalities Using Deep Learning

  • Maity, Sayan;Abdel-Mottaleb, Mohamed;Asfour, Shihab S.
    • Journal of Information Processing Systems
    • /
    • v.16 no.1
    • /
    • pp.6-29
    • /
    • 2020
  • Biometrics identification using multiple modalities has attracted the attention of many researchers as it produces more robust and trustworthy results than single modality biometrics. In this paper, we present a novel multimodal recognition system that trains a deep learning network to automatically learn features after extracting multiple biometric modalities from a single data source, i.e., facial video clips. Utilizing different modalities, i.e., left ear, left profile face, frontal face, right profile face, and right ear, present in the facial video clips, we train supervised denoising auto-encoders to automatically extract robust and non-redundant features. The automatically learned features are then used to train modality specific sparse classifiers to perform the multimodal recognition. Moreover, the proposed technique has proven robust when some of the above modalities were missing during the testing. The proposed system has three main components that are responsible for detection, which consists of modality specific detectors to automatically detect images of different modalities present in facial video clips; feature selection, which uses supervised denoising sparse auto-encoders network to capture discriminative representations that are robust to the illumination and pose variations; and classification, which consists of a set of modality specific sparse representation classifiers for unimodal recognition, followed by score level fusion of the recognition results of the available modalities. Experiments conducted on the constrained facial video dataset (WVU) and the unconstrained facial video dataset (HONDA/UCSD), resulted in a 99.17% and 97.14% Rank-1 recognition rates, respectively. The multimodal recognition accuracy demonstrates the superiority and robustness of the proposed approach irrespective of the illumination, non-planar movement, and pose variations present in the video clips even in the situation of missing modalities.

Selection of ROI for the AF using by Learning Algorithm and Stabilization Method for the Region (학습 알고리즘을 이용한 AF용 ROI 선택과 영역 안정화 방법)

  • Han, Hag-Yong;Jang, Won-Woo;Ha, Joo-Young;Hur, Kang-In;Kang, Bong-Soon
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.10 no.4
    • /
    • pp.233-238
    • /
    • 2009
  • In this paper, we propose the methods to select the stable region for the detect region which is required in the system used the face to the ROI in the auto-focus digital camera. this method regards the face region as the ROI in the progressive input frame and focusing the region in the mobile camera embeded ISP module automatically. The learning algorithm to detect the face is the Adaboost algorithm. we proposed the method to detect the slanted face not participate in the train process and postprocessing method for the results of detection, and then we proposed the stabilization method to sustain the region not shake for the region. we estimated the capability for the stabilization algorithm using the RMS between the trajectory and regression curve.

  • PDF

CALS: Channel State Information Auto-Labeling System for Large-scale Deep Learning-based Wi-Fi Sensing (딥러닝 기반 Wi-Fi 센싱 시스템의 효율적인 구축을 위한 지능형 데이터 수집 기법)

  • Jang, Jung-Ik;Choi, Jaehyuk
    • Journal of IKEEE
    • /
    • v.26 no.3
    • /
    • pp.341-348
    • /
    • 2022
  • Wi-Fi Sensing, which uses Wi-Fi technology to sense the surrounding environments, has strong potentials in a variety of sensing applications. Recently several advanced deep learning-based solutions using CSI (Channel State Information) data have achieved high performance, but it is still difficult to use in practice without explicit data collection, which requires expensive adaptation efforts for model retraining. In this study, we propose a Channel State Information Automatic Labeling System (CALS) that automatically collects and labels training CSI data for deep learning-based Wi-Fi sensing systems. The proposed system allows the CSI data collection process to efficiently collect labeled CSI for labeling for supervised learning using computer vision technologies such as object detection algorithms. We built a prototype of CALS to demonstrate its efficiency and collected data to train deep learning models for detecting the presence of a person in an indoor environment, showing to achieve an accuracy of over 90% with the auto-labeled data sets generated by CALS.

Refinement of damage identification capability of neural network techniques in application to a suspension bridge

  • Wang, J.Y.;Ni, Y.Q.
    • Structural Monitoring and Maintenance
    • /
    • v.2 no.1
    • /
    • pp.77-93
    • /
    • 2015
  • The idea of using measured dynamic characteristics for damage detection is attractive because it allows for a global evaluation of the structural health and condition. However, vibration-based damage detection for complex structures such as long-span cable-supported bridges still remains a challenge. As a suspension or cable-stayed bridge involves in general thousands of structural components, the conventional damage detection methods based on model updating and/or parameter identification might result in ill-conditioning and non-uniqueness in the solution of inverse problems. Alternatively, methods that utilize, to the utmost extent, information from forward problems and avoid direct solution to inverse problems would be more suitable for vibration-based damage detection of long-span cable-supported bridges. The auto-associative neural network (ANN) technique and the probabilistic neural network (PNN) technique, that both eschew inverse problems, have been proposed for identifying and locating damage in suspension and cable-stayed bridges. Without the help of a structural model, ANNs with appropriate configuration can be trained using only the measured modal frequencies from healthy structure under varying environmental conditions, and a new set of modal frequency data acquired from an unknown state of the structure is then fed into the trained ANNs for damage presence identification. With the help of a structural model, PNNs can be configured using the relative changes of modal frequencies before and after damage by assuming damage at different locations, and then the measured modal frequencies from the structure can be presented to locate the damage. However, such formulated ANNs and PNNs may still be incompetent to identify damage occurring at the deck members of a cable-supported bridge because of very low modal sensitivity to the damage. The present study endeavors to enhance the damage identification capability of ANNs and PNNs when being applied for identification of damage incurred at deck members. Effort is first made to construct combined modal parameters which are synthesized from measured modal frequencies and modal shape components to train ANNs for damage alarming. With the purpose of improving identification accuracy, effort is then made to configure PNNs for damage localization by adapting the smoothing parameter in the Bayesian classifier to different values for different pattern classes. The performance of the ANNs with their input being modal frequencies and the combined modal parameters respectively and the PNNs with constant and adaptive smoothing parameters respectively is evaluated through simulation studies of identifying damage inflicted on different deck members of the double-deck suspension Tsing Ma Bridge.

Comparative analysis of Machine-Learning Based Models for Metal Surface Defect Detection (머신러닝 기반 금속외관 결함 검출 비교 분석)

  • Lee, Se-Hun;Kang, Seong-Hwan;Shin, Yo-Seob;Choi, Oh-Kyu;Kim, Sijong;Kang, Jae-Mo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.6
    • /
    • pp.834-841
    • /
    • 2022
  • Recently, applying artificial intelligence technologies in various fields of production has drawn an upsurge of research interest due to the increase for smart factory and artificial intelligence technologies. A great deal of effort is being made to introduce artificial intelligence algorithms into the defect detection task. Particularly, detection of defects on the surface of metal has a higher level of research interest compared to other materials (wood, plastics, fibers, etc.). In this paper, we compare and analyze the speed and performance of defect classification by combining machine learning techniques (Support Vector Machine, Softmax Regression, Decision Tree) with dimensionality reduction algorithms (Principal Component Analysis, AutoEncoders) and two convolutional neural networks (proposed method, ResNet). To validate and compare the performance and speed of the algorithms, we have adopted two datasets ((i) public dataset, (ii) actual dataset), and on the basis of the results, the most efficient algorithm is determined.

Stacked Autoencoder Based Malware Feature Refinement Technology Research (Stacked Autoencoder 기반 악성코드 Feature 정제 기술 연구)

  • Kim, Hong-bi;Lee, Tae-jin
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.30 no.4
    • /
    • pp.593-603
    • /
    • 2020
  • The advent of malicious code has increased exponentially due to the spread of malicious code generation tools in accordance with the development of the network, but there is a limit to the response through existing malicious code detection methods. According to this situation, a machine learning-based malicious code detection method is evolving, and in this paper, the feature of data is extracted from the PE header for machine-learning-based malicious code detection, and then it is used to automate the malware through autoencoder. Research on how to extract the indicated features and feature importance. In this paper, 549 features composed of information such as DLL/API that can be identified from PE files that are commonly used in malware analysis are extracted, and autoencoder is used through the extracted features to improve the performance of malware detection in machine learning. It was proved to be successful in providing excellent accuracy and reducing the processing time by 2 times by effectively extracting the features of the data by compressively storing the data. The test results have been shown to be useful for classifying malware groups, and in the future, a classifier such as SVM will be introduced to continue research for more accurate malware detection.

A Real-time Face Recognition System using Fast Face Detection (빠른 얼굴 검출을 이용한 실시간 얼굴 인식 시스템)

  • Lee Ho-Geun;Jung Sung-Tae
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.12
    • /
    • pp.1247-1259
    • /
    • 2005
  • This paper proposes a real-time face recognition system which detects multiple faces from low resolution video such as web-camera video. Face recognition system consists of the face detection step and the face classification step. At First, it finds face region candidates by using AdaBoost based object detection method which have fast speed and robust performance. It generates reduced feature vector for each face region candidate by using principle component analysis. At Second, Face classification used Principle Component Analysis and multi-SVM. Experimental result shows that the proposed method achieves real-time face detection and face recognition from low resolution video. Additionally, We implement the auto-tracking face recognition system using the Pan-Tilt Web-camera and radio On/Off digital door-lock system with face recognition system.

Abnormal signal detection based on parallel autoencoders (병렬 오토인코더 기반의 비정상 신호 탐지)

  • Lee, Kibae;Lee, Chong Hyun
    • The Journal of the Acoustical Society of Korea
    • /
    • v.40 no.4
    • /
    • pp.337-346
    • /
    • 2021
  • Detection of abnormal signal generally can be done by using features of normal signals as main information because of data imbalance. This paper propose an efficient method for abnormal signal detection using parallel AutoEncoder (AE) which can use features of abnormal signals as well. The proposed Parallel AE (PAE) is composed of a normal and an abnormal reconstructors having identical AE structure and train features of normal and abnormal signals, respectively. The PAE can effectively solve the imbalanced data problem by sequentially training normal and abnormal data. For further detection performance improvement, additional binary classifier can be added to the PAE. Through experiments using public acoustic data, we obtain that the proposed PAE shows Area Under Curve (AUC) improvement of minimum 22 % at the expenses of training time increased by 1.31 ~ 1.61 times to the single AE. Furthermore, the PAE shows 93 % AUC improvement in detecting abnormal underwater acoustic signal when pre-trained PAE is transferred to train open underwater acoustic data.

Anomaly Detection using Geometric Transformation of Normal Sample Images (정상 샘플 이미지의 기하학적 변환을 사용한 이상 징후 검출)

  • Kwon, Yong-Wan;Kang, Dong-Joong
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.4
    • /
    • pp.157-163
    • /
    • 2022
  • Recently, with the development of automation in the industrial field, research on anomaly detection is being actively conducted. An application for anomaly detection used in factory automation is camera-based defect inspection. Vision camera inspection shows high performance and efficiency in factory automation, but it is difficult to overcome the instability of lighting and environmental conditions. Although camera inspection using deep learning can solve the problem of vision camera inspection with much higher performance, it is difficult to apply to actual industrial fields because it requires a huge amount of normal and abnormal data for learning. Therefore, in this study, we propose a network that overcomes the problem of collecting abnormal data with 72 geometric transformation deep learning methods using only normal data and adds an outlier exposure method for performance improvement. By applying and verifying this to the MVTec data set, which is a database for auto-mobile parts data and outlier detection, it is shown that it can be applied in actual industrial sites.