• 제목/요약/키워드: Sensor Fusion System

검색결과 433건 처리시간 0.026초

다중 수동 소나 센서 기반 에너지 인식 분산탐지 체계의 설계 및 성능 분석 (Design and Performance Analysis of Energy-Aware Distributed Detection Systems with Multiple Passive Sonar Sensors)

  • 김송근;홍순목
    • 한국군사과학기술학회지
    • /
    • 제13권1호
    • /
    • pp.9-21
    • /
    • 2010
  • In this paper, optimum design of distributed detection is considered for a parallel sensor network system consisting of a fusion center and multiple passive sonar nodes. Nonrandom fusion rules are employed as the fusion rules of the sensor network. For the nonrandom fusion rules, it is shown that a threshold rule of each sensor node has uniformly most powerful properties. Optimum threshold for each sensor is investigated that maximizes the probability of detection under a constraint on energy consumption due to false alarms. It is also investigated through numerical experiments how signal strength, false alarm probability, and the distance between three sensor nodes affect the system detection performances.

ACC/AEBS 시스템용 센서퓨전을 통한 주행경로 추정 알고리즘 (Development of the Driving path Estimation Algorithm for Adaptive Cruise Control System and Advanced Emergency Braking System Using Multi-sensor Fusion)

  • 이동우;이경수;이재완
    • 자동차안전학회지
    • /
    • 제3권2호
    • /
    • pp.28-33
    • /
    • 2011
  • This paper presents driving path estimation algorithm for adaptive cruise control system and advanced emergency braking system using multi-sensor fusion. Through data collection, yaw rate filtering based road curvature and vision sensor road curvature characteristics are analyzed. Yaw rate filtering based road curvature and vision sensor road curvature are fused into the one curvature by weighting factor which are considering characteristics of each curvature data. The proposed driving path estimation algorithm has been investigated via simulation performed on a vehicle package Carsim and Matlab/Simulink. It has been shown via simulation that the proposed driving path estimation algorithm improves primary target detection rate.

3차원 물체의 인식 성능 향상을 위한 감각 융합 신경망 시스템 (Neural Network Approach to Sensor Fusion System for Improving the Recognition Performance of 3D Objects)

  • 동성수;이종호;김지경
    • 대한전기학회논문지:시스템및제어부문D
    • /
    • 제54권3호
    • /
    • pp.156-165
    • /
    • 2005
  • Human being recognizes the physical world by integrating a great variety of sensory inputs, the information acquired by their own action, and their knowledge of the world using hierarchically parallel-distributed mechanism. In this paper, authors propose the sensor fusion system that can recognize multiple 3D objects from 2D projection images and tactile informations. The proposed system focuses on improving recognition performance of 3D objects. Unlike the conventional object recognition system that uses image sensor alone, the proposed method uses tactual sensors in addition to visual sensor. Neural network is used to fuse the two sensory signals. Tactual signals are obtained from the reaction force of the pressure sensors at the fingertips when unknown objects are grasped by four-fingered robot hand. The experiment evaluates the recognition rate and the number of learning iterations of various objects. The merits of the proposed systems are not only the high performance of the learning ability but also the reliability of the system with tactual information for recognizing various objects even though the visual sensory signals get defects. The experimental results show that the proposed system can improve recognition rate and reduce teeming time. These results verify the effectiveness of the proposed sensor fusion system as recognition scheme for 3D objects.

Control of the Mobile Robot Navigation Using a New Time Sensor Fusion

  • Tack, Han-Ho;Kim, Chang-Geun;Kim, Myeong-Kyu
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • 제4권1호
    • /
    • pp.23-28
    • /
    • 2004
  • This paper proposes a sensor-fusion technique where the data sets for the previous moments are properly transformed and fused into the current data sets to enable accurate measurement, such as, distance to an obstacle and location of the service robot itself. In the conventional fusion schemes, the measurement is dependent on the current data sets. As the results, more of sensors are required to measure a certain physical parameter or to improve the accuracy of the measurement. However, in this approach, instead of adding more sensors to the system, the temporal sequence of the data sets are stored and utilized for the measurement improvement. Theoretical basis is illustrated by examples and the effectiveness is proved through the simulations. Finally, the new space and time sensor fusion(STSF) scheme is applied to the control of a mobile robot in an unstructured environment as well as structured environment.

AVM 카메라와 융합을 위한 다중 상용 레이더 데이터 획득 플랫폼 개발 (Development of Data Logging Platform of Multiple Commercial Radars for Sensor Fusion With AVM Cameras)

  • 진영석;전형철;신영남;현유진
    • 대한임베디드공학회논문지
    • /
    • 제13권4호
    • /
    • pp.169-178
    • /
    • 2018
  • Currently, various sensors have been used for advanced driver assistance systems. In order to overcome the limitations of individual sensors, sensor fusion has recently attracted the attention in the field of intelligence vehicles. Thus, vision and radar based sensor fusion has become a popular concept. The typical method of sensor fusion involves vision sensor that recognizes targets based on ROIs (Regions Of Interest) generated by radar sensors. Especially, because AVM (Around View Monitor) cameras due to their wide-angle lenses have limitations of detection performance over near distance and around the edges of the angle of view, for high performance of sensor fusion using AVM cameras and radar sensors the exact ROI extraction of the radar sensor is very important. In order to resolve this problem, we proposed a sensor fusion scheme based on commercial radar modules of the vendor Delphi. First, we configured multiple radar data logging systems together with AVM cameras. We also designed radar post-processing algorithms to extract the exact ROIs. Finally, using the developed hardware and software platforms, we verified the post-data processing algorithm under indoor and outdoor environments.

센서융합 검증을 위한 실시간 모니터링 및 검증 도구 개발 (Development of a Monitoring and Verification Tool for Sensor Fusion)

  • 김현우;신승환;배상진
    • 한국자동차공학회논문집
    • /
    • 제22권3호
    • /
    • pp.123-129
    • /
    • 2014
  • SCC (Smart Cruise Control) and AEBS (Autonomous Emergency Braking System) are using various types of sensors data, so it is important to consider about sensor data reliability. In this paper, data from radar and vision sensor is fused by applying a Bayesian sensor fusion technique to improve the reliability of sensors data. Then, it presents a sensor fusion verification tool developed to monitor acquired sensors data and to verify sensor fusion results, efficiently. A parallel computing method was applied to reduce verification time and a series of simulation results of this method are discussed in detail.

레이더/카메라 센서융합을 이용한 전방차량 충돌경보 시스템 (Forward Collision Warning System based on Radar driven Fusion with Camera)

  • 문승욱;문일기;신광근
    • 자동차안전학회지
    • /
    • 제5권1호
    • /
    • pp.5-10
    • /
    • 2013
  • This paper describes a Forward Collision Warning (FCW) system based on the radar driven fusion with camera. The objective of FCW system is to provide an appropriate alert with satisfying the evaluation scenarios of US-NCAP and a driver acceptance. For this purpose, this paper proposed a data fusion algorithm and a collision warning algorithm. The data fusion algorithm generates information of fusion target depending on the confidence of camera sensor. The collision warning algorithm calculates indexes and determines an appropriate alert-timing by using analysis results of manual driving data. The FCW system with the proposed data fusion and collision warning algorithm was investigated via scenarios of US-NCAP and a real-road driving. It is shown that the proposed FCW system can improve the accuracy of an alarm-timing and reduce the false alarm in real roads.

청소로봇의 최적비용함수를 고려한 지도 작성에 관한 연구 (A Study on the Map-Building of a Cleaning Robot Base upon the Optimal Cost Function)

  • 강진구
    • 디지털산업정보학회논문지
    • /
    • 제5권3호
    • /
    • pp.39-45
    • /
    • 2009
  • In this paper we present a cleaning robot system for an autonomous mobile robot. Our robot performs goal reaching tasks into unknown indoor environments by using sensor fusion. The robot's operation objective is to clean floor or any other applicable surface and to build a map of the surrounding environment for some further purpose such as finding the shortest path available. Using its cleaning robot system for an autonomous mobile robot can move in various modes and perform dexterous tasks. Performance of the cleaning robot system is better than a fixed base redundant robot in avoiding singularity and obstacle. Sensor fusion using the clean robot improves the performance of the robot with redundant freedom in workspace and Map-Building. In this paper, Map-building of the cleaning robot has been studied using sensor fusion. A sequence of this alternating task execution scheme enables the clean robot to execute various tasks efficiently. The proposed algorithm is experimentally verified and discussed with a cleaning robot, KCCR.

CCD카메라와 적외선 카메라의 융합을 통한 효과적인 객체 추적 시스템 (Efficient Object Tracking System Using the Fusion of a CCD Camera and an Infrared Camera)

  • 김승훈;정일균;박창우;황정훈
    • 제어로봇시스템학회논문지
    • /
    • 제17권3호
    • /
    • pp.229-235
    • /
    • 2011
  • To make a robust object tracking and identifying system for an intelligent robot and/or home system, heterogeneous sensor fusion between visible ray system and infrared ray system is proposed. The proposed system separates the object by combining the ROI (Region of Interest) estimated from two different images based on a heterogeneous sensor that consolidates the ordinary CCD camera and the IR (Infrared) camera. Human's body and face are detected in both images by using different algorithms, such as histogram, optical-flow, skin-color model and Haar model. Also the pose of human body is estimated from the result of body detection in IR image by using PCA algorithm along with AdaBoost algorithm. Then, the results from each detection algorithm are fused to extract the best detection result. To verify the heterogeneous sensor fusion system, few experiments were done in various environments. From the experimental results, the system seems to have good tracking and identification performance regardless of the environmental changes. The application area of the proposed system is not limited to robot or home system but the surveillance system and military system.

Multi-Sensor Data Fusion Model that Uses a B-Spline Fuzzy Inference System

  • Lee, K.S.;S.W. Shin;D.S. Ahn
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2001년도 ICCAS
    • /
    • pp.23.3-23
    • /
    • 2001
  • The main object of this work is the development of an intelligent multi-sensor integration and fusion model that uses fuzzy inference system. Sensor data from different types of sensors are integrated and fused together based on the confidence which is not typically used in traditional data fusion methods. The information is fed as input to a fuzzy inference system(FIS). The output of the FIS is weights that are assigned to the different sensor data reflecting the confidence En the sensor´s behavior and performance. We interpret a type of fuzzy inference system as an interpolator of B-spline hypersurfaces. B-spline basis functions of different orders are regarded as a class of membership functions. This paper presents a model that ...

  • PDF