• Title/Summary/Keyword: Sensor fusion

Search Result 820, Processing Time 0.031 seconds

A Cyber-Physical Information System for Smart Buildings with Collaborative Information Fusion

  • Liu, Qing;Li, Lanlan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.5
    • /
    • pp.1516-1539
    • /
    • 2022
  • This article shows a set of physical information fusion IoT systems that we designed for smart buildings. Its essence is a computer system that combines physical quantities in buildings with quantitative analysis and control. In the part of the Internet of Things, its mechanism is controlled by a monitoring system based on sensor networks and computer-based algorithms. Based on the design idea of the agent, we have realized human-machine interaction (HMI) and machine-machine interaction (MMI). Among them, HMI is realized through human-machine interaction, while MMI is realized through embedded computing, sensors, controllers, and execution. Device and wireless communication network. This article mainly focuses on the function of wireless sensor networks and MMI in environmental monitoring. This function plays a fundamental role in building security, environmental control, HVAC, and other smart building control systems. The article not only discusses various network applications and their implementation based on agent design but also demonstrates our collaborative information fusion strategy. This strategy can provide a stable incentive method for the system through collaborative information fusion when the sensor system is unstable in the physical measurements, thereby preventing system jitter and unstable response caused by uncertain disturbances and environmental factors. This article also gives the results of the system test. The results show that through the CPS interaction of HMI and MMI, the intelligent building IoT system can achieve comprehensive monitoring, thereby providing support and expansion for advanced automation management.

Mobile robot absolute position tracking system (이동 로봇 절대위치 추적 제어 시스템)

  • Choi, Hyun-Seung;Hyun, Woong-Keun
    • Proceedings of the KIEE Conference
    • /
    • 2008.07a
    • /
    • pp.1811-1812
    • /
    • 2008
  • This paper represents a absolute position tracking system with sensor fusion and PD-gain. this paper presents an accurate localization method by relative and absolute sensor fusion and PD control for position tracking of mobile robot. we developed a sensor based absolution position tracking and smooth moving algorithm using this algorithm.

  • PDF

MULTI SENSOR DATA FUSION FOR IMPROVING PERFORMANCE AND RELIABILITY OF FULLY AUTOMATED MULTIPASS WELDING

  • Beattie, R.J.
    • Proceedings of the KWS Conference
    • /
    • 2002.10a
    • /
    • pp.336-341
    • /
    • 2002
  • Recent developments in sensor hardware and in advanced software have made it feasible to consider automating some of the most difficult welding operations. This paper describes some techniques used to automate successfully multipass submerged arc welding operations typically used in pressure vessel manufacture, shipbuilding, production of offshore structures and in pipe mills.

  • PDF

Development of an Obstacle Avoidance Algorithm for a Network-based Autonomous Mobile Robot (네트워크 기반 자율이동로봇을 위한 장애물 회피 알고리즘 개발)

  • Kim Hongryeol;Kim Dae Won;Kim Hong-Seok;Sohn SooKyung
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.54 no.5
    • /
    • pp.291-299
    • /
    • 2005
  • An obstacle avoidance algorithm for a network-based autonomous mobile robot is proposed in this paper. The obstacle avoidance algorithm is based on the VFH(Vector Field Histogram) algorithm and two delay compensation methods with the VFH algorithm are proposed for a network-based robot with distributed environmental sensors, mobile actuators, and the VFH controller. Firstly, the environmental sensor information is compensated by prospection with acquired environmental sensor information, measured network delays, and the kinematic model of the robot. The compensated environmental sensor information is used for building polar histogram with the VFH algorithm. Secondly, a sensor fusion algorithm for localization of the robot is proposed to compensate the delay of odometry sensor information and the delay of environmental sensor information. Through some simulation tests, the performance enhancement of the proposed algorithm in the viewpoint of efficient path generation and accurate goal positioning is shown here.

SWNT-UHF Fusion Sensor for GIS Partial Discharge Detection (가스절연기기의 부분방전검출을 위한 SWNT-UHF 융합센서)

  • Lee, Sang-Wook;Chang, Yong-Moo;Baik, Seung-Hyun;Lee, Jong-Chul
    • Proceedings of the Korean Institute of Electrical and Electronic Material Engineers Conference
    • /
    • 2010.06a
    • /
    • pp.120-120
    • /
    • 2010
  • To detect the PD events, we have studied a fusion sensor, the UHF sensor and the single-walled carbon nanotube(SWNT) gas sensor. We are accustomed to the UHF sensor which have employed to detect the partial discharges in apparatus GIS-like. But the SWNT gas sense is a newly way proposed to detect the partial discharges. In this study, we monitored not only the changes of the electrical conductance of the SWNT sensors in responding to the PD events but also the signal of the UHF sensor at the same time with IEC 60270 standard method for reference on the partial discharge events.

  • PDF

Information-Theoretic Approaches for Sensor Selection and Placement in Sensor Networks for Target Localization and Tracking

  • Wang Hanbiao;Yao Kung;Estrin Deborah
    • Journal of Communications and Networks
    • /
    • v.7 no.4
    • /
    • pp.438-449
    • /
    • 2005
  • In this paper, we describes the information-theoretic approaches to sensor selection and sensor placement in sensor net­works for target localization and tracking. We have developed a sensor selection heuristic to activate the most informative candidate sensor for collaborative target localization and tracking. The fusion of the observation by the selected sensor with the prior target location distribution yields nearly the greatest reduction of the entropy of the expected posterior target location distribution. Our sensor selection heuristic is computationally less complex and thus more suitable to sensor networks with moderate computing power than the mutual information sensor selection criteria. We have also developed a method to compute the posterior target location distribution with the minimum entropy that could be achieved by the fusion of observations of the sensor network with a given deployment geometry. We have found that the covariance matrix of the posterior target location distribution with the minimum entropy is consistent with the Cramer-Rao lower bound (CRB) of the target location estimate. Using the minimum entropy of the posterior target location distribution, we have characterized the effect of the sensor placement geometry on the localization accuracy.

Study on the Localization Improvement of the Dead Reckoning using the INS Calibrated by the Fusion Sensor Network Information (융합 센서 네트워크 정보로 보정된 관성항법센서를 이용한 추측항법의 위치추정 향상에 관한 연구)

  • Choi, Jae-Young;Kim, Sung-Gaun
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.18 no.8
    • /
    • pp.744-749
    • /
    • 2012
  • In this paper, we suggest that how to improve an accuracy of mobile robot's localization by using the sensor network information which fuses the machine vision camera, encoder and IMU sensor. The heading value of IMU sensor is measured using terrestrial magnetism sensor which is based on magnetic field. However, this sensor is constantly affected by its surrounding environment. So, we isolated template of ceiling using vision camera to increase the sensor's accuracy when we use IMU sensor; we measured the angles by pattern matching algorithm; and to calibrate IMU sensor, we compared the obtained values with IMU sensor values and the offset value. The values that were used to obtain information on the robot's position which were of Encoder, IMU sensor, angle sensor of vision camera are transferred to the Host PC by wireless network. Then, the Host PC estimates the location of robot using all these values. As a result, we were able to get more accurate information on estimated positions than when using IMU sensor calibration solely.

Design and Performance Analysis of Energy-Aware Distributed Detection Systems with Two Passive Sonar Sensors (수동 소나 쌍을 이용한 에너지 인식 분산탐지 체계의 설계 및 성능 분석)

  • Do, Joo-Hwan;Kim, Song-Geun;Hong, Sun-Mog
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.4
    • /
    • pp.139-147
    • /
    • 2009
  • In this paper, optimum design of energy-aware distributed detection is considered for a parallel sensor network system consisting of a fusion center and two passive sonar nodes. AND rule and OR rule are employed as the fusion rules of the sensor network. For the fusion rules, it is shown that a threshold rule of each sensor node has uniformly most powerful properties. Optimum threshold for each sensor is investigated that maximizes the probability of detection under a constraint on energy consumption due to false alarms. It is also investigated through numerical experiments how signal strength, an energy constraint, and the distance between two sensor nodes affect the system detection performances.

Intelligent Hexapod Mobile Robot using Image Processing and Sensor Fusion (영상처리와 센서융합을 활용한 지능형 6족 이동 로봇)

  • Lee, Sang-Mu;Kim, Sang-Hoon
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.15 no.4
    • /
    • pp.365-371
    • /
    • 2009
  • A intelligent mobile hexapod robot with various types of sensors and wireless camera is introduced. We show this mobile robot can detect objects well by combining the results of active sensors and image processing algorithm. First, to detect objects, active sensors such as infrared rays sensors and supersonic waves sensors are employed together and calculates the distance in real time between the object and the robot using sensor's output. The difference between the measured value and calculated value is less than 5%. This paper suggests effective visual detecting system for moving objects with specified color and motion information. The proposed method includes the object extraction and definition process which uses color transformation and AWUPC computation to decide the existence of moving object. We add weighing values to each results from sensors and the camera. Final results are combined to only one value which represents the probability of an object in the limited distance. Sensor fusion technique improves the detection rate at least 7% higher than the technique using individual sensor.