• Title/Summary/Keyword: Smart sensors

Search Result 1,513, Processing Time 0.03 seconds

A Secure and Lightweight Authentication Scheme for Ambient Assisted Living Systems (전천 후 생활보조 시스템을 위한 안전하고 경량화 된 인증기법)

  • Yi, Myung-Kyu;Choi, Hyunchul;Whangbo, Taeg-Keun
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.4
    • /
    • pp.77-83
    • /
    • 2019
  • With the increase in population, the number of such senior citizens is increasing day by day. These senior citizens have a variety of care needs, but there are not enough health workers to look after them. Ambient Assisted Living (AAL) aims at ensuring the safety and health quality of the older adults and extending the number of years the senior citizens can live independently in an environment of their own preference. AAL provides a system comprising of smart devices, medical sensors, wireless networks, computer and software applications for healthcare monitoring. AAL can be used for various purposes like preventing, curing, and improving wellness and health conditions of older adults. While information security and privacy are critical to providing assurance that users of AAL systems are protected, few studies take into account this feature. In this paper, we propose a secure and lightweight authentication scheme for the AAL systems. The proposed authentication scheme not only supports several important security requirements needed by the AAL systems, but can also withstand various types of attacks. Also, the security analysis results are presented to show the proposed authentication scheme is more secure and efficient rather than existing authentication schemes.

Modbus TCP based Solar Power Plant Monitoring System using Raspberry Pi (라즈베리파이를 이용한 Modbus TCP 기반 태양광 발전소 모니터링 시스템)

  • Park, Jin-Hwan;Kim, Chang-Bok
    • Journal of Advanced Navigation Technology
    • /
    • v.24 no.6
    • /
    • pp.620-626
    • /
    • 2020
  • This research propose and simulate a solar power generation system monitoring system based on Modbus TCP communication using RaspberryPi, an IOT equipment, as a master and an inverter as a slave. In this model, various sensors are added to the RaspberryPi to add necessary information for monitoring solar power plants, and power generation prediction and monitoring information are transmitted to the smart phone through real-time power generation prediction. In addition, information that is continuously generated by the solar power plant is built on the server as big data, and a deep learning model for predicting power generation is trained and updated. As a result of the study, stable communication was possible based on Modbus TCP with the Raspberry Pi in the inverter, and real-time prediction was possible with the deep learning model learned in the Raspberry Pi. The server was able to train various deep learning models with big data, and it was confirmed that LSTM showed the best error with a learning error of 0.0069, a test error of 0.0075, and an RMSE of 0.0866. This model suggested that it is possible to implement a real-time monitoring system that is simpler, more convenient, and can predict the amount of power generation for inverters of various manufacturers.

A study on the honeycomb entry and exit counting system for measuring the amount of movement of honeybees inside the beehive (벌통 내부 꿀벌 이동량 측정을 위한 벌집 입·출입 계수 시스템 연구)

  • Kim, Joon Ho;Seo, Hee;Han, Wook;Chung, Wonki
    • The Journal of the Convergence on Culture Technology
    • /
    • v.7 no.4
    • /
    • pp.857-862
    • /
    • 2021
  • Recently, rapid climate change has had a significant impact on the bee ecosystem. The decrease in the number of bees and the change in the flowering period have a huge impact on the harvesting of beekeepers. Accordingly, attention is focused on smart beekeeping, which introduces IoT technology to beekeeping. According to the characteristics of beekeeping, it is impossible to continuously observe the beehive in the hive with the naked eye, and the condition of the hive is mostly dependent on knowledge from experience. Although a system that can measure partly through sensors such as temperature/humidity change inside the hive and measurement of the amount of CO2 is applied, there is no research on measuring the movement path and amount of movement of bees inside the beehive. Part of the migration of honeybees inside the hive can provide basic information to predict the most important cleavage time in beekeeping. In this study, we propose a device that detects the movement path of bees and measures and records data entering and exiting the hive in real time. The device proposed in this study was developed according to the honeycomb standard of the existing beehive so that beekeeping farms could use it. The development method used a photodetector that can detect the movement of bees to configure 16 movement paths and to detect the movement of bees in real time. If the measured honeybee movement status is utilized, the problem of directly observing the colony with the naked eye in order not to miss the swarming time can be solved.

Important Facility Guard System Using Edge Computing for LiDAR (LiDAR용 엣지 컴퓨팅을 활용한 중요시설 경계 시스템)

  • Jo, Eun-Kyung;Lee, Eun-Seok;Shin, Byeong-Seok
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.11 no.10
    • /
    • pp.345-352
    • /
    • 2022
  • Recent LiDAR(Light Detection And Ranging) sensor is used for scanning object around in real-time. This sensor can detect movement of the object and how it has changed. As the production cost of the sensors has been decreased, LiDAR begins to be used for various industries such as facility guard, smart city and self-driving car. However, LiDAR has a large input data size due to its real-time scanning process. So another way for processing a large amount of data are needed in LiDAR system because it can cause a bottleneck. This paper proposes edge computing to compress massive point cloud for processing quickly. Since laser's reflection range of LiDAR sensor is limited, multiple LiDAR should be used to scan a large area. In this reason multiple LiDAR sensor's data should be processed at once to detect or recognize object in real-time. Edge computer compress point cloud efficiently to accelerate data processing and decompress every data in the main cloud in real-time. In this way user can control LiDAR sensor in the main system without any bottleneck. The system we suggest solves the bottleneck which was problem on the cloud based method by applying edge computing service.

Composite-Based Material and Process Technology Review for Improving Performance of Piezoelectric Energy Harvester (압전 에너지 수확기의 성능 향상을 위한 복합재료 기반 소재 및 공정 기술 검토)

  • Kim, Geon Su;Jang, Ji-un;Kim, Seong Yun
    • Composites Research
    • /
    • v.34 no.6
    • /
    • pp.357-372
    • /
    • 2021
  • The energy harvesting device is known to be promising as an alternative to solve the resource shortage caused by the depletion of petroleum resources. In order to overcome the limitations (environmental pollution and low mechanical properties) of piezoelectric elements capable of converting mechanical motion into electrical energy, many studies have been conducted on a polymer matrix-based composite piezoelectric energy harvesting device. In this paper, the output performance and related applications of the reported piezoelectric composites are reviewed based on the applied materials and processes. As for the piezoelectric fillers, zinc oxide, which is advantageous in terms of eco-friendliness, biocompatibility, and flexibility, as well as ceramic fillers based on lead zirconate titanate and barium titanate, were reviewed. The polymer matrix was classified into piezoelectric polymers composed of polyvinylidene fluoride and copolymers, and flexible polymers based on epoxy and polydimethylsiloxane, to discuss piezoelectric synergy of composite materials and improvement of piezoelectric output by high external force application, respectively. In addition, the effect of improving the conductivity or the mechanical properties of composite material by the application of a metal or carbon-based secondary filler on the output performance of the piezoelectric harvesting device was explained in terms of the structure of the composite material. Composite material-based piezoelectric harvesting devices, which can be applied to small electronic devices, smart sensors, and medicine with improved performance, can provide potential insights as a power source for wireless electronic devices expected to be encountered in future daily life.

A Study of the Correlation Between Nighttime Light and Individual Land Price by Province in South Korea, Using DMSP OLS Data (야간광과 남한의 시도별 개별 공시지가 총액의 상관관계 연구 - DMSP OLS 자료를 중심으로)

  • Bong Chan Kim ;Seulki Lee ;Chang-Wook Lee
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.5_1
    • /
    • pp.729-741
    • /
    • 2023
  • The Operational Linescan System (OLS)sensor is a sensor aboard satellites launched through the Defense Meteorological Satellite Program (DMSP) that detects light in the visible and infrared bands emitted at night. Studies by several researchers have shown a high correlation between nighttime light data from OLS sensors and gross domestic product values. In this study, we investigated the correlation of nighttime light data with the total amount of individual land prices, which is one of the various indicators related to economic development. The study found that most cities and provinces showed a high correlation with a correlation coefficient of more than 0.7, and the correlation coefficient of 0.7837 between the total amount of individual land price and nighttime light data for the entire South Korea was also high. However, unlike other cities and provinces, Seoul has a low correlation coefficient of 0.5648 between nighttime light and the total amount of individual land price, which is analyzed as a reason that the digital number value of the OLS sensor is close to the maximum value and cannot show further brightness changes. This study is expected to help identify announced land prices in areas where announced land prices are not systematically organized and to analyze land use changes in such areas.

Wearable Computers

  • Cho, Gil-Soo;Barfield, Woodrow;Baird, Kevin
    • Fiber Technology and Industry
    • /
    • v.2 no.4
    • /
    • pp.490-508
    • /
    • 1998
  • One of the latest fields of research in the area of output devices is tactual display devices [13,31]. These tactual or haptic devices allow the user to receive haptic feedback output from a variety of sources. This allows the user to actually feel virtual objects and manipulate them by touch. This is an emerging technology and will be instrumental in enhancing the realism of wearable augmented environments for certain applications. Tactual displays have previously been used for scientific visualization in virtual environments by chemists and engineers to improve perception and understanding of force fields and of world models populated with the impenetrable. In addition to tactual displays, the use of wearable audio displays that allow sound to be spatialized are being developed. With wearable computers, designers will soon be able to pair spatialized sound to virtual representations of objects when appropriate to make the wearable computer experience even more realistic to the user. Furthermore, as the number and complexity of wearable computing applications continues to grow, there will be increasing needs for systems that are faster, lighter, and have higher resolution displays. Better networking technology will also need to be developed to allow all users of wearable computers to have high bandwidth connections for real time information gathering and collaboration. In addition to the technology advances that make users need to wear computers in everyday life, there is also the desire to have users want to wear their computers. In order to do this, wearable computing needs to be unobtrusive and socially acceptable. By making wearables smaller and lighter, or actually embedding them in clothing, users can conceal them easily and wear them comfortably. The military is currently working on the development of the Personal Information Carrier (PIC) or digital dog tag. The PIC is a small electronic storage device containing medical information about the wearer. While old military dog tags contained only 5 lines of information, the digital tags may contain volumes of multi-media information including medical history, X-rays, and cardiograms. Using hand held devices in the field, medics would be able to call this information up in real time for better treatment. A fully functional transmittable device is still years off, but this technology once developed in the military, could be adapted tp civilian users and provide ant information, medical or otherwise, in a portable, not obstructive, and fashionable way. Another future device that could increase safety and well being of its users is the nose on-a-chip developed by the Oak Ridge National Lab in Tennessee. This tiny digital silicon chip about the size of a dime, is capable of 'smelling' natural gas leaks in stoves, heaters, and other appliances. It can also detect dangerous levels of carbon monoxide. This device can also be configured to notify the fire department when a leak is detected. This nose chip should be commercially available within 2 years, and is inexpensive, requires low power, and is very sensitive. Along with gas detection capabilities, this device may someday also be configured to detect smoke and other harmful gases. By embedding this chip into workers uniforms, name tags, etc., this could be a lifesaving computational accessory. In addition to the future safety technology soon to be available as accessories are devices that are for entertainment and security. The LCI computer group is developing a Smartpen, that electronically verifies a user's signature. With the increase in credit card use and the rise in forgeries, is the need for commercial industries to constantly verify signatures. This Smartpen writes like a normal pen but uses sensors to detect the motion of the pen as the user signs their name to authenticate the signature. This computational accessory should be available in 1999, and would bring increased peace of mind to consumers and vendors alike. In the entertainment domain, Panasonic is creating the first portable hand-held DVD player. This device weight less than 3 pounds and has a screen about 6' across. The color LCD has the same 16:9 aspect ratio of a cinema screen and supports a high resolution of 280,000 pixels and stereo sound. The player can play standard DVD movies and has a hour battery life for mobile use. To summarize, in this paper we presented concepts related to the design and use of wearable computers with extensions to smart spaces. For some time, researchers in telerobotics have used computer graphics to enhance remote scenes. Recent advances in augmented reality displays make it possible to enhance the user's local environment with 'information'. As shown in this paper, there are many application areas for this technology such as medicine, manufacturing, training, and recreation. Wearable computers allow a much closer association of information with the user. By embedding sensors in the wearable to allow it to see what the user sees, hear what the user hears, sense the user's physical state, and analyze what the user is typing, an intelligent agent may be able to analyze what the user is doing and try to predict the resources he will need next or in the near future. Using this information, the agent may download files, reserve communications bandwidth, post reminders, or automatically send updates to colleagues to help facilitate the user's daily interactions. This intelligent wearable computer would be able to act as a personal assistant, who is always around, knows the user's personal preferences and tastes, and tries to streamline interactions with the rest of the world.

  • PDF

An Analysis of Big Video Data with Cloud Computing in Ubiquitous City (클라우드 컴퓨팅을 이용한 유시티 비디오 빅데이터 분석)

  • Lee, Hak Geon;Yun, Chang Ho;Park, Jong Won;Lee, Yong Woo
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.45-52
    • /
    • 2014
  • The Ubiquitous-City (U-City) is a smart or intelligent city to satisfy human beings' desire to enjoy IT services with any device, anytime, anywhere. It is a future city model based on Internet of everything or things (IoE or IoT). It includes a lot of video cameras which are networked together. The networked video cameras support a lot of U-City services as one of the main input data together with sensors. They generate huge amount of video information, real big data for the U-City all the time. It is usually required that the U-City manipulates the big data in real-time. And it is not easy at all. Also, many times, it is required that the accumulated video data are analyzed to detect an event or find a figure among them. It requires a lot of computational power and usually takes a lot of time. Currently we can find researches which try to reduce the processing time of the big video data. Cloud computing can be a good solution to address this matter. There are many cloud computing methodologies which can be used to address the matter. MapReduce is an interesting and attractive methodology for it. It has many advantages and is getting popularity in many areas. Video cameras evolve day by day so that the resolution improves sharply. It leads to the exponential growth of the produced data by the networked video cameras. We are coping with real big data when we have to deal with video image data which are produced by the good quality video cameras. A video surveillance system was not useful until we find the cloud computing. But it is now being widely spread in U-Cities since we find some useful methodologies. Video data are unstructured data thus it is not easy to find a good research result of analyzing the data with MapReduce. This paper presents an analyzing system for the video surveillance system, which is a cloud-computing based video data management system. It is easy to deploy, flexible and reliable. It consists of the video manager, the video monitors, the storage for the video images, the storage client and streaming IN component. The "video monitor" for the video images consists of "video translater" and "protocol manager". The "storage" contains MapReduce analyzer. All components were designed according to the functional requirement of video surveillance system. The "streaming IN" component receives the video data from the networked video cameras and delivers them to the "storage client". It also manages the bottleneck of the network to smooth the data stream. The "storage client" receives the video data from the "streaming IN" component and stores them to the storage. It also helps other components to access the storage. The "video monitor" component transfers the video data by smoothly streaming and manages the protocol. The "video translator" sub-component enables users to manage the resolution, the codec and the frame rate of the video image. The "protocol" sub-component manages the Real Time Streaming Protocol (RTSP) and Real Time Messaging Protocol (RTMP). We use Hadoop Distributed File System(HDFS) for the storage of cloud computing. Hadoop stores the data in HDFS and provides the platform that can process data with simple MapReduce programming model. We suggest our own methodology to analyze the video images using MapReduce in this paper. That is, the workflow of video analysis is presented and detailed explanation is given in this paper. The performance evaluation was experiment and we found that our proposed system worked well. The performance evaluation results are presented in this paper with analysis. With our cluster system, we used compressed $1920{\times}1080(FHD)$ resolution video data, H.264 codec and HDFS as video storage. We measured the processing time according to the number of frame per mapper. Tracing the optimal splitting size of input data and the processing time according to the number of node, we found the linearity of the system performance.

Analysis of Emerging Geo-technologies and Markets Focusing on Digital Twin and Environmental Monitoring in Response to Digital and Green New Deal (디지털 트윈, 환경 모니터링 등 디지털·그린 뉴딜 정책 관련 지질자원 유망기술·시장 분석)

  • Ahn, Eun-Young;Lee, Jaewook;Bae, Junhee;Kim, Jung-Min
    • Economic and Environmental Geology
    • /
    • v.53 no.5
    • /
    • pp.609-617
    • /
    • 2020
  • After introducing the industry 4.0 policy, Korean government announced 'Digital New Deal' and 'Green New Deal' as 'Korean New Deal' in 2020. We analyzed Korea Institute of Geoscience and Mineral Resources (KIGAM)'s research projects related to that policy and conducted markets analysis focused on Digital Twin and environmental monitoring technologies. Regarding 'Data Dam' policy, we suggested the digital geo-contents with Augmented Reality (AR) & Virtual Reality (VR) and the public geo-data collection & sharing system. It is necessary to expand and support the smart mining and digital oil fields research for '5th generation mobile communication (5G) and artificial intelligence (AI) convergence into all industries' policy. Korean government is suggesting downtown 3D maps for 'Digital Twin' policy. KIGAM can provide 3D geological maps and Internet of Things (IoT) systems for social overhead capital (SOC) management. 'Green New Deal' proposed developing technologies for green industries including resource circulation, Carbon Capture Utilization and Storage (CCUS), and electric & hydrogen vehicles. KIGAM has carried out related research projects and currently conducts research on domestic energy storage minerals. Oil and gas industries are presented as representative applications of digital twin. Many progress is made in mining automation and digital mapping and Digital Twin Earth (DTE) is a emerging research subject. The emerging research subjects are deeply related to data analysis, simulation, AI, and the IoT, therefore KIGAM should collaborate with sensors and computing software & system companies.

T-Cache: a Fast Cache Manager for Pipeline Time-Series Data (T-Cache: 시계열 배관 데이타를 위한 고성능 캐시 관리자)

  • Shin, Je-Yong;Lee, Jin-Soo;Kim, Won-Sik;Kim, Seon-Hyo;Yoon, Min-A;Han, Wook-Shin;Jung, Soon-Ki;Park, Se-Young
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.13 no.5
    • /
    • pp.293-299
    • /
    • 2007
  • Intelligent pipeline inspection gauges (PIGs) are inspection vehicles that move along within a (gas or oil) pipeline and acquire signals (also called sensor data) from their surrounding rings of sensors. By analyzing the signals captured in intelligent PIGs, we can detect pipeline defects, such as holes and curvatures and other potential causes of gas explosions. There are two major data access patterns apparent when an analyzer accesses the pipeline signal data. The first is a sequential pattern where an analyst reads the sensor data one time only in a sequential fashion. The second is the repetitive pattern where an analyzer repeatedly reads the signal data within a fixed range; this is the dominant pattern in analyzing the signal data. The existing PIG software reads signal data directly from the server at every user#s request, requiring network transfer and disk access cost. It works well only for the sequential pattern, but not for the more dominant repetitive pattern. This problem becomes very serious in a client/server environment where several analysts analyze the signal data concurrently. To tackle this problem, we devise a fast in-memory cache manager, called T-Cache, by considering pipeline sensor data as multiple time-series data and by efficiently caching the time-series data at T-Cache. To the best of the authors# knowledge, this is the first research on caching pipeline signals on the client-side. We propose a new concept of the signal cache line as a caching unit, which is a set of time-series signal data for a fixed distance. We also provide the various data structures including smart cursors and algorithms used in T-Cache. Experimental results show that T-Cache performs much better for the repetitive pattern in terms of disk I/Os and the elapsed time. Even with the sequential pattern, T-Cache shows almost the same performance as a system that does not use any caching, indicating the caching overhead in T-Cache is negligible.