• Title/Summary/Keyword: Building Object Detection

Search Result 128, Processing Time 0.028 seconds

Implementation of Smart Shopping Cart using Object Detection Method based on Deep Learning (딥러닝 객체 탐지 기술을 사용한 스마트 쇼핑카트의 구현)

  • Oh, Jin-Seon;Chun, In-Gook
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.7
    • /
    • pp.262-269
    • /
    • 2020
  • Recently, many attempts have been made to reduce the time required for payment in various shopping environments. In addition, for the Fourth Industrial Revolution era, artificial intelligence is advancing, and Internet of Things (IoT) devices are becoming more compact and cheaper. So, by integrating these two technologies, access to building an unmanned environment to save people time has become easier. In this paper, we propose a smart shopping cart system based on low-cost IoT equipment and deep-learning object-detection technology. The proposed smart cart system consists of a camera for real-time product detection, an ultrasonic sensor that acts as a trigger, a weight sensor to determine whether a product is put into or taken out of the shopping cart, an application for smartphones that provides a user interface for a virtual shopping cart, and a deep learning server where learned product data are stored. Communication between each module is through Transmission Control Protocol/Internet Protocol, a Hypertext Transmission Protocol network, a You Only Look Once darknet library, and an object detection system used by the server to recognize products. The user can check a list of items put into the smart cart via the smartphone app, and can automatically pay for them. The smart cart system proposed in this paper can be applied to unmanned stores with high cost-effectiveness.

Moving Target Tracking and Recognition for Location Based Surveillance Service (위치기반 감시 서비스를 위한 이동 객체 추적 및 인식)

  • Kim, Hyun;Park, Chan-Ho;Woo, Jong-Woo;Doo, Seok-Bae
    • Proceedings of the IEEK Conference
    • /
    • 2008.06a
    • /
    • pp.1211-1212
    • /
    • 2008
  • In this paper, we propose image process modeling as a part of location based surveillance system for unauthorized target recognition and tracking in harbor, airport, military zone. For this, we compress and store background image in lower resolution and perform object extraction and motion tracking by using sobel edge detection and difference picture method between real images and a background image. In addition to, we use Independent Component Analysis Neural Network for moving target recognition. Experiments are performed for object extraction and tracking of moving targets on road by using static camera in 20m height building and it shows the robust results.

  • PDF

A Vehicle Recognition Method based on Radar and Camera Fusion in an Autonomous Driving Environment

  • Park, Mun-Yong;Lee, Suk-Ki;Shin, Dong-Jin
    • International journal of advanced smart convergence
    • /
    • v.10 no.4
    • /
    • pp.263-272
    • /
    • 2021
  • At a time when securing driving safety is the most important in the development and commercialization of autonomous vehicles, AI and big data-based algorithms are being studied to enhance and optimize the recognition and detection performance of various static and dynamic vehicles. However, there are many research cases to recognize it as the same vehicle by utilizing the unique advantages of radar and cameras, but they do not use deep learning image processing technology or detect only short distances as the same target due to radar performance problems. Radars can recognize vehicles without errors in situations such as night and fog, but it is not accurate even if the type of object is determined through RCS values, so accurate classification of the object through images such as cameras is required. Therefore, we propose a fusion-based vehicle recognition method that configures data sets that can be collected by radar device and camera device, calculates errors in the data sets, and recognizes them as the same target.

Detecting and Counting People system based on Vision Sensor (비전 센서 기반의 사람 검출 및 계수 시스템)

  • Park, Ho-Sik
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.6 no.1
    • /
    • pp.1-5
    • /
    • 2013
  • The number of pedestrians is considered essential information which can be used to control a person who makes a entrance or a exit into a building. The number of pedestrians, also, can be used to help to manage pedestrian traffic and the volume of pedestrian flow within the building. Due to the fact there is incorrect detection by occluded, shadows, and illumination, however, difficulty can arise in existing system which is for detection and counts of a person who makes a entrance or a exit into a building. In this paper, it is minimized that the change of illumination and the effect of shadow through the transmitted image from camera which is created and processed with great adaptability. The accuracy of the calculations can be increase as well by using Kalman Filter and Mean-Shift Algorithm in order to avoid overlapped counts. As a result of the test, it is proved that the count method that shows the accuracy of 95.4% should be effective for detection and counts.

A Collision detection from division space for performance improvement of MMORPG game engine (MMORPG 게임엔진의 성능개선을 위한 분할공간에서의 충돌검출)

  • Lee, Sung-Ug
    • The KIPS Transactions:PartB
    • /
    • v.10B no.5
    • /
    • pp.567-574
    • /
    • 2003
  • Application field of third dimension graphic is becoming diversification by the fast development of hardware recently. Various theory of details technology necessary to design game such as 3D MMORPG (Massive Multi-play Online Role Flaying Game) that do with third dimension. Cyber city should be absorbed. It is the detection speed that this treatise is necessary in game engine design. 3D MMORPG game engine has much factor that influence to speed as well as rendering processing because it express huge third dimension city´s grate many building and individual fast effectively by real time. This treatise nay get concept about the collision in 3D MMORPG and detection speed elevation of game engine through improved detection method. Space division is need to process fast dynamically wide outside that is 3D MMORPG´s main detection target. 3D is constructed with tree construct individual that need collision using processing geometry dataset that is given through new graph. We may search individual that need in collision detection and improve the collision detection speed as using hierarchical bounding box that use it with detection volume. Octree that will use by division octree is used mainly to express rightly static object but this paper use limited OSP by limited space division structure to use this in dynamic environment. Limited OSP space use limited space with method that divide square to classify typically complicated 3D space´s object. Through this detection, this paper propose follow contents, first, this detection may judge collision detection at early time without doing all polygon´s collision examination. Second, this paper may improve detection efficiency of game engine through and then reduce detection time because detection time of bounding box´s collision detection.

Automatic assessment of post-earthquake buildings based on multi-task deep learning with auxiliary tasks

  • Zhihang Li;Huamei Zhu;Mengqi Huang;Pengxuan Ji;Hongyu Huang;Qianbing Zhang
    • Smart Structures and Systems
    • /
    • v.31 no.4
    • /
    • pp.383-392
    • /
    • 2023
  • Post-earthquake building condition assessment is crucial for subsequent rescue and remediation and can be automated by emerging computer vision and deep learning technologies. This study is based on an endeavour for the 2nd International Competition of Structural Health Monitoring (IC-SHM 2021). The task package includes five image segmentation objectives - defects (crack/spall/rebar exposure), structural component, and damage state. The structural component and damage state tasks are identified as the priority that can form actionable decisions. A multi-task Convolutional Neural Network (CNN) is proposed to conduct the two major tasks simultaneously. The rest 3 sub-tasks (spall/crack/rebar exposure) were incorporated as auxiliary tasks. By synchronously learning defect information (spall/crack/rebar exposure), the multi-task CNN model outperforms the counterpart single-task models in recognizing structural components and estimating damage states. Particularly, the pixel-level damage state estimation witnesses a mIoU (mean intersection over union) improvement from 0.5855 to 0.6374. For the defect detection tasks, rebar exposure is omitted due to the extremely biased sample distribution. The segmentations of crack and spall are automated by single-task U-Net but with extra efforts to resample the provided data. The segmentation of small objects (spall and crack) benefits from the resampling method, with a substantial IoU increment of nearly 10%.

Information Fusion of Photogrammetric Imagery and Lidar for Reliable Building Extraction (광학 영상과 Lidar의 정보 융합에 의한 신뢰성 있는 구조물 검출)

  • Lee, Dong-Hyuk;Lee, Kyoung-Mu;Lee, Sang-Uk
    • Journal of Broadcast Engineering
    • /
    • v.13 no.2
    • /
    • pp.236-244
    • /
    • 2008
  • We propose a new building detection and description algorithm for Lidar data and photogrammetric imagery using color segmentation, line segments matching, perceptual grouping. Our algorithm consists of two steps. In the first step, from the initial building regions extracted from Lidar data and the color segmentation results from the photogrammetric imagery, we extract coarse building boundaries based on the Lidar results with split and merge technique from aerial imagery. In the secondstep, we extract precise building boundaries based on coarse building boundaries and edges from aerial imagery using line segments matching and perceptual grouping. The contribution of this algorithm is that color information in photogrammetric imagery is used to complement collapsed building boundaries obtained by Lidar. Moreover, linearity of the edges and construction of closed roof form are used to reflect the characteristic of man-made object. Experimental results on multisensor data demonstrate that the proposed algorithm produces more accurate and reliable results than Lidar sensor.

Object detection in financial reporting documents for subsequent recognition

  • Sokerin, Petr;Volkova, Alla;Kushnarev, Kirill
    • International journal of advanced smart convergence
    • /
    • v.10 no.1
    • /
    • pp.1-11
    • /
    • 2021
  • Document page segmentation is an important step in building a quality optical character recognition module. The study examined already existing work on the topic of page segmentation and focused on the development of a segmentation model that has greater functional significance for application in an organization, as well as broad capabilities for managing the quality of the model. The main problems of document segmentation were highlighted, which include a complex background of intersecting objects. As classes for detection, not only classic text, table and figure were selected, but also additional types, such as signature, logo and table without borders (or with partially missing borders). This made it possible to pose a non-trivial task of detecting non-standard document elements. The authors compared existing neural network architectures for object detection based on published research data. The most suitable architecture was RetinaNet. To ensure the possibility of quality control of the model, a method based on neural network modeling using the RetinaNet architecture is proposed. During the study, several models were built, the quality of which was assessed on the test sample using the Mean average Precision metric. The best result among the constructed algorithms was shown by a model that includes four neural networks: the focus of the first neural network on detecting tables and tables without borders, the second - seals and signatures, the third - pictures and logos, and the fourth - text. As a result of the analysis, it was revealed that the approach based on four neural networks showed the best results in accordance with the objectives of the study on the test sample in the context of most classes of detection. The method proposed in the article can be used to recognize other objects. A promising direction in which the analysis can be continued is the segmentation of tables; the areas of the table that differ in function will act as classes: heading, cell with a name, cell with data, empty cell.

Analysis of Utilization Status and Limitations of Intelligent CCTV for Safety Management at Construction Sites (건설현장 안전관리를 위한 지능형 CCTV의 활용 현황 및 한계 분석)

  • Kim, Jae-Min;Yu, Jung-Ho
    • Proceedings of the Korean Institute of Building Construction Conference
    • /
    • 2023.05a
    • /
    • pp.203-204
    • /
    • 2023
  • The construction industry is a hazardous environment in which many field workers work. Therefore, there is a limit to the safety manager's grasp of all situations. In order to solve these problems, the application of automatic control technology in connection with AI and CCTV is being introduced, and the development of intelligent CCTV to reduce the safety accident rate is actively progressing. This study seeks to present future directions by identifying the current status of intelligent CCTV developed to reduce the safety accident rate at construction sites and analyzing its limitations. Through this, the range of accident prevention types of the safety control system at the construction site will be confirmed and the need for future intelligent CCTV function development will be suggested.

  • PDF

Equipment and Worker Recognition of Construction Site with Vision Feature Detection

  • Qi, Shaowen;Shan, Jiazeng;Xu, Lei
    • International Journal of High-Rise Buildings
    • /
    • v.9 no.4
    • /
    • pp.335-342
    • /
    • 2020
  • This article comes up with a new method which is based on the visual characteristic of the objects and machine learning technology to achieve semi-automated recognition of the personnel, machine & materials of the construction sites. Balancing the real-time performance and accuracy, using Faster RCNN (Faster Region-based Convolutional Neural Networks) with transfer learning method appears to be a rational choice. After fine-tuning an ImageNet pre-trained Faster RCNN and testing with it, the result shows that the precision ratio (mAP) has so far reached 67.62%, while the recall ratio (AR) has reached 56.23%. In other word, this recognizing method has achieved rational performance. Further inference with the video of the construction of Huoshenshan Hospital also indicates preliminary success.