• Title/Summary/Keyword: Visual detection

Search Result 866, Processing Time 0.027 seconds

Local and Global Information Exchange for Enhancing Object Detection and Tracking

  • Lee, Jin-Seok;Cho, Shung-Han;Oh, Seong-Jun;Hong, Sang-Jin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.6 no.5
    • /
    • pp.1400-1420
    • /
    • 2012
  • Object detection and tracking using visual sensors is a critical component of surveillance systems, which presents many challenges. This paper addresses the enhancement of object detection and tracking via the combination of multiple visual sensors. The enhancement method we introduce compensates for missed object detection based on the partial detection of objects by multiple visual sensors. When one detects an object or more visual sensors, the detected object's local positions transformed into a global object position. Local and global information exchange allows a missed local object's position to recover. However, the exchange of the information may degrade the detection and tracking performance by incorrectly recovering the local object position, which propagated by false object detection. Furthermore, local object positions corresponding to an identical object can transformed into nonequivalent global object positions because of detection uncertainty such as shadows or other artifacts. We improved the performance by preventing the propagation of false object detection. In addition, we present an evaluation method for the final global object position. The proposed method analyzed and evaluated using case studies.

A Fast and Precise Blob Detection

  • Nguyen, Thanh Binh;Chung, Sun-Tae
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2009.05a
    • /
    • pp.23-29
    • /
    • 2009
  • Blob detection is an essential ingredient process in some computer applications such as intelligent visual surveillance. However, previous blob detection algorithms are still computationally heavy so that supporting real-time multi-channel intelligent visual surveillance in a workstation or even one-channel real-time visual surveillance in a embedded system using them turns out prohibitively difficult. In this paper, we propose a fast and precise blob detection algorithm for visual surveillance. Blob detection in visual surveillance goes through several processing steps: foreground mask extraction, foreground mask correction, and connected component labeling. Foreground mask correction necessary for a precise detection is usually accomplished using morphological operations like opening and closing. Morphological operations are computationally expensive and moreover, they are difficult to run in parallel with connected component labeling routine since they need much different processing from what connected component labeling does. In this paper, we first develop a fast and precise foreground mask correction method utilizing on neighbor pixel checking which is also employed in connected component labeling so that the developed foreground mask correction method can be incorporated into connected component labeling routine. Through experiments, it is verified that our proposed blob detection algorithm based on the foreground mask correction method developed in this paper shows better processing speed and more precise blob detection.

  • PDF

Video smoke detection with block DNCNN and visual change image

  • Liu, Tong;Cheng, Jianghua;Yuan, Zhimin;Hua, Honghu;Zhao, Kangcheng
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.9
    • /
    • pp.3712-3729
    • /
    • 2020
  • Smoke detection is helpful for early fire detection. With its large coverage area and low cost, vision-based smoke detection technology is the main research direction of outdoor smoke detection. We propose a two-stage smoke detection method combined with block Deep Normalization and Convolutional Neural Network (DNCNN) and visual change image. In the first stage, each suspected smoke region is detected from each frame of the images by using block DNCNN. According to the physical characteristics of smoke diffusion, a concept of visual change image is put forward in this paper, which is constructed by the video motion change state of the suspected smoke regions, and can describe the physical diffusion characteristics of smoke in the time and space domains. In the second stage, the Support Vector Machine (SVM) classifier is used to classify the Histogram of Oriented Gradients (HOG) features of visual change images of the suspected smoke regions, in this way to reduce the false alarm caused by the smoke-like objects such as cloud and fog. Simulation experiments are carried out on two public datasets of smoke. Results show that the accuracy and recall rate of smoke detection are high, and the false alarm rate is much lower than that of other comparison methods.

Lip and Voice Synchronization Using Visual Attention (시각적 어텐션을 활용한 입술과 목소리의 동기화 연구)

  • Dongryun Yoon;Hyeonjoong Cho
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.4
    • /
    • pp.166-173
    • /
    • 2024
  • This study explores lip-sync detection, focusing on the synchronization between lip movements and voices in videos. Typically, lip-sync detection techniques involve cropping the facial area of a given video, utilizing the lower half of the cropped box as input for the visual encoder to extract visual features. To enhance the emphasis on the articulatory region of lips for more accurate lip-sync detection, we propose utilizing a pre-trained visual attention-based encoder. The Visual Transformer Pooling (VTP) module is employed as the visual encoder, originally designed for the lip-reading task, predicting the script based solely on visual information without audio. Our experimental results demonstrate that, despite having fewer learning parameters, our proposed method outperforms the latest model, VocaList, on the LRS2 dataset, achieving a lip-sync detection accuracy of 94.5% based on five context frames. Moreover, our approach exhibits an approximately 8% superiority over VocaList in lip-sync detection accuracy, even on an untrained dataset, Acappella.

Evaluation of Application of Possibility of Visual Surveillance System for Cow Heat Detection

  • Park, Heesu;Roy, Pantu Kumar;Noh, Youngju;Park, Hyuk;Lee, Joongho;Shin, Sangtae;Cho, Jongki
    • Journal of Embryo Transfer
    • /
    • v.31 no.2
    • /
    • pp.137-143
    • /
    • 2016
  • This study was conducted to evaluate a visual surveillance system. The advancement of recording technology and network service make it easy to record and transfer the videos. Moreover, progressed recognition technology help to make a distinction each other. Cows show distinguishing behaviors during their estrus period. The mounting is one of the behaviors. The result was different depending on the breed of the cows and the size of the farm. In the case of Korean native cattle, the estrus detection rate was 71.15%, however, dairy cows, the estrus detection rate was 39.38%. At the farms having below 6 modules, the estrus detection rate was 87.41%. On the other hand, at the farms having over 6 modules, the estrus detection rate was 77.78%. With the proper progress, the visual surveillance system can be used to detect heat detection.

Trends on Object Detection Techniques Based on Deep Learning (딥러닝 기반 객체 인식 기술 동향)

  • Lee, J.S.;Lee, S.K.;Kim, D.W.;Hong, S.J.;Yang, S.I.
    • Electronics and Telecommunications Trends
    • /
    • v.33 no.4
    • /
    • pp.23-32
    • /
    • 2018
  • Object detection is a challenging field in the visual understanding research area, detecting objects in visual scenes, and the location of such objects. It has recently been applied in various fields such as autonomous driving, image surveillance, and face recognition. In traditional methods of object detection, handcrafted features have been designed for overcoming various visual environments; however, they have a trade-off issue between accuracy and computational efficiency. Deep learning is a revolutionary paradigm in the machine-learning field. In addition, because deep-learning-based methods, particularly convolutional neural networks (CNNs), have outperformed conventional methods in terms of object detection, they have been studied in recent years. In this article, we provide a brief descriptive summary of several recent deep-learning methods for object detection and deep learning architectures. We also compare the performance of these methods and present a research guide of the object detection field.

A Review of Aircraft Camouflage Techniques to Reduce Visual Detection (항공기 시각 탐지 감소 위장기술 고찰)

  • Jin, Wonjin
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.5
    • /
    • pp.630-636
    • /
    • 2020
  • This study reviewed camouflage techniques to reduce the visual detect-ability of aircraft. Visual camouflage can be defined as the process of making objects less visible. Aircraft visual camouflage delays detection of the aircraft position, speed, and flight direction. Multi-tone and counter-shaded schemes are generally adopted as camouflage patterns for close-air-support aircraft and air-superiority aircraft, respectively. Another study showed that the monotone scheme is also efficient when the hue and brightness of the camouflage color are controlled correctly. Active camouflage techniques for aircraft have been studied to increase the camouflage effectiveness. In particular, counter-illumination techniques using electroluminescence devices can minimize the difference in brightness between the aircraft and sky background. Active camouflage techniques are expected to enhance the survivability of low-altitude UAVs, which are vulnerable to visual detection.

Real-time Human Detection under Omni-dir ectional Camera based on CNN with Unified Detection and AGMM for Visual Surveillance

  • Nguyen, Thanh Binh;Nguyen, Van Tuan;Chung, Sun-Tae;Cho, Seongwon
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.8
    • /
    • pp.1345-1360
    • /
    • 2016
  • In this paper, we propose a new real-time human detection under omni-directional cameras for visual surveillance purpose, based on CNN with unified detection and AGMM. Compared to CNN-based state-of-the-art object detection methods. YOLO model-based object detection method boasts of very fast object detection, but with less accuracy. The proposed method adapts the unified detecting CNN of YOLO model so as to be intensified by the additional foreground contextual information obtained from pre-stage AGMM. Increased computational time incurred by additional AGMM processing is compensated by speed-up gain obtained from utilizing 2-D input data consisting of grey-level image data and foreground context information instead of 3-D color input data. Through various experiments, it is shown that the proposed method performs better with respect to accuracy and more robust to environment changes than YOLO model-based human detection method, but with the similar processing speeds to that of YOLO model-based one. Thus, it can be successfully employed for embedded surveillance application.

Shot Group and Representative Shot Frame Detection using Similarity-based Clustering

  • Lee, Gye-Sung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.21 no.9
    • /
    • pp.37-43
    • /
    • 2016
  • This paper introduces a method for video shot group detection needed for efficient management and summary of video. The proposed method detects shots based on low-level visual properties and performs temporal and spatial clustering based on visual similarity of neighboring shots. Shot groups created from temporal clustering are further clustered into small groups with respect to visual similarity. A set of representative shot frames are selected from each cluster of the smaller groups representing a scene. Shots excluded from temporal clustering are also clustered into groups from which representative shot frames are selected. A number of video clips are collected and applied to the method for accuracy of shot group detection. We achieved 91% of accuracy of the method for shot group detection. The number of representative shot frames is reduced to 1/3 of the total shot frames. The experiment also shows the inverse relationship between accuracy and compression rate.

Scalable Re-detection for Correlation Filter in Visual Tracking

  • Park, Kayoung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.7
    • /
    • pp.57-64
    • /
    • 2020
  • In this paper, we propose an scalable re-detection for correlation filter in visual tracking. In real world, there are lots of target disappearances and reappearances during tracking, thus failure detection and re-detection methods are needed. One of the important point for re-detection is that a searching area must be large enough to find the missing target. For robust visual tracking, we adopt kernelized correlation filter as a baseline. Correlation filters have been extensively studied for visual object tracking in recent years. However conventional correlation filters detect the target in the same size area with the trained filter which is only 2 to 3 times larger than the target. When the target is disappeared for a long time, we need to search a wide area to re-detect the target. Proposed algorithm can search the target in a scalable area, hence the searching area is expanded by 2% in every frame from the target loss. Four datasets are used for experiments and both qualitative and quantitative results are shown in this paper. Our algorithm succeed the target re-detection in challenging datasets while conventional correlation filter fails.