• Title/Summary/Keyword: Visual Detection

Search Result 874, Processing Time 0.024 seconds

An Adaptive ROI Detection System for Spatiotemporal Features (시.공간특징에 대해 적응할 수 있는 ROI 탐지 시스템)

  • Park Min-Chul;Cheoi Kyung-Joo
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.1
    • /
    • pp.41-53
    • /
    • 2006
  • In this paper, an adaptive ROI(region of interest) detection system for spatialtemporal features is proposed. It utilizes spatiotemporal features for the purpose of detecting ROI. It is assumed that motion representing temporal visual conspicuity between adjacent frames takes higher priority over spatial visual conspicuity. Because objects or regions in motion usually draw stronger attention than others in motion pictures. In case of still images visual features that constitute topographic feature maps are used as spatial features. Comparative experiments with a human subjective evaluation show that correct detection rate of visual attention region is improved by exploiting both spatial and temporal features compared to the case of exploiting either feature.

  • PDF

Digital Modelling of Visual Perception in Architectural Environment

  • Seo, Dong-Yeon;Lee, Kyung-Hoi
    • KIEAE Journal
    • /
    • v.3 no.2
    • /
    • pp.59-66
    • /
    • 2003
  • To be the design method supporting aesthetic ability of human, CAAD system should essentially recognize architectural form in the same way of human. In this study, visual perception process of human was analyzed to search proper computational method performing similar step of perception of it. Through the analysis of visual perception, vision was separated to low-level vision and high-level vision. Edge detection and neural network were selected to model after low-level vision and high-level vision. The 24 images of building, tree and landscape were processed by edge detection and trained by neural network. And 24 new images were used to test trained network. The test shows that trained network gives right perception result toward each images with low error rate. This study is on the meaning of artificial intelligence in design process rather than on the design automation strategy through artificial intelligence.

Human-Robot Interaction in Real Environments by Audio-Visual Integration

  • Kim, Hyun-Don;Choi, Jong-Suk;Kim, Mun-Sang
    • International Journal of Control, Automation, and Systems
    • /
    • v.5 no.1
    • /
    • pp.61-69
    • /
    • 2007
  • In this paper, we developed not only a reliable sound localization system including a VAD(Voice Activity Detection) component using three microphones but also a face tracking system using a vision camera. Moreover, we proposed a way to integrate three systems in the human-robot interaction to compensate errors in the localization of a speaker and to reject unnecessary speech or noise signals entering from undesired directions effectively. For the purpose of verifying our system's performances, we installed the proposed audio-visual system in a prototype robot, called IROBAA(Intelligent ROBot for Active Audition), and demonstrated how to integrate the audio-visual system.

Event recognition of entering and exiting (출입 이벤트 인식)

  • Cui, Yaohuan;Lee, Chang-Woo
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2008.06a
    • /
    • pp.199-204
    • /
    • 2008
  • Visual surveillance is an active topic recently in Computer Vision. Event detection and recognition is one important and useful application of visual surveillance system. In this paper, we propose a new method to recognize the entering and exiting events based on the human's movement feature and the door's state. Without sensors, the proposed approach is based on novel and simple vision method as a combination of edge detection, motion history image and geometrical characteristic of the human shape. The proposed method includes several applications such as access control in visual surveillance and computer vision fields.

  • PDF

Visual Positioning System based on Voxel Labeling using Object Simultaneous Localization And Mapping

  • Jung, Tae-Won;Kim, In-Seon;Jung, Kye-Dong
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.302-306
    • /
    • 2021
  • Indoor localization is one of the basic elements of Location-Based Service, such as indoor navigation, location-based precision marketing, spatial recognition of robotics, augmented reality, and mixed reality. We propose a Voxel Labeling-based visual positioning system using object simultaneous localization and mapping (SLAM). Our method is a method of determining a location through single image 3D cuboid object detection and object SLAM for indoor navigation, then mapping to create an indoor map, addressing it with voxels, and matching with a defined space. First, high-quality cuboids are created from sampling 2D bounding boxes and vanishing points for single image object detection. And after jointly optimizing the poses of cameras, objects, and points, it is a Visual Positioning System (VPS) through matching with the pose information of the object in the voxel database. Our method provided the spatial information needed to the user with improved location accuracy and direction estimation.

A Study on Lip Detection based on Eye Localization for Visual Speech Recognition in Mobile Environment (모바일 환경에서의 시각 음성인식을 위한 눈 정위 기반 입술 탐지에 대한 연구)

  • Gyu, Song-Min;Pham, Thanh Trung;Kim, Jin-Young;Taek, Hwang-Sung
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.19 no.4
    • /
    • pp.478-484
    • /
    • 2009
  • Automatic speech recognition(ASR) is attractive technique in trend these day that seek convenient life. Although many approaches have been proposed for ASR but the performance is still not good in noisy environment. Now-a-days in the state of art in speech recognition, ASR uses not only the audio information but also the visual information. In this paper, We present a novel lip detection method for visual speech recognition in mobile environment. In order to apply visual information to speech recognition, we need to extract exact lip regions. Because eye-detection is more easy than lip-detection, we firstly detect positions of left and right eyes, then locate lip region roughly. After that we apply K-means clustering technique to devide that region into groups, than two lip corners and lip center are detected by choosing biggest one among clustered groups. Finally, we have shown the effectiveness of the proposed method through the experiments based on samsung AVSR database.

Deepfake Image Detection based on Visual Saliency (Visual Saliency 기반의 딥페이크 이미지 탐지 기법)

  • Harim Noh;Jehyeok Rew
    • Journal of Platform Technology
    • /
    • v.12 no.1
    • /
    • pp.128-140
    • /
    • 2024
  • 'Deepfake' refers to a video synthesis technique that utilizes various artificial intelligence technologies to create highly realistic fake content, causing serious confusion to individuals and society by being used for generating fake news, fraud, malicious impersonation, and more. To address this issue, there is a need for methods to detect malicious images generated by deepfake accurately. In this paper, we extract and analyze saliency features from deepfake and real images, and detect candidate synthesis regions on the images, and finally construct an automatic deepfake detection model by focusing on the extracted features. The proposed saliency feature-based model can be universally applied in situations where deepfake detection is required, such as synthesized images and videos. To demonstrate the performance of our approach, we conducted several experiments that have shown the effectiveness of the deepfake detection task.

  • PDF

Visual Phrase Detection and Evaluation Technology for Car Front Monitoring (자동차 전방 감시를 위한 영상 구문 검출 평가 기술)

  • OH, Weon Geun;KO, Jong-Gook
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2019.06a
    • /
    • pp.13-16
    • /
    • 2019
  • 영상이 포함하고 있는 풍부한 정보를 검출하고 이해하기 위해서는, 영상속의 일관된 상호관계를 갖는 영상 객체 그룹을 이해하고 표현하는 영상 구문 검출 기술(Visual Phrase Detecting Technology)이 필수적이다. 영상 구문 검출 기술은, 영상이 포함하고 있는 다양하고 풍부한 정보를 추출하고 활용하기 위한 핵심 기술로 이를 이용한 자동차 주행중 전방 감시, 영상의 자동 주석 달기, 동영상의 검색, 쇼핑 공간에서의 부정행위 검출(fraud detection) 등 다양한 분야에 적용할 수 있다.

  • PDF

Visual tracking based Discriminative Correlation Filter Using Target Separation and Detection

  • Lee, Jun-Haeng
    • Journal of the Korea Society of Computer and Information
    • /
    • v.22 no.12
    • /
    • pp.55-61
    • /
    • 2017
  • In this paper, we propose a novel tracking method using target separation and detection that are based on discriminative correlation filter (DCF), which is studied a lot recently. 'Retainability' is one of the most important factor of tracking. There are some factors making retainability of tracking worse. Especially, fast movement and occlusion of a target frequently occur in image data, and when it happens, it would make target lost. As a result, the tracking cannot be retained. For maintaining a robust tracking, in this paper, separation of a target is used so that normal tracking is maintained even though some part of a target is occluded. The detection algorithm is executed and find new location of the target when the target gets out of tracking range due to occlusion of whole part of a target or fast movement speed of a target. A variety of experiments with various image data sets are conducted. The algorithm proposed in this paper showed better performance than other conventional algorithms when fast movement and occlusion of a target occur.