• Title/Summary/Keyword: scene detection

Search Result 519, Processing Time 0.027 seconds

Change Detection of Land-cover from Multi-temporal KOMPSAT-1 EOC Imageries

  • Ha, Sung-Ryong;Ahn, Byung-Woon;Park, Sang-Young
    • Korean Journal of Remote Sensing
    • /
    • v.18 no.1
    • /
    • pp.13-23
    • /
    • 2002
  • A radiometric correction method is developed to apply multi-temporal KOMPSAT-1 EOC satellite images for the detection of land-cover changes b\ulcorner recognizing changes in reflection pattern. Radiometric correction was carried out to eliminate the atmospheric effects that could interfere with the image properly of the satellite data acquired at different multi-times. Four invariant features of water, sand, paved road, and roofs of building are selected and a linear regression relationship among the control set images is used as a correction scheme. It is found that the utilization of panchromatic multi-temporal imagery requires the radiometric scene standardization process to correct radiometric errors that include atmospheric effects and digital image processing errors. Land-cover with specific change pattern such as paddy field is extracted by seasonal change recognition process.

A Study of Medium Shot Detection (미디엄 숏 검출에 관한 연구)

  • Hyung Lee
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2023.01a
    • /
    • pp.93-95
    • /
    • 2023
  • 본 논문에서는 장편의 드라마나 영화에서 스토리 기반의 축약된 요약본을 자동으로 제작하기 위해 미디엄 숏(medium shot) 크기의 숏(shot)들을 추출하기 위한 방법을 고려한다. 미디엄 숏 정도의 크기는 보통 인물에 중심을 둔 숏들로 인물들 간의 관계에서 특히 대사나 표정으로 내용을 전달하기 위한 목적으로 적극 권장된다. 비디오 검색을 위한 인덱싱에서 신(scene) 전환 검출 및 숏 경계 검출, 그리고 이미지에서 심도와 초점기반의 화질 및 피사체 추출 등을 위해 전통적인 신호/영상처리 기법의 활용에서부터 최근의 기계학습 접목 등 다양한 연구들이 진행되고 있다. 영상문법에 근거하여 편집된 영상물에서 미디엄 숏 정도 크기의 숏들을 추출하여 배열한다면 어느 정도 원본 내용을 충실히 전달할 수 있는 축약된 요약본을 제작할 수 있다는 가정하에 해당 샷들을 블러(blur) 기반으로 검출하기 위해 이와 관련된 키워드들을 기반으로 기존 연구들을 살펴보고 적용 방법을 모색한다.

  • PDF

Template Based Object Detection & Tracking by Chamfer Matching in Real Time Video (Chamfer Matching을 이용한 실시간 템플릿 기반 개체 검출 및 추적)

  • Islam, Md. Zahidul;Setiawan, Nurul Arif;Kim, Hyung-Kwan;Lee, Chil-Woo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2008.05a
    • /
    • pp.92-94
    • /
    • 2008
  • In this paper we describe an approach for template based detection and tracking of objects by chamfer matching in real time video. Detecting and tracking of any objects is the key problem in computer vision. In our case we try for hand and head of human for detection and tracking by chamfer matching technique. Matching involves correlating the templates with the distance transformed scene and determining the locations where the mismatch is below a certain user defined threshold.

Analysis of the Fog Detection Algorithm of DCD Method with SST and CALIPSO Data (SST와 CALIPSO 자료를 이용한 DCD 방법으로 정의된 안개화소 분석)

  • Shin, Daegeun;Park, Hyungmin;Kim, Jae Hwan
    • Atmosphere
    • /
    • v.23 no.4
    • /
    • pp.471-483
    • /
    • 2013
  • Nighttime sea fog detection from satellite is very hard due to limitation in using visible channels. Currently, most widely used method for the detection is the Dual Channel Difference (DCD) method based on Brightness Temperature Difference between 3.7 and 11 ${\mu}m$ channel (BTD). However, this method have difficulty in distinguishing between fog and low cloud, and sometimes misjudges middle/high cloud as well as clear scene as fog. Using CALIPSO Lidar Profile measurements, we have analyzed the intrinsic problems in detecting nighttime sea fog from various satellite remote sensing algorithms and suggested the direction for the improvement of the algorithm. From the comparison with CALIPSO measurements for May-July in 2011, the DCD method excessively overestimates foggy pixels (2542 pixels). Among them, only 524 pixel are real foggy pixels, but 331 pixels and 1687 pixels are clear and other type of clouds, respectively. The 514 of real foggy pixels accounts for 70% of 749 foggy pixels identified by CALIPSO. Our proposed new algorithm detects foggy pixels by comparing the difference between cloud top temperature and underneath sea surface temperature from assimilated data along with the DCD method. We have used two types of cloud top temperature, which obtained from 11 ${\mu}m$ brightness temperature (B_S1) and operational COMS algorithm (B_S2). The detected foggy 1794 pixels from B_S1 and 1490 pixel from B_S2 are significantly reduced the overestimation detected by the DCD method. However, 477 and 446 pixels have been found to be real foggy pixels, 329 and 264 pixels be clear, and 989 and 780 pixels be other type of clouds, detected by B_S1 and B_S2 respectively. The analysis of the operational COMS fog detection algorithm reveals that the cloud screening process was strictly enforced, which resulted in underestimation of foggy pixel. The 538 of total detected foggy pixels obtain only 187 of real foggy pixels, but 61 of clear pixels and 290 of other type clouds. Our analysis suggests that there is no winner for nighttime sea fog detection algorithms, but loser because real foggy pixels are less than 30% among the foggy pixels declared by all algorithms. This overwhelming evidence reveals that current nighttime sea fog algorithms have provided a lot of misjudged information, which are mostly originated from difficulty in distinguishing between clear and cloudy scene as well as fog and other type clouds. Therefore, in-depth researches are urgently required to reduce the enormous error in nighttime sea fog detection from satellite.

Performance Analysis of Feature Detection Methods for Topology-Based Feature Description (토폴로지 기반 특징 기술을 위한 특징 검출 방법의 성능 분석)

  • Park, Han-Hoon;Moon, Kwang-Seok
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.16 no.2
    • /
    • pp.44-49
    • /
    • 2015
  • When the scene has less texture or when camera pose largely changes, the existing texture-based feature tracking methods are not reliable. Topology-based feature description methods, which use the geometric relationship between features such as LLAH, is a good alternative. However, they require feature detection methods with high performance. As a basic study on developing an effective feature detection method for topology-based feature description, this paper aims at examining their applicability to topology-based feature description by analyzing the repeatability of several feature detection methods that are included in the OpenCV library. Experimental results show that FAST outperforms the others.

Dense Optical flow based Moving Object Detection at Dynamic Scenes (동적 배경에서의 고밀도 광류 기반 이동 객체 검출)

  • Lim, Hyojin;Choi, Yeongyu;Nguyen Khac, Cuong;Jung, Ho-Youl
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.11 no.5
    • /
    • pp.277-285
    • /
    • 2016
  • Moving object detection system has been an emerging research field in various advanced driver assistance systems (ADAS) and surveillance system. In this paper, we propose two optical flow based moving object detection methods at dynamic scenes. Both proposed methods consist of three successive steps; pre-processing, foreground segmentation, and post-processing steps. Two proposed methods have the same pre-processing and post-processing steps, but different foreground segmentation step. Pre-processing calculates mainly optical flow map of which each pixel has the amplitude of motion vector. Dense optical flows are estimated by using Farneback technique, and the amplitude of the motion normalized into the range from 0 to 255 is assigned to each pixel of optical flow map. In the foreground segmentation step, moving object and background are classified by using the optical flow map. Here, we proposed two algorithms. One is Gaussian mixture model (GMM) based background subtraction, which is applied on optical map. Another is adaptive thresholding based foreground segmentation, which classifies each pixel into object and background by updating threshold value column by column. Through the simulations, we show that both optical flow based methods can achieve good enough object detection performances in dynamic scenes.

Manhole Cover Detection from Natural Scene Based on Imaging Environment Perception

  • Liu, Haoting;Yan, Beibei;Wang, Wei;Li, Xin;Guo, Zhenhui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.10
    • /
    • pp.5095-5111
    • /
    • 2019
  • A multi-rotor Unmanned Aerial Vehicle (UAV) system is developed to solve the manhole cover detection problem for the infrastructure maintenance in the suburbs of big city. The visible light sensor is employed to collect the ground image data and a series of image processing and machine learning methods are used to detect the manhole cover. First, the image enhancement technique is employed to improve the imaging effect of visible light camera. An imaging environment perception method is used to increase the computation robustness: the blind Image Quality Evaluation Metrics (IQEMs) are used to percept the imaging environment and select the images which have a high imaging definition for the following computation. Because of its excellent processing effect the adaptive Multiple Scale Retinex (MSR) is used to enhance the imaging quality. Second, the Single Shot multi-box Detector (SSD) method is utilized to identify the manhole cover for its stable processing effect. Third, the spatial coordinate of manhole cover is also estimated from the ground image. The practical applications have verified the outdoor environment adaptability of proposed algorithm and the target detection correctness of proposed system. The detection accuracy can reach 99% and the positioning accuracy is about 0.7 meters.

Social Pedestrian Group Detection Based on Spatiotemporal-oriented Energy for Crowd Video Understanding

  • Huang, Shaonian;Huang, Dongjun;Khuhroa, Mansoor Ahmed
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.8
    • /
    • pp.3769-3789
    • /
    • 2018
  • Social pedestrian groups are the basic elements that constitute a crowd; therefore, detection of such groups is scientifically important for modeling social behavior, as well as practically useful for crowd video understanding. A social group refers to a cluster of members who tend to keep similar motion state for a sustained period of time. One of the main challenges of social group detection arises from the complex dynamic variations of crowd patterns. Therefore, most works model dynamic groups to analysis the crowd behavior, ignoring the existence of stationary groups in crowd scene. However, in this paper, we propose a novel unified framework for detecting social pedestrian groups in crowd videos, including dynamic and stationary pedestrian groups, based on spatiotemporal-oriented energy measurements. Dynamic pedestrian groups are hierarchically clustered based on energy flow similarities and trajectory motion correlations between the atomic groups extracted from principal spatiotemporal-oriented energies. Furthermore, the probability distribution of static spatiotemporal-oriented energies is modeled to detect stationary pedestrian groups. Extensive experiments on challenging datasets demonstrate that our method can achieve superior results for social pedestrian group detection and crowd video classification.

Salient Region Detection Algorithm for Music Video Browsing (뮤직비디오 브라우징을 위한 중요 구간 검출 알고리즘)

  • Kim, Hyoung-Gook;Shin, Dong
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.2
    • /
    • pp.112-118
    • /
    • 2009
  • This paper proposes a rapid detection algorithm of a salient region for music video browsing system, which can be applied to mobile device and digital video recorder (DVR). The input music video is decomposed into the music and video tracks. For the music track, the music highlight including musical chorus is detected based on structure analysis using energy-based peak position detection. Using the emotional models generated by SVM-AdaBoost learning algorithm, the music signal of the music videos is classified into one of the predefined emotional classes of the music automatically. For the video track, the face scene including the singer or actor/actress is detected based on a boosted cascade of simple features. Finally, the salient region is generated based on the alignment of boundaries of the music highlight and the visual face scene. First, the users select their favorite music videos from various music videos in the mobile devices or DVR with the information of a music video's emotion and thereafter they can browse the salient region with a length of 30-seconds using the proposed algorithm quickly. A mean opinion score (MOS) test with a database of 200 music videos is conducted to compare the detected salient region with the predefined manual part. The MOS test results show that the detected salient region using the proposed method performed much better than the predefined manual part without audiovisual processing.

Mobile Robot Path Finding Using Invariant Landmarks

  • Sharma, Kajal
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.5 no.3
    • /
    • pp.178-184
    • /
    • 2016
  • This paper proposes a new path-finding scheme using viewpoint-invariant landmarks. The scheme introduces the concept of landmark detection in images captured with a vision sensor attached to a mobile robot, and provides landmark clues to determine a path. Experiment results show that the scheme efficiently detects landmarks with changes in scenes due to the robot's movement. The scheme accurately detects landmarks and reduces the overall landmark computation cost. The robot moves in the room to capture different images. It can efficiently detect landmarks in the room from different viewpoints of each scene. The outcome of the proposed scheme results in accurate and obstacle-free path estimation.