• Title/Summary/Keyword: Real scene image

Search Result 223, Processing Time 0.029 seconds

Intelligent Composition of CG and Dynamic Scene (CG와 동영상의 지적합성)

  • 박종일;정경훈;박경세;송재극
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1995.06a
    • /
    • pp.77-81
    • /
    • 1995
  • Video composition is to integrate multiple image materials into one scene. It considerably enhances the degree of freedom in producing various scenes. However, we need to adjust the viewing point sand the image planes of image planes of image materials for high quality video composition. In this paper, were propose an intelligent video composition technique concentrating on the composition of CG and real scene. We first model the camera system. The projection is assumed to be perspective and the camera motion is assumed to be 3D rotational and 3D translational. Then, we automatically extract camera parameters comprising the camera model from real scene by a dedicated algorithm. After that, CG scene is generated according to the camera parameters of the real scene. Finally the two are composed into one scene. Experimental results justify the validity of the proposed method.

Real Scene Text Image Super-Resolution Based on Multi-Scale and Attention Fusion

  • Xinhua Lu;Haihai Wei;Li Ma;Qingji Xue;Yonghui Fu
    • Journal of Information Processing Systems
    • /
    • v.19 no.4
    • /
    • pp.427-438
    • /
    • 2023
  • Plenty of works have indicated that single image super-resolution (SISR) models relying on synthetic datasets are difficult to be applied to real scene text image super-resolution (STISR) for its more complex degradation. The up-to-date dataset for realistic STISR is called TextZoom, while the current methods trained on this dataset have not considered the effect of multi-scale features of text images. In this paper, a multi-scale and attention fusion model for realistic STISR is proposed. The multi-scale learning mechanism is introduced to acquire sophisticated feature representations of text images; The spatial and channel attentions are introduced to capture the local information and inter-channel interaction information of text images; At last, this paper designs a multi-scale residual attention module by skillfully fusing multi-scale learning and attention mechanisms. The experiments on TextZoom demonstrate that the model proposed increases scene text recognition's (ASTER) average recognition accuracy by 1.2% compared to text super-resolution network.

An Onboard Image Processing System for Road Images (도로교통 영상처리를 위한 고속 영상처리시스템의 하드웨어 구현)

  • 이운근;이준웅;조석빈;고덕화;백광렬
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.9 no.7
    • /
    • pp.498-506
    • /
    • 2003
  • A computer vision system applied to an intelligent safety vehicle has been required to be worked on a small sized real time special purposed hardware not on a general purposed computer. In addition, the system should have a high reliability even under the adverse road traffic environment. This paper presents a design and an implementation of an onboard hardware system taking into account for high speed image processing to analyze a road traffic scene. The system is mainly composed of two parts: an early processing module of FPGA and a postprocessing module of DSP. The early processing module is designed to extract several image primitives such as the intensity of a gray level image and edge attributes in a real-time Especially, the module is optimized for the Sobel edge operation. The postprocessing module of DSP utilizes the image features from the early processing module for making image understanding or image analysis of a road traffic scene. The performance of the proposed system is evaluated by an experiment of a lane-related information extraction. The experiment shows the successful results of image processing speed of twenty-five frames of 320$\times$240 pixels per second.

Modeling the Visual Target Search in Natural Scenes

  • Park, Daecheol;Myung, Rohae;Kim, Sang-Hyeob;Jang, Eun-Hye;Park, Byoung-Jun
    • Journal of the Ergonomics Society of Korea
    • /
    • v.31 no.6
    • /
    • pp.705-713
    • /
    • 2012
  • Objective: The aim of this study is to predict human visual target search using ACT-R cognitive architecture in real scene images. Background: Human uses both the method of bottom-up and top-down process at the same time using characteristics of image itself and knowledge about images. Modeling of human visual search also needs to include both processes. Method: In this study, visual target object search performance in real scene images was analyzed comparing experimental data and result of ACT-R model. 10 students participated in this experiment and the model was simulated ten times. This experiment was conducted in two conditions, indoor images and outdoor images. The ACT-R model considering the first saccade region through calculating the saliency map and spatial layout was established. Proposed model in this study used the guide of visual search and adopted visual search strategies according to the guide. Results: In the analysis results, no significant difference on performance time between model prediction and empirical data was found. Conclusion: The proposed ACT-R model is able to predict the human visual search process in real scene images using salience map and spatial layout. Application: This study is useful in conducting model-based evaluation in visual search, particularly in real images. Also, this study is able to adopt in diverse image processing program such as helper of the visually impaired.

Fast Content-Aware Video Retargeting Algorithm (고속 컨텐츠 인식 동영상 리타겟팅 기법)

  • Park, Dae-Hyun;Kim, Yoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.11
    • /
    • pp.77-86
    • /
    • 2013
  • In this paper, we propose a fast video retargeting method which preserves the contents of a video and converts the image size. Since the conventional Seam Carving which is the well-known content-aware image retargeting technique uses the dynamic programming method, the repetitive update procedure of the accumulation energy is absolutely needed to obtain seam. The energy update procedure cannot avoid the processing time delay because of many operations by the image full-searching. By applying the proposed method, frames which have similar features in video are classified into a scene, and the first frame of a scene is resized by the modified Seam Carving where multiple seams are extracted from candidate seams to reduce the repetitive update procedure. After resizing the first frame of a scene, all continuous frames of the same scene are resized with reference to the seam information stored in the previous frame without the calculation of the accumulation energy. Therefore, although the fast processing is possible with reducing complexity and without analyzing all frames of scene, the quality of an image can be analogously maintained with an existing method. The experimental results show that the proposed method can preserve the contents of an image and can be practically applied to retarget the image on real time.

A Calibration Algorithm Using Known Angle (각도 정보를 이용한 카메라 보정 알고리듬)

  • 권인소;하종은
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.10 no.5
    • /
    • pp.415-420
    • /
    • 2004
  • We present a new algorithm for the calibration of a camera and the recovery of 3D scene structure up to a scale from image sequences using known angles between lines in the scene. Traditional method for calibration using scene constraints requires various scene constraints due to the stratified approach. Proposed method requires only one type of scene constraint of known angle and also it directly recovers metric structure up to an unknown scale from projective structure. Specifically, we recover the matrix that is the homography between the projective structure and the Euclidean structure using angles. Since this matrix is a unique one in the given set of image sequences, we can easily deal with the problem of varying intrinsic parameters of the camera. Experimental results on the synthetic and real images demonstrate the feasibility of the proposed algorithm.

An effective background subtraction in dynamic scene. (동적 환경에서의 효과적인 움직이는 객체 추출)

  • Han, Jae-Hyek;Kim, Yong-Jin;Ryu, Sae-Woon;Lee, Sang-Hwa;Park, Jong-Il
    • 한국HCI학회:학술대회논문집
    • /
    • 2009.02a
    • /
    • pp.631-636
    • /
    • 2009
  • Foreground segmentation methods have steadily been researched in the field of computer vision. Especially, background subtraction which extracts a foreground image from the difference between the current frame and a reference image, called as "background image" have been widely used for a variety of real-time applications because of low computation and high-quality. However, if the background scene was dynamically changed, the background subtraction causes lots of errors. In this paper, we propose an efficient background subtraction method in dynamic environment with both static and dynamic scene. The proposed method is a hybrid method that uses the conventional background subtraction for static scene and depth information for dynamic scene. Its validity and efficiency are verified by demonstration in dynamic environment, where a video projector projects various images in the background.

  • PDF

Development of 3D Stereoscopic Image Generation System Using Real-time Preview Function in 3D Modeling Tools

  • Yun, Chang-Ok;Yun, Tae-Soo;Lee, Dong-Hoon
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.6
    • /
    • pp.746-754
    • /
    • 2008
  • A 3D stereoscopic image is generated by interdigitating every scene with video editing tools that are rendered by two cameras' views in 3D modeling tools, like Autodesk MAX(R) and Autodesk MAYA(R). However, the depth of object from a static scene and the continuous stereo effect in the view of transformation, are not represented in a natural method. This is because after choosing the settings of arbitrary angle of convergence and the distance between the modeling and those two cameras, the user needs to render the view from both cameras. So, the user needs a process of controlling the camera's interval and rendering repetitively, which takes too much time. Therefore, in this paper, we will propose the 3D stereoscopic image editing system for solving such problems as well as exposing the system's inherent limitations. We can generate the view of two cameras and can confirm the stereo effect in real-time on 3D modeling tools. Then, we can intuitively determine immersion of 3D stereoscopic image in real-time, by using the 3D stereoscopic image preview function.

  • PDF

Fast Object Recognition using Local Energy Propagation from Combination of Saline Line Groups (직선 조합의 에너지 전파를 이용한 고속 물체인식)

  • 강동중
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2000.10a
    • /
    • pp.311-311
    • /
    • 2000
  • We propose a DP-based formulation for matching line patterns by defining a robust and stable geometric representation that is based on the conceptual organizations. Usually, the endpoint proximity and collinearity of image lines, as two main conceptual organization groups, are useful cues to match the model shape in the scene. As the endpoint proximity, we detect junctions from image lines. We then search for junction groups by using geometric constraint between the junctions. A junction chain similar to the model chain is searched in the scene, based on a local comparison. A Dynamic Programming-based search algorithm reduces the time complexity for the search of the model chain in the scene. Our system can find a reasonable matching, although there exist severely distorted objects in the scene. We demonstrate the feasibility of the DP-based matching method using both synthetic and real images.

  • PDF