• Title/Summary/Keyword: Video matching

Search Result 445, Processing Time 0.026 seconds

MRBR-based JPEG2000 Codec for Stereoscopic Image Compression of 3-Dimensional Digital Cinema (3차원 디지털 시네마의 스테레오 영상 압축을 위한 MRBR기반의 JPEG2000 코덱)

  • Seo, Young-Ho;Sin, Wan-Soo;Choi, Hyun-Jun;Yoo, Ji-Sang;Kim, Dong-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.12 no.12
    • /
    • pp.2146-2152
    • /
    • 2008
  • In In this paper, we proposed a new JPEG2000 codec using multiresolution-based rendering (MRBR) technique for video compression of 3-dimensional digital cinema. We introduced discrete wavelet transform (DWT) for stereoscopic image and stereo matching technique in the wavelet domain. The disparity was extracted using stereo matching and transmitted with the reference (left) image. Since the generated right image was degraded by the occlusion lesion, the residual image which is generated from difference between the original right image and the generated one was transmitted at the same tine. The disparity data was extracted using the dynamic programming method in the disparity domain. There is high correlation between the higher and lower subbands. Therefore we decreased the calculation amount and enhanced accuracy by restricting the search window and applying the disparity information generated from higher subband.

Proposal of a Hand Motion and Control Method Matching System for Interaction Optimized for Mixed Reality Control - Focusing on Meta Interface and Augmented Behavior (혼합현실 컨트롤에 최적화된 인터랙션을 위한 손 동작과 컨트롤 방식 Matching체계의 제안 - Augmented Behavior와 Meta Interface를 중심으로)

  • Lee, SaYa;Lee, EunJong
    • Smart Media Journal
    • /
    • v.11 no.9
    • /
    • pp.81-93
    • /
    • 2022
  • In an era where non-face-to-face meetings become common, eXtended Reality(XR) is rapidly developing and filling areas that are not satisfied in online meetings based on existing photos/video method. In particular, general users are also able to easily access and use HMD-type mixed reality devices. However, the basic operations applied in HMD-type in Mixed Reality(MR) with hands as the main input tool do not have a standardized system, and each manufacturer operates in a separate response to each other's hand movements. Therefore, this study considered that a systematic hand motion matching system considering the usability and efficiency of operations performed in mixed reality was necessary, and conducted a study to clarify this. First, the basic operation performed in the MR environment and its attributes were investigated, and at the same time, the structure of the hand and the attributes of the possible hand movements were identified. Based on the identified properties, it is intended to present a system that can intuitively and efficiently match basic operation properties in the MR environment with subtle operation properties according to the structure/context of the hand.

A Research on Cylindrical Pill Bottle Recognition with YOLOv8 and ORB

  • Dae-Hyun Kim;Hyo Hyun Choi
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.2
    • /
    • pp.13-20
    • /
    • 2024
  • This paper introduces a method for generating model images that can identify specific cylindrical medicine containers in videos and investigates data collection techniques. Previous research had separated object detection from specific object recognition, making it challenging to apply automated image stitching. A significant issue was that the coordinate-based object detection method included extraneous information from outside the object area during the image stitching process. To overcome these challenges, this study applies the newly released YOLOv8 (You Only Look Once) segmentation technique to vertically rotating pill bottles video and employs the ORB (Oriented FAST and Rotated BRIEF) feature matching algorithm to automate model image generation. The research findings demonstrate that applying segmentation techniques improves recognition accuracy when identifying specific pill bottles. The model images created with the feature matching algorithm could accurately identify the specific pill bottles.

A Hybrid Approach of Efficient Facial Feature Detection and Tracking for Real-time Face Direction Estimation (실시간 얼굴 방향성 추정을 위한 효율적인 얼굴 특성 검출과 추적의 결합방법)

  • Kim, Woonggi;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.117-124
    • /
    • 2013
  • In this paper, we present a new method which efficiently estimates a face direction from a sequences of input video images in real time fashion. For this work, the proposed method performs detecting the facial region and major facial features such as both eyes, nose and mouth by using the Haar-like feature, which is relatively not sensitive against light variation, from the detected facial area. Then, it becomes able to track the feature points from every frame using optical flow in real time fashion, and determine the direction of the face based on the feature points tracked. Further, in order to prevent the erroneously recognizing the false positions of the facial features when if the coordinates of the features are lost during the tracking by using optical flow, the proposed method determines the validity of locations of the facial features using the template matching of detected facial features in real time. Depending on the correlation rate of re-considering the detection of the features by the template matching, the face direction estimation process is divided into detecting the facial features again or tracking features while determining the direction of the face. The template matching initially saves the location information of 4 facial features such as the left and right eye, the end of nose and mouse in facial feature detection phase and reevaluated these information when the similarity measure between the stored information and the traced facial information by optical flow is exceed a certain level of threshold by detecting the new facial features from the input image. The proposed approach automatically combines the phase of detecting facial features and the phase of tracking features reciprocally and enables to estimate face pose stably in a real-time fashion. From the experiment, we can prove that the proposed method efficiently estimates face direction.

A Study on Improved Image Matching Method using the CUDA Computing (CUDA 연산을 이용한 개선된 영상 매칭 방법에 관한 연구)

  • Cho, Kyeongrae;Park, Byungjoon;Yoon, Taebok
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.16 no.4
    • /
    • pp.2749-2756
    • /
    • 2015
  • Recently, Depending on the quality of data increases, the problem of time-consuming to process the image is raised by being required to accelerate the image processing algorithms, in a traditional CPU and CUDA(Compute Unified Device Architecture) based recognition system for computing speed and performance gains compared to OpenMP When character recognition has been learned by the system to measure the input by the character data matching is implemented in an environment that recognizes the region of the well, so that the font of the characters image learning English alphabet are each constant and standardized in size and character an image matching method for calculating the matching has also been implemented. GPGPU (General Purpose GPU) programming platform technology when using the CUDA computing techniques to recognize and use the four cores of Intel i5 2500 with OpenMP to deal quickly and efficiently an algorithm, than the performance of existing CPU does not produce the rate of four times due to the delay of the data of the partition and merge operation proposed a method of improving the rate of speed of about 3.2 times, and the parallel processing of the video card that processes a result, the sequential operation of the process compared to CPU-based who performed the performance gain is about 21 tiems improvement in was confirmed.

An Efficient Hardware-Software Co-Implementation of an H.263 Video Codec (하드웨어 소프트웨어 통합 설계에 의한 H.263 동영상 코덱 구현)

  • 장성규;김성득;이재헌;정의철;최건영;김종대;나종범
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.25 no.4B
    • /
    • pp.771-782
    • /
    • 2000
  • In this paper, an H.263 video codec is implemented by adopting the concept of hardware and software co-design. Each module of the codec is investigated to find which approach between hardware and software is better to achieve real-time processing speed as well as flexibility. The hardware portion includes motion-related engines, such as motion estimation and compensation, and a memory control part. The remaining portion of theH.263 video codec is implemented in software using a RISC processor. This paper also introduces efficient design methods for hardware and software modules. In hardware, an area-efficient architecture for the motion estimator of a multi-resolution block matching algorithm using multiple candidates and spatial correlation in motion vector fields (MRMCS), is suggested to reduce the chip size. Software optimization techniques are also explored by using the statistics of transformed coefficients and the minimum sum of absolute difference (SAD)obtained from the motion estimator.

  • PDF

Stereo-To-Multiview Conversion System Using FPGA and GPU Device (FPGA와 GPU를 이용한 스테레오/다시점 변환 시스템)

  • Shin, Hong-Chang;Lee, Jinwhan;Lee, Gwangsoon;Hur, Namho
    • Journal of Broadcast Engineering
    • /
    • v.19 no.5
    • /
    • pp.616-626
    • /
    • 2014
  • In this paper, we introduce a real-time stereo-to-multiview conversion system using FPGA and GPU. The system is based on two different devices so that it consists of two major blocks. The first block is a disparity estimation block that is implemented on FPGA. In this block, each disparity map of stereoscopic video is estimated by DP(dynamic programming)-based stereo matching. And then the estimated disparity maps are refined by post-processing. The refined disparity map is transferred to the GPU device through USB 3.0 and PCI-express interfaces. Stereoscopic video is also transferred to the GPU device. These data are used to render arbitrary number of virtual views in next block. In the second block, disparity-based view interpolation is performed to generate virtual multi-view video. As a final step, all generated views have to be re-arranged into a single image at full resolution for presenting on the target autostereoscopic 3D display. All these steps of the second block are performed in parallel on the GPU device.

Design of Moving Picture Retrieval System using Scene Change Technique (장면 전환 기법을 이용한 동영상 검색 시스템 설계)

  • Kim, Jang-Hui;Kang, Dae-Seong
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.3
    • /
    • pp.8-15
    • /
    • 2007
  • Recently, it is important to process multimedia data efficiently. Especially, in case of retrieval of multimedia information, technique of user interface and retrieval technique are necessary. This paper proposes a new technique which detects cuts effectively in compressed image information by MPEG. A cut is a turning point of scenes. The cut-detection is the basic work and the first-step for video indexing and retrieval. Existing methods have a weak point that they detect wrong cuts according to change of a screen such as fast motion of an object, movement of a camera and a flash. Because they compare between previous frame and present frame. The proposed technique detects shots at first using DC(Direct Current) coefficient of DCT(Discrete Cosine Transform). The database is composed of these detected shots. Features are extracted by HMMD color model and edge histogram descriptor(EHD) among the MPEG-7 visual descriptors. And detections are performed in sequence by the proposed matching technique. Through this experiments, an improved video segmentation system is implemented that it performs more quickly and precisely than existing techniques have.

Rain Detection and Removal Algorithm using Motion-Compensated Non-local Means Filter for Video Sequences (동영상을 위한 움직임 보상 기반 Non-Local Means 필터를 이용한 우적 검출 및 제거 알고리즘)

  • Seo, Seung Ji;Song, Byung Cheol
    • Journal of Broadcast Engineering
    • /
    • v.20 no.1
    • /
    • pp.153-163
    • /
    • 2015
  • This paper proposes a rain detection and removal algorithm that is robust against camera motion in video sequences. In detection part, the proposed algorithm initially detects possible rain streaks by using intensity properties and spatial properties. Then, the rain streak candidates are selected based on Gaussian distribution model. In removal part, a non-rain block matching algorithm is performed between adjacent frames to find similar blocks to the block that has rain pixels. If the similar blocks to the block are obtained, the rain region of the block is reconstructed by non-local means (NLM) filter using the similar neighbors. Experimental results show that the proposed algorithm outperforms the previous works in terms of subjective visual quality of de-rained video sequences.

Efficient Representation and Matching of Object Movement using Shape Sequence Descriptor (모양 시퀀스 기술자를 이용한 효과적인 동작 표현 및 검색 방법)

  • Choi, Min-Seok
    • The KIPS Transactions:PartB
    • /
    • v.15B no.5
    • /
    • pp.391-396
    • /
    • 2008
  • Motion of object in a video clip often plays an important role in characterizing the content of the clip. A number of methods have been developed to analyze and retrieve video contents using motion information. However, most of these methods focused more on the analysis of direction or trajectory of motion but less on the analysis of the movement of an object itself. In this paper, we propose the shape sequence descriptor to describe and compare the movement based on the shape deformation caused by object motion along the time. A movement information is first represented a sequence of 2D shape of object extracted from input image sequence, and then 2D shape information is converted 1D shape feature using the shape descriptor. The shape sequence descriptor is obtained from the shape descriptor sequence by frequency transform along the time. Our experiment results show that the proposed method can be very simple and effective to describe the object movement and can be applicable to semantic applications such as content-based video retrieval and human movement recognition.