• Title/Summary/Keyword: Optical flow estimation

Search Result 122, Processing Time 0.027 seconds

Motion Depth Generation Using MHI for 3D Video Conversion (3D 동영상 변환을 위한 MHI 기반 모션 깊이맵 생성)

  • Kim, Won Hoi;Gil, Jong In;Choi, Changyeol;Kim, Manbae
    • Journal of Broadcast Engineering
    • /
    • v.22 no.4
    • /
    • pp.429-437
    • /
    • 2017
  • 2D-to-3D conversion technology has been studied over past decades and integrated to commercial 3D displays and 3DTVs. Generally, depth cues extracted from a static image is used for generating a depth map followed by DIBR (Depth Image Based Rendering) for producing a stereoscopic image. Further, motion is also an important cue for depth estimation and is estimated by block-based motion estimation, optical flow and so forth. This papers proposes a new method for motion depth generation using Motion History Image (MHI) and evaluates the feasiblity of the MHI utilization. In the experiments, the proposed method was performed on eight video clips with a variety of motion classes. From a qualitative test on motion depth maps as well as the comparison of the processing time, we validated the feasibility of the proposed method.

Estimation of Urban Traffic State Using Black Box Camera (차량 블랙박스 카메라를 이용한 도시부 교통상태 추정)

  • Haechan Cho;Yeohwan Yoon;Hwasoo Yeo
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.22 no.2
    • /
    • pp.133-146
    • /
    • 2023
  • Traffic states in urban areas are essential to implement effective traffic operation and traffic control. However, installing traffic sensors on numerous road sections is extremely expensive. Accordingly, estimating the traffic state using a vehicle-mounted camera, which shows a high penetration rate, is a more effective solution. However, the previously proposed methodology using object tracking or optical flow has a high computational cost and requires consecutive frames to obtain traffic states. Accordingly, we propose a method to detect vehicles and lanes by object detection networks and set the region between lanes as a region of interest to estimate the traffic density of the corresponding area. The proposed method only uses less computationally expensive object detection models and can estimate traffic states from sampled frames rather than consecutive frames. In addition, the traffic density estimation accuracy was over 90% on the black box videos collected from two buses having different characteristics.

Estimation of liquid limit of cohesive soil using video-based vibration measurement

  • Matthew Sands;Evan Hayes;Soonkie Nam;Jinki Kim
    • Geomechanics and Engineering
    • /
    • v.33 no.2
    • /
    • pp.175-182
    • /
    • 2023
  • In general, the design of structures and its construction processes are fundamentally dependent on their foundation and supporting ground. Thus, it is imperative to understand the behavior of the soil under certain stress and drainage conditions. As it is well known that certain characteristics and behaviors of soils with fines are highly dependent on water content, it is critical to accurately measure and identify the status of the soils in terms of water contents. Liquid limit is one of the important soil index properties to define such characteristics. However, liquid limit measurement can be affected by the proficiency of the operator. On the other hand, dynamic properties of soils are also necessary in many different applications and current testing methods often require special equipment in the laboratory, which is often expensive and sensitive to test conditions. In order to address these concerns and advance the state of the art, this study explores a novel method to determine the liquid limit of cohesive soil by employing video-based vibration analysis. In this research, the modal characteristics of cohesive soil columns are extracted from videos by utilizing phase-based motion estimation. By utilizing the proposed method that analyzes the optical flow in every pixel of the series of frames that effectively represents the motion of corresponding points of the soil specimen, the vibration characteristics of the entire soil specimen could be assessed in a non-contact and non-destructive manner. The experimental investigation results compared with the liquid limit determined by the standard method verify that the proposed method reliably and straightforwardly identifies the liquid limit of clay. It is envisioned that the proposed approach could be applied to measuring liquid limit of soil in practical field, entertaining its simple implementation that only requires a digital camera or even a smartphone without the need for special equipment that may be subject to the proficiency of the operator.

Estimation of channel morphology using RGB orthomosaic images from drone - focusing on the Naesung stream - (드론 RGB 정사영상 기반 하도 지형 공간 추정 방법 - 내성천 중심으로 -)

  • Woo-Chul, KANG;Kyng-Su, LEE;Eun-Kyung, JANG
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.25 no.4
    • /
    • pp.136-150
    • /
    • 2022
  • In this study, a comparative review was conducted on how to use RGB images to obtain river topographic information, which is one of the most essential data for eco-friendly river management and flood level analysis. In terms of the topographic information of river zone, to obtain the topographic information of flow section is one of the difficult topic, therefore, this study focused on estimating the river topographic information of flow section through RGB images. For this study, the river topography surveying was directly conducted using ADCP and RTK-GPS, and at the same time, and orthomosiac image were created using high-resolution images obtained by drone photography. And then, the existing developed regression equations were applied to the result of channel topography surveying by ADCP and the band values of the RGB images, and the channel bathymetry in the study area was estimated using the regression equation that showed the best predictability. In addition, CCHE2D flow modeling was simulated to perform comparative verification of the topographical informations. The modeling result with the image-based topographical information provided better water depth and current velocity simulation results, when it compared to the directly measured topographical information for which measurement of the sub-section was not performed. It is concluded that river topographic information could be obtained from RGB images, and if additional research was conducted, it could be used as a method of obtaining efficient river topographic information for river management.

ESTIMATION OF ERRORS IN THE TRANSVERSE VELOCITY VECTORS DETERMINED FROM HINODE/SOT MAGNETOGRAMS USING THE NAVE TECHNIQUE

  • Chae, Jong-Chul;Moon, Yong-Jae
    • Journal of The Korean Astronomical Society
    • /
    • v.42 no.3
    • /
    • pp.61-69
    • /
    • 2009
  • Transverse velocity vectors can be determined from a pair of images successively taken with a time interval using an optical flow technique. We have tested the performance of the new technique called NAVE (non-linear affine velocity estimator) recently implemented by Chae & Sakurai using real image data taken by the Narrowband Filter Imager (NFI) of the Solar Optical Telescope (SOT) aboard the Hinode satellite. We have developed two methods of estimating the errors in the determination of velocity vectors, one resulting from the non-linear fitting ${\sigma}_{\upsilon}$ and the other ${\epsilon}_u$ resulting from the statistics of the determined velocity vectors. The real error is expected to be somewhere between ${\sigma}_{\upsilon}$ and ${\epsilon}_u$. We have investigated the dependence of the determined velocity vectors and their errors on the different parameters such as the critical speed for the subsonic filtering, the width of the localizing window, the time interval between two successive images, and the signal-to-noise ratio of the feature. With the choice of $v_{crit}$ = 2 pixel/step for the subsonic filtering, and the window FWHM of 16 pixels, and the time interval of one step (2 minutes), we find that the errors of velocity vectors determined using the NAVE range from around 0.04 pixel/step in high signal-to-noise ratio features (S/N $\sim$ 10), to 0.1 pixel/step in low signa-to-noise ratio features (S/N $\sim$ 3) with the mean of about 0.06 pixel/step where 1 pixel/step corresponds roughly to 1 km/s in our case.

3D Ultrasound Panoramic Image Reconstruction using Deep Learning (딥러닝을 활용한 3차원 초음파 파노라마 영상 복원)

  • SiYeoul Lee;Seonho Kim;Dongeon Lee;ChunSu Park;MinWoo Kim
    • Journal of Biomedical Engineering Research
    • /
    • v.44 no.4
    • /
    • pp.255-263
    • /
    • 2023
  • Clinical ultrasound (US) is a widely used imaging modality with various clinical applications. However, capturing a large field of view often requires specialized transducers which have limitations for specific clinical scenarios. Panoramic imaging offers an alternative approach by sequentially aligning image sections acquired from freehand sweeps using a standard transducer. To reconstruct a 3D volume from these 2D sections, an external device can be employed to track the transducer's motion accurately. However, the presence of optical or electrical interferences in a clinical setting often leads to incorrect measurements from such sensors. In this paper, we propose a deep learning (DL) framework that enables the prediction of scan trajectories using only US data, eliminating the need for an external tracking device. Our approach incorporates diverse data types, including correlation volume, optical flow, B-mode images, and rawer data (IQ data). We develop a DL network capable of effectively handling these data types and introduce an attention technique to emphasize crucial local areas for precise trajectory prediction. Through extensive experimentation, we demonstrate the superiority of our proposed method over other DL-based approaches in terms of long trajectory prediction performance. Our findings highlight the potential of employing DL techniques for trajectory estimation in clinical ultrasound, offering a promising alternative for panoramic imaging.

Video Deblurring using Camera Motion Estimation and Patch-wise Deconvolution (카메라 움직임 추정 및 패치 기반 디컨볼루션을 이용한 동영상의 번짐 현상 제거 방법)

  • Jeong, Woojin;Park, Jin Wook;Lee, Jong Min;Song, Tae Eun;Choi, Wonju;Moon, Young Shik
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.51 no.12
    • /
    • pp.130-139
    • /
    • 2014
  • Undesired camera shaking can make a blur effect, which causes a degradation of video quality. We propose an efficient method of removing the blur effects on video captured from a single camera. The proposed method has a sequential process that is applied to each frame. The first stage is to estimate the camera motion for each frame. In order to estimate the camera motion, we compute the optical flow using 3 consecutive frames. Then a patch-wise image deconvolution is applied. During the deconvolution, edge prediction is used to improve the quality of image deconvolution. After patch-wise image deconvolution, deblurred patches are integrated into an image to produce a deblurred frame. The above process is performed for each frame. The experimental result shows that the proposed method removes the blur effect efficiently.

Development of a Cost-Effective Tele-Robot System Delivering Speaker's Affirmative and Negative Intentions (화자의 긍정·부정 의도를 전달하는 실용적 텔레프레즌스 로봇 시스템의 개발)

  • Jin, Yong-Kyu;You, Su-Jeong;Cho, Hye-Kyung
    • The Journal of Korea Robotics Society
    • /
    • v.10 no.3
    • /
    • pp.171-177
    • /
    • 2015
  • A telerobot offers a more engaging and enjoyable interaction with people at a distance by communicating via audio, video, expressive gestures, body pose and proxemics. To provide its potential benefits at a reasonable cost, this paper presents a telepresence robot system for video communication which can deliver speaker's head motion through its display stanchion. Head gestures such as nodding and head-shaking can give crucial information during conversation. We also can assume a speaker's eye-gaze, which is known as one of the key non-verbal signals for interaction, from his/her head pose. In order to develop an efficient head tracking method, a 3D cylinder-like head model is employed and the Harris corner detector is combined with the Lucas-Kanade optical flow that is known to be suitable for extracting 3D motion information of the model. Especially, a skin color-based face detection algorithm is proposed to achieve robust performance upon variant directions while maintaining reasonable computational cost. The performance of the proposed head tracking algorithm is verified through the experiments using BU's standard data sets. A design of robot platform is also described as well as the design of supporting systems such as video transmission and robot control interfaces.

In-Car Video Stabilization using Focus of Expansion

  • Kim, Jin-Hyun;Baek, Yeul-Min;Yun, Jea-Ho;Kim, Whoi-Yul
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.12
    • /
    • pp.1536-1543
    • /
    • 2011
  • Video stabilization is a very important step for vision based applications in the vehicular technology because the accuracy of these applications such as obstacle distance estimation, lane detection and tracking can be affected by bumpy roads and oscillation of vehicle. Conventional methods suffer from either the zooming effect which caused by a camera movement or some motion of surrounding vehicles. In order to overcome this problem, we propose a novel video stabilization method using FOE(Focus of Expansion). When a vehicle moves, optical flow diffuses from the FOE and the FOE is equal to an epipole. If a vehicle moves with vibration, the position of the epipole in the two consecutive frames is changed by oscillation of the vehicle. Therefore, we carry out video stabilization using motion vector estimated from the amount of change of the epipoles. Experiment results show that the proposed method is more efficient than conventional methods.

Multi-robot Formation based on Object Tracking Method using Fisheye Images (어안 영상을 이용한 물체 추적 기반의 한 멀티로봇의 대형 제어)

  • Choi, Yun Won;Kim, Jong Uk;Choi, Jeong Won;Lee, Suk Gyu
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.6
    • /
    • pp.547-554
    • /
    • 2013
  • This paper proposes a novel formation algorithm of identical robots based on object tracking method using omni-directional images obtained through fisheye lenses which are mounted on the robots. Conventional formation methods of multi-robots often use stereo vision system or vision system with reflector instead of general purpose camera which has small angle of view to enlarge view angle of camera. In addition, to make up the lack of image information on the environment, robots share the information on their positions through communication. The proposed system estimates the region of robots using SURF in fisheye images that have $360^{\circ}$ of image information without merging images. The whole system controls formation of robots based on moving directions and velocities of robots which can be obtained by applying Lucas-Kanade Optical Flow Estimation for the estimated region of robots. We confirmed the reliability of the proposed formation control strategy for multi-robots through both simulation and experiment.