• Title/Summary/Keyword: Monocular

Search Result 236, Processing Time 0.04 seconds

A Study on the Test Evaluation Method of LKAS Using a Monocular Camera (단안 카메라를 이용한 LKAS 시험평가 방법에 관한 연구)

  • Bae, Geon Hwan;Lee, Seon Bong
    • Journal of Auto-vehicle Safety Association
    • /
    • v.12 no.3
    • /
    • pp.34-42
    • /
    • 2020
  • ADAS (Advanced Driver Assistance Systems) uses sensors such as camera, radar, lidar and GPS (Global Positioning System). Among these sensors, the camera has many advantages compared with other sensors. The reason is that it is cheap, easy to use and can identify objects. In this paper, therefore, a theoretical formula was proposed to obtain the distance from the vehicle's front wheel to the lane using a monocular camera. And the validity of the theoretical formula was verified through the actual vehicle test. The results of the actual vehicle test in scenario 4 resulted in a maximum error of 0.21 m. The reason is that it is difficult to detect the lane in the curved road, and it is judged that errors occurred due to the occurrence of significant yaw rates. The maximum error occurred in curve road condition, but the error decreased after lane return. Therefore, the proposed theoretical formula makes it possible to assess the safety of the LKA system.

Multi-focus 3D Display (다초점 3차원 영상 표시 장치)

  • Kim, Seong-Gyu;Kim, Dong-Uk;Gwon, Yong-Mu;Son, Jeong-Yeong
    • Proceedings of the Optical Society of Korea Conference
    • /
    • 2008.07a
    • /
    • pp.119-120
    • /
    • 2008
  • A HMD type multi-focus 3D display system is developed and proof about satisfaction of eye accommodation is tested. Four LEDs(Light Emitting Diode) and a DMD are used to generate four parallax images at single eye and any mechanical part is not included in this system. The multi-focus means the ability of monocular depth cue to various depth levels. By achieving multi-focus function, we developed a 3D display system for only one eye, which can satisfy the accommodation to displayed virtual objects within defined depth. We could achieve a result that focus adjustment is possible at 5 step depths in sequence within 2m depth for only one eye. Additionally, the change level of burring depending on the focusing depth is tested by captured photos and moving pictures of video camera and several subjects. And the HMD type multi-focus 3D display can be applied to a monocular 3D display and monocular AR 3D display.

  • PDF

Deep Learning Based Monocular Depth Estimation: Survey

  • Lee, Chungkeun;Shim, Dongseok;Kim, H. Jin
    • Journal of Positioning, Navigation, and Timing
    • /
    • v.10 no.4
    • /
    • pp.297-305
    • /
    • 2021
  • Monocular depth estimation helps the robot to understand the surrounding environments in 3D. Especially, deep-learning-based monocular depth estimation has been widely researched, because it may overcome the scale ambiguity problem, which is a main issue in classical methods. Those learning based methods can be mainly divided into three parts: supervised learning, unsupervised learning, and semi-supervised learning. Supervised learning trains the network from dense ground-truth depth information, unsupervised one trains it from images sequences and semi-supervised one trains it from stereo images and sparse ground-truth depth. We describe the basics of each method, and then explain the recent research efforts to enhance the depth estimation performance.

Performance Analysis of Optimization Method and Filtering Method for Feature-based Monocular Visual SLAM (특징점 기반 단안 영상 SLAM의 최적화 기법 및 필터링 기법 성능 분석)

  • Jeon, Jin-Seok;Kim, Hyo-Joong;Shim, Duk-Sun
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.68 no.1
    • /
    • pp.182-188
    • /
    • 2019
  • Autonomous mobile robots need SLAM (simultaneous localization and mapping) to look for the location and simultaneously to make the map around the location. In order to achieve visual SLAM, it is necessary to form an algorithm that detects and extracts feature points from camera images, and gets the camera pose and 3D points of the features. In this paper, we propose MPROSAC algorithm which combines MSAC and PROSAC, and compare the performance of optimization method and the filtering method for feature-based monocular visual SLAM. Sparse Bundle Adjustment (SBA) is used for the optimization method and the extended Kalman filter is used for the filtering method.

A Survey for 3D Object Detection Algorithms from Images

  • Lee, Han-Lim;Kim, Ye-ji;Kim, Byung-Gyu
    • Journal of Multimedia Information System
    • /
    • v.9 no.3
    • /
    • pp.183-190
    • /
    • 2022
  • Image-based 3D object detection is one of the important and difficult problems in autonomous driving and robotics, and aims to find and represent the location, dimension and orientation of the object of interest. It generates three dimensional (3D) bounding boxes with only 2D images obtained from cameras, so there is no need for devices that provide accurate depth information such as LiDAR or Radar. Image-based methods can be divided into three main categories: monocular, stereo, and multi-view 3D object detection. In this paper, we investigate the recent state-of-the-art models of the above three categories. In the multi-view 3D object detection, which appeared together with the release of the new benchmark datasets, NuScenes and Waymo, we discuss the differences from the existing monocular and stereo methods. Also, we analyze their performance and discuss the advantages and disadvantages of them. Finally, we conclude the remaining challenges and a future direction in this field.

Bayesian Sensor Fusion of Monocular Vision and Laser Structured Light Sensor for Robust Localization of a Mobile Robot (이동 로봇의 강인 위치 추정을 위한 단안 비젼 센서와 레이저 구조광 센서의 베이시안 센서융합)

  • Kim, Min-Young;Ahn, Sang-Tae;Cho, Hyung-Suck
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.4
    • /
    • pp.381-390
    • /
    • 2010
  • This paper describes a procedure of the map-based localization for mobile robots by using a sensor fusion technique in structured environments. A combination of various sensors with different characteristics and limited sensibility has advantages in view of complementariness and cooperation to obtain better information on the environment. In this paper, for robust self-localization of a mobile robot with a monocular camera and a laser structured light sensor, environment information acquired from two sensors is combined and fused by a Bayesian sensor fusion technique based on the probabilistic reliability function of each sensor predefined through experiments. For the self-localization using the monocular vision, the robot utilizes image features consisting of vertical edge lines from input camera images, and they are used as natural landmark points in self-localization process. However, in case of using the laser structured light sensor, it utilizes geometrical features composed of corners and planes as natural landmark shapes during this process, which are extracted from range data at a constant height from the navigation floor. Although only each feature group of them is sometimes useful to localize mobile robots, all features from the two sensors are simultaneously used and fused in term of information for reliable localization under various environment conditions. To verify the advantage of using multi-sensor fusion, a series of experiments are performed, and experimental results are discussed in detail.

Tracking a Walking Motion Based on Dynamics Using a Monocular Camera (단일 카메라를 이용한 동역학 기반의 보행 동작 추적)

  • Yoo, Tae-Keun;Choi, Jae-Lim;Kim, Deok-Won
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.49 no.1
    • /
    • pp.20-28
    • /
    • 2012
  • Gait analysis is an examination which extracts objective information from observing human gait and assesses the function. The equipments used recently for gait analysis are expensive due to multiple cameras and force plates, and require the large space to set up the system. In this paper, we proposed a method to measure human gait motions in 3D from a monocular video. Our approach was based on particle filtering to track human motion without training data and previous information about a gait. We used dynamics to make physics-based motions with the consideration of contacts between feet and base. In a walking sequence, our approach showed the mean angular error of $12.4^{\circ}$ over all joints, which was much smaller than the error of $34.6^{\circ}$ with the conventional particle filter. These results showed that a monocular camera is able to replace the existing complicated system for measuring human gait quantitatively.

Method to Improve Localization and Mapping Accuracy on the Urban Road Using GPS, Monocular Camera and HD Map (GPS와 단안카메라, HD Map을 이용한 도심 도로상에서의 위치측정 및 맵핑 정확도 향상 방안)

  • Kim, Young-Hun;Kim, Jae-Myeong;Kim, Gi-Chang;Choi, Yun-Soo
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.5_1
    • /
    • pp.1095-1109
    • /
    • 2021
  • The technology used to recognize the location and surroundings of autonomous vehicles is called SLAM. SLAM standsfor Simultaneously Localization and Mapping and hasrecently been actively utilized in research on autonomous vehicles,starting with robotic research. Expensive GPS, INS, LiDAR, RADAR, and Wheel Odometry allow precise magnetic positioning and mapping in centimeters. However, if it can secure similar accuracy as using cheaper Cameras and GPS data, it will contribute to advancing the era of autonomous driving. In this paper, we present a method for converging monocular camera with RTK-enabled GPS data to perform RMSE 33.7 cm localization and mapping on the urban road.

Effects of visual information on Y-Balance Test (시각정보가 Y-Balance Test에 미치는 영향)

  • Byung-Hoon Woo
    • Journal of the Korean Applied Science and Technology
    • /
    • v.40 no.5
    • /
    • pp.977-987
    • /
    • 2023
  • The purpose of this study was to investigate the effect of visual information on the dynamic balance on Y-balance Test(YBT). The subjects of the study were 18 male and female adults in their 20s and 30s (age: 23.17±1.72 years, height: 172.46±9.84 cm, weight: 73.39±11.44 kg, leg length: 88.89±5.69 cm) who participated in the study. To measure dynamic balance between binocular and monocular use, absolute reach distance, composite score, and COP variables were measured on left and right feet through YBT and results were derived. As a result of the study, monocular block(left and right eye block) showed higher absolute reach and composite scores than binocular use in posterolateral, posteromedial, and composite scores during YBT. As a result of COP, there was no difference in anterior and posteromedial reach. When reaching posterolateral, AP COP velocity of left foot in monocular block appeared slower than that in binocular vision, and in COP velocity, COP velocity of left foot in monocular block appeared slower than binocular vision.

Application of Virtual Studio Technology and Digital Human Monocular Motion Capture Technology -Based on <Beast Town> as an Example-

  • YuanZi Sang;KiHong Kim;JuneSok Lee;JiChu Tang;GaoHe Zhang;ZhengRan Liu;QianRu Liu;ShiJie Sun;YuTing Wang;KaiXing Wang
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.1
    • /
    • pp.106-123
    • /
    • 2024
  • This article takes the talk show "Beast Town" as an example to introduce the overall technical solution, technical difficulties and countermeasures for the combination of cartoon virtual characters and virtual studio technology, providing reference and experience for the multi-scenario application of digital humans. Compared with the live broadcast that combines reality and reality, we have further upgraded our virtual production technology and digital human-driven technology, adopted industry-leading real-time virtual production technology and monocular camera driving technology, and launched a virtual cartoon character talk show - "Beast Town" to achieve real Perfectly combined with virtuality, it further enhances program immersion and audio-visual experience, and expands infinite boundaries for virtual manufacturing. In the talk show, motion capture shooting technology is used for final picture synthesis. The virtual scene needs to present dynamic effects, and at the same time realize the driving of the digital human and the movement with the push, pull and pan of the overall picture. This puts forward very high requirements for multi-party data synchronization, real-time driving of digital people, and synthetic picture rendering. We focus on issues such as virtual and real data docking and monocular camera motion capture effects. We combine camera outward tracking, multi-scene picture perspective, multi-machine rendering and other solutions to effectively solve picture linkage and rendering quality problems in a deeply immersive space environment. , presenting users with visual effects of linkage between digital people and live guests.