• Title/Summary/Keyword: vision model

Search Result 1,320, Processing Time 0.037 seconds

Model-Based Pose Estimation for High-Precise Underwater Navigation Using Monocular Vision (단안 카메라를 이용한 수중 정밀 항법을 위한 모델 기반 포즈 추정)

  • Park, JiSung;Kim, JinWhan
    • The Journal of Korea Robotics Society
    • /
    • v.11 no.4
    • /
    • pp.226-234
    • /
    • 2016
  • In this study, a model-referenced underwater navigation algorithm is proposed for high-precise underwater navigation using monocular vision near underwater structures. The main idea of this navigation algorithm is that a 3D model-based pose estimation is combined with the inertial navigation using an extended Kalman filter (EKF). The spatial information obtained from the navigation algorithm is utilized for enabling the underwater robot to navigate near underwater structures whose geometric models are known a priori. For investigating the performance of the proposed approach the model-referenced navigation algorithm was applied to an underwater robot and a set of experiments was carried out in a water tank.

The Optimal Grasp Planning by Using a 3-D Computer Vision Technique (3차원 영상처리 기술을 이용한 Grasp planning의 최적화)

  • 이현기;김성환;최상균;이상룡
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.19 no.11
    • /
    • pp.54-64
    • /
    • 2002
  • This paper deals with the problem of synthesis of stable and optimal grasps with unknown objects by 3-finger hand. Previous robot grasp research has mainly analyzed with either unknown objects 2-dimensionally by vision sensor or known objects, such as cylindrical objects, 3-dimensionally. As extending the previous work, in this study we propose an algorithm to analyze grasp of unknown objects 3-dimensionally by using vision sensor. This is archived by two steps. The first step is to make a 3-dimensional geometrical model for unknown objects by using stereo matching. The second step is to find the optimal grasping points. In this step, we choose the 3-finger hand which has the characteristic of multi-finger hand and is easy to model. To find the optimal grasping points, genetic algorithm is employed and objective function minimizes the admissible force of finger tip applied to the objects. The algorithm is verified by computer simulation by which optimal grasping points of known objects with different angle are checked.

MPC-based Active Steering Control using Multi-rate Kalman Filter for Autonomous Vehicle Systems with Vision (비젼 기반 자율주행을 위한 다중비율 예측기 설계와 모델예측 기반 능동조향 제어)

  • Kim, Bo-Ah;Lee, Young-Ok;Lee, Seung-Hi;Chung, Chung-Choo
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.61 no.5
    • /
    • pp.735-743
    • /
    • 2012
  • In this paper, we present model predictive control (MPC) applied to lane keeping system (LKS) based on a vision module. Due to a slow sampling rate of the vision system, the conventional LKS using single rate control may result in uncomfortable steering control rate in a high vehicle speed. By applying MPC using multi-rate Kalman filter to active steering control, the proposed MPC-based active steering control system prevents undesirable saturated steering control command. The effectiveness of the MPC is validated by simulations for the LKS equipped with a camera module having a slow sampling rate on the curved lane with the minimum radius of 250[m] at a vehicle speed of 30[m/s].

Robust Camera Calibration using TSK Fuzzy Modeling

  • Lee, Hee-Sung;Hong, Sung-Jun;Kim, Eun-Tai
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.7 no.3
    • /
    • pp.216-220
    • /
    • 2007
  • Camera calibration in machine vision is the process of determining the intrinsic camera parameters and the three-dimensional (3D) position and orientation of the camera frame relative to a certain world coordinate system. On the other hand, Takagi-Sugeno-Kang (TSK) fuzzy system is a very popular fuzzy system and approximates any nonlinear function to arbitrary accuracy with only a small number of fuzzy rules. It demonstrates not only nonlinear behavior but also transparent structure. In this paper, we present a novel and simple technique for camera calibration for machine vision using TSK fuzzy model. The proposed method divides the world into some regions according to camera view and uses the clustered 3D geometric knowledge. TSK fuzzy system is employed to estimate the camera parameters by combining partial information into complete 3D information. The experiments are performed to verify the proposed camera calibration.

Identifying Barriers to Big Data Analytics: Design-Reality Gap Analysis in Saudi Higher Education

  • AlMobark, Bandar Abdullah
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.9
    • /
    • pp.261-266
    • /
    • 2021
  • The spread of cloud computing, digital computing, and the popular social media platforms have led to increased growth of data. That growth of data results in what is known as big data (BD), which seen as one of the most strategic resources. The analysis of these BD has allowed generating value from massive raw data that helps in making effective decisions and providing quality of service. With Vision 2030, Saudi Arabia seeks to invest in BD technologies, but many challenges and barriers have led to delays in adopting BD. This research paper aims to search in the state of Big Data Analytics (BDA) in Saudi higher education sector, identify the barriers by reviewing the literature, and then to apply the design-reality gap model to assess these barriers that prevent effective use of big data and highlights priority areas for action to accelerate the application of BD to comply with Vision 2030.

A Computer Vision Approach for Identifying Acupuncture Points on the Face and Hand Using the MediaPipe Framework (MediaPipe Framework를 이용한 얼굴과 손의 경혈 판별을 위한 Computer Vision 접근법)

  • Hadi S. Malekroodi;Myunggi Yi;Byeong-il Lee
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.563-565
    • /
    • 2023
  • Acupuncture and acupressure apply needles or pressure to anatomical points for therapeutic benefit. The over 350 mapped acupuncture points in the human body can each treat various conditions, but anatomical variations make precisely locating these acupoints difficult. We propose a computer vision technique using the real-time hand and face tracking capabilities of the MediaPipe framework to identify acupoint locations. Our model detects anatomical facial and hand landmarks, and then maps these to corresponding acupoint regions. In summary, our proposed model facilitates precise acupoint localization for self-treatment and enhances practitioners' abilities to deliver targeted acupuncture and acupressure therapies.

Reflectance estimation for infrared and visible image fusion

  • Gu, Yan;Yang, Feng;Zhao, Weijun;Guo, Yiliang;Min, Chaobo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.8
    • /
    • pp.2749-2763
    • /
    • 2021
  • The desirable result of infrared (IR) and visible (VIS) image fusion should have textural details from VIS images and salient targets from IR images. However, detail information in the dark regions of VIS image has low contrast and blurry edges, resulting in performance degradation in image fusion. To resolve the troubles of fuzzy details in dark regions of VIS image fusion, we have proposed a method of reflectance estimation for IR and VIS image fusion. In order to maintain and enhance details in these dark regions, dark region approximation (DRA) is proposed to optimize the Retinex model. With the improved Retinex model based on DRA, quasi-Newton method is adopted to estimate the reflectance of a VIS image. The final fusion outcome is obtained by fusing the DRA-based reflectance of VIS image with IR image. Our method could simultaneously retain the low visibility details in VIS images and the high contrast targets in IR images. Experiment statistic shows that compared to some advanced approaches, the proposed method has superiority on detail preservation and visual quality.

Particle Filters using Gaussian Mixture Models for Vision-Based Navigation (영상 기반 항법을 위한 가우시안 혼합 모델 기반 파티클 필터)

  • Hong, Kyungwoo;Kim, Sungjoong;Bang, Hyochoong;Kim, Jin-Won;Seo, Ilwon;Pak, Chang-Ho
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.47 no.4
    • /
    • pp.274-282
    • /
    • 2019
  • Vision-based navigation of unmaned aerial vehicle is a significant technology that can reinforce the vulnerability of the widely used GPS/INS integrated navigation system. However, the existing image matching algorithms are not suitable for matching the aerial image with the database. For the reason, this paper proposes particle filters using Gaussian mixture models to deal with matching between aerial image and database for vision-based navigation. The particle filters estimate the position of the aircraft by comparing the correspondences of aerial image and database under the assumption of Gaussian mixture model. Finally, Monte Carlo simulation is presented to demonstrate performance of the proposed method.

Vision and Lidar Sensor Fusion for VRU Classification and Tracking in the Urban Environment (카메라-라이다 센서 융합을 통한 VRU 분류 및 추적 알고리즘 개발)

  • Kim, Yujin;Lee, Hojun;Yi, Kyongsu
    • Journal of Auto-vehicle Safety Association
    • /
    • v.13 no.4
    • /
    • pp.7-13
    • /
    • 2021
  • This paper presents an vulnerable road user (VRU) classification and tracking algorithm using vision and LiDAR sensor fusion method for urban autonomous driving. The classification and tracking for vulnerable road users such as pedestrian, bicycle, and motorcycle are essential for autonomous driving in complex urban environments. In this paper, a real-time object image detection algorithm called Yolo and object tracking algorithm from LiDAR point cloud are fused in the high level. The proposed algorithm consists of four parts. First, the object bounding boxes on the pixel coordinate, which is obtained from YOLO, are transformed into the local coordinate of subject vehicle using the homography matrix. Second, a LiDAR point cloud is clustered based on Euclidean distance and the clusters are associated using GNN. In addition, the states of clusters including position, heading angle, velocity and acceleration information are estimated using geometric model free approach (GMFA) in real-time. Finally, the each LiDAR track is matched with a vision track using angle information of transformed vision track and assigned a classification id. The proposed fusion algorithm is evaluated via real vehicle test in the urban environment.