• Title/Summary/Keyword: object coordinates

Search Result 299, Processing Time 0.022 seconds

A New Object Region Detection and Classification Method using Multiple Sensors on the Driving Environment (다중 센서를 사용한 주행 환경에서의 객체 검출 및 분류 방법)

  • Kim, Jung-Un;Kang, Hang-Bong
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.8
    • /
    • pp.1271-1281
    • /
    • 2017
  • It is essential to collect and analyze target information around the vehicle for autonomous driving of the vehicle. Based on the analysis, environmental information such as location and direction should be analyzed in real time to control the vehicle. In particular, obstruction or cutting of objects in the image must be handled to provide accurate information about the vehicle environment and to facilitate safe operation. In this paper, we propose a method to simultaneously generate 2D and 3D bounding box proposals using LiDAR Edge generated by filtering LiDAR sensor information. We classify the classes of each proposal by connecting them with Region-based Fully-Covolutional Networks (R-FCN), which is an object classifier based on Deep Learning, which uses two-dimensional images as inputs. Each 3D box is rearranged by using the class label and the subcategory information of each class to finally complete the 3D bounding box corresponding to the object. Because 3D bounding boxes are created in 3D space, object information such as space coordinates and object size can be obtained at once, and 2D bounding boxes associated with 3D boxes do not have problems such as occlusion.

Position Detection and Gathering Swimming Control of Fish Robot Using Color Detection Algorithm (색상 검출 알고리즘을 활용한 물고기로봇의 위치인식과 군집 유영제어)

  • Akbar, Muhammad;Shin, Kyoo Jae
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2016.10a
    • /
    • pp.510-513
    • /
    • 2016
  • Detecting of the object in image processing is substantial but it depends on the object itself and the environment. An object can be detected either by its shape or color. Color is an essential for pattern recognition and computer vision. It is an attractive feature because of its simplicity and its robustness to scale changes and to detect the positions of the object. Generally, color of an object depends on its characteristics of the perceiving eye and brain. Physically, objects can be said to have color because of the light leaving their surfaces. Here, we conducted experiment in the aquarium fish tank. Different color of fish robots are mimic the natural swim of fish. Unfortunately, in the underwater medium, the colors are modified by attenuation and difficult to identify the color for moving objects. We consider the fish motion as a moving object and coordinates are found at every instinct of the aquarium to detect the position of the fish robot using OpenCV color detection. In this paper, we proposed to identify the position of the fish robot by their color and use the position data to control the fish robot gathering in one point in the fish tank through serial communication using RF module. It was verified by the performance test of detecting the position of the fish robot.

Development of a Remote Object's 3D Position Measuring System (원격지 물체의 삼차원 위치 측정시스템의 개발)

  • Park, Kang
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.17 no.8
    • /
    • pp.60-70
    • /
    • 2000
  • In this paper a 3D position measuring device that finds the 3D position of an arbitarily placed object using a camersa system is introduced. The camera system consists of three stepping motors and a CCD camera and a laser. The viewing direction of the camera is controlled by two stepping motors (pan and tilt motors) and the direction of a laser is also controlled by a stepping motors(laser motor). If an object in a remote place is selected from a live video image the x,y,z coordinates of the object with respect to the reference coordinate system can be obtained by calculating the distance from the camera to the object using a structured light scheme and by obtaining the orientation of the camera that is controlled by two stepping motors. The angles o f stepping motors are controlled by a SGI O2 workstation through a parallel port. The mathematical model of the camera and the distance measuring system are calibrated to calculate an accurate position of the object. This 3D position measuring device can be used to acquire information that is necessary to monitor a remote place.

  • PDF

Coordinate Determination for Texture Mapping using Camera Calibration Method (카메라 보정을 이용한 텍스쳐 좌표 결정에 관한 연구)

  • Jeong K. W.;Lee Y.Y.;Ha S.;Park S.H.;Kim J. J.
    • Korean Journal of Computational Design and Engineering
    • /
    • v.9 no.4
    • /
    • pp.397-405
    • /
    • 2004
  • Texture mapping is the process of covering 3D models with texture images in order to increase the visual realism of the models. For proper mapping the coordinates of texture images need to coincide with those of the 3D models. When projective images from the camera are used as texture images, the texture image coordinates are defined by a camera calibration method. The texture image coordinates are determined by the relation between the coordinate systems of the camera image and the 3D object. With the projective camera images, the distortion effect caused by the camera lenses should be compensated in order to get accurate texture coordinates. The distortion effect problem has been dealt with iterative methods, where the camera calibration coefficients are computed first without considering the distortion effect and then modified properly. The methods not only cause to change the position of the camera perspective line in the image plane, but also require more control points. In this paper, a new iterative method is suggested for reducing the error by fixing the principal points in the image plane. The method considers the image distortion effect independently and fixes the values of correction coefficients, with which the distortion coefficients can be computed with fewer control points. It is shown that the camera distortion effects are compensated with fewer numbers of control points than the previous methods and the projective texture mapping results in more realistic image.

A Study on the Use of Photogrammetry in Survey and Recording of Modern Architectural Properties (근대건축문화재 기록에서의 사진실측기법 활용)

  • Kwon, Soon-Chan
    • Journal of the Korean Institute of Rural Architecture
    • /
    • v.15 no.2
    • /
    • pp.9-16
    • /
    • 2013
  • This paper is to discover how the photogrammetry method can be utilized for recording modern architectural properties. As a result of conducting photogrammetry aimed at the main building and water supply tower of Goesan Middle School, it was possible to measure actually either complicated patterns or a damaged area on the facade. However, there appeared errors in the actual measurement of the hollow or protruded area on the facade to measure actually or of an object in the cylinder form. First, for the actual measurement of the protruded area, it is possible to complete a facade by measuring actually the protruded area separately after dividing it and complementing it on the facade, Second, concerning an object in the cylinder form, it is possible to complete one facade by matching a facade which is measured actually with a Diaver method among the MSR photogrammetry methods, and three dimensional standard coordinates which are obtained through Totalstation.

PROGRESSIVE ALGORITHM FOR RECONSTRUCTING A 3D STRUCTURE FROM A 2D SKETCH DRAWING

  • Oh, Beom-Soo;Kim, Chang-Hun
    • Proceedings of the Korea Society for Simulation Conference
    • /
    • 2001.10a
    • /
    • pp.248-254
    • /
    • 2001
  • This paper presents a progressive algorithm for reconstructing a 3D structure from a given 2D sketch drawing (edge-vertex graph without hidden line removal) according to the user's sketch order. While previous methods reconstruct a 3D structure at once, the proposed method progressively calculate a 3D structure by optimizing the coordinates of vertices of an object according to the sketch order. The progressive method reconstructs the most plausible 3D object quickly by applying 3D constraints that are derived from the relationship between the object and the sketch drawing in the optimization process. The progressive reconstruction algorithm is discussed, and examples from a working implementation are given.

  • PDF

3-Dimensional Measurement of the Cylindrical Object Using Computer Vision (컴퓨터 비젼을 이용한 원기둥형 물체의 3차원 측정)

  • Jang, TaekJun;Joo, Kisee;Han, Min-Hong
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.12 no.12
    • /
    • pp.38-44
    • /
    • 1995
  • This paper presents a method to measure the position and orientation of a Cylindrical Object(unknown the eiameter and length) lying on a floor, using a camera. The two extreme cross section of the cylinder will be viewed as distorted ellipese or circular are, while its limb edge will be shown as two straight lines. The diameter of the cylinder is determined from the geometric properties of the two straight lines, which in turn provides information regarding the length of the cylinder. From the 3-dimensional measurement, the 3D coordinates of the center points of the two extreme cross sections are determined to give the position and orientation of the cylinder. This method is used for automated pick-and-place operations of cylinder, such as sheet coils, or drums in warehouses.

  • PDF

Constructing 3D Outlines of Objects based on Feature Points using Monocular Camera (단일카메라를 사용한 특징점 기반 물체 3차원 윤곽선 구성)

  • Park, Sang-Heon;Lee, Jeong-Oog;Baik, Doo-Kwon
    • The KIPS Transactions:PartB
    • /
    • v.17B no.6
    • /
    • pp.429-436
    • /
    • 2010
  • This paper presents a method to extract 3D outlines of objects in an image obtained from a monocular vision. After detecting the general outlines of the object by MOPS(Multi-Scale Oriented Patches) -algorithm and we obtain their spatial coordinates. Simultaneously, it obtains the space-coordinates with feature points to be immanent within the outlines of objects through SIFT(Scale Invariant Feature Transform)-algorithm. It grasps a form of objects to join the space-coordinates of outlines and SIFT feature points. The method which is proposed in this paper, it forms general outlines of objects, so that it enables a rapid calculation, and also it has the advantage capable of collecting a detailed data because it supplies the internal-data of outlines through SIFT feature points.

A Study on the 3-D Information Abstraction of object using Triangulation System (물체의 3-D 형상 복원을 위한 삼각측량 시스템)

  • Kim, Kuk-Se;Lee, Jeong-Ki;Cho, Ai-Ri;Ba, Il-Ho;Lee, Joon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2003.05a
    • /
    • pp.409-412
    • /
    • 2003
  • The 3-D shape use to effect of movie, animation, industrial design, medical treatment service, education, engineering etc... But it is not easy to make 3-D shape from the information of 2-D image. There are two methods in restoring 3-D video image through 2-D image; First the method of using a laser; Second, the method of acquiring 3-D image through stereo vision. Instead of doing two methods with many difficulties, I study the method of simple 3-D image in this research paper. We present here a simple and efficient method, called direct calibration, which does not require any equations at all. The direct calibration procedure builds a lookup table(LUT) linking image and 3-D coordinates by a real 3-D triangulation system. The LUT is built by measuring the image coordinates of a grid of known 3-D points, and recording both image and world coordinates for each point; the depth values of all other visible points are obtained by interpolation.

  • PDF

Real-time Rebar Injection Endpoints Tracking Method to Improve the Straightness of Rebars (철근 직진도 개선을 위한 실시간 철근 사출 끝점 추적 방법)

  • Kim, Jong-Sik;Kang, Dae-Seong
    • The Journal of Korean Institute of Information Technology
    • /
    • v.17 no.8
    • /
    • pp.75-83
    • /
    • 2019
  • In this paper, we propose a method that can detect and trace the end point of real - time reinforcement steel to various environmental conditions of industrial field by using Median flow and Depth information. We proposed a method to derive two steel end points by using Median filter, Binarization, Morphology, and Blob algorithm on image depth information. The coordinates of the final position were determined by comparing the coordinates of the reinforcement steel endpoints detected in the Depth image and the position tracking coordinates of the reinforcement steel using Median Flow. As a result, when the existing Median Flow method was used, the success rate of the final position determination of reinforcement steel of 75% was increased to 95% when the Depth of reinforcement steel was used.