• Title/Summary/Keyword: location of camera

Search Result 589, Processing Time 0.03 seconds

The Design and Implementation Navigation System For Visually Impaired Person (시각 장애인을 위한 Navigation System의 설계 및 구현)

  • Kong, Sung-Hun;Kim, Young-Kil
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.12
    • /
    • pp.2702-2707
    • /
    • 2012
  • In the rapid growth of cities, road has heavy traffic and many buildings are under constructions. These kinds of environments make more difficulty for a person who is visually handicapped to walk comfortable. To alleviate the problem, we introduce Navigation System to help walking for Visually Impaired Person. It follows, service center give instant real time monitoring to visually impaired person for their convenient by this system. This Navigation System has GPS, Camera, Audio and Wi-Fi(wireless fidelity) available. It means that GPS location and Camera image information can be sent to service center by Wi-Fi network. To be specific, transmitted GPS location information enables service center to figure out the visually impaired person's whereabouts and mark the location on the map. By delivered Camera image information, service center monitors the visually impaired person's view. Also, they can offer live guidance to visually impaired person by equipped Audio with live talking. To sum up, Android based Portable Navigation System is a specialized navigation system that gives practical effect to realize more comfortable walking for visually impaired person.

A Study on Estimating Skill of Smartphone Camera Position using Essential Matrix (필수 행렬을 이용한 카메라 이동 위치 추정 기술 연구)

  • Oh, Jongtaek;Kim, Hogyeom
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.6
    • /
    • pp.143-148
    • /
    • 2022
  • It is very important for metaverse, mobile robot, and user location services to analyze the images continuously taken using a mobile smartphone or robot's monocular camera to estimate the camera's location. So far, PnP-related techniques have been applied to calculate the position. In this paper, the camera's moving direction is obtained using the essential matrix in the epipolar geometry applied to successive images, and the camera's continuous moving position is calculated through geometrical equations. A new estimation method was proposed, and its accuracy was verified through simulation. This method is completely different from the existing method and has a feature that it can be applied even if there is only one or more matching feature points in two or more images.

Multi-camera-based 3D Human Pose Estimation for Close-Proximity Human-robot Collaboration in Construction

  • Sarkar, Sajib;Jang, Youjin;Jeong, Inbae
    • International conference on construction engineering and project management
    • /
    • 2022.06a
    • /
    • pp.328-335
    • /
    • 2022
  • With the advance of robot capabilities and functionalities, construction robots assisting construction workers have been increasingly deployed on construction sites to improve safety, efficiency and productivity. For close-proximity human-robot collaboration in construction sites, robots need to be aware of the context, especially construction worker's behavior, in real-time to avoid collision with workers. To recognize human behavior, most previous studies obtained 3D human poses using a single camera or an RGB-depth (RGB-D) camera. However, single-camera detection has limitations such as occlusions, detection failure, and sensor malfunction, and an RGB-D camera may suffer from interference from lighting conditions and surface material. To address these issues, this study proposes a novel method of 3D human pose estimation by extracting 2D location of each joint from multiple images captured at the same time from different viewpoints, fusing each joint's 2D locations, and estimating the 3D joint location. For higher accuracy, the probabilistic representation is used to extract the 2D location of the joints, considering each joint location extracted from images as a noisy partial observation. Then, this study estimates the 3D human pose by fusing the probabilistic 2D joint locations to maximize the likelihood. The proposed method was evaluated in both simulation and laboratory settings, and the results demonstrated the accuracy of estimation and the feasibility in practice. This study contributes to ensuring human safety in close-proximity human-robot collaboration by providing a novel method of 3D human pose estimation.

  • PDF

A Study on the HMD-AR Interaction System Combining Optical Camera Communication to Provide Location-based Service for Tourist in Jeonju Hanok Village (전주 한옥마을 관광객의 위치 기반 서비스 제공을 위한 광카메라통신 접목형 HMD-AR 인터렉션 시스템에 관한 연구)

  • Min, Byung-Jun;Choi, Jin-Yeong;Cha, Jae-Sang;Choi, Bang-Ho;Cho, Ju-Phil
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.11 no.4
    • /
    • pp.445-451
    • /
    • 2018
  • In this paper, we propose an HMD-AR interaction system that combines optical camera communication to provide location-based service for tourists in Jeonju Hanok Village. The proposed system receives optical camera communication data from Light infrastructure existing in Jeonju Hanok Village and provides service through ID information along with its own location information. We researched optical camera communication technology and smart device based HMD-AR system and constructed the actual HMD-AR system and tested it. The proposed system is expected that it will be applied to various tourist attractions by utilizing the proposed system in the future, and it is expected to be used as a valuable feedback for smart device based HMD-AR.

Malfunction Detection of High Voltage Equipment Using Microphone Array and Infrared Thermal Imaging Camera (Microphone Array와 열화상 카메라를 이용한 고압설비 고장검출)

  • Han, Sun-Sin;Choi, Jae-Young;Lee, Jang-Myung
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.1
    • /
    • pp.25-32
    • /
    • 2010
  • The paper proposes a hierarchical fault detection method for the high voltage equipment using a microphone array which detects the location of fault and the thermal imaging and CCD cameras which verifies the fault and stores the image, respectively. There are partial arc discharges on the faulty insulators, which generates a specific pattern of sound. Detecting the signal using the microphone array, the location of the faulty insulator can be estimated. The 6th band-pass filter was applied to remove noise signal from wind or external influence. When the mobile robot carries the thermal and CCD cameras to the possible place of the fault insulator, the fault insulators or power transmission wires can be detected by the thermal images, which are caused by the aging or natural erosion. Finally, the CCD camera captures the image of the fault insulator for the record. The detection scheme of fault location using the microphone array and the thermal images have been proved to be effective through the real experiments. As a result of this research, it becomes possible to use a mobile robot with the integrated sensors to detect the fault insulators instead of a human being.

Compressed Sensing-based Multiple-target Tracking Algorithm for Ad Hoc Camera Sensor Networks

  • Lu, Xu;Cheng, Lianglun;Liu, Jun;Chen, Rongjun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.3
    • /
    • pp.1287-1300
    • /
    • 2018
  • Target-tracking algorithm based on ad hoc camera sensor networks (ACSNs) utilizes the distributed observation capability of nodes to achieve accurate target tracking. A compressed sensing-based multiple-target tracking algorithm (CSMTTA) for ACSNs is proposed in this work based on the study of camera node observation projection model and compressed sensing model. The proposed algorithm includes reconfiguration of observed signals and evaluation of target locations. It reconfigures observed signals by solving the convex optimization of L1-norm least and forecasts node group to evaluate a target location by the motion features of the target. Simulation results show that CSMTTA can recover the subtracted observation information accurately under the condition of sparse sampling to a high target-tracking accuracy and accomplish the distributed tracking task of multiple mobile targets.

Localization System for Mobile Robot Using Electric Compass and Tracking IR Light Source (전자 나침반과 적외선 광원 추적을 이용한 이동로봇용 위치 인식 시스템)

  • Son, Chang-Woo;Lee, Seung-Heui;Lee, Min-Cheol
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.14 no.8
    • /
    • pp.767-773
    • /
    • 2008
  • This paper presents a localization system based on the use of electric compass and tracking IR light source. Digital RGB(Red, Green, Blue)signal of digital CMOS Camera is sent to CPLD which converts the color image to binary image at 30 frames per second. CMOS camera has IR filter and UV filter in front of CMOS cell. The filters cut off above 720nm light source. Binary output data of CPLD is sent to DSP that rapidly tracks the IR light source by moving Camera tilt DC motor. At a robot toward north, electric compass signals and IR light source angles which are used for calculating the data of the location system. Because geomagnetic field is linear in local position, this location system is possible. Finally, it is shown that position error is within ${\pm}1.3cm$ in this system.

Virtual portraits from rotating selfies

  • Yongsik Lee;Jinhyuk Jang;SeungjoonYang
    • ETRI Journal
    • /
    • v.45 no.2
    • /
    • pp.291-303
    • /
    • 2023
  • Selfies are a popular form of photography. However, due to physical constraints, the compositions of selfies are limited. We present algorithms for creating virtual portraits with interesting compositions from a set of selfies. The selfies are taken at the same location while the user spins around. The scene is analyzed using multiple selfies to determine the locations of the camera, subject, and background. Then, a view from a virtual camera is synthesized. We present two use cases. After rearranging the distances between the camera, subject, and background, we render a virtual view from a camera with a longer focal length. Following that, changes in perspective and lens characteristics caused by new compositions and focal lengths are simulated. Second, a virtual panoramic view with a larger field of view is rendered, with the user's image placed in a preferred location. In our experiments, virtual portraits with a wide range of focal lengths were obtained using a device equipped with a lens that has only one focal length. The rendered portraits included compositions that would be photographed with actual lenses. Our proposed algorithms can provide new use cases in which selfie compositions are not limited by a camera's focal length or distance from the camera.

Human-Computer Interaction Based Only on Auditory and Visual Information

  • Sha, Hui;Agah, Arvin
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.2 no.4
    • /
    • pp.285-297
    • /
    • 2000
  • One of the research objectives in the area of multimedia human-computer interaction is the application of artificial intelligence and robotics technologies to the development of computer interfaces. This involves utilizing many forms of media, integrating speed input, natural language, graphics, hand pointing gestures, and other methods for interactive dialogues. Although current human-computer communication methods include computer keyboards, mice, and other traditional devices, the two basic ways by which people communicate with each other are voice and gesture. This paper reports on research focusing on the development of an intelligent multimedia interface system modeled based on the manner in which people communicate. This work explores the interaction between humans and computers based only on the processing of speech(Work uttered by the person) and processing of images(hand pointing gestures). The purpose of the interface is to control a pan/tilt camera to point it to a location specified by the user through utterance of words and pointing of the hand, The systems utilizes another stationary camera to capture images of the users hand and a microphone to capture the users words. Upon processing of the images and sounds, the systems responds by pointing the camera. Initially, the interface uses hand pointing to locate the general position which user is referring to and then the interface uses voice command provided by user to fine-the location, and change the zooming of the camera, if requested. The image of the location is captured by the pan/tilt camera and sent to a color TV monitor to be displayed. This type of system has applications in tele-conferencing and other rmote operations, where the system must respond to users command, in a manner similar to how the user would communicate with another person. The advantage of this approach is the elimination of the traditional input devices that the user must utilize in order to control a pan/tillt camera, replacing them with more "natural" means of interaction. A number of experiments were performed to evaluate the interface system with respect to its accuracy, efficiency, reliability, and limitation.

  • PDF