• Title/Summary/Keyword: Human Tracking

Search Result 654, Processing Time 0.027 seconds

Smart composite repetitive-control design for nonlinear perturbation

  • ZY Chen;Ruei-Yuan Wang;Yahui Meng;Timothy Chen
    • Steel and Composite Structures
    • /
    • v.51 no.5
    • /
    • pp.473-485
    • /
    • 2024
  • This paper proposes a composite form of fuzzy adaptive control plan based on a robust observer. The fuzzy 2D control gains are regulated by the parameters in the LMIs. Then, control and learning performance indices with weight matrices are constructed as the cost functions, which allows the regulation of the trade-off between the two performance by setting appropriate weight matrices. The design of 2D control gains is equivalent to the LMIs-constrained multi-objective optimization problem under dual performance indices. By using this proposed smart tracking design via fuzzy nonlinear criterion, the data link can be further extended. To evaluate the performance of the controller, the proposed controller was compared with other control technologies. This ensures the execution of the control program used to track position and trajectory in the presence of great model uncertainty and external disturbances. The performance of monitoring and control is verified by quantitative analysis. The goals of this paper are towards access to adequate, safe and affordable housing and basic services, promotion of inclusive and sustainable urbanization and participation, implementation of sustainable and disaster-resilient buildings, sustainable human settlement planning and manage. Therefore, the goal is believed to achieved in the near future by the ongoing development of AI and control theory.

Vision-based Motion Control for the Immersive Interaction with a Mobile Augmented Reality Object (모바일 증강현실 물체와 몰입형 상호작용을 위한 비전기반 동작제어)

  • Chun, Jun-Chul
    • Journal of Internet Computing and Services
    • /
    • v.12 no.3
    • /
    • pp.119-129
    • /
    • 2011
  • Vision-based Human computer interaction is an emerging field of science and industry to provide natural way to communicate with human and computer. Especially, recent increasing demands for mobile augmented reality require the development of efficient interactive technologies between the augmented virtual object and users. This paper presents a novel approach to construct marker-less mobile augmented reality object and control the object. Replacing a traditional market, the human hand interface is used for marker-less mobile augmented reality system. In order to implement the marker-less mobile augmented system in the limited resources of mobile device compared with the desktop environments, we proposed a method to extract an optimal hand region which plays a role of the marker and augment object in a realtime fashion by using the camera attached on mobile device. The optimal hand region detection can be composed of detecting hand region with YCbCr skin color model and extracting the optimal rectangle region with Rotating Calipers Algorithm. The extracted optimal rectangle region takes a role of traditional marker. The proposed method resolved the problem of missing the track of fingertips when the hand is rotated or occluded in the hand marker system. From the experiment, we can prove that the proposed framework can effectively construct and control the augmented virtual object in the mobile environments.

Interactive Motion Retargeting for Humanoid in Constrained Environment (제한된 환경 속에서 휴머노이드를 위한 인터랙티브 모션 리타겟팅)

  • Nam, Ha Jong;Lee, Ji Hye;Choi, Myung Geol
    • Journal of the Korea Computer Graphics Society
    • /
    • v.23 no.3
    • /
    • pp.1-8
    • /
    • 2017
  • In this paper, we introduce a technique to retarget human motion data to the humanoid body in a constrained environment. We assume that the given motion data includes detailed interactions such as holding the object by hand or avoiding obstacles. In addition, we assume that the humanoid joint structure is different from the human joint structure, and the shape of the surrounding environment is different from that at the time of the original motion. Under such a condition, it is also difficult to preserve the context of the interaction shown in the original motion data, if the retargeting technique that considers only the change of the body shape. Our approach is to separate the problem into two smaller problems and solve them independently. One is to retarget motion data to a new skeleton, and the other is to preserve the context of interactions. We first retarget the given human motion data to the target humanoid body ignoring the interaction with the environment. Then, we precisely deform the shape of the environmental model to match with the humanoid motion so that the original interaction is reproduced. Finally, we set spatial constraints between the humanoid body and the environmental model, and restore the environmental model to the original shape. To demonstrate the usefulness of our method, we conducted an experiment by using the Boston Dynamic's Atlas robot. We expected that out method can help the humanoid motion tracking problem in the future.

Object Detection Algorithm Using Edge Information on the Sea Environment (해양 환경에서 에지 정보를 이용한 물표 추출 알고리즘)

  • Jeong, Jong-Myeon;Park, Gyei-Kark
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.9
    • /
    • pp.69-76
    • /
    • 2011
  • According to the related reports, about 60 percents of ship collisions have resulted from operating mistake caused by human factor. Specially, the report said that negligence of observation caused 66.8 percents of the accidents due to a human factor. Hence automatic detection and tracking of an object from an IR images are crucial for safety navigation because it can relieve officer's burden and remedies imperfections of human visual system. In this paper, we present a method to detect an object such as ship, rock and buoy from a sea IR image. Most edge directions of the sea image are horizontal and most vertical edges come out from the object areas. The presented method uses them as a characteristic for the object detection. Vertical edges are extracted from the input image and isolated edges are eliminated. Then morphological closing operation is performed on the vertical edges. This caused vertical edges that actually compose an object be connected and become an object candidate region. Next, reference object regions are extracted using horizontal edges, which appear on the boundaries between surface of the sea and the objects. Finally, object regions are acquired by sequentially integrating reference region and object candidate regions.

Augmented Reality Game Interface Using Hand Gestures Tracking (사용자 손동작 추적에 기반한 증강현실 게임 인터페이스)

  • Yoon, Jong-Hyun;Park, Jong-Seung
    • Journal of Korea Game Society
    • /
    • v.6 no.2
    • /
    • pp.3-12
    • /
    • 2006
  • Recently, Many 3D augmented reality games that provide strengthened immersive have appeared in the 3D game environment. In this article, we describe a barehanded interaction method based on human hand gestures for augmented reality games. First, feature points are extracted from input video streams. Point features are tracked and motion of moving objects are computed. The shape of the motion trajectories are used to determine whether the motion is intended gestures. A long smooth trajectory toward one of virtual objects or menus is classified as an intended gesture and the corresponding action is invoked. To prove the validity of the proposed method, we implemented two simple augmented reality applications: a gesture-based music player and a virtual basketball game. In the music player, several menu icons are displayed on the top of the screen and an user can activate a menu by hand gestures. In the virtual basketball game, a virtual ball is bouncing in a virtual cube space and the real video stream is shown in the background. An user can hit the virtual ball with his hand gestures. From the experiments for three untrained users, it is shown that the accuracy of menu activation according to the intended gestures is 94% for normal speed gestures and 84% for fast and abrupt gestures.

  • PDF

A Study on the Characteristics of Consumer Visual-Perceptional Information Acquisition in Commercial Facilities in Regard to its Construction of Space from Real-Time Eye Gaze Tracking (상업시설 공간구성의 실시간 시선추적에 나타난 소비자 시지각 정보획득 특성 연구)

  • Park, Sunmyung
    • Science of Emotion and Sensibility
    • /
    • v.21 no.2
    • /
    • pp.3-14
    • /
    • 2018
  • For satisfying consumer needs, commercial facilities require a variety of sale-related space expressions and eye-catching product arrangements; space composition can also be a direct marketing strategy. The human eye is the sensory organ that acquires the largest amount of information, and an analysis of visual information helps in understanding visual relations between . However, the existing studies are mostly focused on analysis of still frames in experimental images, and there is a lack of studies analyzing gaze information based on mobile images of commercial spaces. Therefore, this study analyzed emotional responses through gaze information of space users in reality using a video of a movement route through a commercial facility. The analysis targeted straight sections of the moving route; based on the data acquired, sectional characteristics of five gaze intensity ranges were examined. As a result, section A, the starting point of the route, had a low gaze intensity, while section B had the highest gaze intensity. This indicates that, starting in section B, the subjects explored the space in a stable way and needed time to adapt to the experimental video. In relation to space characteristics of the gaze-concentrated area, display formats of the right stores in 4 of 6 sections received greater attention. The gaze of consumers was mostly focused on props, and big gaze information was revealed in showcase display formats of the stores. In conclusion, this analysis method can provide highly useful direct design data about merchandise display and merchandise component arrangement based on consumer visual preference.

Hand Gesture Recognition Algorithm Robust to Complex Image (복잡한 영상에 강인한 손동작 인식 방법)

  • Park, Sang-Yun;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.7
    • /
    • pp.1000-1015
    • /
    • 2010
  • In this paper, we propose a novel algorithm for hand gesture recognition. The hand detection method is based on human skin color, and we use the boundary energy information to locate the hand region accurately, then the moment method will be employed to locate the hand palm center. Hand gesture recognition can be separated into 2 step: firstly, the hand posture recognition: we employ the parallel NNs to deal with problem of hand posture recognition, pattern of a hand posture can be extracted by utilize the fitting ellipses method, which separates the detected hand region by 12 ellipses and calculates the white pixels rate in ellipse line. the pattern will be input to the NNs with 12 input nodes, the NNs contains 4 output nodes, each output node out a value within 0~1, the posture is then represented by composed of the 4 output codes. Secondly, the hand gesture tracking and recognition: we employed the Kalman filter to predict the position information of gesture to create the position sequence, distance relationship between positions will be used to confirm the gesture. The simulation have been performed on Windows XP to evaluate the efficiency of the algorithm, for recognizing the hand posture, we used 300 training images to train the recognizing machine and used 200 images to test the machine, the correct number is up to 194. And for testing the hand tracking recognition part, we make 1200 times gesture (each gesture 400 times), the total correct number is 1002 times. These results shows that the proposed gesture recognition algorithm can achieve an endurable job for detecting the hand and its' gesture.

Design and Performance Analysis of a new MAC Protocol for Providing Real-time Traffic Information using USN (USN 기반 실시간 주행 상황 정보 제공을 위한 MAC 설계 및 성능 분석)

  • Park, Man-Kyu;So, Sang-Ho;Lee, Jae-Yong;Lim, Jae-Han;Son, Myung-Hee;Kim, Byung-Chul
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.44 no.5
    • /
    • pp.38-48
    • /
    • 2007
  • In ubiquitous environment, sensor networks that sense and transmit surrounding data without human intervention will become more important. If sensors are installed for detecting vehicles and measuring their speed in the road and that real-time information is given to drivers, it will be very effective for enhancing safety and controlling traffic in the road. In this paper, we proposed a new reliable and real-time sensor MAC protocol between AP and sensor nodes in order to provide real-time traffic flow information based on ubiquitous sensor networks. The proposed MAC allocates one TDMA slot for each sensor node on the IEEE 802.15.4 based channel structure, introduces relayed communication for distant sensors, and adopts a frame structure that supports retransmission for the case of errors. In addition, the proposed MAC synchronizes with AP by using beacon and adopts a hybrid tracking mode that supports economic power consumption according to various traffic situations, We implemented a simulator for the proposed MAC by using sim++ and evaluated various performances. The simulation results show that the proposed MAC reduces the power consumption and reveals excellent performance in real-time application systems.

A Study on Manipulating Method of 3D Game in HMD Environment by using Eye Tracking (HMD(Head Mounted Display)에서 시선 추적을 통한 3차원 게임 조작 방법 연구)

  • Park, Kang-Ryoung;Lee, Eui-Chul
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.2
    • /
    • pp.49-64
    • /
    • 2008
  • Recently, many researches about making more comfortable input device based on gaze detection technology have been done in human computer interface. However, the system cost becomes high due to the complicated hardware and there is difficulty to use the gaze detection system due to the complicated user calibration procedure. In this paper, we propose a new gaze detection method based on the 2D analysis and a simple user calibration. Our method used a small USB (Universal Serial Bus) camera attached on a HMD (Head-Mounted Display), hot-mirror and IR (Infra-Red) light illuminator. Because the HMD is moved according to user's facial movement, we can implement the gaze detection system of which performance is not affected by facial movement. In addition, we apply our gaze detection system to 3D first person shooting game. From that, the gaze direction of game character is controlled by our gaze detection method and it can target the enemy character and shoot, which can increase the immersion and interest of game. Experimental results showed that the game and gaze detection system could be operated at real-time speed in one desktop computer and we could obtain the gaze detection accuracy of 0.88 degrees. In addition, we could know our gaze detection technology could replace the conventional mouse in the 3D first person shooting game.

Development of CCTV Cooperation Tracking System for Real-Time Crime Monitoring (실시간 범죄 모니터링을 위한 CCTV 협업 추적시스템 개발 연구)

  • Choi, Woo-Chul;Na, Joon-Yeop
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.20 no.12
    • /
    • pp.546-554
    • /
    • 2019
  • Typically, closed-circuit television (CCTV) monitoring is mainly used for post-processes (i.e. to provide evidence after an incident has occurred), but by using a streaming video feed, machine-based learning, and advanced image recognition techniques, current technology can be extended to respond to crimes or reports of missing persons in real time. The multi-CCTV cooperation technique developed in this study is a program model that delivers similarity information about a suspect (or moving object) extracted via CCTV at one location and sent to a monitoring agent to track the selected suspect or object when he, she, or it moves out of range to another CCTV camera. To improve the operating efficiency of local government CCTV control centers, we describe here the partial automation of a CCTV control system that currently relies upon monitoring by human agents. We envisage an integrated crime prevention service, which incorporates the cooperative CCTV network suggested in this study and that can easily be experienced by citizens in ways such as determining a precise individual location in real time and providing a crime prevention service linked to smartphones and/or crime prevention/safety information.