• Title/Summary/Keyword: pose estimation

Search Result 388, Processing Time 0.041 seconds

Violence Detection System in Streaming Service and SNS Using Artificial Intelligence Technologies (인공지능을 활용한 스트리밍 서비스/SNS 내에서의 폭력 감지 시스템)

  • Kim, Seon-Min;Lee, Seok-Won;Lim, Seung-Su;Choi, Sangil
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2020.05a
    • /
    • pp.442-445
    • /
    • 2020
  • 인터넷 및 IT 기술의 발전과 더불어 미디어산업에도 큰 변화가 일어나고 있다. TV 를 대신하여 스트리밍 서비스를 이용하는 사람들이 늘고 있으며 SNS 를 활용하여 서로의 경험을 간접적으로 공유하는 형태의 새로운 문화 컨텐츠가 자리잡아가고 있다. 하지만 이러한 컨텐츠를 소비하는 주요 계층 중에는 초중고 학생들도 포함되어 있다. 인터넷 혹은 SNS 에서 소비되는 컨텐츠들을 관리 감독하는 컨트롤 타워가 부족하거나 전무하기 때문에 폭력, 음주, 흡연 등 사회적으로 악영향을 줄 수 있는 영상 또는 사진이 무분별하게 생산되어 청소년들에 의해 소비되고 있으며 더 나아가 이것이 사회적 문제로까지 대두되고 있다. 이러한 문제를 해결하기 위해 인공지능 기술을 활용한 여러 다양한 감시 시스템 개발을 위한 연구가 한창이다. 본 연구에서는 SNS 및 스트리밍 서비스에서 제공되는 영상 및 사진을 Pose Estimation 및 표정 인식 기술을 활용하여 폭력을 자동적으로 감지할 수 있는 폭력 감지 시스템을 개발하는데 그 목적이 있다.

Pose estimation-based 3D model motion control using low-performance devices (저성능 디바이스를 이용한 자세추정 기반 3D 모델 움직임 제어)

  • Jae-Hoon Jang;Yoo-Joo Choi
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.763-765
    • /
    • 2023
  • 본 논문에서는 저성능 컴퓨터나 스마트폰의 카메라를 통해 입력받은 영상을 기반으로 사용자의 포즈를 추정하고, 실시간으로 사용자의 포즈에 따라 3D 모델의 모션이 제어되어 가시화 될 수 있는 클라이어트-서버 구조의 "자세추정 및 3D 모델 모션 제어 시스템"을 제안한다. 제안 시스템은 소켓통신 기반의 클라이언트-서버구조로 구성되어, 서버에서는 실시간 자세 추정을 위한 딥러닝 모델이 수행되고, 저성능 클라이언트에서는 실시간으로 카메라 영상을 획득하여 영상을 서버에 전송하고, 서버로부터 자세 추정 정보를 받아 이를 3D 모델에 반영하고 렌더링 함으로써 사용자와 함께 3D 모델이 같은 동작을 수행하는 증강현실 화면을 생성한다. 고성능을 요구하는 객체 자세 추정 모듈은 서버에서 실행하고, 클라이언트에서는 영상 획득 및 렌더링만을 실행하기 때문에, 모바일 앱에서의 실시간 증강현실을 위한 자세 추정 및 3D 모델 모션 제어가 가능하다. 제안 시스템은 "증강현실 기반 영상 찍기 앱" 에 반영되어 사용자의 움직임을 따라하는 3D 캐릭터들의 영상을 쉽게 생성할 수 있도록 할 수 있다.

Motion-Based User Authentication for Enhanced Metaverse Security (메타버스 보안 강화를 위한 동작 기반 사용자 인증)

  • Seonggyu Park;Gwonsang Ryu
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.34 no.3
    • /
    • pp.493-503
    • /
    • 2024
  • This paper addresses the issue of continuous user authentication within the metaverse environment. Recently, the metaverse plays a vital role in personal interaction, entertainment, education, and business, bringing forth significant security concerns. Particularly, vulnerabilities related to user identity verification have emerged as a major issue. This research proposes a novel method to verify identities by analyzing users' character movements in the metaverse through a pose estimation model. This method uses only video data for authentication, allowing flexibility in limited environments, and investigates how character movements contribute to user identification through various experiments. Furthermore, it explores the potential for extending this approach to other digital platforms. This research is expected to significantly contribute to enhancing security and innovating user identity verification methods in the metaverse environment.

Improved CS-RANSAC Algorithm Using K-Means Clustering (K-Means 클러스터링을 적용한 향상된 CS-RANSAC 알고리즘)

  • Ko, Seunghyun;Yoon, Ui-Nyoung;Alikhanov, Jumabek;Jo, Geun-Sik
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.6 no.6
    • /
    • pp.315-320
    • /
    • 2017
  • Estimating the correct pose of augmented objects on the real camera view efficiently is one of the most important questions in image tracking area. In computer vision, Homography is used for camera pose estimation in augmented reality system with markerless. To estimating Homography, several algorithm like SURF features which extracted from images are used. Based on extracted features, Homography is estimated. For this purpose, RANSAC algorithm is well used to estimate homography and DCS-RANSAC algorithm is researched which apply constraints dynamically based on Constraint Satisfaction Problem to improve performance. In DCS-RANSAC, however, the dataset is based on pattern of feature distribution of images manually, so this algorithm cannot classify the input image, pattern of feature distribution is not recognized in DCS-RANSAC algorithm, which lead to reduce it's performance. To improve this problem, we suggest the KCS-RANSAC algorithm using K-means clustering in CS-RANSAC to cluster the images automatically based on pattern of feature distribution and apply constraints to each image groups. The suggested algorithm cluster the images automatically and apply the constraints to each clustered image groups. The experiment result shows that our KCS-RANSAC algorithm outperformed the DCS-RANSAC algorithm in terms of speed, accuracy, and inlier rate.

Estimation of Manhattan Coordinate System using Convolutional Neural Network (합성곱 신경망 기반 맨하탄 좌표계 추정)

  • Lee, Jinwoo;Lee, Hyunjoon;Kim, Junho
    • Journal of the Korea Computer Graphics Society
    • /
    • v.23 no.3
    • /
    • pp.31-38
    • /
    • 2017
  • In this paper, we propose a system which estimates Manhattan coordinate systems for urban scene images using a convolutional neural network (CNN). Estimating the Manhattan coordinate system from an image under the Manhattan world assumption is the basis for solving computer graphics and vision problems such as image adjustment and 3D scene reconstruction. We construct a CNN that estimates Manhattan coordinate systems based on GoogLeNet [1]. To train the CNN, we collect about 155,000 images under the Manhattan world assumption by using the Google Street View APIs and calculate Manhattan coordinate systems using existing calibration methods to generate dataset. In contrast to PoseNet [2] that trains per-scene CNNs, our method learns from images under the Manhattan world assumption and thus estimates Manhattan coordinate systems for new images that have not been learned. Experimental results show that our method estimates Manhattan coordinate systems with the median error of $3.157^{\circ}$ for the Google Street View images of non-trained scenes, as test set. In addition, compared to an existing calibration method [3], the proposed method shows lower intermediate errors for the test set.

Driver Assistance System for Integration Interpretation of Driver's Gaze and Selective Attention Model (운전자 시선 및 선택적 주의 집중 모델 통합 해석을 통한 운전자 보조 시스템)

  • Kim, Jihun;Jo, Hyunrae;Jang, Giljin;Lee, Minho
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.16 no.3
    • /
    • pp.115-122
    • /
    • 2016
  • This paper proposes a system to detect driver's cognitive state by internal and external information of vehicle. The proposed system can measure driver's eye gaze. This is done by concept of information delivery and mutual information measure. For this study, we set up two web-cameras at vehicles to obtain visual information of the driver and front of the vehicle. We propose Gestalt principle based selective attention model to define information quantity of road scene. The saliency map based on gestalt principle is prominently represented by stimulus such as traffic signals. The proposed system assumes driver's cognitive resource allocation on the front scene by gaze analysis and head pose direction information. Then we use several feature algorithms for detecting driver's characteristics in real time. Modified census transform (MCT) based Adaboost is used to detect driver's face and its component whereas POSIT algorithms are used for eye detection and 3D head pose estimation. Experimental results show that the proposed system works well in real environment and confirm its usability.

Camera calibration parameters estimation using perspective variation ratio of grid type line widths (격자형 선폭들의 투영변화비를 이용한 카메라 교정 파라메터 추정)

  • Jeong, Jun-Ik;Choi, Seong-Gu;Rho, Do-Hwan
    • Proceedings of the KIEE Conference
    • /
    • 2004.11c
    • /
    • pp.30-32
    • /
    • 2004
  • With 3-D vision measuring, camera calibration is necessary to calculate parameters accurately. Camera calibration was developed widely in two categories. The first establishes reference points in space, and the second uses a grid type frame and statistical method. But, the former has difficulty to setup reference points and the latter has low accuracy. In this paper we present an algorithm for camera calibration using perspective ratio of the grid type frame with different line widths. It can easily estimate camera calibration parameters such as lens distortion, focal length, scale factor, pose, orientations, and distance. The advantage of this algorithm is that it can estimate the distance of the object. Also, the proposed camera calibration method is possible estimate distance in dynamic environment such as autonomous navigation. To validate proposed method, we set up the experiments with a frame on rotator at a distance of 1, 2, 3, 4[m] from camera and rotate the frame from -60 to 60 degrees. Both computer simulation and real data have been used to test the proposed method and very good results have been obtained. We have investigated the distance error affected by scale factor or different line widths and experimentally found an average scale factor that includes the least distance error with each image. The average scale factor tends to fluctuate with small variation and makes distance error decrease. Compared with classical methods that use stereo camera or two or three orthogonal planes, the proposed method is easy to use and flexible. It advances camera calibration one more step from static environments to real world such as autonomous land vehicle use.

  • PDF

Uncertainty Characteristics in Future Prediction of Agrometeorological Indicators using a Climatic Water Budget Approach (기후학적 물수지를 적용한 기후변화에 따른 농업기상지표 변동예측의 불확실성)

  • Nam, Won-Ho;Hong, Eun-Mi;Choi, Jin-Yong;Cho, Jaepil;Hayes, Michael J.
    • Journal of The Korean Society of Agricultural Engineers
    • /
    • v.57 no.2
    • /
    • pp.1-13
    • /
    • 2015
  • The Coupled Model Intercomparison Project Phase 5 (CMIP5), coordinated by the World Climate Research Programme in support of the Intergovernmental Panel on Climate Change (IPCC) AR5, is the most recent, provides projections of future climate change using various global climate models under four major greenhouse gas emission scenarios. There is a wide selection of climate models available to provide projections of future climate change. These provide for a wide range of possible outcomes when trying to inform managers about possible climate changes. Hence, future agrometeorological indicators estimation will be much impacted by which global climate model and climate change scenarios are used. Decision makers are increasingly expected to use climate information, but the uncertainties associated with global climate models pose substantial hurdles for agricultural resources planning. Although it is the most reasonable that quantifying of the future uncertainty using climate change scenarios, preliminary analysis using reasonable factors for selecting a subset for decision making are needed. In order to narrow the projections to a handful of models that could be used in a climate change impact study, we could provide effective information for selecting climate model and scenarios for climate change impact assessment using maximum/minimum temperature, precipitation, reference evapotranspiration, and moisture index of nine Representative Concentration Pathways (RCP) scenarios.

Zoom Lens Distortion Correction Of Video Sequence Using Nonlinear Zoom Lens Distortion Model (비선형 줌-렌즈 왜곡 모델을 이용한 비디오 영상에서의 줌-렌즈 왜곡 보정)

  • Kim, Dae-Hyun;Shin, Hyoung-Chul;Oh, Ju-Hyun;Nam, Seung-Jin;Sohn, Kwang-Hoon
    • Journal of Broadcast Engineering
    • /
    • v.14 no.3
    • /
    • pp.299-310
    • /
    • 2009
  • In this paper, we proposed a new method to correct the zoom lens distortion for the video sequence captured by the zoom lens. First, we defined the nonlinear zoom lens distortion model which is represented by the focal length and the lens distortion using the characteristic that lens distortion parameters are nonlinearly and monotonically changed while the focal length is increased. Then, we chose some sample images from the video sequence and estimated a focal length and a lens distortion parameter for each sample image. Using these estimated parameters, we were able to optimize the zoom lens distortion model. Once the zoom lens distortion model was obtained, lens distortion parameters of other images were able to be computed as their focal lengths were input. The proposed method has been made experiments with many real images and videos. As a result, accurate distortion parameters were estimated from the zoom lens distortion model and distorted images were well corrected without any visual artifacts.

Motion Plane Estimation for Real-Time Hand Motion Recognition (실시간 손동작 인식을 위한 동작 평면 추정)

  • Jeong, Seung-Dae;Jang, Kyung-Ho;Jung, Soon-Ki
    • The KIPS Transactions:PartB
    • /
    • v.16B no.5
    • /
    • pp.347-358
    • /
    • 2009
  • In this thesis, we develop a vision based hand motion recognition system using a camera with two rotational motors. Existing systems were implemented using a range camera or multiple cameras and have a limited working area. In contrast, we use an uncalibrated camera and get more wide working area by pan-tilt motion. Given an image sequence provided by the pan-tilt camera, color and pattern information are integrated into a tracking system in order to find the 2D position and direction of the hand. With these pose information, we estimate 3D motion plane on which the gesture motion trajectory from approximately forms. The 3D trajectory of the moving finger tip is projected into the motion plane, so that the resolving power of the linear gesture patterns is enhanced. We have tested the proposed approach in terms of the accuracy of trace angle and the dimension of the working volume.