• Title/Summary/Keyword: video recognition

Search Result 696, Processing Time 0.028 seconds

Automatic Person Identification using Multiple Cues

  • Swangpol, Danuwat;Chalidabhongse, Thanarat
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1202-1205
    • /
    • 2005
  • This paper describes a method for vision-based person identification that can detect, track, and recognize person from video using multiple cues: height and dressing colors. The method does not require constrained target's pose or fully frontal face image to identify the person. First, the system, which is connected to a pan-tilt-zoom camera, detects target using motion detection and human cardboard model. The system keeps tracking the moving target while it is trying to identify whether it is a human and identify who it is among the registered persons in the database. To segment the moving target from the background scene, we employ a version of background subtraction technique and some spatial filtering. Once the target is segmented, we then align the target with the generic human cardboard model to verify whether the detected target is a human. If the target is identified as a human, the card board model is also used to segment the body parts to obtain some salient features such as head, torso, and legs. The whole body silhouette is also analyzed to obtain the target's shape information such as height and slimness. We then use these multiple cues (at present, we uses shirt color, trousers color, and body height) to recognize the target using a supervised self-organization process. We preliminary tested the system on a set of 5 subjects with multiple clothes. The recognition rate is 100% if the person is wearing the clothes that were learned before. In case a person wears new dresses the system fail to identify. This means height is not enough to classify persons. We plan to extend the work by adding more cues such as skin color, and face recognition by utilizing the zoom capability of the camera to obtain high resolution view of face; then, evaluate the system with more subjects.

  • PDF

Video-based Facial Emotion Recognition using Active Shape Models and Statistical Pattern Recognizers (Active Shape Model과 통계적 패턴인식기를 이용한 얼굴 영상 기반 감정인식)

  • Jang, Gil-Jin;Jo, Ahra;Park, Jeong-Sik;Seo, Yong-Ho
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.14 no.3
    • /
    • pp.139-146
    • /
    • 2014
  • This paper proposes an efficient method for automatically distinguishing various facial expressions. To recognize the emotions from facial expressions, the facial images are obtained by digital cameras, and a number of feature points were extracted. The extracted feature points are then transformed to 49-dimensional feature vectors which are robust to scale and translational variations, and the facial emotions are recognized by statistical pattern classifiers such Naive Bayes, MLP (multi-layer perceptron), and SVM (support vector machine). Based on the experimental results with 5-fold cross validation, SVM was the best among the classifiers, whose performance was obtained by 50.8% for 6 emotion classification, and 78.0% for 3 emotions.

Development of Interactive Media Player for Kiosk with User Motion Detection (사용자 모션 인식 기반 키오스크 전용 인터랙티브 미디어 플레이어 개발)

  • Song, Bok Deuk;Kim, Hyeong-Jin;Jeong, Hyeon-Jae;Choi, Yeon Jun
    • The Journal of the Korea Contents Association
    • /
    • v.19 no.11
    • /
    • pp.270-277
    • /
    • 2019
  • These days, with the advent of digital broadcasting, media environment offers users an opportunity to enjoy differentiated contents in a more aggressive fashion through user-media interactions based on computer technology. In fact, the development of contents which can induce spontaneous acts from users such as outdoor ads which use certain sensors and devices and exhibition halls has been active. With the development of low-price motion recognition devices, people have been able to enjoy diverse interaction-applied media by recognizing users' motion data without body contact. In this paper, we developed an interactive media player that can recognize the user's motion and control the video in the web service environment without installing a specific program. In addition, we set user motion recognition range and developed a user motion recognition algorithm suitable for the Leap Motion equipment installed in the kiosk. The results of this study can be experienced by various interactive media such as interactive tourism, education, and movie contents in kiosks that can be installed in public places.

Determining Method of Factors for Effective Real Time Background Modeling (효과적인 실시간 배경 모델링을 위한 환경 변수 결정 방법)

  • Lee, Jun-Cheol;Ryu, Sang-Ryul;Kang, Sung-Hwan;Kim, Sung-Ho
    • Journal of KIISE:Software and Applications
    • /
    • v.34 no.1
    • /
    • pp.59-69
    • /
    • 2007
  • In the video with a various environment, background modeling is important for extraction and recognition the moving object. For this object recognition, many methods of the background modeling are proposed in a process of preprocess. Among these there is a Kumar method which represents the Queue-based background modeling. Because this has a fixed period of updating examination of the frame, there is a limit for various system. This paper use a background modeling based on the queue. We propose the method that major parameters are decided as adaptive by background model. They are the queue size of the sliding window, the sire of grouping by the brightness of the visual and the period of updating examination of the frame. In order to determine the factors, in every process, RCO (Ratio of Correct Object), REO (Ratio of Error Object) and UR (Update Ratio) are considered to be the standard of evaluation. The proposed method can improve the existing techniques of the background modeling which is unfit for the real-time processing and recognize the object more efficient.

The Character Recognition System of Mobile Camera Based Image (모바일 이미지 기반의 문자인식 시스템)

  • Park, Young-Hyun;Lee, Hyung-Jin;Baek, Joong-Hwan
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.11 no.5
    • /
    • pp.1677-1684
    • /
    • 2010
  • Recently, due to the development of mobile phone and supply of smart phone, many contents have been developed. Especially, since the small-sized cameras are equiped in mobile devices, people are interested in the image based contents development, and it also becomes important part in their practical use. Among them, the character recognition system can be widely used in the applications such as blind people guidance systems, automatic robot navigation systems, automatic video retrieval and indexing systems, automatic text translation systems. Therefore, this paper proposes a system that is able to extract text area from the natural images captured by smart phone camera. The individual characters are recognized and result is output in voice. Text areas are extracted using Adaboost algorithm and individual characters are recognized using error back propagated neural network.

Bio-mimetic Recognition of Action Sequence using Unsupervised Learning (비지도 학습을 이용한 생체 모방 동작 인지 기반의 동작 순서 인식)

  • Kim, Jin Ok
    • Journal of Internet Computing and Services
    • /
    • v.15 no.4
    • /
    • pp.9-20
    • /
    • 2014
  • Making good predictions about the outcome of one's actions would seem to be essential in the context of social interaction and decision-making. This paper proposes a computational model for learning articulated motion patterns for action recognition, which mimics biological-inspired visual perception processing of human brain. Developed model of cortical architecture for the unsupervised learning of motion sequence, builds upon neurophysiological knowledge about the cortical sites such as IT, MT, STS and specific neuronal representation which contribute to articulated motion perception. Experiments show how the model automatically selects significant motion patterns as well as meaningful static snapshot categories from continuous video input. Such key poses correspond to articulated postures which are utilized in probing the trained network to impose implied motion perception from static views. We also present how sequence selective representations are learned in STS by fusing snapshot and motion input and how learned feedback connections enable making predictions about future input sequence. Network simulations demonstrate the computational capacity of the proposed model for motion recognition.

Research on the Convergence of CCTV Video Information with Disaster Recognition and Real-time Crisis Response System (CCTV 영상 정보와 재난재해 인식 및 실시간 위기 대응 시스템의 융합에 관한 연구)

  • Kim, Ki-Bong;Geum, Gi-Moon;Jang, Chang-Bok
    • Journal of the Korea Convergence Society
    • /
    • v.8 no.3
    • /
    • pp.15-22
    • /
    • 2017
  • People generally believe that disaster forecast and warning systems and response systems are well established in the age of cutting edge technology. As a matter of fact, reliable systems to respond to disasters are not properly equipped, as we witnessed the Sewol ferry disaster in 2014. The existing forecast and warning systems are based on sensor information with low efficiency, and image information is only operated by monitoring staff manually. In addition, the interconnection between a warning system and a response system in order to decide how to cope with the recognized disaster is very insufficient. This paper introduces the CCTV based disaster recognition and real time crisis response system composed of the CCTV image recognition engine and the crisis response technique. This system has brought the possibility to overcome the limitations of existing sensor based forecast and warning systems, and to resolve the problems in the absence of monitoring staff when responding to crisis.

A Speech Recognition System based on a New Endpoint Estimation Method jointly using Audio/Video Informations (음성/영상 정보를 이용한 새로운 끝점추정 방식에 기반을 둔 음성인식 시스템)

  • 이동근;김성준;계영철
    • Journal of Broadcast Engineering
    • /
    • v.8 no.2
    • /
    • pp.198-203
    • /
    • 2003
  • We develop the method of estimating the endpoints of speech by jointly using the lip motion (visual speech) and speech being included in multimedia data and then propose a new speech recognition system (SRS) based on that method. The endpoints of noisy speech are estimated as follows : For each test word, two kinds of endpoints are detected from visual speech and clean speech, respectively Their difference is made and then added to the endpoints of visual speech to estimate those for noisy speech. This estimation method for endpoints (i.e. speech interval) is applied to form a new SRS. The SRS differs from the convention alone in that each word model in the recognizer is provided an interval of speech not Identical but estimated respectively for the corresponding word. Simulation results show that the proposed method enables the endpoints to be accurately estimated regardless of the amount of noise and consequently achieves 8 o/o improvement in recognition rate.

Development of Lane and Vehicle Headway Direction Recognition System for Military Heavy Equipment's Safe Transport - Based on Kalman Filter and Neural Network - (안전한 군용 중장비 수송을 위한 차선 및 차량 진행 방향 인식 시스템 개발 - 칼만 필터와 신경망을 기반으로 -)

  • Choi, Yeong-Yoon;Choi, Kwang-Mo;Moon, Ho-Seok
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.10 no.3
    • /
    • pp.139-147
    • /
    • 2007
  • In military transportation, the use of wide trailer for transporting the large and heavy weight equipments such as tank, armoured vehicle, and mobile gunnery is quite common. So, the vulnerability of causing traffic accidents for these wide military trailer to bump or collide with another car in adjacent lane is very high due to its broad width in excess of its own lane's width. Also, the possibility of these strayed accidents can be increased especially by the careless driver. In this paper, the recognition system of lane and vehicle headway direction is developed to detect the possible collision and warn the driver to prevent the fatal accident. In the system development, Kalman filtering is used first to extract the border of driving lane from the video images supplied by the CCD camera attached to the vehicle and the driving lane detection is completed with regression analysis. Next, the vehicle headway direction is recognized by using neural network scheme with the extracted parameters of the detected driving lane feature. The practical experiments for the developed system are also carried out in the real traffic road of Seoul city area and the results show us the more than 90% accuracy in recognizing the driving lane and vehicle headway direction.

User interface of Home-Automation for the physically handicapped Person in wearable computing environment (웨어러블 환경에서의 수족사용 불능자를 위한 홈오토메이션 사용자 인터페이스)

  • Kang, Sun-Kyung;Kim, Young-Un;Han, Dae-Kyung;Jung, Sung-Tae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.5
    • /
    • pp.187-193
    • /
    • 2008
  • Interface technologies for a user to control home automation system in wearable computing environment has been studied recently. This paper proposes a new interface method for a disabled person to control home automation system in wearable computing environment by using EOG sensing circuit and marker recognition. In the proposed interface method, the operations of a home network device are represented with human readable markers and displayed around the device. A user wearing a HMD, a video camera, and a computer selects the desired operation by seeing the markers and selecting one of them with eye movement from the HMD display The requested operation is executed by sending the control command for the selected marker to the home network control device. By using the EOG sensing circuit and the marker recognition system a user having problem with moving hands and fit can manipulate a home automation system with only eye movement.

  • PDF