• Title/Summary/Keyword: Vision recognition

Search Result 1,033, Processing Time 0.03 seconds

Egocentric Vision for Human Activity Recognition Using Deep Learning

  • Malika Douache;Badra Nawal Benmoussat
    • Journal of Information Processing Systems
    • /
    • v.19 no.6
    • /
    • pp.730-744
    • /
    • 2023
  • The topic of this paper is the recognition of human activities using egocentric vision, particularly captured by body-worn cameras, which could be helpful for video surveillance, automatic search and video indexing. This being the case, it could also be helpful in assistance to elderly and frail persons for revolutionizing and improving their lives. The process throws up the task of human activities recognition remaining problematic, because of the important variations, where it is realized through the use of an external device, similar to a robot, as a personal assistant. The inferred information is used both online to assist the person, and offline to support the personal assistant. With our proposed method being robust against the various factors of variability problem in action executions, the major purpose of this paper is to perform an efficient and simple recognition method from egocentric camera data only using convolutional neural network and deep learning. In terms of accuracy improvement, simulation results outperform the current state of the art by a significant margin of 61% when using egocentric camera data only, more than 44% when using egocentric camera and several stationary cameras data and more than 12% when using both inertial measurement unit (IMU) and egocentric camera data.

Stereo Vision Neural Networks with Competition and Cooperation for Phoneme Recognition

  • Kim, Sung-Ill;Chung, Hyun-Yeol
    • The Journal of the Acoustical Society of Korea
    • /
    • v.22 no.1E
    • /
    • pp.3-10
    • /
    • 2003
  • This paper describes two kinds of neural networks for stereoscopic vision, which have been applied to an identification of human speech. In speech recognition based on the stereoscopic vision neural networks (SVNN), the similarities are first obtained by comparing input vocal signals with standard models. They are then given to a dynamic process in which both competitive and cooperative processes are conducted among neighboring similarities. Through the dynamic processes, only one winner neuron is finally detected. In a comparative study, with, the average phoneme recognition accuracy on the two-layered SVNN was 7.7% higher than the Hidden Markov Model (HMM) recognizer with the structure of a single mixture and three states, and the three-layered was 6.6% higher. Therefore, it was noticed that SVNN outperformed the existing HMM recognizer in phoneme recognition.

Automatic Recognition of In-Process mold Dies Based on Reverse Engineering Technology (형상 역공학을 통한 공정중 금형 가공물의 자동인식)

  • 김정권;윤길상;최진화;김동우;조명우;박균명
    • Proceedings of the Korean Society of Machine Tool Engineers Conference
    • /
    • 2003.10a
    • /
    • pp.420-425
    • /
    • 2003
  • Generally, reverse engineering means getting CAD data from unidentified shape using vision or 3D laser scanner system. In this paper, we studied unidentified model by machine vision based reverse engineering system to get information about in-processing model. Recently, vision technology is widely used in current factories, because it could inspect the in-process object easily, quickly, accurately. The following tasks were mainly investigated and implemented. We obtained more precise data by corning camera's distortion, compensating slit-beam error and revising acquired image. Much more, we made similar curves or surface with B-spline approximation for precision. Until now, there have been many case study of shape recognition. But it was uncompatible to apply to the field, because it had taken too many processing time and has frequent recognition failure. This paper propose recognition algorithm that prevent such errors and give applications to the field.

  • PDF

Computer Vision Platform Design with MEAN Stack Basis (MEAN Stack 기반의 컴퓨터 비전 플랫폼 설계)

  • Hong, Seonhack;Cho, Kyungsoon;Yun, Jinseob
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.11 no.3
    • /
    • pp.1-9
    • /
    • 2015
  • In this paper, we implemented the computer vision platform design with MEAN Stack through Raspberry PI 2 model which is an open source platform. we experimented the face recognition, temperature and humidity sensor data logging with WiFi communication under Raspberry Pi 2 model. Especially we directly made the shape of platform with 3D printing design. In this paper, we used the face recognition algorithm with OpenCV software through haarcascade feature extraction machine learning algorithm, and extended the functionality of wireless communication function ability with Bluetooth technology for the purpose of making Android Mobile devices interface. And therefore we implemented the functions of the vision platform for identifying the face recognition characteristics of scanning with PI camera with gathering the temperature and humidity sensor data under IoT environment. and made the vision platform with 3D printing technology. Especially we used MongoDB for developing the performance of vision platform because the MongoDB is more akin to working with objects in a programming language than what we know of as a database. Afterwards, we would enhance the performance of vision platform for clouding functionalities.

A Computer Vision-Based Banknote Recognition System for the Blind with an Accuracy of 98% on Smartphone Videos

  • Sanchez, Gustavo Adrian Ruiz
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.6
    • /
    • pp.67-72
    • /
    • 2019
  • This paper proposes a computer vision-based banknote recognition system intended to assist the blind. This system is robust and fast in recognizing banknotes on videos recorded with a smartphone on real-life scenarios. To reduce the computation time and enable a robust recognition in cluttered environments, this study segments the banknote candidate area from the background utilizing a technique called Pixel-Based Adaptive Segmenter (PBAS). The Speeded-Up Robust Features (SURF) interest point detector is used, and SURF feature vectors are computed only when sufficient interest points are found. The proposed algorithm achieves a recognition accuracy of 98%, a 100% true recognition rate and a 0% false recognition rate. Although Korean banknotes are used as a working example, the proposed system can be applied to recognize other countries' banknotes.

Analysis of Distance Error of Stereo Vision System for Obstacle Recognition System of AGV (AGV의 장애물 판별을 위한 스테레오 비젼시스템의 거리오차 해석)

  • 조연상;배효준;원두원;박흥식
    • Proceedings of the Korean Society of Precision Engineering Conference
    • /
    • 2001.04a
    • /
    • pp.170-173
    • /
    • 2001
  • To apply stereo vision system to obstacle recognition system of AGV, we constructed algorithm of stereo matching and distance measuring with stereo image for positioning of object in area. And using this system, we look into the error between real position and measured position, and studied relationship of compensation.

  • PDF

ADD-Net: Attention Based 3D Dense Network for Action Recognition

  • Man, Qiaoyue;Cho, Young Im
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.6
    • /
    • pp.21-28
    • /
    • 2019
  • Recent years with the development of artificial intelligence and the success of the deep model, they have been deployed in all fields of computer vision. Action recognition, as an important branch of human perception and computer vision system research, has attracted more and more attention. Action recognition is a challenging task due to the special complexity of human movement, the same movement may exist between multiple individuals. The human action exists as a continuous image frame in the video, so action recognition requires more computational power than processing static images. And the simple use of the CNN network cannot achieve the desired results. Recently, the attention model has achieved good results in computer vision and natural language processing. In particular, for video action classification, after adding the attention model, it is more effective to focus on motion features and improve performance. It intuitively explains which part the model attends to when making a particular decision, which is very helpful in real applications. In this paper, we proposed a 3D dense convolutional network based on attention mechanism(ADD-Net), recognition of human motion behavior in the video.

A Study on Weldability Estirmtion of Laser Welded Specimens by Vision Sensor (비전 센서를 이용한 레이져 용접물의 용접성 평가에 관한 연구)

  • 엄기원;이세헌;이정익
    • Proceedings of the Korean Society of Precision Engineering Conference
    • /
    • 1995.10a
    • /
    • pp.1101-1104
    • /
    • 1995
  • Through welding fabrication, user can feel an surficaial and capable unsatisfaction because of welded defects, Generally speaking, these are called weld defects. For checking these defects effectively without time loss effectively, weldability estimation system setup isan urgent thing for detecting whole specimen quality. In this study, by laser vision camera, catching a rawdata on welded specimen profiles, treating vision processing with these data, qualititative defects are estimated from getting these information at first. At the same time, for detecting quantitative defects, whole specimen weldability estimation is pursued by multifeature pattern recognition, which is a kind of fuzzy pattern recognition. For user friendly, by weldability estimation results are shown each profiles, final reports and visual graphics method, user can easily determined weldability. By applying these system to welding fabrication, these technologies are contribution to on-line weldability estimation.

  • PDF

Real-Time Facial Recognition Using the Geometric Informations

  • Lee, Seong-Cheol;Kang, E-Sok
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2001.10a
    • /
    • pp.55.3-55
    • /
    • 2001
  • The implementation of human-like robot has been advanced in various parts such as mechanic arms, legs, and applications of five senses. The vision applications have been developed in several decades and especially the face recognition have become a prominent issue. In addition, the development of computer systems makes it possible to process complex algorithms in realtime. The most of human recognition systems adopt the discerning method using fingerprint, iris, and etc. These methods restrict the motion of the person to be discriminated. Recently, the researchers of human recognition systems are interested in facial recognition by using machine vision. Thus, the object of this paper is the implementation of the realtime ...

  • PDF

Color Vision Abnormality of Elementary School Students in Kwang Ju Area (광주지역 초등학생들의 색각이상에 관한 연구)

  • Ryu, Geun-Chang;Yoon, Young;Seong, Jeong-Sub
    • Journal of Korean Ophthalmic Optics Society
    • /
    • v.12 no.3
    • /
    • pp.89-91
    • /
    • 2007
  • Color vision test was conducted to elementary school students at age of 9 to 13, with total subject number of 598 which live in Gwang Ju area. 325 (54.3%) boys and 273 (45.7%) girls were subjected using Hahn Color Vision test to find out color recognition problems. 1. Ratios of color vision abnomality were 6.25% in 9 year old students, 9.2% in 10 year old students, 8.4% in 12 year old students, 7.8% in 13 year old students, which means 7.9% of the total 598 subjects had color recognition matters. 2. Red-Green dyschromatopsia was 7.8% of the total 598 subjects which includes most of subjects. None of them had green-yellow recognition problem. Full dyschromatopsia had frequency of 0.2%. 3. 10.7% of boys had color vision abnormality, while 10.7% of girls had color vision abnormality, which means that boys have color recognition problems with higher rate than girls.

  • PDF