• Title/Summary/Keyword: vision-based recognition

Search Result 633, Processing Time 0.036 seconds

Posture Stabilization Control for Mobile Robot using Marker Recognition and Hybrid Visual Servoing (마커인식과 혼합 비주얼 서보잉 기법을 통한 이동로봇의 자세 안정화 제어)

  • Lee, Sung-Goo;Kwon, Ji-Wook;Hong, Suk-Kyo;Chwa, Dong-Kyoung
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.60 no.8
    • /
    • pp.1577-1585
    • /
    • 2011
  • This paper proposes a posture stabilization control algorithm for a wheeled mobile robot using hybrid visual servo control method with a position based and an image based visual servoing (PBVS and IBVS). To overcome chattering phenomena which were shown in the previous researches using a simple switching function based on a threshold, the proposed hybrid visual servo control law introduces the fusion function based on a blending function. Then, the chattering problem and rapid motion of the mobile robot can be eliminated. Also, we consider the nonlinearity of the wheeled mobile robot unlike the previous visual servo control laws using linear control methods to improve the performances of the visual servo control law. The proposed posture stabilization control law using hybrid visual servoing is verified by a theoretical analysis and simulation and experimental results.

Analysis of Feature Extraction Algorithms Based on Deep Learning (Deep Learning을 기반으로 한 Feature Extraction 알고리즘의 분석)

  • Kim, Gyung Tae;Lee, Yong Hwan;Kim, Yeong Seop
    • Journal of the Semiconductor & Display Technology
    • /
    • v.19 no.2
    • /
    • pp.60-67
    • /
    • 2020
  • Recently, artificial intelligence related technologies including machine learning are being applied to various fields, and the demand is also increasing. In particular, with the development of AR, VR, and MR technologies related to image processing, the utilization of computer vision based on deep learning has increased. The algorithms for object recognition and detection based on deep learning required for image processing are diversified and advanced. Accordingly, problems that were difficult to solve with the existing methodology were solved more simply and easily by using deep learning. This paper introduces various deep learning-based object recognition and extraction algorithms used to detect and recognize various objects in an image and analyzes the technologies that attract attention.

A Study on Alignment Correction Algorithm for Detecting Specific Areas of Video Images (영상 이미지의 특정 영역 검출을 위한 정렬 보정 알고리즘 연구)

  • Jin, Go-Whan
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.11
    • /
    • pp.9-14
    • /
    • 2018
  • The vision system is a device for acquiring images and analyzing and discriminating inspection areas. Demand for use in the automation process has increased, and the introduction of a vision-based inspection system has emerged as a very important issue. These vision systems are used for everyday life and used as inspection equipment in production processes. Image processing technology is actively being studied. However, there is little research on the area definition for extracting objects such as character recognition or semiconductor packages. In this paper, define a region of interest and perform edge extraction to prevent the user from judging noise as an edge. We propose a noise-robust alignment correction model that can extract the edge of a region to be inspected using the distribution of edges in a specific region even if noise exists in the image. Through the proposed model, it is expected that the product production efficiency will be improved if it is applied to production field such as character recognition of tire or inspection of semiconductor packages.

Volume Control using Gesture Recognition System

  • Shreyansh Gupta;Samyak Barnwal
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.6
    • /
    • pp.161-170
    • /
    • 2024
  • With the technological advances, the humans have made so much progress in the ease of living and now incorporating the use of sight, motion, sound, speech etc. for various application and software controls. In this paper, we have explored the project in which gestures plays a very significant role in the project. The topic of gesture control which has been researched a lot and is just getting evolved every day. We see the usage of computer vision in this project. The main objective that we achieved in this project is controlling the computer settings with hand gestures using computer vision. In this project we are creating a module which acts a volume controlling program in which we use hand gestures to control the computer system volume. We have included the use of OpenCV. This module is used in the implementation of hand gestures in computer controls. The module in execution uses the web camera of the computer to record the images or videos and then processes them to find the needed information and then based on the input, performs the action on the volume settings if that computer. The program has the functionality of increasing and decreasing the volume of the computer. The setup needed for the program execution is a web camera to record the input images and videos which will be given by the user. The program will perform gesture recognition with the help of OpenCV and python and its libraries and them it will recognize or identify the specified human gestures and use them to perform or carry out the changes in the device setting. The objective is to adjust the volume of a computer device without the need for physical interaction using a mouse or keyboard. OpenCV, a widely utilized tool for image processing and computer vision applications in this domain, enjoys extensive popularity. The OpenCV community consists of over 47,000 individuals, and as of a survey conducted in 2020, the estimated number of downloads exceeds 18 million.

Multiple Camera-Based Real-Time Long Queue Vision Algorithm for Public Safety and Efficiency

  • Tae-hoon Kim;Ji-young Na;Ji-won Yoon;Se-Hun Lee;Jun-ho Ahn
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.10
    • /
    • pp.47-57
    • /
    • 2024
  • This paper proposes a system to efficiently manage delays caused by unmanaged and congested queues in crowded environments. Such queues not only cause inconvenience but also pose safety risks. Existing systems, relying on single-camera feeds, are inadequate for complex scenarios requiring multiple cameras. To address this, we developed a multi-vision long queue detection system that integrates multiple vision algorithms to accurately detect various types of queues. The algorithm processes real-time video data from multiple cameras, stitching overlapping segments into a single panoramic image. By combining object detection, tracking, and position variation analysis, the system recognizes long queues in crowded environments. The algorithm was validated with 96% accuracy and a 92% F1-score across diverse settings.

FRS-OCC: Face Recognition System for Surveillance Based on Occlusion Invariant Technique

  • Abbas, Qaisar
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.8
    • /
    • pp.288-296
    • /
    • 2021
  • Automated face recognition in a runtime environment is gaining more and more important in the fields of surveillance and urban security. This is a difficult task keeping in mind the constantly volatile image landscape with varying features and attributes. For a system to be beneficial in industrial settings, it is pertinent that its efficiency isn't compromised when running on roads, intersections, and busy streets. However, recognition in such uncontrolled circumstances is a major problem in real-life applications. In this paper, the main problem of face recognition in which full face is not visible (Occlusion). This is a common occurrence as any person can change his features by wearing a scarf, sunglass or by merely growing a mustache or beard. Such types of discrepancies in facial appearance are frequently stumbled upon in an uncontrolled circumstance and possibly will be a reason to the security systems which are based upon face recognition. These types of variations are very common in a real-life environment. It has been analyzed that it has been studied less in literature but now researchers have a major focus on this type of variation. Existing state-of-the-art techniques suffer from several limitations. Most significant amongst them are low level of usability and poor response time in case of any calamity. In this paper, an improved face recognition system is developed to solve the problem of occlusion known as FRS-OCC. To build the FRS-OCC system, the color and texture features are used and then an incremental learning algorithm (Learn++) to select more informative features. Afterward, the trained stack-based autoencoder (SAE) deep learning algorithm is used to recognize a human face. Overall, the FRS-OCC system is used to introduce such algorithms which enhance the response time to guarantee a benchmark quality of service in any situation. To test and evaluate the performance of the proposed FRS-OCC system, the AR face dataset is utilized. On average, the FRS-OCC system is outperformed and achieved SE of 98.82%, SP of 98.49%, AC of 98.76% and AUC of 0.9995 compared to other state-of-the-art methods. The obtained results indicate that the FRS-OCC system can be used in any surveillance application.

A Study on the Five Senses Information Processing for HCI (HCI를 위한 오감정보처리에 관한 연구)

  • Lee, Hyeon Gu;Kim, Dong Kyu
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.5 no.2
    • /
    • pp.77-85
    • /
    • 2009
  • In this paper, we propose data format for smell, taste, touch with speech and vision which can be transmitted and implement a floral scent detection and recognition system. We provide representation method of data of smell, taste, and touch. Also, proposed floral scent recognition system consists of three module such as floral scent acquisition module using Metal Oxide Semiconductor (MOS) sensor array, entropy-based floral scent detection module, and floral scent recognition module using correlation coefficients. The proposed system calculates correlation coefficients of the individual sensor between feature vector(16 sensors) from floral scent input point until the stable region and 12 types of reference models. Then, this system selects the floral scent with the maximum similarity to the calculated average of individual correlation coefficients. To evaluate the floral scent recognition system using correlation coefficients, we implemented an individual floral scent recognition system using K-NN with PCA and LDA that are generally used in conventional electronic noses. In the experimental results, the proposed system performs approximately 95.7% average recognition rate.

Emotion Recognition Method using Gestures and EEG Signals (제스처와 EEG 신호를 이용한 감정인식 방법)

  • Kim, Ho-Duck;Jung, Tae-Min;Yang, Hyun-Chang;Sim, Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.13 no.9
    • /
    • pp.832-837
    • /
    • 2007
  • Electroencephalographic(EEG) is used to record activities of human brain in the area of psychology for many years. As technology develope, neural basis of functional areas of emotion processing is revealed gradually. So we measure fundamental areas of human brain that controls emotion of human by using EEG. Hands gestures such as shaking and head gesture such as nodding are often used as human body languages for communication with each other, and their recognition is important that it is a useful communication medium between human and computers. Research methods about gesture recognition are used of computer vision. Many researchers study Emotion Recognition method which uses one of EEG signals and Gestures in the existing research. In this paper, we use together EEG signals and Gestures for Emotion Recognition of human. And we select the driver emotion as a specific target. The experimental result shows that using of both EEG signals and gestures gets high recognition rates better than using EEG signals or gestures. Both EEG signals and gestures use Interactive Feature Selection(IFS) for the feature selection whose method is based on a reinforcement learning.

Face Recognition using Extended Center-Symmetric Pattern and 2D-PCA (Extended Center-Symmetric Pattern과 2D-PCA를 이용한 얼굴인식)

  • Lee, Hyeon Gu;Kim, Dong Ju
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.9 no.2
    • /
    • pp.111-119
    • /
    • 2013
  • Face recognition has recently become one of the most popular research areas in the fields of computer vision, machine learning, and pattern recognition because it spans numerous applications, such as access control, surveillance, security, credit-card verification, and criminal identification. In this paper, we propose a simple descriptor called an ECSP(Extended Center-Symmetric Pattern) for illumination-robust face recognition. The ECSP operator encodes the texture information of a local face region by emphasizing diagonal components of a previous CS-LBP(Center-Symmetric Local Binary Pattern). Here, the diagonal components are emphasized because facial textures along the diagonal direction contain much more information than those of other directions. The facial texture information of the ECSP operator is then used as the input image of an image covariance-based feature extraction algorithm such as 2D-PCA(Two-Dimensional Principal Component Analysis). Performance evaluation of the proposed approach was carried out using various binary pattern operators and recognition algorithms on the Yale B database. The experimental results demonstrated that the proposed approach achieved better recognition accuracy than other approaches, and we confirmed that the proposed approach is effective against illumination variation.

Real-time traffic light information recognition based on object detection models (객체 인식 모델 기반 실시간 교통신호 정보 인식)

  • Joo, eun-oh;Kim, Min-Soo
    • Journal of Cadastre & Land InformatiX
    • /
    • v.52 no.1
    • /
    • pp.81-93
    • /
    • 2022
  • Recently, there have been many studies on object recognition around the vehicle and recognition of traffic signs and traffic lights in autonomous driving. In particular, such the recognition of traffic lights is one of the core technologies in autonomous driving. Therefore, many studies for such the recognition of traffic lights have been performed, the studies based on various deep learning models have increased significantly in recent. In addition, as a high-quality AI training data set for voice, vision, and autonomous driving is released on AIHub, it makes it possible to develop a recognition model for traffic lights suitable for the domestic environment using the data set. In this study, we developed a recognition model for traffic lights that can be used in Korea using the AIHub's training data set. In particular, in order to improve the recognition performance, we used various models of YOLOv4 and YOLOv5, and performed our recognition experiments by defining various classes for the training data. In conclusion, we could see that YOLOv5 shows better performance in the recognition than YOLOv4 and could confirm the reason from the architecture comparison of the two models.