• Title/Summary/Keyword: recognition of object

Search Result 1,722, Processing Time 0.031 seconds

Discriminant Analysis of Human's Implicit Intent based on Eyeball Movement (안구운동 기반의 사용자 묵시적 의도 판별 분석 모델)

  • Jang, Young-Min;Mallipeddi, Rammohan;Kim, Cheol-Su;Lee, Minho
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.6
    • /
    • pp.212-220
    • /
    • 2013
  • Recently, there has been tremendous increase in human-computer/machine interaction system, where the goal is to provide with an appropriate service to the user at the right time with minimal human inputs for human augmented cognition system. To develop an efficient human augmented cognition system based on human computer/machine interaction, it is important to interpret the user's implicit intention, which is vague, in addition to the explicit intention. According to cognitive visual-motor theory, human eye movements and pupillary responses are rich sources of information about human intention and behavior. In this paper, we propose a novel approach for the identification of human implicit visual search intention based on eye movement pattern and pupillary analysis such as pupil size, gradient of pupil size variation, fixation length/count for the area of interest. The proposed model identifies the human's implicit intention into three types such as navigational intent generation, informational intent generation, and informational intent disappearance. Navigational intent refers to the search to find something interesting in an input scene with no specific instructions, while informational intent refers to the search to find a particular target object at a specific location in the input scene. In the present study, based on the human eye movement pattern and pupillary analysis, we used a hierarchical support vector machine which can detect the transitions between the different implicit intents - navigational intent generation to informational intent generation and informational intent disappearance.

Performance Enhancement of the Attitude Estimation using Small Quadrotor by Vision-based Marker Tracking (영상기반 물체추적에 의한 소형 쿼드로터의 자세추정 성능향상)

  • Kang, Seokyong;Choi, Jongwhan;Jin, Taeseok
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.25 no.5
    • /
    • pp.444-450
    • /
    • 2015
  • The accuracy of small and low cost CCD camera is insufficient to provide data for precisely tracking unmanned aerial vehicles(UAVs). This study shows how UAV can hover on a human targeted tracking object by using CCD camera rather than imprecise GPS data. To realize this, UAVs need to recognize their attitude and position in known environment as well as unknown environment. Moreover, it is necessary for their localization to occur naturally. It is desirable for an UAV to estimate of his attitude by environment recognition for UAV hovering, as one of the best important problems. In this paper, we describe a method for the attitude of an UAV using image information of a maker on the floor. This method combines the observed position from GPS sensors and the estimated attitude from the images captured by a fixed camera to estimate an UAV. Using the a priori known path of an UAV in the world coordinates and a perspective camera model, we derive the geometric constraint equations which represent the relation between image frame coordinates for a marker on the floor and the estimated UAV's attitude. Since the equations are based on the estimated position, the measurement error may exist all the time. The proposed method utilizes the error between the observed and estimated image coordinates to localize the UAV. The Kalman filter scheme is applied for this method. its performance is verified by the image processing results and the experiment.

A Study on Similar Trademark Search Model Using Convolutional Neural Networks (합성곱 신경망(Convolutional Neural Network)을 활용한 지능형 유사상표 검색 모형 개발)

  • Yoon, Jae-Woong;Lee, Suk-Jun;Song, Chil-Yong;Kim, Yeon-Sik;Jung, Mi-Young;Jeong, Sang-Il
    • Management & Information Systems Review
    • /
    • v.38 no.3
    • /
    • pp.55-80
    • /
    • 2019
  • Recently, many companies improving their management performance by building a powerful brand value which is recognized for trademark rights. However, as growing up the size of online commerce market, the infringement of trademark rights is increasing. According to various studies and reports, cases of foreign and domestic companies infringing on their trademark rights are increased. As the manpower and the cost required for the protection of trademark are enormous, small and medium enterprises(SMEs) could not conduct preliminary investigations to protect their trademark rights. Besides, due to the trademark image search service does not exist, many domestic companies have a problem that investigating huge amounts of trademarks manually when conducting preliminary investigations to protect their rights of trademark. Therefore, we develop an intelligent similar trademark search model to reduce the manpower and cost for preliminary investigation. To measure the performance of the model which is developed in this study, test data selected by intellectual property experts was used, and the performance of ResNet V1 101 was the highest. The significance of this study is as follows. The experimental results empirically demonstrate that the image classification algorithm shows high performance not only object recognition but also image retrieval. Since the model that developed in this study was learned through actual trademark image data, it is expected that it can be applied in the real industrial environment.

Training Performance Analysis of Semantic Segmentation Deep Learning Model by Progressive Combining Multi-modal Spatial Information Datasets (다중 공간정보 데이터의 점진적 조합에 의한 의미적 분류 딥러닝 모델 학습 성능 분석)

  • Lee, Dae-Geon;Shin, Young-Ha;Lee, Dong-Cheon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.40 no.2
    • /
    • pp.91-108
    • /
    • 2022
  • In most cases, optical images have been used as training data of DL (Deep Learning) models for object detection, recognition, identification, classification, semantic segmentation, and instance segmentation. However, properties of 3D objects in the real-world could not be fully explored with 2D images. One of the major sources of the 3D geospatial information is DSM (Digital Surface Model). In this matter, characteristic information derived from DSM would be effective to analyze 3D terrain features. Especially, man-made objects such as buildings having geometrically unique shape could be described by geometric elements that are obtained from 3D geospatial data. The background and motivation of this paper were drawn from concept of the intrinsic image that is involved in high-level visual information processing. This paper aims to extract buildings after classifying terrain features by training DL model with DSM-derived information including slope, aspect, and SRI (Shaded Relief Image). The experiments were carried out using DSM and label dataset provided by ISPRS (International Society for Photogrammetry and Remote Sensing) for CNN-based SegNet model. In particular, experiments focus on combining multi-source information to improve training performance and synergistic effect of the DL model. The results demonstrate that buildings were effectively classified and extracted by the proposed approach.

An Image Processing System for the Harvesting robot$^{1)}$ (포도수확용 로봇 개발을 위한 영상처리시스템)

  • Lee, Dae-Weon;Kim, Dong-Woo;Kim, Hyun-Tae;Lee, Yong-Kuk;Si-Heung
    • Journal of Bio-Environment Control
    • /
    • v.10 no.3
    • /
    • pp.172-180
    • /
    • 2001
  • A grape fruit is required for a lot of labor to harvest in time in Korea, since the fruit is cut and grabbed currently by hand. In foreign country, especially France, a grape harvester has been developed for processing to make wine out of a grape, not to eat a fresh grape fruit. However, a harvester which harvests to eat a fresh grape fruit has not been developed yet. Therefore, this study was designed and constructed to develope a image processing system for a fresh grape harvester. Its development involved the integration of a vision system along with an personal computer and two cameras. Grape recognition, which was able to found the accurate cutting position in three dimension by the end-effector, needed to find out the object from the background by using two different images from two cameras. Based on the results of this research the following conclusions were made: The model grape was located and measured within less than 1,100 mm from camera center, which means center between two cameras. The distance error of the calculated distance had the distance error within 5mm by using model image in the laboratory. The image processing system proved to be a reliable system for measuring the accurate distance between the camera center and the grape fruit. Also, difference between actual distance and calculated distance was found within 5 mm using stereo vision system in the field. Therefore, the image processing system would be mounted on a grape harvester to be founded to the position of the a grape fruit.

  • PDF

Image Processing System based on Deep Learning for Safety of Heat Treatment Equipment (열처리 장비의 Safety를 위한 딥러닝 기반 영상처리 시스템)

  • Lee, Jeong-Hoon;Lee, Ro-Woon;Hong, Seung-Taek;Kim, Young-Gon
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.6
    • /
    • pp.77-83
    • /
    • 2020
  • The heat treatment facility is in a situation where the scope of application of the remote IOT system is expanding due to the harsh environment caused by high heat and long working hours among the root industries. In this heat treatment process environment, the IOT middleware is required to play a pivotal role in interpreting, managing and controlling data information of IoT devices (sensors, etc.). Until now, the system controlled by the heat treatment remotely was operated with the command of the operator's batch system without overall monitoring of the site situation. However, for the safety and precise control of the heat treatment facility, it is necessary to control various sensors and recognize the surrounding work environment. As a solution to this, the heat treatment safety support system presented in this paper proposes a support system that can detect the access of the work manpower to the heat treatment furnace through thermal image detection and operate safely when ordering work from a remote location. In addition, an OPEN CV-based deterioration analysis system using DNN deep learning network was constructed for faster and more accurate recognition than general fixed hot spot monitoring-based thermal image analysis. Through this, we would like to propose a system that can be used universally in the heat treatment environment and support the safety management specialized in the heat treatment industry.

Ontology Modeling and Rule-based Reasoning for Automatic Classification of Personal Media (미디어 영상 자동 분류를 위한 온톨로지 모델링 및 규칙 기반 추론)

  • Park, Hyun-Kyu;So, Chi-Seung;Park, Young-Tack
    • Journal of KIISE
    • /
    • v.43 no.3
    • /
    • pp.370-379
    • /
    • 2016
  • Recently personal media were produced in a variety of ways as a lot of smart devices have been spread and services using these data have been desired. Therefore, research has been actively conducted for the media analysis and recognition technology and we can recognize the meaningful object from the media. The system using the media ontology has the disadvantage that can't classify the media appearing in the video because of the use of a video title, tags, and script information. In this paper, we propose a system to automatically classify video using the objects shown in the media data. To do this, we use a description logic-based reasoning and a rule-based inference for event processing which may vary in order. Description logic-based reasoning system proposed in this paper represents the relation of the objects in the media as activity ontology. We describe how to another rule-based reasoning system defines an event according to the order of the inference activity and order based reasoning system automatically classify the appropriate event to the category. To evaluate the efficiency of the proposed approach, we conducted an experiment using the media data classified as a valid category by the analysis of the Youtube video.

Limitations of neurobiological approach convergent to neuropsychiatry: DCD and two visual systems theory (신경정신학에 융복합되는 신경생물학적 접근법의 한계점: 발달성 협응장애와 두 시각 이론에 관한 종설)

  • Lee, Young-Lim
    • Journal of Digital Convergence
    • /
    • v.13 no.6
    • /
    • pp.225-234
    • /
    • 2015
  • Neurobiological approach helps to resolve the mind-body dualism and develop new assessment and treatment approaches in psychiatry. However, it could be a problem to place too much emphasis on certain aspects of neurobiology, specifically structural neuroanatomy, because of the complexity or comorbidity of neuropsychiatric disorders. Developmental Coordination Disorder (DCD), for instance, is generally related to problems in motor skills and this movement disability is often related to perception. One account, two visual systems theory, relied on functional distinction in brain; ventral stream is responsible for visual recognition, and dorsal stream is responsible for the guidance of actions. However, Studies are now showing that shape perception is relevant to visually guided action, such as reaching-to-grasp an object. In this article, I reviewed fundamental findings of two-visual system theory and suggested problems of visually guided action to consider what shape perception implies for the two visual systems. Questions raised highlight possible limitations of adopting a structural neuroanatomical approach to account for perception and action effects, and by extent related psychiatric conditions such as DCD. In conclusion, neurobiological approach converging to neuropsychiatry, while useful, would be limited if it focuses too much on anatomical distinction.

A Study on Virtual Environment Platform for Autonomous Tower Crane (타워크레인 자율화를 위한 가상환경 플랫폼 개발에 관한 연구)

  • Kim, Myeongjun;Yoon, Inseok;Kim, Namkyoun;Park, Moonseo;Ahn, Changbum;Jung, Minhyuk
    • Korean Journal of Construction Engineering and Management
    • /
    • v.23 no.4
    • /
    • pp.3-14
    • /
    • 2022
  • Autonomous equipment requires a large amount of data from various environments. However, it takes a lot of time and cost for an experiment in a real construction sites, which are difficulties in data collection and processing. Therefore, this study aims to develop a virtual environment for autonomous tower cranes technology development and validation. The authors defined automation functions and operation conditions of tower cranes with three performance criteria: operational design domain, object and event detection and response, and minimum functional conditions. Afterward, this study developed a virtual environment for learning and validation for autonomous functions such as recognition, decision making, and control using the Unity game engine. Validation was conducted by construction industry experts with a fidelity which is the representative matrix for virtual environment assessment. Through the virtual environment platform developed in this study, it will be possible to reduce the cost and time for data collection and technology development. Also, it is also expected to contribute to autonomous driving for not only tower cranes but also other construction equipment.

Implementation of A Vibration Notification System to Support Driving for Drivers with Cognitive Delay Impairment

  • Gyu-Seok Lee;Tae-Sung Kim;Myeong-Chul Park
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.4
    • /
    • pp.115-123
    • /
    • 2024
  • In this paper, we propose a vibration notification system that combines navigation information and wearable bands to ensure safe driving for the transportation vulnerable. This system transmits navigation driving information to a linked application, converts it into a vibration signal, and provides notifications through a wearable band. Existing navigation systems focus on providing route guidance and location information, so the driver's concentration is dispersed, and safety and convenience are deteriorated, especially for those with mobility impairments, due to standard vision and delayed recognition of stimuli, resulting in an increasingly high traffic accident rate. To solve this problem, navigation driving information is converted into vibration signals through a linked application, and vibration notifications for events, left turns, right turns, and speeding are provided through a wearable band to ensure driver safety and convenience. In the future, we will use cameras and vehicle sensors to increase awareness of safety inside and outside the vehicle by adding a function that provides notifications with vibration and LED when the vehicle approaches or recognizes an object, and we will continue to conduct research to build a safer driving environment. plan.