• 제목/요약/키워드: Scene Recognition

검색결과 193건 처리시간 0.028초

중년 및 고령운전자의 운전장면 개별요소에 대한 재인기억 차이: 민감도, 반응기준 및 반응시간 비교 (Comparisons of Middle- and Old-Aged Drivers' Recognition for Driving Scene Elements using Sensitivity, Response Bias, and Response Time)

  • 이재식
    • Journal of the Korean Data Analysis Society
    • /
    • 제20권6호
    • /
    • pp.3185-3199
    • /
    • 2018
  • 본 연구에서는 운전장면 개별요소들에 대한 중년 및 고령운전자의 작업기억의 특성을 반응민감도, 반응편파 및 반응시간을 통해 비교하였다. 본 연구의 결과를 요약하면 다음과 같다. 첫째, 고령운전자 집단은 중년운전자 집단에 비해 민감도가 낮았고 특히 중년운전자 집단의 경우는 중심자극과 주변자극에 대한 민감도에서의 차이가 유의하지 않았던 반면, 고령운전자 집단의 경우는 중심자극에 비해 주변자극에 대한 민감도가 현저하게 더 낮았다. 둘째, 중년운전자 집단은 전반적으로 보수적으로 반응한 반면 고령운전자 집단은 대체적으로 관대한 방향으로 반응하는 경향이 상대적으로 더 컸다. 셋째, 중년운전자들에 비해 고령운전자들의 반응시간이 더 느렸고, 중심자극보다는 주변자극에 대한 반응시간이 더 느렸으며, 반응의 유형에 따라서는 적중의 반응시간이 가장 빨랐던 반면 헛경보의 반응시간이 가장 느렸다. 특히, 탈루과 정기각 반응에 대해서는 두 집단 사이의 반응시간 차이가 더 증가하였다. 이러한 결과는 중년운전자들에 비해 고령운전자들은 운전장면 개별요소에 대한 작업기억에서의 파지가 상대적으로 더 불완전하고, 이에 따라 작업기억에 기초한 신속한 판단에서 어려움이 증가할 수 있음을 시사한다.

Touch TT: Scene Text Extractor Using Touchscreen Interface

  • Jung, Je-Hyun;Lee, Seong-Hun;Cho, Min-Su;Kim, Jin-Hyung
    • ETRI Journal
    • /
    • 제33권1호
    • /
    • pp.78-88
    • /
    • 2011
  • In this paper, we present the Touch Text exTractor (Touch TT), an interactive text segmentation tool for the extraction of scene text from camera-based images. Touch TT provides a natural interface for a user to simply indicate the location of text regions with a simple touchline. Touch TT then automatically estimates the text color and roughly locates the text regions. By inferring text characteristics from the estimated text color and text region, Touch TT can extract text components. Touch TT can also handle partially drawn lines which cover only a small section of text area. The proposed system achieves reasonable accuracy for text extraction from moderately difficult examples from the ICDAR 2003 database and our own database.

자동 잔향 편집을 위한 컬러 및 깊이 정보 기반 실내 장면 분류 (Indoor Scene Classification based on Color and Depth Images for Automated Reverberation Sound Editing)

  • 정민혁;유용현;박성준;황승준;백중환
    • 한국정보통신학회논문지
    • /
    • 제24권3호
    • /
    • pp.384-390
    • /
    • 2020
  • 영화나 VR 콘텐츠 제작 시 음향에 잔향 효과를 주는 것은 현장감과 생동감을 느끼게 하는데 매우 중요한 요소이다. 공간에 따른 음향의 잔향 시간은 RT60(Reverberation Time 60dB)이라는 표준에서 권고된다. 본 논문에서는 음향 편집 시 자동 잔향 편집을 위한 장면 인식 기법을 제안한다. 이를 위해 컬러 이미지와 예측된 깊이 이미지를 동일한 모델에 독립적으로 학습하는 분류 모델을 설계하였다. 실내 장면 분류는 내부 구조가 유사한 클래스가 존재하여 컬러 정보 학습만으로는 인식률의 한계가 존재한다. 공간의 깊이 정보를 사용하기 위해 딥러닝 기반의 깊이 정보 추출 기술을 사용하였다. RT60을 기반으로 총 10개의 장면 클래스를 구성하고 모델 학습 및 평가를 진행하였다. 최종적으로 제안하는 SCR+DNet(Scene Classification for Reverb+Depth Net) 분류기는 92.4%의 정확도로 기존의 CNN 분류기들보다 더 높은 성능을 달성하였다.

Robust Sign Recognition System at Subway Stations Using Verification Knowledge

  • Lee, Dongjin;Yoon, Hosub;Chung, Myung-Ae;Kim, Jaehong
    • ETRI Journal
    • /
    • 제36권5호
    • /
    • pp.696-703
    • /
    • 2014
  • In this paper, we present a walking guidance system for the visually impaired for use at subway stations. This system, which is based on environmental knowledge, automatically detects and recognizes both exit numbers and arrow signs from natural outdoor scenes. The visually impaired can, therefore, utilize the system to find their own way (for example, using exit numbers and the directions provided) through a subway station. The proposed walking guidance system consists mainly of three stages: (a) sign detection using the MCT-based AdaBoost technique, (b) sign recognition using support vector machines and hidden Markov models, and (c) three verification techniques to discriminate between signs and non-signs. The experimental results indicate that our sign recognition system has a high performance with a detection rate of 98%, a recognition rate of 99.5%, and a false-positive error rate of 0.152.

A Survey of Human Action Recognition Approaches that use an RGB-D Sensor

  • Farooq, Adnan;Won, Chee Sun
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제4권4호
    • /
    • pp.281-290
    • /
    • 2015
  • Human action recognition from a video scene has remained a challenging problem in the area of computer vision and pattern recognition. The development of the low-cost RGB depth camera (RGB-D) allows new opportunities to solve the problem of human action recognition. In this paper, we present a comprehensive review of recent approaches to human action recognition based on depth maps, skeleton joints, and other hybrid approaches. In particular, we focus on the advantages and limitations of the existing approaches and on future directions.

동적 도시 환경에서 의미론적 시각적 장소 인식 (Semantic Visual Place Recognition in Dynamic Urban Environment)

  • 사바 아르샤드;김곤우
    • 로봇학회논문지
    • /
    • 제17권3호
    • /
    • pp.334-338
    • /
    • 2022
  • In visual simultaneous localization and mapping (vSLAM), the correct recognition of a place benefits in relocalization and improved map accuracy. However, its performance is significantly affected by the environmental conditions such as variation in light, viewpoints, seasons, and presence of dynamic objects. This research addresses the problem of feature occlusion caused by interference of dynamic objects leading to the poor performance of visual place recognition algorithm. To overcome the aforementioned problem, this research analyzes the role of scene semantics in correct detection of a place in challenging environments and presents a semantics aided visual place recognition method. Semantics being invariant to viewpoint changes and dynamic environment can improve the overall performance of the place matching method. The proposed method is evaluated on the two benchmark datasets with dynamic environment and seasonal changes. Experimental results show the improved performance of the visual place recognition method for vSLAM.

무인 자동차의 주변 환경 인식을 위한 도시 환경에서의 그래프 기반 물체 분할 방법 (Graph-based Segmentation for Scene Understanding of an Autonomous Vehicle in Urban Environments)

  • 서보길;최윤근;노현철;정명진
    • 로봇학회논문지
    • /
    • 제9권1호
    • /
    • pp.1-10
    • /
    • 2014
  • In recent years, the research of 3D mapping technique in urban environments obtained by mobile robots equipped with multiple sensors for recognizing the robot's surroundings is being studied actively. However, the map generated by simple integration of multiple sensors data only gives spatial information to robots. To get a semantic knowledge to help an autonomous mobile robot from the map, the robot has to convert low-level map representations to higher-level ones containing semantic knowledge of a scene. Given a 3D point cloud of an urban scene, this research proposes a method to recognize the objects effectively using 3D graph model for autonomous mobile robots. The proposed method is decomposed into three steps: sequential range data acquisition, normal vector estimation and incremental graph-based segmentation. This method guarantees the both real-time performance and accuracy of recognizing the objects in real urban environments. Also, it can provide plentiful data for classifying the objects. To evaluate a performance of proposed method, computation time and recognition rate of objects are analyzed. Experimental results show that the proposed method has efficiently in understanding the semantic knowledge of an urban environment.

Salient Region Extraction based on Global Contrast Enhancement and Saliency Cut for Image Information Recognition of the Visually Impaired

  • Yoon, Hongchan;Kim, Baek-Hyun;Mukhriddin, Mukhiddinov;Cho, Jinsoo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제12권5호
    • /
    • pp.2287-2312
    • /
    • 2018
  • Extracting key visual information from images containing natural scene is a challenging task and an important step for the visually impaired to recognize information based on tactile graphics. In this study, a novel method is proposed for extracting salient regions based on global contrast enhancement and saliency cuts in order to improve the process of recognizing images for the visually impaired. To accomplish this, an image enhancement technique is applied to natural scene images, and a saliency map is acquired to measure the color contrast of homogeneous regions against other areas of the image. The saliency maps also help automatic salient region extraction, referred to as saliency cuts, and assist in obtaining a binary mask of high quality. Finally, outer boundaries and inner edges are detected in images with natural scene to identify edges that are visually significant. Experimental results indicate that the method we propose in this paper extracts salient objects effectively and achieves remarkable performance compared to conventional methods. Our method offers benefits in extracting salient objects and generating simple but important edges from images containing natural scene and for providing information to the visually impaired.

겹쳐진 물체의 인식을 위한 정합 알고리즘 (ON THE MATCHING ALGORITHM FOR THE RECOGNITION OF THE OCCLUDED OBJECTS)

  • 남기곤;박의열;이양성
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 1988년도 전기.전자공학 학술대회 논문집
    • /
    • pp.671-674
    • /
    • 1988
  • This paper describes a matching method to solve the problem of occlusion in a two dimensional scene. The technique consist of three steps: generation of hypotheses, clustering of hypotheses by matching probability, updating of hypotheses. Using this algorithm, simulation results have been tested for 20 scenes contained the 80 models, and have obtained 95% of properly correct recognition rate in average.

  • PDF

비디오에서 양방향 문맥 정보를 이용한 상호 협력적인 위치 및 물체 인식 (Collaborative Place and Object Recognition in Video using Bidirectional Context Information)

  • 김성호;권인소
    • 로봇학회논문지
    • /
    • 제1권2호
    • /
    • pp.172-179
    • /
    • 2006
  • In this paper, we present a practical place and object recognition method for guiding visitors in building environments. Recognizing places or objects in real world can be a difficult problem due to motion blur and camera noise. In this work, we present a modeling method based on the bidirectional interaction between places and objects for simultaneous reinforcement for the robust recognition. The unification of visual context including scene context, object context, and temporal context is also. The proposed system has been tested to guide visitors in a large scale building environment (10 topological places, 80 3D objects).

  • PDF