• 제목/요약/키워드: visual features

검색결과 1,073건 처리시간 0.027초

말지각의 기초표상: 음소 또는 변별자질 (The Primitive Representation in Speech Perception: Phoneme or Distinctive Features)

  • 배문정
    • 말소리와 음성과학
    • /
    • 제5권4호
    • /
    • pp.157-169
    • /
    • 2013
  • Using a target detection task, this study compared the processing automaticity of phonemes and features in spoken syllable stimuli to determine the primitive representation in speech perception, phoneme or distinctive feature. For this, we modified the visual search task(Treisman et al., 1992) developed to investigate the processing of visual features(ex. color, shape or their conjunction) for auditory stimuli. In our task, the distinctive features(ex. aspiration or coronal) corresponded to visual primitive features(ex. color and shape), and the phonemes(ex. /$t^h$/) to visual conjunctive features(ex. colored shapes). The automaticity is measured by the set size effect that was the increasing amount of reaction time when the number of distracters increased. Three experiments were conducted. The laryngeal features(experiment 1), the manner features(experiment 2), and the place features(experiment 3) were compared with phonemes. The results showed that the distinctive features are consistently processed faster and automatically than the phonemes. Additionally there were differences in the processing automaticity among the classes of distinctive features. The laryngeal features are the most automatic, the manner features are moderately automatic and the place features are the least automatic. These results are consistent with the previous studies(Bae et al., 2002; Bae, 2010) that showed the perceptual hierarchy of distinctive features.

자율적인 시각 센서 피드백 기능을 갖는 원격 로보트 시스템교환 제어 (Traded control of telerobot system with an autonomous visual sensor feedback)

  • 김주곤;차동혁;김승호
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1996년도 한국자동제어학술회의논문집(국내학술편); 포항공과대학교, 포항; 24-26 Oct. 1996
    • /
    • pp.940-943
    • /
    • 1996
  • In teleoperating, as seeing the monitor screen obtained from a camera instituted in the working environment, human operator generally controls the slave arm. Because we can see only 2-D image in a monitor, human operator does not know the depth information and can not work with high accuracy. In this paper, we proposed a traded control method using an visual sensor for the purpose of solving this problem. We can control a teleoperation system with precision when we use the proposed algorithm. Not only a human operator command but also an autonomous visual sensor feedback command is given to a slave arm for the purpose of coincidence current image features and target image features. When the slave arm place in a distant place from the target position, human operator can know very well the difference between the desired image features and the current image features, but calculated visual sensor command have big errors. And when the slave arm is near the target position, the state of affairs is changed conversely. With this visual sensor feedback, human does not need coincide the detail difference between the desired image features and the current image features and proposed method can work with higher accuracy than other method without, sensor feedback. The effectiveness of the proposed control method is verified through series of experiments.

  • PDF

Video Captioning with Visual and Semantic Features

  • Lee, Sujin;Kim, Incheol
    • Journal of Information Processing Systems
    • /
    • 제14권6호
    • /
    • pp.1318-1330
    • /
    • 2018
  • Video captioning refers to the process of extracting features from a video and generating video captions using the extracted features. This paper introduces a deep neural network model and its learning method for effective video captioning. In this study, visual features as well as semantic features, which effectively express the video, are also used. The visual features of the video are extracted using convolutional neural networks, such as C3D and ResNet, while the semantic features are extracted using a semantic feature extraction network proposed in this paper. Further, an attention-based caption generation network is proposed for effective generation of video captions using the extracted features. The performance and effectiveness of the proposed model is verified through various experiments using two large-scale video benchmarks such as the Microsoft Video Description (MSVD) and the Microsoft Research Video-To-Text (MSR-VTT).

A Study on Feature-Based Visual Servoing Control of Robot System by Utilizing Redundant Feature

  • Han, Sung-Hyun;Hideki Hashimoto
    • Journal of Mechanical Science and Technology
    • /
    • 제16권6호
    • /
    • pp.762-769
    • /
    • 2002
  • This paper presents how effective it is to use many features for improving the speed and accuracy of visual servo systems. Some rank conditions which relate the image Jacobian to the control performance are derived. The focus is to describe that the accuracy of the camera position control in the world coordinate system is increased by utilizing redundant features in this paper. It is also proven that the accuracy is improved by increasing the number of features involved. Effectiveness of the redundant features is evaluated by the smallest singular value of the image Jacobian which is closely related to the accuracy with respect to the world coordinate system. Usefulness of the redundant features is verified by the real time experiments on a Dual-Arm robot manipulator made by Samsung Electronic Co. Ltd..

시계열 스트리트뷰 데이터베이스를 이용한 시각적 위치 인식 알고리즘 (Visual Location Recognition Using Time-Series Streetview Database)

  • 박천수;최준연
    • 반도체디스플레이기술학회지
    • /
    • 제18권4호
    • /
    • pp.57-61
    • /
    • 2019
  • Nowadays, portable digital cameras such as smart phone cameras are being popularly used for entertainment and visual information recording. Given a database of geo-tagged images, a visual location recognition system can determine the place depicted in a query photo. One of the most common visual location recognition approaches is the bag-of-words method where local image features are clustered into visual words. In this paper, we propose a new bag-of-words-based visual location recognition algorithm using time-series streetview database. The proposed algorithm selects only a small subset of image features which will be used in image retrieval process. By reducing the number of features to be used, the proposed algorithm can reduce the memory requirement of the image database and accelerate the retrieval process.

A Study on Visual Feedback Control of Industrial Articulated Robot

  • 심병균;이우송;박인만;황원준;최영식
    • 한국산업융합학회 논문집
    • /
    • 제17권1호
    • /
    • pp.27-34
    • /
    • 2014
  • This paper proposes a new approach to the designed of visual feedback control system based on visual servoing method. The main focus of this paper is presented how it is effective to use many features for improving the accuracy of the visual feedback control of industrial articulated robot for assembling and inspection of parts. Some rank conditions, which relate the image Jacobian, and the control performance are derived. It is also proven that the accuracy is improved by increasing the number of features. The effectiveness of redundant features is verified by the real time experiments on a SCARA type robot(FARA) made in samsung electronics company.

산업용 다관절 로봇의 비주얼 피드백 제어에 관한 연구 (A Study on Visual Feedback Control of Industrial Articulated Robot)

  • 심병균;한성현
    • 한국기계가공학회지
    • /
    • 제12권1호
    • /
    • pp.35-42
    • /
    • 2013
  • This paper proposes a new approach to the designed of visual feedback control system based on visual servoing method. The main focus of this paper is presented how it is effective to use many features for improving the accuracy of the visual feedback control of industrial articulated robot for assembling and inspection of parts. Some rank conditions, which relate the image Jacobian, and the control performance are derived. It is also proven that the accuracy is improved by increasing the number of features. The effectiveness of redundant features is verified by the real time experiments on a SCARA type robot(FARA) made in samsung electronics company.

고속 이미지 검색을 위한 2진 시각 단어 생성 기법 (Binary Visual Word Generation Techniques for A Fast Image Search)

  • 이수원
    • 정보과학회 논문지
    • /
    • 제44권12호
    • /
    • pp.1313-1318
    • /
    • 2017
  • 다수의 지역 특징들을 취합하여 하나의 벡터로 표현하는 것은 이미지 검색의 핵심 기술이다. 이 과정에서 경사도 기반 특징에 비해 수십 배 빠르게 추출되는 2진 특징이 활용된다면 이미지 검색의 고속화가 가능하다. 이를 위해서는 2진 특징들을 군집하여 2진 시각 단어를 생성하는 기법에 대한 연구가 선행되어야 한다. 기존의 경사도 기반 특징들을 군집하는 전통적인 방식으로는 2진 특징들을 군집할 수 없기 때문이다. 이를 위해 본 논문은 2진 특징들을 군집하여 2진 시각 단어를 생성하는 기법들에 대해 연구한다. 실험을 통해 2진 특징의 활용이 이미지 검색에 미치는 정확도와 연산효율 사이의 상충관계에 대해 분석한 후, 제안한 기법들을 비교한다. 본 연구는 고속 이미지 검색을 필요로 하는 모바일 응용, 리얼 타임 응용, 웹 스케일 응용 등에 활용될 것으로 기대된다.

Novel Intent based Dimension Reduction and Visual Features Semi-Supervised Learning for Automatic Visual Media Retrieval

  • kunisetti, Subramanyam;Ravichandran, Suban
    • International Journal of Computer Science & Network Security
    • /
    • 제22권6호
    • /
    • pp.230-240
    • /
    • 2022
  • Sharing of online videos via internet is an emerging and important concept in different types of applications like surveillance and video mobile search in different web related applications. So there is need to manage personalized web video retrieval system necessary to explore relevant videos and it helps to peoples who are searching for efficient video relates to specific big data content. To evaluate this process, attributes/features with reduction of dimensionality are computed from videos to explore discriminative aspects of scene in video based on shape, histogram, and texture, annotation of object, co-ordination, color and contour data. Dimensionality reduction is mainly depends on extraction of feature and selection of feature in multi labeled data retrieval from multimedia related data. Many of the researchers are implemented different techniques/approaches to reduce dimensionality based on visual features of video data. But all the techniques have disadvantages and advantages in reduction of dimensionality with advanced features in video retrieval. In this research, we present a Novel Intent based Dimension Reduction Semi-Supervised Learning Approach (NIDRSLA) that examine the reduction of dimensionality with explore exact and fast video retrieval based on different visual features. For dimensionality reduction, NIDRSLA learns the matrix of projection by increasing the dependence between enlarged data and projected space features. Proposed approach also addressed the aforementioned issue (i.e. Segmentation of video with frame selection using low level features and high level features) with efficient object annotation for video representation. Experiments performed on synthetic data set, it demonstrate the efficiency of proposed approach with traditional state-of-the-art video retrieval methodologies.

Implementation of Real Time Visual Servoing Control for Robot Manipulator

  • Han, Sung-Hyun;Jung, Ding-Yean;Kim, Hong-Rae;Hashmoto, Hideki
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2004년도 ICCAS
    • /
    • pp.1650-1654
    • /
    • 2004
  • This paper presents how it is effective to use many features for improving the speed and the accuracy of the visual servo systems. Some rank conditions which relate the image Jacobian and the control performance are derived. It is also proven that the accuracy is improved by increasing the number of features. Effectiveness of the redundant features is evaluated by the smallest singular value of the image Jacobian which is closely related to the accuracy with respect to the world coordinate system. Usefulness of the redundant features is verified by the real time experiments on a Dual-Arm Robot manipulator made in Samsung Electronic Co. Ltd.

  • PDF