• 제목/요약/키워드: Feature Fusion

검색결과 294건 처리시간 0.027초

다중 센서 융합 알고리즘을 이용한 사용자의 감정 인식 및 표현 시스템 (Emotion Recognition and Expression System of User using Multi-Modal Sensor Fusion Algorithm)

  • 염홍기;주종태;심귀보
    • 한국지능시스템학회논문지
    • /
    • 제18권1호
    • /
    • pp.20-26
    • /
    • 2008
  • 지능형 로봇이나 컴퓨터가 일상생활 속에서 차지하는 비중이 점점 높아짐에 따라 인간과의 상호교류도 점점 중요시되고 있다. 이렇게 지능형 로봇(컴퓨터) - 인간의 상호 교류하는데 있어서 감정 인식 및 표현은 필수라 할 수 있겠다. 본 논문에서는 음성 신호와 얼굴 영상에서 감정적인 특징들을 추출한 후 이것을 Bayesian Learning과 Principal Component Analysis에 적용하여 5가지 감정(평활, 기쁨, 슬픔, 화남, 놀람)으로 패턴을 분류하였다. 그리고 각각 매개체의 단점을 보완하고 인식률을 높이기 위해서 결정 융합 방법과 특징 융합 방법을 적용하여 감정 인식 실험을 하였다. 결정 융합 방법은 각각 인식 시스템을 통해 얻어진 인식 결과 값을 퍼지 소속 함수에 적용하여 감정 인식 실험을 하였으며, 특징 융합 방법은 SFS(Sequential Forward Selection) 특징 선택 방법을 통해 우수한 특징들을 선택한 후 MLP(Multi Layer Perceptron) 기반 신경망(Neural Networks)에 적용하여 감정 인식 실험을 실행하였다. 그리고 인식된 결과 값을 2D 얼굴 형태에 적용하여 감정을 표현하였다.

Video Expression Recognition Method Based on Spatiotemporal Recurrent Neural Network and Feature Fusion

  • Zhou, Xuan
    • Journal of Information Processing Systems
    • /
    • 제17권2호
    • /
    • pp.337-351
    • /
    • 2021
  • Automatically recognizing facial expressions in video sequences is a challenging task because there is little direct correlation between facial features and subjective emotions in video. To overcome the problem, a video facial expression recognition method using spatiotemporal recurrent neural network and feature fusion is proposed. Firstly, the video is preprocessed. Then, the double-layer cascade structure is used to detect a face in a video image. In addition, two deep convolutional neural networks are used to extract the time-domain and airspace facial features in the video. The spatial convolutional neural network is used to extract the spatial information features from each frame of the static expression images in the video. The temporal convolutional neural network is used to extract the dynamic information features from the optical flow information from multiple frames of expression images in the video. A multiplication fusion is performed with the spatiotemporal features learned by the two deep convolutional neural networks. Finally, the fused features are input to the support vector machine to realize the facial expression classification task. The experimental results on cNTERFACE, RML, and AFEW6.0 datasets show that the recognition rates obtained by the proposed method are as high as 88.67%, 70.32%, and 63.84%, respectively. Comparative experiments show that the proposed method obtains higher recognition accuracy than other recently reported methods.

Reliability improvement of nonlinear ultrasonic modulation based fatigue crack detection using feature-level data fusion

  • Lim, Hyung Jin;Kim, Yongtak;Sohn, Hoon;Jeon, Ikgeun;Liu, Peipei
    • Smart Structures and Systems
    • /
    • 제20권6호
    • /
    • pp.683-696
    • /
    • 2017
  • In this study, the reliability of nonlinear ultrasonic modulation based fatigue crack detection is improved using a feature-level data fusion approach. When two ultrasonic inputs at two distinct frequencies are applied to a specimen with a fatigue crack, modulation components at the summation and difference of these two input frequencies appear. First, the spectral amplitudes of the modulation components and their spectral correlations are defined as individual features. Then, a 2D feature space is constructed by combining these two features, and the presence of a fatigue crack is identified in the feature space. The effectiveness of the proposed fatigue crack detection technique is experimentally validated through cyclic loading tests of aluminum plates, full-scale steel girders and a rotating shaft component. Subsequently, the improved reliability of the proposed technique is quantitatively investigated using receiver operating characteristic analysis. The uniqueness of this study lies in (1) improvement of nonlinear ultrasonic modulation based fatigue crack detection reliability using feature-level data fusion, (2) reference-free fatigue crack diagnosis without using the baseline data obtained from the intact condition of the structure, (3) application to full-scale steel girders and shaft component, and (4) quantitative investigation of the improved reliability using receiver operating characteristic analysis.

특징 융합을 이용한 농작물 다중 분광 이미지의 의미론적 분할 (Semantic Segmentation of Agricultural Crop Multispectral Image Using Feature Fusion)

  • 문준렬;박성준;백중환
    • 한국항행학회논문지
    • /
    • 제28권2호
    • /
    • pp.238-245
    • /
    • 2024
  • 본 논문에서는 농작물 다중 분광 이미지에 대해 특징 융합 기법을 이용하여 의미론적 분할 성능을 향상시키기 위한 프레임워크를 제안한다. 스마트팜 분야에서 연구 중인 딥러닝 기술 중 의미론적 분할 모델 대부분은 RGB(red-green-blue)로 학습을 진행하고 있고 성능을 높이기 위해 모델의 깊이와 복잡성을 증가시키는 데에 집중하고 있다. 본 연구는 기존 방식과 달리 다중 분광과 어텐션 메커니즘을 통해 모델을 최적화하여 설계한다. 제안하는 방식은 RGB 단일 이미지와 함께 UAV (unmanned aerial vehicle)에서 수집된 여러 채널의 특징을 융합하여 특징 추출 성능을 높이고 상호보완적인 특징을 인식하여 학습 효과를 증대시킨다. 특징 융합에 집중할 수 있도록 모델 구조를 개선하고, 작물 이미지에 유리한 채널 및 조합을 실험하여 다른 모델과의 성능을 비교한다. 실험 결과 RGB와 NDVI (normalized difference vegetation index)가 융합된 모델이 다른 채널과의 조합보다 성능이 우수함을 보였다.

A Survey of Fusion Techniques for Multi-spectral Images

  • Achalakul, Tiranee
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2002년도 ITC-CSCC -2
    • /
    • pp.1244-1247
    • /
    • 2002
  • This paper discusses various algorithms to the fusion of multi-spectral image. These fusion techniques have a wide variety of applications that range from hospital pathology to battlefield management. Different algorithms in each fusion level, namely data, feature, and decision are compared. The PCT-Based algorithm, which has the characteristic of data compression, is described. The algorithm is experimented on a foliated aerial scene and the fusion result is presented.

  • PDF

Finger Vein Recognition based on Matching Score-Level Fusion of Gabor Features

  • Lu, Yu;Yoon, Sook;Park, Dong Sun
    • 한국통신학회논문지
    • /
    • 제38A권2호
    • /
    • pp.174-182
    • /
    • 2013
  • Most methods for fusion-based finger vein recognition were to fuse different features or matching scores from more than one trait to improve performance. To overcome the shortcomings of "the curse of dimensionality" and additional running time in feature extraction, in this paper, we propose a finger vein recognition technology based on matching score-level fusion of a single trait. To enhance the quality of finger vein image, the contrast-limited adaptive histogram equalization (CLAHE) method is utilized and it improves the local contrast of normalized image after ROI detection. Gabor features are then extracted from eight channels based on a bank of Gabor filters. Instead of using the features for the recognition directly, we analyze the contributions of Gabor feature from each channel and apply a weighted matching score-level fusion rule to get the final matching score, which will be used for the last recognition. Experimental results demonstrate the CLAHE method is effective to enhance the finger vein image quality and the proposed matching score-level fusion shows better recognition performance.

Virtual Environment Building and Navigation of Mobile Robot using Command Fusion and Fuzzy Inference

  • Jin, Taeseok
    • 한국산업융합학회 논문집
    • /
    • 제22권4호
    • /
    • pp.427-433
    • /
    • 2019
  • This paper propose a fuzzy inference model for map building and navigation for a mobile robot with an active camera, which is intelligently navigating to the goal location in unknown environments using sensor fusion, based on situational command using an active camera sensor. Active cameras provide a mobile robot with the capability to estimate and track feature images over a hallway field of view. In this paper, instead of using "physical sensor fusion" method which generates the trajectory of a robot based upon the environment model and sensory data. Command fusion method is used to govern the robot navigation. The navigation strategy is based on the combination of fuzzy rules tuned for both goal-approach and obstacle-avoidance. To identify the environments, a command fusion technique is introduced, where the sensory data of active camera sensor for navigation experiments are fused into the identification process. Navigation performance improves on that achieved using fuzzy inference alone and shows significant advantages over command fusion techniques. Experimental evidences are provided, demonstrating that the proposed method can be reliably used over a wide range of relative positions between the active camera and the feature images.

센서 융합을 통한 환경지도 기반의 강인한 전역 위치추정 (Robust Global Localization based on Environment map through Sensor Fusion)

  • 정민국;송재복
    • 로봇학회논문지
    • /
    • 제9권2호
    • /
    • pp.96-103
    • /
    • 2014
  • Global localization is one of the essential issues for mobile robot navigation. In this study, an indoor global localization method is proposed which uses a Kinect sensor and a monocular upward-looking camera. The proposed method generates an environment map which consists of a grid map, a ceiling feature map from the upward-looking camera, and a spatial feature map obtained from the Kinect sensor. The method selects robot pose candidates using the spatial feature map and updates sample poses by particle filter based on the grid map. Localization success is determined by calculating the matching error from the ceiling feature map. In various experiments, the proposed method achieved a position accuracy of 0.12m and a position update speed of 10.4s, which is robust enough for real-world applications.

다중 신경회로망을 이용한 특징정보 융합과 적외선영상에서의 표적식별에의 응용 (Feature information fusion using multiple neural networks and target identification application of FLIR image)

  • 선선구;박현욱
    • 대한전자공학회논문지SP
    • /
    • 제40권4호
    • /
    • pp.266-274
    • /
    • 2003
  • 전방 관측 적외선 영상에서 가려짐이 없는 표적과 부분적으로 가려진 표적을 식별하기 위해 국부적 표적 경계선에 대한 거리함수의 푸리에기술자와 다중의 다층 퍼셉트론을 사용한 특징정보 융합 방법을 제안한다. 표적을 배경으로부터 분리한 후에 표적 경계선의 중심을 기준으로 푸리에 기술자를 구해 전역적 특징으로 사용한다. 국부적인 형상 특징을 찾기 위해 표적 경계선을 분할하여 4개의 국부적 경계선을 만들고, 각 국부적 경계선에서 두 개의 극단점이 이루는 직선과 경계선 픽셀로부터 거리함수를 정의한다. 거리함수에 대한 푸리에 기술자를 국부적 형상특징으로 사용한다. 1개의 광역적 특징 백터와 4개의 국부적 특징 백터를 정의하고 다중의 다층 퍼셉트론을 사용하여 특징정보들을 융합함으로써 최종 표적식별 결과를 얻는다. 실험을 통해 기존의 특징벡터들에 의한 표적식별 방법과 비교하여 제안한 방법의 우수성을 입증한다.

A Multimodal Fusion Method Based on a Rotation Invariant Hierarchical Model for Finger-based Recognition

  • Zhong, Zhen;Gao, Wanlin;Wang, Minjuan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제15권1호
    • /
    • pp.131-146
    • /
    • 2021
  • Multimodal biometric-based recognition has been an active topic because of its higher convenience in recent years. Due to high user convenience of finger, finger-based personal identification has been widely used in practice. Hence, taking Finger-Print (FP), Finger-Vein (FV) and Finger-Knuckle-Print (FKP) as the ingredients of characteristic, their feature representation were helpful for improving the universality and reliability in identification. To usefully fuse the multimodal finger-features together, a new robust representation algorithm was proposed based on hierarchical model. Firstly, to obtain more robust features, the feature maps were obtained by Gabor magnitude feature coding and then described by Local Binary Pattern (LBP). Secondly, the LGBP-based feature maps were processed hierarchically in bottom-up mode by variable rectangle and circle granules, respectively. Finally, the intension of each granule was represented by Local-invariant Gray Features (LGFs) and called Hierarchical Local-Gabor-based Gray Invariant Features (HLGGIFs). Experiment results revealed that the proposed algorithm is capable of improving rotation variation of finger-pose, and achieving lower Equal Error Rate (EER) in our homemade database.