• Title/Summary/Keyword: 특징 집합 선택

Search Result 112, Processing Time 0.024 seconds

An Adaptive Method For Face Recognition Based Filters and Selection of Features (필터 및 특징 선택 기반의 적응형 얼굴 인식 방법)

  • Cho, Byoung-Mo;Kim, Gi-Han;Rhee, Phill-Kyu
    • The Journal of the Korea Contents Association
    • /
    • v.9 no.6
    • /
    • pp.1-8
    • /
    • 2009
  • There are a lot of influences, such as location of camera, luminosity, brightness, and direction of light, which affect the performance of 2-dimensional image recognition. This paper suggests an adaptive method for face-image recognition in noisy environments using evolvable filtering and feature extraction which uses one sample image from camera. This suggested method consists of two main parts. One is the environmental-adjustment module which determines optimum sets of filters, filter parameters, and dimensions of features by using "steady state genetic algorithm". The other another part is for face recognition module which performs recognition of face-image using the previous results. In the processing, we used Gabor wavelet for extracting features in the images and k-Nearest Neighbor method for the classification. For testing of the adaptive face recognition method, we tested the adaptive method in the brightness noise, in the impulse noise and in the composite noise and verified that the adaptive method protects face recognition-rate's rapidly decrease which can be occurred generally in the noisy environments.

Extraction of Classes and Hierarchy from Procedural Software (절차지향 소프트웨어로부터 클래스와 상속성 추출)

  • Choi, Jeong-Ran;Park, Sung-Og;Lee, Moon-Kun
    • Journal of KIISE:Software and Applications
    • /
    • v.28 no.9
    • /
    • pp.612-628
    • /
    • 2001
  • This paper presents a methodology to extract classes and inheritance relations from procedural software. The methodology is based on the idea of generating all groups of class candidates, based on the combinatorial groups of object candidates, and their inheritance with all possible combinations and selecting a group of object candidates, and their inheritance with all possible combinations and selecting a group with the best or optimal combination of candidates with respect to the degree of relativity and similarity between class candidates in the group and classes in a domain model. The methodology has innovative features in class candidates in the group and classes in a domain model. The methodology has innovative features in class and inheritance extraction: a clustering method based on both static (attribute) and dynamic (method) clustering, the combinatorial cases of grouping class candidate cases based on abstraction, a signature similarity measurement for inheritance relations among n class candidates or m classes, two-dimensional similarity measurement for inheritance relations among n class candidates or m classes, two-dimensional similarity measurement, that is, the horizontal measurement for overall group similarity between n class candidates and m classes, and the vertical measurement for specific similarity between a set of classes in a group of class candidates and a set of classes with the same class hierarchy in a domain model, etc. This methodology provides reengineering experts with a comprehensive and integrated environment to select the best or optimal group of class candidates.

  • PDF

Generation Method of Expression Space for Realtime Facial Expression Control of 3D Avatar (3차원 아바타의 실시간 얼굴표정제어를 위한 표정공간의 생성방법)

  • Kim Sung-Ho
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2006.06a
    • /
    • pp.109-111
    • /
    • 2006
  • 본 논문은 애니메이터로 하여금 얼굴 표정들의 공간으로부터 일련의 표정을 선택하게 함으로써 3차원 아바타의 표정을 실시간적으로 제어할 수 있도록 하기 위한 표정공간의 생성방법에 관하여 기술한다. 본 시스템에서는 약 2400여개의 얼굴 표정 프레임을 이용하여 표정공간을 구성하였다. 본 기법에서는 한 표정을 표시하는 상태표현으로 얼굴특징 점들 간의 상호거리를 표시하는 거리행렬을 사용한다. 이 거리행렬의 집합을 표정공간으로 한다. 그러나 이 표정공간은 한 표정이 다른 표정까지 이동할 때 두 표정간의 직선경로를 통해 이동할 수 있는 그런 공간이 아니다. 본 기법에서는 한 표정에서 다른 표정까지 거쳐 갈 수 있는 경로를 캡쳐된 표정 데이터로부터 근사적으로 유추한다. 우선, 각 표정상태를 표현하는 거리행렬간의 거리가 일정 값 이하인 경우 두 표정을 인접해 있다고 간주한다. 임의의 두 표정 상태가 일련의 인접표정들의 집합으로 연결되어 있으면 두 표정간에 경로가 존재한다고 간주한다. 한 표정에서 다른 표정으로 변화할 때 두 표정간의 최단경로를 통해 이동한다고 가정한다. 두 표정간의 최단거리를 구하기 위해 다이내믹 프로그래밍 기법을 이용한다. 이 거리행렬의 집합인 표정공간은 다차원 공간이다. 3차원 아바타의 얼굴 표정 제어는 애니메이터들이 표정공간을 항해할 때 실시간적으로 수행된다. 이를 도와주기 위해 표정공간을 차원 스케일링 기법을 이용하여 2차원 공간으로 가시화하였고, 애니메이터들로 하여금 본 시스템을 사용하여 실시간 표정 제어를 수행하게 했는데, 본 논문은 그 결과를 평가한다.참여하는 빈들 간의 관계를 분석하여 워크플로우에 대한 성능 측정이 가능하도록 한다. 또한 제안된 메트릭을 통하여 EJB 어플리케이션의 성능 향상을 도모할 수 있도록 한다.로 표면 위로 자라났고, 부종은 창상 밑 조직까지 감소하였으며, 육아조직은 교원질 섬유로 대체되었다. 창상 유발 21일 후, 다른 창상에 비해, HG 처치창은 유의적으로 창상 표면이 거의 재생성 상피로 덮였으며, 육아조직은 창상 유발 14일 후와 비교해서 유의적으로 교원질 섬유로 대체되었다. 위의 모든 결과에서 보듯이, 개에서 전층피부 창상의 처치 시 HG의 사용은 HC와 생리 식염수에 비해 창상치유 복구기의 치유 속도를 촉진하는 것으로 사료된다.시범학교의 자녀를 들 부모들은 환경관련문제에 대한 의식 및 환경관련 제품에 대한 구매행동의 변화가 두드러지게 나타났다.EX>$d^{2+}$ + SA처리구는 두 화합물의 이중 효과에 의해 전반적인 생리 활성을 억제하여, 결국에는 식물의 고사를 유도하는 것으로 사료된다.목에 대한 보안'이 가장 중요한 것으로 나타났다. 본 연구에서는 솔루션 선정요인에 관한 중요도를 e-마켓플레이스의 유형과 산업 별로 평가해보았는데, 여기에서 밝혀진 중요도를 통해 e 마켓플레이스를 구축하고 자 하는 기업은 솔루션을 자체 개발하거나 구입할 때 올바른 의사결정을 할 수 있다. 그리고 솔루션을 제공하려는 기업측면에서는 e-마켓플레이스를 구축하고자 하는 기업에게 유형과 산업별로 적절한 솔루션을 제공할 수 있다.순환이 뇌대사 및 수술 후 신경학적 결과에 보다 유익한 효과를 제공해 줄 수 있음

  • PDF

Facial Expression Control of 3D Avatar using Motion Data (모션 데이터를 이용한 3차원 아바타 얼굴 표정 제어)

  • Kim Sung-Ho;Jung Moon-Ryul
    • The KIPS Transactions:PartA
    • /
    • v.11A no.5
    • /
    • pp.383-390
    • /
    • 2004
  • This paper propose a method that controls facial expression of 3D avatar by having the user select a sequence of facial expressions in the space of facial expressions. And we setup its system. The space of expression is created from about 2400 frames consist of motion captured data of facial expressions. To represent the state of each expression, we use the distance matrix that represents the distances between pairs of feature points on the face. The set of distance matrices is used as the space of expressions. But this space is not such a space where one state can go to another state via the straight trajectory between them. We derive trajectories between two states from the captured set of expressions in an approximate manner. First, two states are regarded adjacent if the distance between their distance matrices is below a given threshold. Any two states are considered to have a trajectory between them If there is a sequence of adjacent states between them. It is assumed . that one states goes to another state via the shortest trajectory between them. The shortest trajectories are found by dynamic programming. The space of facial expressions, as the set of distance matrices, is multidimensional. Facial expression of 3D avatar Is controled in real time as the user navigates the space. To help this process, we visualized the space of expressions in 2D space by using the multidimensional scaling(MDS). To see how effective this system is, we had users control facial expressions of 3D avatar by using the system. As a result of that, users estimate that system is very useful to control facial expression of 3D avatar in real-time.

VILODE : A Real-Time Visual Loop Closure Detector Using Key Frames and Bag of Words (VILODE : 키 프레임 영상과 시각 단어들을 이용한 실시간 시각 루프 결합 탐지기)

  • Kim, Hyesuk;Kim, Incheol
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.5
    • /
    • pp.225-230
    • /
    • 2015
  • In this paper, we propose an effective real-time visual loop closure detector, VILODE, which makes use of key frames and bag of visual words (BoW) based on SURF feature points. In order to determine whether the camera has re-visited one of the previously visited places, a loop closure detector has to compare an incoming new image with all previous images collected at every visited place. As the camera passes through new places or locations, the amount of images to be compared continues growing. For this reason, it is difficult for a visual loop closure detector to meet both real-time constraint and high detection accuracy. To address the problem, the proposed system adopts an effective key frame selection strategy which selects and compares only distinct meaningful ones from continuously incoming images during navigation, and so it can reduce greatly image comparisons for loop detection. Moreover, in order to improve detection accuracy and efficiency, the system represents each key frame image as a bag of visual words, and maintains indexes for them using DBoW database system. The experiments with TUM benchmark datasets demonstrates high performance of the proposed visual loop closure detector.

Transfer Learning using Multiple ConvNet Layers Activation Features with Principal Component Analysis for Image Classification (전이학습 기반 다중 컨볼류션 신경망 레이어의 활성화 특징과 주성분 분석을 이용한 이미지 분류 방법)

  • Byambajav, Batkhuu;Alikhanov, Jumabek;Fang, Yang;Ko, Seunghyun;Jo, Geun Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.205-225
    • /
    • 2018
  • Convolutional Neural Network (ConvNet) is one class of the powerful Deep Neural Network that can analyze and learn hierarchies of visual features. Originally, first neural network (Neocognitron) was introduced in the 80s. At that time, the neural network was not broadly used in both industry and academic field by cause of large-scale dataset shortage and low computational power. However, after a few decades later in 2012, Krizhevsky made a breakthrough on ILSVRC-12 visual recognition competition using Convolutional Neural Network. That breakthrough revived people interest in the neural network. The success of Convolutional Neural Network is achieved with two main factors. First of them is the emergence of advanced hardware (GPUs) for sufficient parallel computation. Second is the availability of large-scale datasets such as ImageNet (ILSVRC) dataset for training. Unfortunately, many new domains are bottlenecked by these factors. For most domains, it is difficult and requires lots of effort to gather large-scale dataset to train a ConvNet. Moreover, even if we have a large-scale dataset, training ConvNet from scratch is required expensive resource and time-consuming. These two obstacles can be solved by using transfer learning. Transfer learning is a method for transferring the knowledge from a source domain to new domain. There are two major Transfer learning cases. First one is ConvNet as fixed feature extractor, and the second one is Fine-tune the ConvNet on a new dataset. In the first case, using pre-trained ConvNet (such as on ImageNet) to compute feed-forward activations of the image into the ConvNet and extract activation features from specific layers. In the second case, replacing and retraining the ConvNet classifier on the new dataset, then fine-tune the weights of the pre-trained network with the backpropagation. In this paper, we focus on using multiple ConvNet layers as a fixed feature extractor only. However, applying features with high dimensional complexity that is directly extracted from multiple ConvNet layers is still a challenging problem. We observe that features extracted from multiple ConvNet layers address the different characteristics of the image which means better representation could be obtained by finding the optimal combination of multiple ConvNet layers. Based on that observation, we propose to employ multiple ConvNet layer representations for transfer learning instead of a single ConvNet layer representation. Overall, our primary pipeline has three steps. Firstly, images from target task are given as input to ConvNet, then that image will be feed-forwarded into pre-trained AlexNet, and the activation features from three fully connected convolutional layers are extracted. Secondly, activation features of three ConvNet layers are concatenated to obtain multiple ConvNet layers representation because it will gain more information about an image. When three fully connected layer features concatenated, the occurring image representation would have 9192 (4096+4096+1000) dimension features. However, features extracted from multiple ConvNet layers are redundant and noisy since they are extracted from the same ConvNet. Thus, a third step, we will use Principal Component Analysis (PCA) to select salient features before the training phase. When salient features are obtained, the classifier can classify image more accurately, and the performance of transfer learning can be improved. To evaluate proposed method, experiments are conducted in three standard datasets (Caltech-256, VOC07, and SUN397) to compare multiple ConvNet layer representations against single ConvNet layer representation by using PCA for feature selection and dimension reduction. Our experiments demonstrated the importance of feature selection for multiple ConvNet layer representation. Moreover, our proposed approach achieved 75.6% accuracy compared to 73.9% accuracy achieved by FC7 layer on the Caltech-256 dataset, 73.1% accuracy compared to 69.2% accuracy achieved by FC8 layer on the VOC07 dataset, 52.2% accuracy compared to 48.7% accuracy achieved by FC7 layer on the SUN397 dataset. We also showed that our proposed approach achieved superior performance, 2.8%, 2.1% and 3.1% accuracy improvement on Caltech-256, VOC07, and SUN397 dataset respectively compare to existing work.

Real-time Color Recognition Based on Graphic Hardware Acceleration (그래픽 하드웨어 가속을 이용한 실시간 색상 인식)

  • Kim, Ku-Jin;Yoon, Ji-Young;Choi, Yoo-Joo
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.1
    • /
    • pp.1-12
    • /
    • 2008
  • In this paper, we present a real-time algorithm for recognizing the vehicle color from the indoor and outdoor vehicle images based on GPU (Graphics Processing Unit) acceleration. In the preprocessing step, we construct feature victors from the sample vehicle images with different colors. Then, we combine the feature vectors for each color and store them as a reference texture that would be used in the GPU. Given an input vehicle image, the CPU constructs its feature Hector, and then the GPU compares it with the sample feature vectors in the reference texture. The similarities between the input feature vector and the sample feature vectors for each color are measured, and then the result is transferred to the CPU to recognize the vehicle color. The output colors are categorized into seven colors that include three achromatic colors: black, silver, and white and four chromatic colors: red, yellow, blue, and green. We construct feature vectors by using the histograms which consist of hue-saturation pairs and hue-intensity pairs. The weight factor is given to the saturation values. Our algorithm shows 94.67% of successful color recognition rate, by using a large number of sample images captured in various environments, by generating feature vectors that distinguish different colors, and by utilizing an appropriate likelihood function. We also accelerate the speed of color recognition by utilizing the parallel computation functionality in the GPU. In the experiments, we constructed a reference texture from 7,168 sample images, where 1,024 images were used for each color. The average time for generating a feature vector is 0.509ms for the $150{\times}113$ resolution image. After the feature vector is constructed, the execution time for GPU-based color recognition is 2.316ms in average, and this is 5.47 times faster than the case when the algorithm is executed in the CPU. Our experiments were limited to the vehicle images only, but our algorithm can be extended to the input images of the general objects.

A Study of Textured Image Segmentation using Phase Information (페이즈 정보를 이용한 텍스처 영상 분할 연구)

  • Oh, Suk
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.2
    • /
    • pp.249-256
    • /
    • 2011
  • Finding a new set of features representing textured images is one of the most important studies in textured image analysis. This is because it is impossible to construct a perfect set of features representing every textured image, and it is inevitable to choose some relevant features which are efficient to on-going image processing jobs. This paper intends to find relevant features which are efficient to textured image segmentation. In this regards, this paper presents a different method for the segmentation of textured images based on the Gabor filter. Gabor filter is known to be a very efficient and effective tool which represents human visual system for texture analysis. Filtering a real-valued input image by the Gabor filter results in complex-valued output data defined in the spatial frequency domain. This complex value, as usual, gives the module and the phase. This paper focused its attention on the phase information, rather than the module information. In fact, the module information is considered very useful at region analysis in texture, while the phase information was considered almost of no use. But this paper shows that the phase information can also be fully useful and effective at region analysis in texture, once a good method introduced. We now propose "phase derivated method", which is an efficient and effective way to compute the useful phase information directly from the filtered value. This new method reduces effectively computing burden and widen applicable textured images.

Realtime Facial Expression Control of 3D Avatar by PCA Projection of Motion Data (모션 데이터의 PCA투영에 의한 3차원 아바타의 실시간 표정 제어)

  • Kim Sung-Ho
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.10
    • /
    • pp.1478-1484
    • /
    • 2004
  • This paper presents a method that controls facial expression in realtime of 3D avatar by having the user select a sequence of facial expressions in the space of facial expressions. The space of expression is created from about 2400 frames of facial expressions. To represent the state of each expression, we use the distance matrix that represents the distances between pairs of feature points on the face. The set of distance matrices is used as the space of expressions. Facial expression of 3D avatar is controled in real time as the user navigates the space. To help this process, we visualized the space of expressions in 2D space by using the Principal Component Analysis(PCA) projection. To see how effective this system is, we had users control facial expressions of 3D avatar by using the system. This paper evaluates the results.

  • PDF

A SVM-based Method for Classifying Tagged Web Resources using Tag Stability of Folksonomy in Categories (범주별 태그 안정성을 이용한 태그 부착 자원의 SVM 기반 분류 기법)

  • Koh, Byung-Gul;Lee, Kang-Pyo;Kim, Hyoung-Joo
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.6
    • /
    • pp.414-423
    • /
    • 2009
  • Folksonomy, which is collaborative classification created by freely selected keywords, is one of the driving factors of the web 2.0. Folksonomy has advantage of being built at low cost while its weakness is lack of hierarchical or systematic structure in comparison with taxonomy. If we can build classifier that is able to classify web resources from collective intelligence in taxonomy, we can build taxonomy at low cost. In this paper, targeting folksonomy in Slashdot.org, we define a general model and show that collective intelligence, which can build classifier, really exists in folksonomy using a stability value. We suggest method that builds SVM classifier using stability that is result from this collective intelligence. The experiment shows that our proposed method managed to build taxonomy from folksonomy with high accuracy.