• Title/Summary/Keyword: 모양기반 추출

Search Result 201, Processing Time 0.023 seconds

Measurement-based Face Rendering reflecting Positional Scattering Properties (위치별 산란특성을 반영한 측정기반 얼굴 렌더링)

  • Park, Sun-Yong;Oh, Kyoung-Su
    • Journal of Korea Game Society
    • /
    • v.9 no.5
    • /
    • pp.137-144
    • /
    • 2009
  • This paper predicts 6 facial regions that may have sharply different scattering properties, rendering the face more realistically based on their diffusion profiles. The scattering properties are acquired in the form of high dynamic range by photographing the pattern formed around an unit ray incident on facial skin. The acquired data are fitted to a 'linear combination of Gaussian functions', which well approximates the original diffusion profile of skin and has good characteristics as the filter. During the process, to prevent its solutions from converging into local minima, we take advantage of the genetic algorithm to set up the initial value. Each Gaussian term is applied to the irradiance map as a filter, expressing subsurface scattering effect. In this paper, to efficiently handle the maximum 12 Gaussian filterings, we make use of the parallel capacity of CUDA.

  • PDF

Position Estimation of Wheeled Mobile Robot in a Corridor Using Neural Network (신경망을 이용한 복도에서의 구륜이동로봇의 위치추정)

  • Choi, Kyung-Jin;Lee, Young-Hyun;Park, Chong-Kug
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.5
    • /
    • pp.577-582
    • /
    • 2004
  • This paper describes position estimation algorithm using neural network for the navigation of the vision-based Wheeled Mobile Robot (WMR) in a corridor with taking ceiling lamps as landmark. From images of a corridor the lamp's line on the ceiling in corridor has a specific slope to the lateral position of the WMR. The vanishing point produced by the lamp's line also has a specific position to the orientation of WMR. The ceiling lamps has a limited size and shape like a circle in image. Simple image processing algorithms are used to extract lamps from the corridor image. Then the lamp's line and vanishing point's position are defined and calculated at known position of WMR in a corridor To estimate the lateral position and orientation of WMR from an image, the relationship between the position of WMR and the features of ceiling lamps have to be defined. Data set between position of WMR and features of lamps are configured. Neural network are composed and teamed with data set. Back propagation algorithm(BPN) is used for learning. And it is applied in navigation of WMR in a corridor.

3D Model Retrieval Using Geometric Information (기하학 정보를 이용한 3차원 모델 검색)

  • Lee Kee-Ho;Kim Nac-Woo;Kim Tae-Yong;Choi Jong-Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.10C
    • /
    • pp.1007-1016
    • /
    • 2005
  • This paper presents a feature extraction method for shape based retrieval of 3D models. Since the feature descriptor of 3D model should be invariant to translation, rotation and scaling, it is necessary to preprocess the 3D models to represent them in a canonical coordinate system. We use the PCA(Principal Component Analysis) method to preprocess the 3D models. Also, we apply that to make a MBR(Minimum Boundary Rectangle) and a circumsphere. The proposed algorithm is as follows. We generate a circumsphere around 3D models, where radius equals 1(r=1) and locate each model in the center of the circumsphere. We produce the concentric spheres with a different radius($r_i=i/n,\;i=1,2,{\ldots},n$). After looking for meshes intersected with the concentric spheres, we compute the curvature of the meshes. We use these curvatures as the model descriptor. Experimental results numerically show the performance improvement of proposed algorithm from min. 0.1 to max. 0.6 in comparison with conventional methods by ANMRR, although our method uses .relatively small bins. This paper uses $R{^*}-tree$ as the indexing.

Text Detection and Recognition in Outdoor Korean Signboards for Mobile System Applications (모바일 시스템 응용을 위한 실외 한국어 간판 영상에서 텍스트 검출 및 인식)

  • Park, J.H.;Lee, G.S.;Kim, S.H.;Lee, M.H.;Toan, N.D.
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.46 no.2
    • /
    • pp.44-51
    • /
    • 2009
  • Text understand in natural images has become an active research field in the past few decades. In this paper, we present an automatic recognition system in Korean signboards with a complex background. The proposed algorithm includes detection, binarization and extraction of text for the recognition of shop names. First, we utilize an elaborate detection algorithm to detect possible text region based on edge histogram of vertical and horizontal direction. And detected text region is segmented by clustering method. Second, the text is divided into individual characters based on connected components whose center of mass lie below the center line, which are recognized by using a minimum distance classifier. A shape-based statistical feature is adopted, which is adequate for Korean character recognition. The system has been implemented in a mobile phone and is demonstrated to show acceptable performance.

Visual Landmark based Parking Assistance System in Constrained Environment (제한된 환경에서 시각적 랜드마크를 기반으로 한 주차 보조 시스템)

  • Park, Soon-Young;Song, Young-Sub;Kim, Hang-Joon
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.49 no.1
    • /
    • pp.31-40
    • /
    • 2012
  • This paper proposes a visual landmark, and presents a parking assistance system using the landmarks. The visual landmark is a feature corresponding to the parking slots, it must be selected considering the parking lot's environment. The parking lot has simple repetitive pattern environment without noticeable features. The previous landmarks are not proper to the parking lot's environment. We propose the visual landmark for this environment. We estimate the vehicle's location using the proposed landmarks, and expect the vehicle's trajectory according to the vehicle's state. The system's inputs are images from the camera fixed to the vehicle. The presented system estimates the vehicle's location using the input images, and assists a driver through displaying the expected vehicle's trajectory from the steering angle. The experimental results showed the proposed landmark's performance and the parking assistance system's performance.

Mobile Finger Signature Verification Robust to Skilled Forgery (모바일환경에서 위조서명에 강건한 딥러닝 기반의 핑거서명검증 연구)

  • Nam, Seng-soo;Seo, Chang-ho;Choi, Dae-seon
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.26 no.5
    • /
    • pp.1161-1170
    • /
    • 2016
  • In this paper, we provide an authentication technology for verifying dynamic signature made by finger on smart phone. In the proposed method, we are using the Auto-Encoder-based 1 class model in order to effectively distinguish skilled forgery signature. In addition to the basic dynamic signature characteristic information such as appearance and velocity of a signature, we use accelerometer value supported by most of the smartphone. Signed data is re-sampled to give the same length and is normalized to a constant size. We built a test set for evaluation and conducted experiment in three ways. As results of the experiment, the proposed acceleration sensor value and 1 class model shows 6.9% less EER than previous method.

Adverse Effects on EEGs and Bio-Signals Coupling on Improving Machine Learning-Based Classification Performances

  • SuJin Bak
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.10
    • /
    • pp.133-153
    • /
    • 2023
  • In this paper, we propose a novel approach to investigating brain-signal measurement technology using Electroencephalography (EEG). Traditionally, researchers have combined EEG signals with bio-signals (BSs) to enhance the classification performance of emotional states. Our objective was to explore the synergistic effects of coupling EEG and BSs, and determine whether the combination of EEG+BS improves the classification accuracy of emotional states compared to using EEG alone or combining EEG with pseudo-random signals (PS) generated arbitrarily by random generators. Employing four feature extraction methods, we examined four combinations: EEG alone, EG+BS, EEG+BS+PS, and EEG+PS, utilizing data from two widely-used open datasets. Emotional states (task versus rest states) were classified using Support Vector Machine (SVM) and Long Short-Term Memory (LSTM) classifiers. Our results revealed that when using the highest accuracy SVM-FFT, the average error rates of EEG+BS were 4.7% and 6.5% higher than those of EEG+PS and EEG alone, respectively. We also conducted a thorough analysis of EEG+BS by combining numerous PSs. The error rate of EEG+BS+PS displayed a V-shaped curve, initially decreasing due to the deep double descent phenomenon, followed by an increase attributed to the curse of dimensionality. Consequently, our findings suggest that the combination of EEG+BS may not always yield promising classification performance.

3D Models Retrieval Using Shape Index and Curvedness (형태 인덱스와 정규 곡률을 이용한 3차원 모델 검색)

  • Park, Ki-Tae;Hwang, Hae-Jung;Moon, Young-Shik
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.3
    • /
    • pp.33-41
    • /
    • 2007
  • Owing to the development of multimedia and communication technologies, multimedia data become a common feature of the information systems and are on the increase. This has led to the need of 3D shape retrieval systems that, given a query object, retrieve similar 3D objects. Therefore, shape descriptor required to describe a 3D object effectively and efficiently. In this paper, a new descriptor for 3D model retrieval based on shape information is proposed. The proposed descriptor utilizes the curvedness together with the shape index that provides local geometry information. The existing 3D Shape Spectrum Descriptor (3D SSD), which is defined as the histogram of shape index values, represents the characteristics of local shapes of the 3D surface. However, it does not properly represent the local shape characteristics, because many points with different curvedness may have the same shape index value. Therefore, we add a new feature that represents the degree of curvedness, thereby improving the discriminating power of the shape descriptor. We evaluate the performance of the proposed method, compared with the previous method. The experimental results have shown that the performance of retrieval has been improved by 23.6%.

Lip Contour Detection by Multi-Threshold (다중 문턱치를 이용한 입술 윤곽 검출 방법)

  • Kim, Jeong Yeop
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.9 no.12
    • /
    • pp.431-438
    • /
    • 2020
  • In this paper, the method to extract lip contour by multiple threshold is proposed. Spyridonos et. el. proposed a method to extract lip contour. First step is get Q image from transform of RGB into YIQ. Second step is to find lip corner points by change point detection and split Q image into upper and lower part by corner points. The candidate lip contour can be obtained by apply threshold to Q image. From the candidate contour, feature variance is calculated and the contour with maximum variance is adopted as final contour. The feature variance 'D' is based on the absolute difference near the contour points. The conventional method has 3 problems. The first one is related to lip corner point. Calculation of variance depends on much skin pixels and therefore the accuracy decreases and have effect on the split for Q image. Second, there is no analysis for color systems except YIQ. YIQ is a good however, other color systems such as HVS, CIELUV, YCrCb would be considered. Final problem is related to selection of optimal contour. In selection process, they used maximum of average feature variance for the pixels near the contour points. The maximum of variance causes reduction of extracted contour compared to ground contours. To solve the first problem, the proposed method excludes some of skin pixels and got 30% performance increase. For the second problem, HSV, CIELUV, YCrCb coordinate systems are tested and found there is no relation between the conventional method and dependency to color systems. For the final problem, maximum of total sum for the feature variance is adopted rather than the maximum of average feature variance and got 46% performance increase. By combine all the solutions, the proposed method gives 2 times in accuracy and stability than conventional method.

Story-based Information Retrieval (스토리 기반의 정보 검색 연구)

  • You, Eun-Soon;Park, Seung-Bo
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.81-96
    • /
    • 2013
  • Video information retrieval has become a very important issue because of the explosive increase in video data from Web content development. Meanwhile, content-based video analysis using visual features has been the main source for video information retrieval and browsing. Content in video can be represented with content-based analysis techniques, which can extract various features from audio-visual data such as frames, shots, colors, texture, or shape. Moreover, similarity between videos can be measured through content-based analysis. However, a movie that is one of typical types of video data is organized by story as well as audio-visual data. This causes a semantic gap between significant information recognized by people and information resulting from content-based analysis, when content-based video analysis using only audio-visual data of low level is applied to information retrieval of movie. The reason for this semantic gap is that the story line for a movie is high level information, with relationships in the content that changes as the movie progresses. Information retrieval related to the story line of a movie cannot be executed by only content-based analysis techniques. A formal model is needed, which can determine relationships among movie contents, or track meaning changes, in order to accurately retrieve the story information. Recently, story-based video analysis techniques have emerged using a social network concept for story information retrieval. These approaches represent a story by using the relationships between characters in a movie, but these approaches have problems. First, they do not express dynamic changes in relationships between characters according to story development. Second, they miss profound information, such as emotions indicating the identities and psychological states of the characters. Emotion is essential to understanding a character's motivation, conflict, and resolution. Third, they do not take account of events and background that contribute to the story. As a result, this paper reviews the importance and weaknesses of previous video analysis methods ranging from content-based approaches to story analysis based on social network. Also, we suggest necessary elements, such as character, background, and events, based on narrative structures introduced in the literature. We extract characters' emotional words from the script of the movie Pretty Woman by using the hierarchical attribute of WordNet, which is an extensive English thesaurus. WordNet offers relationships between words (e.g., synonyms, hypernyms, hyponyms, antonyms). We present a method to visualize the emotional pattern of a character over time. Second, a character's inner nature must be predetermined in order to model a character arc that can depict the character's growth and development. To this end, we analyze the amount of the character's dialogue in the script and track the character's inner nature using social network concepts, such as in-degree (incoming links) and out-degree (outgoing links). Additionally, we propose a method that can track a character's inner nature by tracing indices such as degree, in-degree, and out-degree of the character network in a movie through its progression. Finally, the spatial background where characters meet and where events take place is an important element in the story. We take advantage of the movie script to extracting significant spatial background and suggest a scene map describing spatial arrangements and distances in the movie. Important places where main characters first meet or where they stay during long periods of time can be extracted through this scene map. In view of the aforementioned three elements (character, event, background), we extract a variety of information related to the story and evaluate the performance of the proposed method. We can track story information extracted over time and detect a change in the character's emotion or inner nature, spatial movement, and conflicts and resolutions in the story.