• Title/Summary/Keyword: 얼굴 영역 분할

Search Result 119, Processing Time 0.035 seconds

Wavelet-Based Face Recognition by Divided Area (웨이브렛을 이용한 공간적 영역분할에 의한 얼굴 인식)

  • 이성록;이상효;조창호;조도현;이상철
    • Proceedings of the IEEK Conference
    • /
    • 2003.07e
    • /
    • pp.2307-2310
    • /
    • 2003
  • In this paper, a method for face recognition based on the wavelet packet decomposition is proposed. In the proposed method, the input image is decomposed by the 2-level wavelet packet transformation and then the face areas are defined by the Integral Projection technique applied to each of the 1-level subband images, HL and LH. After the defined face areas are divided into three areas, called top, bottom, and border, the mean and the variance of the three areas of the approximation image are computed, and the variance of the single predetermined face area for the rest of 15 detail images, from which the feature vectors of statistical measure are extracted. In this paper we use the wavelet packet decomposition, a generalization of the classical wavelet decomposition, to obtain its richer signal analysis features such as discontinuity in higher derivatives, self-similarity, etc. And we have shown that even with very simple statistical features such as mean values and variance we can make an excellent basis for face classification, if an appropriate probability distance is used.

  • PDF

Posture Recognition for a Bi-directional Participatory TV Program based on Face Color Region and Motion Map (시청자 참여형 양방향 TV 방송을 위한 얼굴색 영역 및 모션맵 기반 포스처 인식)

  • Hwang, Sunhee;Lim, Kwangyong;Lee, Suwoong;Yoo, Hoyoung;Byun, Hyeran
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.8
    • /
    • pp.549-554
    • /
    • 2015
  • As intuitive hardware interfaces continue to be developed, it has become more important to recognize the posture of the user. An efficient alternative to adding expensive sensors is to implement computer vision systems. This paper proposes a method to recognize a user's postured in a live broadcast bi-directional participatory TV program. The proposed method first estimates the position of the user's hands by generation a facial color map for the user and a motion map. The posture is then recognized by computing the relative position of the face and the hands. This method exhibited 90% accuracy in an experiment to recognize three defined postures during the live broadcast bi-directional participatory TV program, even when the input images contained a complex background.

A Method of Auto Photography Composition Suggestion (사진의 자동 구도 보정 제시 기법)

  • Choi, Yong-Sub;Park, Dae-Hyun;Kim, Yoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.1
    • /
    • pp.9-21
    • /
    • 2014
  • In this paper, we propose the auto correction technique of photography composition by which the eye line is concentrated and the stable image of the structure can be obtained in case the general user takes a picture. Because the general user photographs in most case without background knowledge about the composition of the photo, the subject location is not appropriate and the unstable composition is contrasted with the stable composition of pictures which the experts take. Therefore, we provide not the method processing the image after photographing, but he method presenting automatically the stable composition when the general users take a photograph. The proposed method analyze the subject through Saliency Map, Image Segmentation, Edge Detection, etc. and outputs the subject at the location where the stable composition can be comprised along with the guideline of the Rule of Thirds. The experimental result shows that the good composition was presented to the user automatically.

Application of Support Vector Regression for Improving the Performance of the Emotion Prediction Model (감정예측모형의 성과개선을 위한 Support Vector Regression 응용)

  • Kim, Seongjin;Ryoo, Eunchung;Jung, Min Kyu;Kim, Jae Kyeong;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.185-202
    • /
    • 2012
  • .Since the value of information has been realized in the information society, the usage and collection of information has become important. A facial expression that contains thousands of information as an artistic painting can be described in thousands of words. Followed by the idea, there has recently been a number of attempts to provide customers and companies with an intelligent service, which enables the perception of human emotions through one's facial expressions. For example, MIT Media Lab, the leading organization in this research area, has developed the human emotion prediction model, and has applied their studies to the commercial business. In the academic area, a number of the conventional methods such as Multiple Regression Analysis (MRA) or Artificial Neural Networks (ANN) have been applied to predict human emotion in prior studies. However, MRA is generally criticized because of its low prediction accuracy. This is inevitable since MRA can only explain the linear relationship between the dependent variables and the independent variable. To mitigate the limitations of MRA, some studies like Jung and Kim (2012) have used ANN as the alternative, and they reported that ANN generated more accurate prediction than the statistical methods like MRA. However, it has also been criticized due to over fitting and the difficulty of the network design (e.g. setting the number of the layers and the number of the nodes in the hidden layers). Under this background, we propose a novel model using Support Vector Regression (SVR) in order to increase the prediction accuracy. SVR is an extensive version of Support Vector Machine (SVM) designated to solve the regression problems. The model produced by SVR only depends on a subset of the training data, because the cost function for building the model ignores any training data that is close (within a threshold ${\varepsilon}$) to the model prediction. Using SVR, we tried to build a model that can measure the level of arousal and valence from the facial features. To validate the usefulness of the proposed model, we collected the data of facial reactions when providing appropriate visual stimulating contents, and extracted the features from the data. Next, the steps of the preprocessing were taken to choose statistically significant variables. In total, 297 cases were used for the experiment. As the comparative models, we also applied MRA and ANN to the same data set. For SVR, we adopted '${\varepsilon}$-insensitive loss function', and 'grid search' technique to find the optimal values of the parameters like C, d, ${\sigma}^2$, and ${\varepsilon}$. In the case of ANN, we adopted a standard three-layer backpropagation network, which has a single hidden layer. The learning rate and momentum rate of ANN were set to 10%, and we used sigmoid function as the transfer function of hidden and output nodes. We performed the experiments repeatedly by varying the number of nodes in the hidden layer to n/2, n, 3n/2, and 2n, where n is the number of the input variables. The stopping condition for ANN was set to 50,000 learning events. And, we used MAE (Mean Absolute Error) as the measure for performance comparison. From the experiment, we found that SVR achieved the highest prediction accuracy for the hold-out data set compared to MRA and ANN. Regardless of the target variables (the level of arousal, or the level of positive / negative valence), SVR showed the best performance for the hold-out data set. ANN also outperformed MRA, however, it showed the considerably lower prediction accuracy than SVR for both target variables. The findings of our research are expected to be useful to the researchers or practitioners who are willing to build the models for recognizing human emotions.

Human Gesture Recognition Technology Based on User Experience for Multimedia Contents Control (멀티미디어 콘텐츠 제어를 위한 사용자 경험 기반 동작 인식 기술)

  • Kim, Yun-Sik;Park, Sang-Yun;Ok, Soo-Yol;Lee, Suk-Hwan;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.15 no.10
    • /
    • pp.1196-1204
    • /
    • 2012
  • In this paper, a series of algorithms are proposed for controlling different kinds of multimedia contents and realizing interact between human and computer by using single input device. Human gesture recognition based on NUI is presented firstly in my paper. Since the image information we get it from camera is not sensitive for further processing, we transform it to YCbCr color space, and then morphological processing algorithm is used to delete unuseful noise. Boundary Energy and depth information is extracted for hand detection. After we receive the image of hand detection, PCA algorithm is used to recognize hand posture, difference image and moment method are used to detect hand centroid and extract trajectory of hand movement. 8 direction codes are defined for quantifying gesture trajectory, so the symbol value will be affirmed. Furthermore, HMM algorithm is used for hand gesture recognition based on the symbol value. According to series of methods we presented, we can control multimedia contents by using human gesture recognition. Through large numbers of experiments, the algorithms we presented have satisfying performance, hand detection rate is up to 94.25%, gesture recognition rate exceed 92.6%, hand posture recognition rate can achieve 85.86%, and face detection rate is up to 89.58%. According to these experiment results, we can control many kinds of multimedia contents on computer effectively, such as video player, MP3, e-book and so on.

Anchor Frame Detection Using Anchor Object Extraction (앵커 객체 추출을 이용한 앵커 프레임 검출)

  • Park Ki-Tae;Hwang Doo-Sun;Moon Young-Shik
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.3 s.309
    • /
    • pp.17-24
    • /
    • 2006
  • In this paper, an algorithm for anchor frame detection in news video is proposed, which consists of four steps. In the first step, the cumulative histogram method is used to detect shot boundaries in order to segment a news video into video shots. In the second step, skin color information is used to detect face regions in each shot boundary. In the third step, color information of upper body regions is used to extract anchor object, which produces candidate anchor frames. Then, from the candidate anchor frames, a graph-theoretic cluster analysis algorithm is utilized to classify the news video into anchor-person frames and non-anchor frames. Experiment results have shown the effectiveness of the proposed algorithm.

Spatiotemporal Saliency-Based Video Summarization on a Smartphone (스마트폰에서의 시공간적 중요도 기반의 비디오 요약)

  • Lee, Won Beom;Williem, Williem;Park, In Kyu
    • Journal of Broadcast Engineering
    • /
    • v.18 no.2
    • /
    • pp.185-195
    • /
    • 2013
  • In this paper, we propose a video summarization technique on a smartphone, based on spatiotemporal saliency. The proposed technique detects scene changes by computing the difference of the color histogram, which is robust to camera and object motion. Then the similarity between adjacent frames, face region, and frame saliency are computed to analyze the spatiotemporal saliency in a video clip. Over-segmented hierarchical tree is created using scene changes and is updated iteratively using mergence and maintenance energies computed during the analysis procedure. In the updated hierarchical tree, segmented frames are extracted by applying a greedy algorithm on the node with high saliency when it satisfies the reduction ratio and the minimum interval requested by the user. Experimental result shows that the proposed method summaries a 2 minute-length video in about 10 seconds on a commercial smartphone. The summarization quality is superior to the commercial video editing software, Muvee.

Wrinkle Pattern in Korean and Mongolian Women Population (한국인과 몽골인의 주름 패턴분석)

  • Seo, Young kyoung;Kim, Minji;Kim, So jeong;Baek, Ji hwoon;Koh, Jae sook;Yang, Sung Min;Kim, Jong Hyun;Lim, Yoo Ree;Choi, Sung Won
    • Journal of the Society of Cosmetic Scientists of Korea
    • /
    • v.44 no.3
    • /
    • pp.259-266
    • /
    • 2018
  • In today's society, where people look younger than their chronological age due to improvements in the quality of life, there is a gaining interest in anti-aging and how people compare to those in the same age group. We evaluated the wrinkle index, which is the most important attribute amongst skin parameters, to evaluate external age (wrinkle age). The wrinkles of the whole face were scored by divided 8 areas (forehead, glabella, nasal root, upper eyelid, lower eyelid, crow's feet, nasolabial groove and perioral skin) and analyzed the correlation between chronological age and skin parameters. 206 subjects (Korean female, n = 105 and Mongolians female, n = 101) were enrolled. Subjects were divided into four groups by ages: 20s, 30s, 40s, and 50s. Wrinkle scores of 8 areas were evaluated and developed a calculation formula based on the wrinkle scores. Skin characteristic parameters were measured about skin elasticity, pore, wrinkle, sebum secretion. There was no difference between the calculated ages and the chronological ages in Korean women. On the other hand, Mongolian looked older than chronological age by 9 years. The correlation between the facial wrinkle ages and skin physiology parameters was presented in the order of skin elasticity > pore or crow's feet > skin tone > sebum secretion in both countries. Skin elasticity represented the most related parameter with the facial wrinkle ages. This study identified the skin wrinkle patterns of Korean and Mongolian women and the wrinkle age calculation formula developed from this study can be used as a tool for calculating the facial wrinkle ages in cosmetic studies.

A Study on Self-Expression Improvement of Children through Orff Activities (유아의 자기표현능력 증진을 위한 오르프 음악활동의 적용)

  • Kwon, Se mi
    • Journal of Music and Human Behavior
    • /
    • v.6 no.1
    • /
    • pp.55-80
    • /
    • 2009
  • The objective of this study was to improve the self-expression of children through Orff activities. In this study, three (3) children from D day care center in Seoul who demonstrated withdrawn behaviors were chosen as research subjects, based on a self-expression test score of 50 points. The activities were conducted for 6 weeks, totaling fourteen (14) sessions, with each session being scheduled for forty (40) minutes. Across 14 sessions, the researcher conducted, analyzed and compared the self-expression scale of subjects, measured during the third and the last session. The researcher then qualitatively analyzed verbal and non-verbal self-expression behaviors of subjects by video recording the session. The analysis results shown by the study are as follows. First, the results of a quantitative analysis of the self-expression scale showed significant changes in self expression. Furthermore, the results of a qualitative analysis of verbal self-expression showed positive changes in self-perception and an increase in feelings of independence and activity than that of initial sessions.

  • PDF