• Title/Summary/Keyword: Image features

Search Result 3,385, Processing Time 0.037 seconds

Tongue Image Segmentation Using CNN and Various Image Augmentation Techniques (콘볼루션 신경망(CNN)과 다양한 이미지 증강기법을 이용한 혀 영역 분할)

  • Ahn, Ilkoo;Bae, Kwang-Ho;Lee, Siwoo
    • Journal of Biomedical Engineering Research
    • /
    • v.42 no.5
    • /
    • pp.201-210
    • /
    • 2021
  • In Korean medicine, tongue diagnosis is one of the important diagnostic methods for diagnosing abnormalities in the body. Representative features that are used in the tongue diagnosis include color, shape, texture, cracks, and tooth marks. When diagnosing a patient through these features, the diagnosis criteria may be different for each oriental medical doctor, and even the same person may have different diagnosis results depending on time and work environment. In order to overcome this problem, recent studies to automate and standardize tongue diagnosis using machine learning are continuing and the basic process of such a machine learning-based tongue diagnosis system is tongue segmentation. In this paper, image data is augmented based on the main tongue features, and backbones of various famous deep learning architecture models are used for automatic tongue segmentation. The experimental results show that the proposed augmentation technique improves the accuracy of tongue segmentation, and that automatic tongue segmentation can be performed with a high accuracy of 99.12%.

Histogram Equalized Eigen Co-occurrence Features for Color Image Classification (컬러이미지 검색을 위한 히스토그램 평활화 기반 고유 병발 특징에 관한 연구)

  • Yoon, TaeBok;Choi, YoungMee;Choo, MoonWon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2010.11a
    • /
    • pp.705-708
    • /
    • 2010
  • An eigen color co-occurrence approach is proposed that exploits the correlation between color channels to identify the degree of image similarity. This method is based on traditional co-occurrence matrix method and histogram equalization. On the purpose of feature extraction, eigen color co-occurrence matrices are computed for extracting the statistical relationships embedded in color images by applying Principal Component Analysis (PCA) on a set of color co-occurrence matrices, which are computed on the histogram equalized images. That eigen space is created with a set of orthogonal axes to gain the essential structures of color co-occurrence matrices, which is used to identify the degree of similarity to classify an input image to be tested for various purposes. In this paper RGB, Gaussian color space are compared with grayscale image in terms of PCA eigen features embedded in histogram equalized co-occurrence features. The experimental results are presented.

Automatic Detection of Absorption Features for Hyperspectral Images

  • Hsu, Pai-Hui;Tseng, Yi-Hsing
    • Proceedings of the KSRS Conference
    • /
    • 2003.11a
    • /
    • pp.700-702
    • /
    • 2003
  • A new method for automatic detection of absorption features is proposed. This method is based on the modulus maximum of the scale-space image calculated by continuous wavelet transform. This method is computationally efficient as compared to traditional methods. The continuum removal algorithm is than implemented on the detected absorption features to reduce some additive factors caused by other absorbing of materials. The results show that the chlorophyll absorption features are detected exactly.

  • PDF

Region-Based Facial Expression Recognition in Still Images

  • Nagi, Gawed M.;Rahmat, Rahmita O.K.;Khalid, Fatimah;Taufik, Muhamad
    • Journal of Information Processing Systems
    • /
    • v.9 no.1
    • /
    • pp.173-188
    • /
    • 2013
  • In Facial Expression Recognition Systems (FERS), only particular regions of the face are utilized for discrimination. The areas of the eyes, eyebrows, nose, and mouth are the most important features in any FERS. Applying facial features descriptors such as the local binary pattern (LBP) on such areas results in an effective and efficient FERS. In this paper, we propose an automatic facial expression recognition system. Unlike other systems, it detects and extracts the informative and discriminant regions of the face (i.e., eyes, nose, and mouth areas) using Haar-feature based cascade classifiers and these region-based features are stored into separate image files as a preprocessing step. Then, LBP is applied to these image files for facial texture representation and a feature-vector per subject is obtained by concatenating the resulting LBP histograms of the decomposed region-based features. The one-vs.-rest SVM, which is a popular multi-classification method, is employed with the Radial Basis Function (RBF) for facial expression classification. Experimental results show that this approach yields good performance for both frontal and near-frontal facial images in terms of accuracy and time complexity. Cohn-Kanade and JAFFE, which are benchmark facial expression datasets, are used to evaluate this approach.

A Study on Mouth Features Detection in Face using HMM (HMM을 이용한 얼굴에서 입 특징점 검출에 관한 연구)

  • Kim, Hea-Chel;Jung, Chan-Ju;Kwag, Jong-Se;Kim, Mun-Hwan;Bae, Chul-Soo;Ra, Snag-Dong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2002.04a
    • /
    • pp.647-650
    • /
    • 2002
  • The human faces do not have distinct features unlike other general objects. In general the features of eyes, nose and mouth which are first recognized when human being see the face are defined. These features have different characteristics depending on different human face. In this paper, We propose a face recognition algorithm using the hidden Markov model(HMM). In the preprocessing stage, we find edges of a face using the locally adaptive threshold scheme and extract features based on generic knowledge of a face, then construct a database with extracted features. In training stage, we generate HMM parameters for each person by using the forward-backward algorithm. In the recognition stage, we apply probability values calculated by the HMM to input data. Then the input face is recognized by the euclidean distance of face feature vector and the cross-correlation between the input image and the database image. Computer simulation shows that the proposed HMM algorithm gives higher recognition rate compared with conventional face recognition algorithms.

  • PDF

Video Expression Recognition Method Based on Spatiotemporal Recurrent Neural Network and Feature Fusion

  • Zhou, Xuan
    • Journal of Information Processing Systems
    • /
    • v.17 no.2
    • /
    • pp.337-351
    • /
    • 2021
  • Automatically recognizing facial expressions in video sequences is a challenging task because there is little direct correlation between facial features and subjective emotions in video. To overcome the problem, a video facial expression recognition method using spatiotemporal recurrent neural network and feature fusion is proposed. Firstly, the video is preprocessed. Then, the double-layer cascade structure is used to detect a face in a video image. In addition, two deep convolutional neural networks are used to extract the time-domain and airspace facial features in the video. The spatial convolutional neural network is used to extract the spatial information features from each frame of the static expression images in the video. The temporal convolutional neural network is used to extract the dynamic information features from the optical flow information from multiple frames of expression images in the video. A multiplication fusion is performed with the spatiotemporal features learned by the two deep convolutional neural networks. Finally, the fused features are input to the support vector machine to realize the facial expression classification task. The experimental results on cNTERFACE, RML, and AFEW6.0 datasets show that the recognition rates obtained by the proposed method are as high as 88.67%, 70.32%, and 63.84%, respectively. Comparative experiments show that the proposed method obtains higher recognition accuracy than other recently reported methods.

An Analysis on the Properties of Features against Various Distortions in Deep Neural Networks

  • Kang, Jung Heum;Jeong, Hye Won;Choi, Chang Kyun;Ali, Muhammad Salman;Bae, Sung-Ho;Kim, Hui Yong
    • Journal of Broadcast Engineering
    • /
    • v.26 no.7
    • /
    • pp.868-876
    • /
    • 2021
  • Deploying deep neural network model training performs remarkable performance in the fields of Object detection and Instance segmentation. To train these models, features are first extracted from the input image using a backbone network. The extracted features can be reused by various tasks. Research has been actively conducted to serve various tasks by using these learned features. In this process, standardization discussions about encoding, decoding, and transmission methods are proceeding actively. In this scenario, it is necessary to analyze the response characteristics of features against various distortions that may occur in the data transmission or data compression process. In this paper, experiment was conducted to inject various distortions into the feature in the object recognition task. And analyze the mAP (mean Average Precision) metric between the predicted value output from the neural network and the target value as the intensity of various distortions was increased. Experiments have shown that features are more robust to distortion than images. And this points out that using the feature as transmission means can prevent the loss of information against the various distortions during data transmission and compression process.

Masked Face Recognition via a Combined SIFT and DLBP Features Trained in CNN Model

  • Aljarallah, Nahla Fahad;Uliyan, Diaa Mohammed
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.319-331
    • /
    • 2022
  • The latest global COVID-19 pandemic has made the use of facial masks an important aspect of our lives. People are advised to cover their faces in public spaces to discourage illness from spreading. Using these face masks posed a significant concern about the exactness of the face identification method used to search and unlock telephones at the school/office. Many companies have already built the requisite data in-house to incorporate such a scheme, using face recognition as an authentication. Unfortunately, veiled faces hinder the detection and acknowledgment of these facial identity schemes and seek to invalidate the internal data collection. Biometric systems that use the face as authentication cause problems with detection or recognition (face or persons). In this research, a novel model has been developed to detect and recognize faces and persons for authentication using scale invariant features (SIFT) for the whole segmented face with an efficient local binary texture features (DLBP) in region of eyes in the masked face. The Fuzzy C means is utilized to segment the image. These mixed features are trained significantly in a convolution neural network (CNN) model. The main advantage of this model is that can detect and recognizing faces by assigning weights to the selected features aimed to grant or provoke permissions with high accuracy.

An Automatic Object Extraction Method Using Color Features Of Object And Background In Image (영상에서 객체와 배경의 색상 특징을 이용한 자동 객체 추출 기법)

  • Lee, Sung Kap;Park, Young Soo;Lee, Gang Seong;Lee, Jong Yong;Lee, Sang Hun
    • Journal of Digital Convergence
    • /
    • v.11 no.12
    • /
    • pp.459-465
    • /
    • 2013
  • This paper is a study on an object extraction method which using color features of an object and background in the image. A human recognizes an object through the color difference of object and background in the image. So we must to emphasize the color's difference that apply to extraction result in this image. Therefore, we have converted to HSV color images which similar to human visual system from original RGB images, and have created two each other images that applied Median Filter and we merged two Median filtered images. And we have applied the Mean Shift algorithm which a data clustering method for clustering color features. Finally, we have normalized 3 image channels to 1 image channel for binarization process. And we have created object map through the binarization which using average value of whole pixels as a threshold. Then, have extracted major object from original image use that object map.

Segmentation and estimation of surfaces from statistical probability of texture features

  • Terauchi, Mutsuhiro;Nagamachi, Mitsuo;Koji-Ito;Tsuji, Toshio
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1988.10b
    • /
    • pp.826-831
    • /
    • 1988
  • This paper presents an approach to segment an image into areas of surfaces, and to compute the surface properties from a gray-scale image in order to describe the surfaces for reconstruction of the 3-D shape of the objects. In general, an rigid body has several surfaces and many edges. But if it is not polyhedoron, it is necessary not only to describe the relation between surfaces, i.e. its line drawings but also to represent the surfaces' equations itself. In order to compute the surfaces' equation we use a probability of edge distribution. At first it is extracted edges from a gray-level image as much as possible. These are not only the points that maximize the change of an image intensuty but candidates which can be seemed to be edges. Next, other character of a surface (color, coordinates and image intensity) are extracted. In our study, we call the all feature of a surface as "texture", for example color, intensity level, orientation of an edge, shape of a surface and so on. These features of a surface on a pixel of an image plane are mapped to a point of the feature space, and segmented to each groups by cluster analysis on this space. These groups are considered to represent object surface in an image plane. Finally, the states of object surface in 3-D space are computed from distributional probability of local and overall statistical features of a surface, and from shape of a surface.a surface.

  • PDF