• Title/Summary/Keyword: texture extraction

Search Result 267, Processing Time 0.029 seconds

Use of Crown Feature Analysis to Separate the Two Pine Species in QuickBird Imagery

  • Kim, Cheon
    • Korean Journal of Remote Sensing
    • /
    • v.24 no.3
    • /
    • pp.267-272
    • /
    • 2008
  • Tree species-specific estimates with spacebome high-resolution imagery improve estimation of forest biomass which is needed to predict the long term planning for the sustainable forest management(SFM). This paper is a contribution to develop crown distinguishing coniferous species, Pinus densiflora and Pinus koraiensis, from QuickBird imagery. The proposed feature analysis derived from shape parameters and first and second-order statistical texture features of the same test area were compared for the two species separation and delineation. As expected, initial studies have shown that both formfactor and compactness shape parameters provided the successful differentiating method between the pine species within the compartment for single crown identification from spaceborne high resolution imagery. Another result revealed that the selected texture parameters - the mean, variance, angular second moment(ASM) - in the infrared band image could produce good subset combination of texture features for representing detailed tree crown outline.

Music Genre Classification Based on Timbral Texture and Rhythmic Content Features

  • Baniya, Babu Kaji;Ghimire, Deepak;Lee, Joonwhon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2013.05a
    • /
    • pp.204-207
    • /
    • 2013
  • Music genre classification is an essential component for music information retrieval system. There are two important components to be considered for better genre classification, which are audio feature extraction and classifier. This paper incorporates two different kinds of features for genre classification, timbral texture and rhythmic content features. Timbral texture contains several spectral and Mel-frequency Cepstral Coefficient (MFCC) features. Before choosing a timbral feature we explore which feature contributes less significant role on genre discrimination. This facilitates the reduction of feature dimension. For the timbral features up to the 4-th order central moments and the covariance components of mutual features are considered to improve the overall classification result. For the rhythmic content the features extracted from beat histogram are selected. In the paper Extreme Learning Machine (ELM) with bagging is used as classifier for classifying the genres. Based on the proposed feature sets and classifier, experiment is performed with well-known datasets: GTZAN databases with ten different music genres, respectively. The proposed method acquires the better classification accuracy than the existing approaches.

A Generalized Method for Extracting Characters and Video Captions (일반화된 문자 및 비디오 자막 영역 추출 방법)

  • Chun, Byung-Tae;Bae, Young-Lae;Kim, Tai-Yun
    • Journal of KIISE:Software and Applications
    • /
    • v.27 no.6
    • /
    • pp.632-641
    • /
    • 2000
  • Conventional character extraction methods extract character regions using methods such as color reduction, region split and merge and texture analysis from the whole image. Because these methods use many heuristic variables and thresholding values derived from a priori knowledge, it is difficult to generalize them algorithmically. In this paper, we propose a method that can extract character regions using a topographical feature extraction method and a point-line-region extension method. The proposed method can also solve the problems of conventional methods by reducing heuristic variables and generalizing thresholding values. We see that character regions can be extracted by generalized variables and thresolding values without using a priori knowledge of character region. Experimental results show that the candidate region extraction rate is 100%, and the character region extraction rate is over 98%.

  • PDF

PDA-based Text Extraction System using Client/Server Architecture (Client/Server구조를 이용한 PDA기반의 문자 추출 시스템)

  • Park Anjin;Jung Keechul
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.2
    • /
    • pp.85-98
    • /
    • 2005
  • Recently, a lot of researches about mobile vision using Personal Digital Assistant(PDA) has been attempted. Many CPUs for PDA are integer CPUs, which have no floating-computation component. It results in slow computation of the algorithms peformed by vision system or image processing, which have much floating-computation. In this paper, in order to resolve this weakness, we propose the Client(PDA)/server(PC) architecture which is connected to each other with a wireless LAN, and we construct the system with pipelining processing using two CPUs of the Client(PDA) and the Server(PC) in image sequence. The Client(PDA) extracts tentative text regions using Edge Density(ED). The Server(PC) uses both the Multi-1.aver Perceptron(MLP)-based texture classifier and Connected Component(CC)-based filtering for a definite text extraction based on the Client(PDA)'s tentativel99-y extracted results. The proposed method leads to not only efficient text extraction by using both the MLP and the CC, but also fast running time using Client(PDA)/server(PC) architecture with the pipelining processing.

Smoke detection in video sequences based on dynamic texture using volume local binary patterns

  • Lin, Gaohua;Zhang, Yongming;Zhang, Qixing;Jia, Yang;Xu, Gao;Wang, Jinjun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.11
    • /
    • pp.5522-5536
    • /
    • 2017
  • In this paper, a video based smoke detection method using dynamic texture feature extraction with volume local binary patterns is studied. Block based method was used to distinguish smoke frames in high definition videos obtained by experiments firstly. Then we propose a method that directly extracts dynamic texture features based on irregular motion regions to reduce adverse impacts of block size and motion area ratio threshold. Several general volume local binary patterns were used to extract dynamic texture, including LBPTOP, VLBP, CLBPTOP and CVLBP, to study the effect of the number of sample points, frame interval and modes of the operator on smoke detection. Support vector machine was used as the classifier for dynamic texture features. The results show that dynamic texture is a reliable clue for video based smoke detection. It is generally conducive to reducing the false alarm rate by increasing the dimension of the feature vector. However, it does not always contribute to the improvement of the detection rate. Additionally, it is found that the feature computing time is not directly related to the vector dimension in our experiments, which is important for the realization of real-time detection.

Visual Semantic Based 3D Video Retrieval System Using HDFS

  • Ranjith Kumar, C.;Suguna, S.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.8
    • /
    • pp.3806-3825
    • /
    • 2016
  • This paper brings out a neoteric frame of reference for visual semantic based 3d video search and retrieval applications. Newfangled 3D retrieval application spotlight on shape analysis like object matching, classification and retrieval not only sticking up entirely with video retrieval. In this ambit, we delve into 3D-CBVR (Content Based Video Retrieval) concept for the first time. For this purpose we intent to hitch on BOVW and Mapreduce in 3D framework. Here, we tried to coalesce shape, color and texture for feature extraction. For this purpose, we have used combination of geometric & topological features for shape and 3D co-occurrence matrix for color and texture. After thriving extraction of local descriptors, TB-PCT (Threshold Based- Predictive Clustering Tree) algorithm is used to generate visual codebook. Further, matching is performed using soft weighting scheme with L2 distance function. As a final step, retrieved results are ranked according to the Index value and produce results .In order to handle prodigious amount of data and Efficacious retrieval, we have incorporated HDFS in our Intellection. Using 3D video dataset, we fiture the performance of our proposed system which can pan out that the proposed work gives meticulous result and also reduce the time intricacy.

A TRUS Prostate Segmentation using Gabor Texture Features and Snake-like Contour

  • Kim, Sung Gyun;Seo, Yeong Geon
    • Journal of Information Processing Systems
    • /
    • v.9 no.1
    • /
    • pp.103-116
    • /
    • 2013
  • Prostate cancer is one of the most frequent cancers in men and is a major cause of mortality in the most of countries. In many diagnostic and treatment procedures for prostate disease accurate detection of prostate boundaries in transrectal ultrasound(TRUS) images is required. This is a challenging and difficult task due to weak prostate boundaries, speckle noise and the short range of gray levels. In this paper a method for automatic prostate segmentation in TRUS images using Gabor feature extraction and snake-like contour is presented. This method involves preprocessing, extracting Gabor feature, training, and prostate segmentation. The speckle reduction for preprocessing step has been achieved by using stick filter and top-hat transform has been implemented for smoothing the contour. A Gabor filter bank for extraction of rotation-invariant texture features has been implemented. A support vector machine(SVM) for training step has been used to get each feature of prostate and nonprostate. Finally, the boundary of prostate is extracted by the snake-like contour algorithm. A number of experiments are conducted to validate this method and results showed that this new algorithm extracted the prostate boundary with less than 10.2% of the accuracy which is relative to boundary provided manually by experts.

Application of YOLOv5 Neural Network Based on Improved Attention Mechanism in Recognition of Thangka Image Defects

  • Fan, Yao;Li, Yubo;Shi, Yingnan;Wang, Shuaishuai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.1
    • /
    • pp.245-265
    • /
    • 2022
  • In response to problems such as insufficient extraction information, low detection accuracy, and frequent misdetection in the field of Thangka image defects, this paper proposes a YOLOv5 prediction algorithm fused with the attention mechanism. Firstly, the Backbone network is used for feature extraction, and the attention mechanism is fused to represent different features, so that the network can fully extract the texture and semantic features of the defect area. The extracted features are then weighted and fused, so as to reduce the loss of information. Next, the weighted fused features are transferred to the Neck network, the semantic features and texture features of different layers are fused by FPN, and the defect target is located more accurately by PAN. In the detection network, the CIOU loss function is used to replace the GIOU loss function to locate the image defect area quickly and accurately, generate the bounding box, and predict the defect category. The results show that compared with the original network, YOLOv5-SE and YOLOv5-CBAM achieve an improvement of 8.95% and 12.87% in detection accuracy respectively. The improved networks can identify the location and category of defects more accurately, and greatly improve the accuracy of defect detection of Thangka images.

A Region-based Image Retrieval System using Salient Point Extraction and Image Segmentation (영상분할과 특징점 추출을 이용한 영역기반 영상검색 시스템)

  • 이희경;호요성
    • Journal of Broadcast Engineering
    • /
    • v.7 no.3
    • /
    • pp.262-270
    • /
    • 2002
  • Although most image indexing schemes ate based on global image features, they have limited discrimination capability because they cannot capture local variations of the image. In this paper, we propose a new region-based image retrieval system that can extract important regions in the image using salient point extraction and image segmentation techniques. Our experimental results show that color and texture information in the region provide a significantly improved retrieval performances compared to the global feature extraction methods.

Enhanced Object Extraction Method Based on Multi-channel Saliency Map (Saliency Map 다중 채널을 기반으로 한 개선된 객체 추출 방법)

  • Choi, Young-jin;Cui, Run;Kim, Kwang-Rag;Kim, Hyoung Joong
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.53 no.2
    • /
    • pp.53-61
    • /
    • 2016
  • Extracting focused object with saliency map is still remaining as one of the most highly tasked research area around computer vision for it is hard to estimate. Through this paper, we propose enhanced object extraction method based on multi-channel saliency map which could be done automatically without machine learning. Proposed Method shows a higher accuracy than Itti method using SLIC, Euclidean, and LBP algorithm as for object extraction. Experiments result shows that our approach is possible to be used for automatic object extraction without any previous training procedure through focusing on the main object from the image instead of estimating the whole image from background to foreground.