• Title/Summary/Keyword: Color and Texture Feature

Search Result 137, Processing Time 0.023 seconds

A Generalized Method for Extracting Characters and Video Captions (일반화된 문자 및 비디오 자막 영역 추출 방법)

  • Chun, Byung-Tae;Bae, Young-Lae;Kim, Tai-Yun
    • Journal of KIISE:Software and Applications
    • /
    • v.27 no.6
    • /
    • pp.632-641
    • /
    • 2000
  • Conventional character extraction methods extract character regions using methods such as color reduction, region split and merge and texture analysis from the whole image. Because these methods use many heuristic variables and thresholding values derived from a priori knowledge, it is difficult to generalize them algorithmically. In this paper, we propose a method that can extract character regions using a topographical feature extraction method and a point-line-region extension method. The proposed method can also solve the problems of conventional methods by reducing heuristic variables and generalizing thresholding values. We see that character regions can be extracted by generalized variables and thresolding values without using a priori knowledge of character region. Experimental results show that the candidate region extraction rate is 100%, and the character region extraction rate is over 98%.

  • PDF

A Novel Face Recognition Algorithm based on the Deep Convolution Neural Network and Key Points Detection Jointed Local Binary Pattern Methodology

  • Huang, Wen-zhun;Zhang, Shan-wen
    • Journal of Electrical Engineering and Technology
    • /
    • v.12 no.1
    • /
    • pp.363-372
    • /
    • 2017
  • This paper presents a novel face recognition algorithm based on the deep convolution neural network and key point detection jointed local binary pattern methodology to enhance the accuracy of face recognition. We firstly propose the modified face key feature point location detection method to enhance the traditional localization algorithm to better pre-process the original face images. We put forward the grey information and the color information with combination of a composite model of local information. Then, we optimize the multi-layer network structure deep learning algorithm using the Fisher criterion as reference to adjust the network structure more accurately. Furthermore, we modify the local binary pattern texture description operator and combine it with the neural network to overcome drawbacks that deep neural network could not learn to face image and the local characteristics. Simulation results demonstrate that the proposed algorithm obtains stronger robustness and feasibility compared with the other state-of-the-art algorithms. The proposed algorithm also provides the novel paradigm for the application of deep learning in the field of face recognition which sets the milestone for further research.

A Object-Based Image Retrieval Using Feature Analysis and Fractal Dimension (특징 분석과 프랙탈 차원을 이용한 객체 기반 영상검색)

  • 이정봉;박장춘
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.2
    • /
    • pp.173-186
    • /
    • 2004
  • This paper proposed the content-based retrieval system as a method for performing image retrieval through the effective feature extraction of the object of significant meaning based on the characteristics of man's visual system. To allow the object region of interest to be primarily detected, the region, being comparatively large size, greatly different from the background color and located in the middle of the image, was judged as the major object with a meaning. To get the original features of the image, the cumulative sum of tile declination difference vector the segment of the object contour had and the signature of the bipartite object were extracted and used in the form of being applied to the rotation of the object and the change of the size after partition of the total length of the object contour of the image into the normalized segment. Starting with this form feature, it was possible to make a retrieval robust to any change in translation, rotation and scaling by combining information on the texture sample, color and eccentricity and measuring the degree of similarity. It responded less sensitively to the phenomenon of distortion of the object feature due to the partial change or damage of the region. Also, the method of imposing a different weight of similarity on the image feature based on the relationship of complexity between measured objects using the fractal dimension by the Boxing-Counting Dimension minimized the wrong retrieval and showed more efficient retrieval rate.

  • PDF

Image Retrieval for Electronic illustrated Fish Book (전자어류도감을 위한 영상검색)

  • Ahn, Soo-Hong;Oh, Jeong-Su
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.4C
    • /
    • pp.226-231
    • /
    • 2011
  • To improve the conventional illustrated fish book, this paper introduces the concept of an electronic illustrated fish book which applies IT techniques to the conventional one, and proposes the image retrieval for it. The image retrieval is a core technology of the electronic illustrated fish book and make it overwhelm the conventional one. Since fishes, even if the same kind, have different features in shape, color, and texture and the same fish can even have different features by its pose or environment at that time for taking a picture, the conventional image retrieval, that uses simple features in shape, color, and texture, is not suitable for the electronic illustrated fish book. The proposed image retrieval adopts detail shape features extracted from head, body, and tail of a fish and different weights are given to the features depending on their invariability. The simulation results show that the proposed algorithm is far superior to the conventional algorithm.

Region-based Image Retrieval using Wavelet Transform and Image Segmentation (웨이브릿 변환과 영상 분할을 이용한 영역기반 영상 검색)

  • 이상훈;홍충선;곽윤식;이대영
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.25 no.8B
    • /
    • pp.1391-1399
    • /
    • 2000
  • In this paper, we discussed the region-based image retrieval method using image segmentation. We proposed a segmentation method which can reduce the effect of a irregular light sources. The image segmentation method uses a region-merging, and candidate regions which are merged were selected by the energy values of high frequency bands in discrete wavelet transform. The content-based image retrieval is executed by using the segmented region information, and the images are retrieved by a color, texture, shape feature vector. The similarity measure between regions is processed by the Euclidean distance of the feature vectors. The simulation results shows that the proposed method is reasonable.

  • PDF

Emotion from Color images and Its Application to Content-based Image Retrievals (칼라영상의 감성평가와 이를 이용한 내용기반 영상검색)

  • Park, Joong-Soo;Eum, Kyoung-Bae;Shin, Kyung-Hae;Lee, Joon-Whoan;Park, Dong-Sun
    • The KIPS Transactions:PartB
    • /
    • v.10B no.2
    • /
    • pp.179-188
    • /
    • 2003
  • In content-based image retrieval, the query is an image itself and the retrieval process is the process that seeking the similar images to the given query image. In this way of retrieval, the user has to know the basic physical features of target images that he wants to retrieve. But it has some restriction because to retrieve the target image he has to know the basic physical feature space such as color, texture, shape and spatial relationship. In this paper, we propose an emotion-based retrieval system. It uses the emotion that color images have. It is different from past emotion-based image retrieval in point of view that it uses relevance feedback to estimate the users intend and it is easily combined with past content-based image retrieval system. To test the performance of our proposed system, we use MPEG-7 color descriptor and emotion language such as "warm", "clean", "bright" and "delight" We test about 1500 wallpaper images and get successful result.lpaper images and get successful result.

A Survey on Image Emotion Recognition

  • Zhao, Guangzhe;Yang, Hanting;Tu, Bing;Zhang, Lei
    • Journal of Information Processing Systems
    • /
    • v.17 no.6
    • /
    • pp.1138-1156
    • /
    • 2021
  • Emotional semantics are the highest level of semantics that can be extracted from an image. Constructing a system that can automatically recognize the emotional semantics from images will be significant for marketing, smart healthcare, and deep human-computer interaction. To understand the direction of image emotion recognition as well as the general research methods, we summarize the current development trends and shed light on potential future research. The primary contributions of this paper are as follows. We investigate the color, texture, shape and contour features used for emotional semantics extraction. We establish two models that map images into emotional space and introduce in detail the various processes in the image emotional semantic recognition framework. We also discuss important datasets and useful applications in the field such as garment image and image retrieval. We conclude with a brief discussion about future research trends.

A Study on Mouth Mouse

  • Han, Chan-Myung;Park, Joon-Ho;Kim, Hwi-Won;Yoon, Young-Woo
    • 한국정보컨버전스학회:학술대회논문집
    • /
    • 2008.06a
    • /
    • pp.173-176
    • /
    • 2008
  • Among human body parts, the human face has been studied the most actively for the interlace between humans and computers because face has statistic consistency in color, shape and texture. Those characteristics make computers detect and track human faces in images robustly and accurately. The human face consists of eyes, nose, mouth, eyebrows and other features, Detecting and tracking each feature have been researched. The open mouth is the largest in size and the easiest to detect among them, In this study, we present a system which can move mouse pointer using the position and state of the mouth.

  • PDF

Parallel Injection Method for Improving Descriptive Performance of Bi-GRU Image Captions (Bi-GRU 이미지 캡션의 서술 성능 향상을 위한 Parallel Injection 기법 연구)

  • Lee, Jun Hee;Lee, Soo Hwan;Tae, Soo Ho;Seo, Dong Hoan
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.11
    • /
    • pp.1223-1232
    • /
    • 2019
  • The injection is the input method of the image feature vector from the encoder to the decoder. Since the image feature vector contains object details such as color and texture, it is essential to generate image captions. However, the bidirectional decoder model using the existing injection method only inputs the image feature vector in the first step, so image feature vectors of the backward sequence are vanishing. This problem makes it difficult to describe the context in detail. Therefore, in this paper, we propose the parallel injection method to improve the description performance of image captions. The proposed Injection method fuses all embeddings and image vectors to preserve the context. Also, We optimize our image caption model with Bidirectional Gated Recurrent Unit (Bi-GRU) to reduce the amount of computation of the decoder. To validate the proposed model, experiments were conducted with a certified image caption dataset, demonstrating excellence in comparison with the latest models using BLEU and METEOR scores. The proposed model improved the BLEU score up to 20.2 points and the METEOR score up to 3.65 points compared to the existing caption model.

Novel Intent based Dimension Reduction and Visual Features Semi-Supervised Learning for Automatic Visual Media Retrieval

  • kunisetti, Subramanyam;Ravichandran, Suban
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.230-240
    • /
    • 2022
  • Sharing of online videos via internet is an emerging and important concept in different types of applications like surveillance and video mobile search in different web related applications. So there is need to manage personalized web video retrieval system necessary to explore relevant videos and it helps to peoples who are searching for efficient video relates to specific big data content. To evaluate this process, attributes/features with reduction of dimensionality are computed from videos to explore discriminative aspects of scene in video based on shape, histogram, and texture, annotation of object, co-ordination, color and contour data. Dimensionality reduction is mainly depends on extraction of feature and selection of feature in multi labeled data retrieval from multimedia related data. Many of the researchers are implemented different techniques/approaches to reduce dimensionality based on visual features of video data. But all the techniques have disadvantages and advantages in reduction of dimensionality with advanced features in video retrieval. In this research, we present a Novel Intent based Dimension Reduction Semi-Supervised Learning Approach (NIDRSLA) that examine the reduction of dimensionality with explore exact and fast video retrieval based on different visual features. For dimensionality reduction, NIDRSLA learns the matrix of projection by increasing the dependence between enlarged data and projected space features. Proposed approach also addressed the aforementioned issue (i.e. Segmentation of video with frame selection using low level features and high level features) with efficient object annotation for video representation. Experiments performed on synthetic data set, it demonstrate the efficiency of proposed approach with traditional state-of-the-art video retrieval methodologies.