• 제목/요약/키워드: image features

검색결과 3,371건 처리시간 0.03초

확장된 개념 기반 이미지 검색 시스템 (An Extended Concept-based Image Retrieval System : E-COIRS)

  • 김용일;양재동;양형정
    • 한국정보과학회논문지:컴퓨팅의 실제 및 레터
    • /
    • 제8권3호
    • /
    • pp.303-317
    • /
    • 2002
  • In this paper, we design and implement E-COIRS enabling users to query with concepts and image features used for further refining the concepts. For example, E-COIRS supports the query "retrieve images containing black home appliance to north of reception set. "The query includes two types of concepts: IS-A and composite. "home appliance"is an IS-A concept, and "reception set" is a composite concept. For evaluating such a query. E-COIRS includes three important components: a visual image indexer, thesauri and a query processor. Each pair of objects in an mage captured by the visual image indexer is converted into a triple. The triple consists of the two object identifiers (oids) and their spatial relationship. All the features of an object is referenced by its old. A composite concept is detected by the triple thesaurus and IS-A concept is recolonized by the fuzzy term thesaurus. The query processor obtains an image set by matching each triple in a user with an inverted file and CS-Tree. To support efficient storage use and fast retrieval on high-dimensional feature vectors, E-COIRS uses Cell-based Signature tree(CS-Tree). E-COIRS is a more advanced content-based image retrieval system than other systems which support only concepts or image features.

An Image Retrieving Scheme Using Salient Features and Annotation Watermarking

  • Wang, Jenq-Haur;Liu, Chuan-Ming;Syu, Jhih-Siang;Chen, Yen-Lin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제8권1호
    • /
    • pp.213-231
    • /
    • 2014
  • Existing image search systems allow users to search images by keywords, or by example images through content-based image retrieval (CBIR). On the other hand, users might learn more relevant textual information about an image from its text captions or surrounding contexts within documents or Web pages. Without such contexts, it's difficult to extract semantic description directly from the image content. In this paper, we propose an annotation watermarking system for users to embed text descriptions, and retrieve more relevant textual information from similar images. First, tags associated with an image are converted by two-dimensional code and embedded into the image by discrete wavelet transform (DWT). Next, for images without annotations, similar images can be obtained by CBIR techniques and embedded annotations can be extracted. Specifically, we use global features such as color ratios and dominant sub-image colors for preliminary filtering. Then, local features such as Scale-Invariant Feature Transform (SIFT) descriptors are extracted for similarity matching. This design can achieve good effectiveness with reasonable processing time in practical systems. Our experimental results showed good accuracy in retrieving similar images and extracting relevant tags from similar images.

Identification of Transformed Image Using the Composition of Features

  • Yang, Won-Keun;Cho, A-Young;Cho, Ik-Hwan;Oh, Weon-Geun;Jeong, Dong-Seok
    • 한국멀티미디어학회논문지
    • /
    • 제11권6호
    • /
    • pp.764-776
    • /
    • 2008
  • Image identification is the process of checking whether the query image is the transformed version of the specific original image or not. In this paper, image identification method based on feature composition is proposed. Used features include color distance, texture information and average pixel intensity. We extract color characteristics using color distance and texture information by Modified Generalized Symmetry Transform as well as average intensity of each pixel as features. Individual feature is quantized adaptively to be used as bins of histogram. The histogram is normalized according to data type and it is used as the signature in comparing the query image with database images. In matching part, Manhattan distance is used for measuring distance between two signatures. To evaluate the performance of the proposed method, independent test and accuracy test are achieved. In independent test, 60,433 images are used to evaluate the ability of discrimination between different images. And 4,002 original images and its 29 transformed versions are used in accuracy test, which evaluate the ability that the proposed algorithm can find the original image correctly when some transforms was applied in original image. Experiment results show that the proposed identification method has good performance in accuracy test. And the proposed method is very useful in real environment because of its high accuracy and fast matching capacity.

  • PDF

행렬 속성을 이용하는 질감 영상 분별기 (A Classifier for Textured Images Based on Matrix Feature)

  • 김준철;이준환
    • 전자공학회논문지B
    • /
    • 제31B권3호
    • /
    • pp.91-102
    • /
    • 1994
  • For the analysis of textured image, it requires large storage space and computation time to calculate the matrix features such as SGLDM(Spatial Gray Level Dependence Matrix). NGLDM(Neighboring Gray Level Dependence Matrix). NSGLDM(Neighboring Spatial Gray Level Dependence Matrix) and GLRLM(Gray Level Run Length Matrix). In spite of a large amount of information that each matrix contains, a set of several correlated scalar features calculated from the matrix is not sufficient to approximate it. In this paper, we propose a new classifier for textured images based on these matrices in which the projected vectors of each matrix on the meaningful directions are used as features. In the proposed method, an unknown image is classified to the class of a known image that gives the maximum similarity between the projected model vector from the known image and the vector from the unknown image. In the experiment to classify images of agricultural products, the proposed method shows good performance as much as 85-95% of correct classification ratio.

  • PDF

축소변환된 의료 이미지의 질감 특징 추출과 인덱싱 (An Extracting and Indexing Schema of Compressed Medical Images)

  • 위희정;엄기현
    • 한국멀티미디어학회:학술대회논문집
    • /
    • 한국멀티미디어학회 2000년도 춘계학술발표논문집
    • /
    • pp.328-331
    • /
    • 2000
  • In this paper , we propose a texture feature extraction method of reduce the massive computational time on extracting texture, features of large sized medical such as MRI, CT-scan , and an index structure, called GLTFT, to speed up the retrieval performance. For these, the original image is transformed into a compressed image by Wavelet transform , and textural features such as contrast, energy, entropy, and homogeneity of the compressed image is extracted by using GLCM(Gray Level Co-occurrence Metrix) . The proposed index structure is organized by using the textural features. The processing in compressed domain can give the solution of storage space and the reduction of computational time of feature extracting . And , by GLTFT index structure, image retrieval performance can be expected to be improved by reducing the retrieval range . Our experiment on 270 MRIs as image database shows that shows that such expectation can be got.

  • PDF

Self-Organizing Neural Network를 이용한 임펄스 노이즈 검출과 선택적 미디언 필터 적용 (Impulse Noise Detection Using Self-Organizing Neural Network and Its Application to Selective Median Filtering)

  • 이종호;동성수;위재우;송승민
    • 대한전기학회논문지:시스템및제어부문D
    • /
    • 제54권3호
    • /
    • pp.166-173
    • /
    • 2005
  • Preserving image features, edges and details in the process of impulsive noise filtering is an important problem. To avoid image blurring, only corrupted pixels must be filtered. In this paper, we propose an effective impulse noise detection method using Self-Organizing Neural Network(SONN) which applies median filter selectively for removing random-valued impulse noises while preserving image features, edges and details. Using a $3\times3$ window, we obtain useful local features with which impulse noise patterns are classified. SONN is trained with sample image patterns and each pixel pattern is classified by its local information in the image. The results of the experiments with various images which are the noise range of $5-15\%$ show that our method performs better than other methods which use multiple threshold values for impulse noise detection.

텍스처 정보 기반의 PCA를 이용한 문서 영상의 분석 (Texture-based PCA for Analyzing Document Image)

  • 김보람;김욱현
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2006년도 하계종합학술대회
    • /
    • pp.283-284
    • /
    • 2006
  • In this paper, we propose a novel segmentation and classification method using texture features for the document image. First, we extract the local entropy and then segment the document image to separate the background and the foreground using the Otsu's method. Finally, we classify the segmented regions into each component using PCA(principle component analysis) algorithm based on the texture features that are extracted from the co-occurrence matrix for the entropy image. The entropy-based segmentation is robust to not only noise and the change of light, but also skew and rotation. Texture features are not restricted from any form of the document image and have a superior discrimination for each component. In addition, PCA algorithm used for the classifier can classify the components more robustly than neural network.

  • PDF

시계열 스트리트뷰 데이터베이스를 이용한 시각적 위치 인식 알고리즘 (Visual Location Recognition Using Time-Series Streetview Database)

  • 박천수;최준연
    • 반도체디스플레이기술학회지
    • /
    • 제18권4호
    • /
    • pp.57-61
    • /
    • 2019
  • Nowadays, portable digital cameras such as smart phone cameras are being popularly used for entertainment and visual information recording. Given a database of geo-tagged images, a visual location recognition system can determine the place depicted in a query photo. One of the most common visual location recognition approaches is the bag-of-words method where local image features are clustered into visual words. In this paper, we propose a new bag-of-words-based visual location recognition algorithm using time-series streetview database. The proposed algorithm selects only a small subset of image features which will be used in image retrieval process. By reducing the number of features to be used, the proposed algorithm can reduce the memory requirement of the image database and accelerate the retrieval process.

영상 검색을 위한 Radon 변형의 이용 (Using Radon Transform for Image Retrieval)

  • 서정만
    • 한국컴퓨터정보학회논문지
    • /
    • 제14권6호
    • /
    • pp.65-71
    • /
    • 2009
  • 전통적인 영상 검색 방법은 영상의 색인화와 검색에서 기본적인 특징으로 컬러, 모양, 그리고 질감 들을 사용한다. 우리는 이러한 특징들을 사용하지 않는 새로운 방법을 제시한다. 내용 기반 영상의 색인화와 검색을 위한 유사성 측정에 기하학적 방법을 사용한 시각적 특징을 제시한다. 이 방법은 Radon 변형이라고 한다. 이 방법은 복잡한 분리 방법이 없이 영상의 기하학적 분포에 따라 계산한다. 실험에서도 매우 뛰어난 검색 효과를 보이고 있다.

Spam Image Detection Model based on Deep Learning for Improving Spam Filter

  • Seong-Guk Nam;Dong-Gun Lee;Yeong-Seok Seo
    • Journal of Information Processing Systems
    • /
    • 제19권3호
    • /
    • pp.289-301
    • /
    • 2023
  • Due to the development and dissemination of modern technology, anyone can easily communicate using services such as social network service (SNS) through a personal computer (PC) or smartphone. The development of these technologies has caused many beneficial effects. At the same time, bad effects also occurred, one of which was the spam problem. Spam refers to unwanted or rejected information received by unspecified users. The continuous exposure of such information to service users creates inconvenience in the user's use of the service, and if filtering is not performed correctly, the quality of service deteriorates. Recently, spammers are creating more malicious spam by distorting the image of spam text so that optical character recognition (OCR)-based spam filters cannot easily detect it. Fortunately, the level of transformation of image spam circulated on social media is not serious yet. However, in the mail system, spammers (the person who sends spam) showed various modifications to the spam image for neutralizing OCR, and therefore, the same situation can happen with spam images on social media. Spammers have been shown to interfere with OCR reading through geometric transformations such as image distortion, noise addition, and blurring. Various techniques have been studied to filter image spam, but at the same time, methods of interfering with image spam identification using obfuscated images are also continuously developing. In this paper, we propose a deep learning-based spam image detection model to improve the existing OCR-based spam image detection performance and compensate for vulnerabilities. The proposed model extracts text features and image features from the image using four sub-models. First, the OCR-based text model extracts the text-related features, whether the image contains spam words, and the word embedding vector from the input image. Then, the convolution neural network-based image model extracts image obfuscation and image feature vectors from the input image. The extracted feature is determined whether it is a spam image by the final spam image classifier. As a result of evaluating the F1-score of the proposed model, the performance was about 14 points higher than the OCR-based spam image detection performance.