• Title/Summary/Keyword: 유사 이미지 검색

Search Result 184, Processing Time 0.026 seconds

A Feature Re-weighting Approach for the Non-Metric Feature Space (가변적인 길이의 특성 정보를 지원하는 특성 가중치 조정 기법)

  • Lee Robert-Samuel;Kim Sang-Hee;Park Ho-Hyun;Lee Seok-Lyong;Chung Chin-Wan
    • Journal of KIISE:Databases
    • /
    • v.33 no.4
    • /
    • pp.372-383
    • /
    • 2006
  • Among the approaches to image database management, content-based image retrieval (CBIR) is viewed as having the best support for effective searching and browsing of large digital image libraries. Typical CBIR systems allow a user to provide a query image, from which low-level features are extracted and used to find 'similar' images in a database. However, there exists the semantic gap between human visual perception and low-level representations. An effective methodology for overcoming this semantic gap involves relevance feedback to perform feature re-weighting. Current approaches to feature re-weighting require the number of components for a feature representation to be the same for every image in consideration. Following this assumption, they map each component to an axis in the n-dimensional space, which we call the metric space; likewise the feature representation is stored in a fixed-length vector. However, with the emergence of features that do not have a fixed number of components in their representation, existing feature re-weighting approaches are invalidated. In this paper we propose a feature re-weighting technique that supports features regardless of whether or not they can be mapped into a metric space. Our approach analyses the feature distances calculated between the query image and the images in the database. Two-sided confidence intervals are used with the distances to obtain the information for feature re-weighting. There is no restriction on how the distances are calculated for each feature. This provides freedom for how feature representations are structured, i.e. there is no requirement for features to be represented in fixed-length vectors or metric space. Our experimental results show the effectiveness of our approach and in a comparison with other work, we can see how it outperforms previous work.

A Review of Access Conditions of the W3 and the Inline Image/Sound Processing of HTML Document for Utilizing of the Virtual Library (W3 가상도서관 활용을 위한 HTML 문서작성과 이미지/사운드 처리)

  • 유사라
    • Journal of the Korean Society for information Management
    • /
    • v.12 no.1
    • /
    • pp.45-66
    • /
    • 1995
  • The information users of the middle of 1990s. who know the Internet as well as its useful information services, are now expecting the virtual library services. Especially the increasing demands on hypertext and hypermedia information in the internet settings have been centered on the W3 with the man-page information. In this manner, the paper describes the access methods with brief concepts of the W3 and explains URLs and HTML. It also gives the retrieval layouts of unformatted data including images and sounds and then provides the information sources and software of W3 Clients and Servers in order to catch up the most recently post version of W3.

  • PDF

GB-Index: An Indexing Method for High Dimensional Complex Similarity Queries with Relevance Feedback (GB-색인: 고차원 데이타의 복합 유사 질의 및 적합성 피드백을 위한 색인 기법)

  • Cha Guang-Ho
    • Journal of KIISE:Databases
    • /
    • v.32 no.4
    • /
    • pp.362-371
    • /
    • 2005
  • Similarity indexing and searching are well known to be difficult in high-dimensional applications such as multimedia databases. Especially, they become more difficult when multiple features have to be indexed together. In this paper, we propose a novel indexing method called the GB-index that is designed to efficiently handle complex similarity queries as well as relevance feedback in high-dimensional image databases. In order to provide the flexibility in controlling multiple features and query objects, the GB-index treats each dimension independently The efficiency of the GB-index is realized by specialized bitmap indexing that represents all objects in a database as a set of bitmaps. Main contributions of the GB-index are three-fold: (1) It provides a novel way to index high-dimensional data; (2) It efficiently handles complex similarity queries; and (3) Disjunctive queries driven by relevance feedback are efficiently treated. Empirical results demonstrate that the GB-index achieves great speedups over the sequential scan and the VA-file.

Design and Implementation of Manufacturing Resource Management System as Open Type based on Mold Image Data (금형 이미지 데이터 기반 개방형 제조 자원 관리 시스템의 설계 및 구현)

  • Yoon, Yeo-Been;Cho, Jung-Hyun;Song, Je-O;Lee, Sang-Moon
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2017.01a
    • /
    • pp.227-228
    • /
    • 2017
  • 본 논문에서는 견적 단계부터 공정별 제조 현황을 고객과 공유하고 소통할 수 있는 MRP 시스템과 스마트 사이니지를 이용하여 제조현장 작업자들 중심의 실시간 제조 현황 데이터를 관리 및 활용하여 생산 납품에 대한 예측 관리하며, BOM 중심의 소요 자재 관리는 물론, 소모성 자재를 위한 Part List 개념이 도입된 MRP 기반의 MRO 관리 체계를 마련하는 동시에 고객의 발주 정보를 기초로 유휴 및 유사 금형을 검색하여, 금형의 재활용 및 생산원가와 납기일 등의 코스트 절감에 도움을 줄 수 있는 고객 개방형 제조 관리 시스템 구축을 제안한다.

  • PDF

Object Recognition using SIFT and Tree Structure (SIFT와 트리구조를 이용한 내용기반 물체인식)

  • Joo, Jung-Kyoung;Lee, Hyun-Chang
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2008.06a
    • /
    • pp.33-38
    • /
    • 2008
  • 최근 컴퓨터비전이나 로봇 공학 분야에서 가격이 저렴한 웹캠을 이용한 영상, 즉 2차원 영상으로부터 물체를 인식하는 연구가 활발히 이루어지고 있다. 이러한 로봇이나 비전에서 물체를 찾아내는 여러 가지 방향들이 제시되고 있으며, 지속적으로 로봇은 사람과 유사해져가고 있다. 이를 실현하기 위해서는 사람이 사과를 보고 사과라고 알기 때문에 사과라고 인식하듯이 로봇 또한 미리 알고 있어야 한다는 가정 하에 내용기반의 물체인식이 필요하다. 그러나 엄청난 양의 내용의 데이터베이스가 필요하다. 그래서 용량은 하드웨어기술로 커버가 가능하지만 화면상에 있는 물체들을 빠르게 데이터베이스상의 자료와 매칭이 되어야한다. 본 논문에서는 이미지를 SIFT(Scale Invariant Feature Transform)알고리즘으로 BTS(Binary Search Tree)로 트리구조의 데이터베이스를 구축하여 많은 양의 데이터베이스 중 빠르게 검색하여 화면에 있는 물체를 인식하는 방법을 제안하였다.

  • PDF

Facial Features Detection for Facial Caricaturing System (캐리커처 실성 시스템을 위한 얼굴 특징 추출 연구)

  • Lee, Ok-Kyoung;Park, Yeun-Chool;Oh, Hae-Seok
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2000.10b
    • /
    • pp.1329-1332
    • /
    • 2000
  • 캐리커처 생성 시스템은 입력된 인물 사진을 세그먼테이션을 통하여 특징(이목구비)을 추출하고, 추출된 특징정보를 이용하여 기와 유사한 특징정보를 가지는 캐리커처 이미지를 검색하여 매핑시키는 시스템이다. 캐리커처 생성 시스템에 얼굴 특징정보 추출은 색상과 모양에 대한 정보를 이용한다. 본 논문은 캐리커처생성을 위한 인물 사진을 세그멘테이션 처리하여 부분 영역 특징정보를 추출하는데 그 목적이 있다. 이때 사용하는 이목구비의 특징정보를 위해 수직, 수평의 히스토그램이 주요하게 사용된다. 또한 인물 사진에서 위치정보를 이용하여 얼굴내의 이목구비를 확인하고, 추출하므로 정확한 정보를 이용할 수 있다.

  • PDF

Color Component Analysis For Image Retrieval (이미지 검색을 위한 색상 성분 분석)

  • Choi, Young-Kwan;Choi, Chul;Park, Jang-Chun
    • The KIPS Transactions:PartB
    • /
    • v.11B no.4
    • /
    • pp.403-410
    • /
    • 2004
  • Recently, studies of image analysis, as the preprocessing stage for medical image analysis or image retrieval, are actively carried out. This paper intends to propose a way of utilizing color components for image retrieval. For image retrieval, it is based on color components, and for analysis of color, CLCM (Color Level Co-occurrence Matrix) and statistical techniques are used. CLCM proposed in this paper is to project color components on 3D space through geometric rotate transform and then, to interpret distribution that is made from the spatial relationship. CLCM is 2D histogram that is made in color model, which is created through geometric rotate transform of a color model. In order to analyze it, a statistical technique is used. Like CLCM, GLCM (Gray Level Co-occurrence Matrix)[1] and Invariant Moment [2,3] use 2D distribution chart, which use basic statistical techniques in order to interpret 2D data. However, even though GLCM and Invariant Moment are optimized in each domain, it is impossible to perfectly interpret irregular data available on the spatial coordinates. That is, GLCM and Invariant Moment use only the basic statistical techniques so reliability of the extracted features is low. In order to interpret the spatial relationship and weight of data, this study has used Principal Component Analysis [4,5] that is used in multivariate statistics. In order to increase accuracy of data, it has proposed a way to project color components on 3D space, to rotate it and then, to extract features of data from all angles.

Design and Implementation of Luo-kuan Recognition Application (낙관 인식을 위한 애플리케이션의 설계 및 구현)

  • Kim, Han-Syel;Seo, Kwi-Bin;Kang, Mingoo;Ryu, Gee Soo;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.19 no.1
    • /
    • pp.97-103
    • /
    • 2018
  • In oriental paintings, there is Luo-kuan that expressed in a single picture by compressing the artist's information. Such Luo-kuan includes various information such as the title of the work or the name of the artist. Therefore, information about Luo-kuan is considered important to those who collect or enjoy oriental paintings. However, most of the letters in the Luo-kuan are difficult kanji, kanzai, or various shapes, so it is difficult for the ordinary people to interpret. In this paper, we developed an Luo-kuan search application to easily check the information of the Luo-kuan. The application uses a search algorithm that analyzes the captured Luo-kuan image and sends it to the server to output information about the Luo-kuan candidates that are most similar to the Luo-kuan images taken from the database in the server. We also compared and analyzed the accuracy of the algorithm based on 170 Luo-kuan data in order to find out the ranking of the Luo-kuan that matched the Luo-kuan among the candidates. Accuracy Analysis Experimental Results The accuracy of the search algorithm of this application is confirmed to be about 90%, and it is anticipated that it will be possible to develop a platform to automatically analyze and search images in a big data environment by supplementing the optimizing algorithm and multi-threading algorithm.

Business Application of Convolutional Neural Networks for Apparel Classification Using Runway Image (합성곱 신경망의 비지니스 응용: 런웨이 이미지를 사용한 의류 분류를 중심으로)

  • Seo, Yian;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.1-19
    • /
    • 2018
  • Large amount of data is now available for research and business sectors to extract knowledge from it. This data can be in the form of unstructured data such as audio, text, and image data and can be analyzed by deep learning methodology. Deep learning is now widely used for various estimation, classification, and prediction problems. Especially, fashion business adopts deep learning techniques for apparel recognition, apparel search and retrieval engine, and automatic product recommendation. The core model of these applications is the image classification using Convolutional Neural Networks (CNN). CNN is made up of neurons which learn parameters such as weights while inputs come through and reach outputs. CNN has layer structure which is best suited for image classification as it is comprised of convolutional layer for generating feature maps, pooling layer for reducing the dimensionality of feature maps, and fully-connected layer for classifying the extracted features. However, most of the classification models have been trained using online product image, which is taken under controlled situation such as apparel image itself or professional model wearing apparel. This image may not be an effective way to train the classification model considering the situation when one might want to classify street fashion image or walking image, which is taken in uncontrolled situation and involves people's movement and unexpected pose. Therefore, we propose to train the model with runway apparel image dataset which captures mobility. This will allow the classification model to be trained with far more variable data and enhance the adaptation with diverse query image. To achieve both convergence and generalization of the model, we apply Transfer Learning on our training network. As Transfer Learning in CNN is composed of pre-training and fine-tuning stages, we divide the training step into two. First, we pre-train our architecture with large-scale dataset, ImageNet dataset, which consists of 1.2 million images with 1000 categories including animals, plants, activities, materials, instrumentations, scenes, and foods. We use GoogLeNet for our main architecture as it has achieved great accuracy with efficiency in ImageNet Large Scale Visual Recognition Challenge (ILSVRC). Second, we fine-tune the network with our own runway image dataset. For the runway image dataset, we could not find any previously and publicly made dataset, so we collect the dataset from Google Image Search attaining 2426 images of 32 major fashion brands including Anna Molinari, Balenciaga, Balmain, Brioni, Burberry, Celine, Chanel, Chloe, Christian Dior, Cividini, Dolce and Gabbana, Emilio Pucci, Ermenegildo, Fendi, Giuliana Teso, Gucci, Issey Miyake, Kenzo, Leonard, Louis Vuitton, Marc Jacobs, Marni, Max Mara, Missoni, Moschino, Ralph Lauren, Roberto Cavalli, Sonia Rykiel, Stella McCartney, Valentino, Versace, and Yve Saint Laurent. We perform 10-folded experiments to consider the random generation of training data, and our proposed model has achieved accuracy of 67.2% on final test. Our research suggests several advantages over previous related studies as to our best knowledge, there haven't been any previous studies which trained the network for apparel image classification based on runway image dataset. We suggest the idea of training model with image capturing all the possible postures, which is denoted as mobility, by using our own runway apparel image dataset. Moreover, by applying Transfer Learning and using checkpoint and parameters provided by Tensorflow Slim, we could save time spent on training the classification model as taking 6 minutes per experiment to train the classifier. This model can be used in many business applications where the query image can be runway image, product image, or street fashion image. To be specific, runway query image can be used for mobile application service during fashion week to facilitate brand search, street style query image can be classified during fashion editorial task to classify and label the brand or style, and website query image can be processed by e-commerce multi-complex service providing item information or recommending similar item.

Tilt-based Photo Browsing UI on Mobile Devices (휴대기기에서의 기울임 기반 사진 감상 UI)

  • Jo, Seong-Jeong;Murray-Smith, Roderick;Choe, Chang-Gyu;Seong, Yeong-Hun;Lee, Gwang-Hyeon;Kim, Yeon-Bae
    • 한국HCI학회:학술대회논문집
    • /
    • 2007.02a
    • /
    • pp.429-434
    • /
    • 2007
  • 본 논문에서는 기울임 동작에 기반한 휴대기기 상의 사진 감상 UI를 제시하고 사용성을 평가한다. 기존의 기울임 입력 방식의 세가지 조작성 문제(overshooting, fluctuation, 부분 이미지 제시)를 개선하기 위하여, 사진 위치 및 기울임에 의존하는 사진 이동 제어 dynamics 모델을 제안한다. 본 시스템은 기울임 감지용 가속도 센서, 기울임에 의한 사진 이동 제어 dynamics 모델, 다중모달(시각, 청각, 촉각)을 통한 모델 상태 출력부로 구성된다. 센서 입력과 다중 모달 출력을 위하여, 삼성 MITs 4300 PDA의 배터리 팩을 개조하여 3축 가속도 센서와 진동 출력장치 (VBW32)를 장착하였다. 제안하는 시스템은 기존의 대표적인 사진 감상 입력 방법인 버튼과 iPod wheel과 비교하였다. 정량적 비교를 위하여 7명의 사용자에게 100장의 사진 중 20장을 차례로 검색하는 과제를 부여하면서 수집한 로그를 분석하였으며, 정성적인 비교를 위하여 설문 조사를 실시하였다. 실험 결과 제안한 방법이 기존 기울임 기반 dynamics에 비하여 overshooting 횟수를 30%, 사진간 이동 거리를 25%, 이동 시간을 17% 감소하였다. 또한 제안한 방법이 버튼과 유사한 조작성을 갖고 있으며, 버튼과 iPod보다 더욱 흥미성이 뛰어났다. 상업적으로 뛰어난 성공을 거둔 iPod이 다수의 overshooting 발생으로 실제로는 사용성이 떨어진다는 점이 예상치 못한 흥미로운 발견이었다.

  • PDF