• Title/Summary/Keyword: 내용기반이미지검색

Search Result 244, Processing Time 0.036 seconds

Reliable Image-Text Fusion CAPTCHA to Improve User-Friendliness and Efficiency (사용자 편의성과 효율성을 증진하기 위한 신뢰도 높은 이미지-텍스트 융합 CAPTCHA)

  • Moon, Kwang-Ho;Kim, Yoo-Sung
    • The KIPS Transactions:PartC
    • /
    • v.17C no.1
    • /
    • pp.27-36
    • /
    • 2010
  • In Web registration pages and online polling applications, CAPTCHA(Completely Automated Public Turing Test To Tell Computers and Human Apart) is used for distinguishing human users from automated programs. Text-based CAPTCHAs have been widely used in many popular Web sites in which distorted text is used. However, because the advanced optical character recognition techniques can recognize the distorted texts, the reliability becomes low. Image-based CAPTCHAs have been proposed to improve the reliability of the text-based CAPTCHAs. However, these systems also are known as having some drawbacks. First, some image-based CAPTCHA systems with small number of image files in their image dictionary is not so reliable since attacker can recognize images by repeated executions of machine learning programs. Second, users may feel uncomfortable since they have to try CAPTCHA tests repeatedly when they fail to input a correct keyword. Third, some image-base CAPTCHAs require high communication cost since they should send several image files for one CAPTCHA. To solve these problems of image-based CAPTCHA, this paper proposes a new CAPTCHA based on both image and text. In this system, an image and keywords are integrated into one CAPTCHA image to give user a hint for the answer keyword. The proposed CAPTCHA can help users to input easily the answer keyword with the hint in the fused image. Also, the proposed system can reduce the communication costs since it uses only a fused image file for one CAPTCHA. To improve the reliability of the image-text fusion CAPTCHA, we also propose a dynamic building method of large image dictionary from gathering huge amount of images from theinternet with filtering phase for preserving the correctness of CAPTCHA images. In this paper, we proved that the proposed image-text fusion CAPTCHA provides users more convenience and high reliability than the image-based CAPTCHA through experiments.

Image Retrieval using Distribution Block Signature of Main Colors' Set and Performance Boosting via Relevance feedback (주요 색상의 분포 블록기호를 이용한 영상검색과 유사도 피드백을 통한 이미지 검색)

  • 박한수;유헌우;장동식
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.2
    • /
    • pp.126-136
    • /
    • 2004
  • This paper proposes a new content-based image retrieval algorithm using color-spatial information. For the purpose, the paper suggests two kinds of indexing key to prune away irrelevant images to a given query image; MCS(Main Colors' Set), which is related with color information and DBS (Distribution Block Signature), which is related with spatial information. After successively applying these filters to a database, we could get a small amount of high potential candidates that are somewhat similar to the query image. Then we would make use of new QM(Quad modeling) and relevance feedback mechanism to obtain more accurate retrieval. It would enhance the retrieval effectiveness by dynamically modulating the weights of color-spatial information. Experiments show that the proposed algorithm can apply successfully image retrieval applications.

A Study of Medium Shot Detection (미디엄 숏 검출에 관한 연구)

  • Hyung Lee
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2023.01a
    • /
    • pp.93-95
    • /
    • 2023
  • 본 논문에서는 장편의 드라마나 영화에서 스토리 기반의 축약된 요약본을 자동으로 제작하기 위해 미디엄 숏(medium shot) 크기의 숏(shot)들을 추출하기 위한 방법을 고려한다. 미디엄 숏 정도의 크기는 보통 인물에 중심을 둔 숏들로 인물들 간의 관계에서 특히 대사나 표정으로 내용을 전달하기 위한 목적으로 적극 권장된다. 비디오 검색을 위한 인덱싱에서 신(scene) 전환 검출 및 숏 경계 검출, 그리고 이미지에서 심도와 초점기반의 화질 및 피사체 추출 등을 위해 전통적인 신호/영상처리 기법의 활용에서부터 최근의 기계학습 접목 등 다양한 연구들이 진행되고 있다. 영상문법에 근거하여 편집된 영상물에서 미디엄 숏 정도 크기의 숏들을 추출하여 배열한다면 어느 정도 원본 내용을 충실히 전달할 수 있는 축약된 요약본을 제작할 수 있다는 가정하에 해당 샷들을 블러(blur) 기반으로 검출하기 위해 이와 관련된 키워드들을 기반으로 기존 연구들을 살펴보고 적용 방법을 모색한다.

  • PDF

A Mobile Landmarks Guide : Outdoor Augmented Reality based on LOD and Contextual Device (모바일 랜드마크 가이드 : LOD와 문맥적 장치 기반의 실외 증강현실)

  • Zhao, Bi-Cheng;Rosli, Ahmad Nurzid;Jang, Chol-Hee;Lee, Kee-Sung;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.1
    • /
    • pp.1-21
    • /
    • 2012
  • In recent years, mobile phone has experienced an extremely fast evolution. It is equipped with high-quality color displays, high resolution cameras, and real-time accelerated 3D graphics. In addition, some other features are includes GPS sensor and Digital Compass, etc. This evolution advent significantly helps the application developers to use the power of smart-phones, to create a rich environment that offers a wide range of services and exciting possibilities. To date mobile AR in outdoor research there are many popular location-based AR services, such Layar and Wikitude. These systems have big limitation the AR contents hardly overlaid on the real target. Another research is context-based AR services using image recognition and tracking. The AR contents are precisely overlaid on the real target. But the real-time performance is restricted by the retrieval time and hardly implement in large scale area. In our work, we exploit to combine advantages of location-based AR with context-based AR. The system can easily find out surrounding landmarks first and then do the recognition and tracking with them. The proposed system mainly consists of two major parts-landmark browsing module and annotation module. In landmark browsing module, user can view an augmented virtual information (information media), such as text, picture and video on their smart-phone viewfinder, when they pointing out their smart-phone to a certain building or landmark. For this, landmark recognition technique is applied in this work. SURF point-based features are used in the matching process due to their robustness. To ensure the image retrieval and matching processes is fast enough for real time tracking, we exploit the contextual device (GPS and digital compass) information. This is necessary to select the nearest and pointed orientation landmarks from the database. The queried image is only matched with this selected data. Therefore, the speed for matching will be significantly increased. Secondly is the annotation module. Instead of viewing only the augmented information media, user can create virtual annotation based on linked data. Having to know a full knowledge about the landmark, are not necessary required. They can simply look for the appropriate topic by searching it with a keyword in linked data. With this, it helps the system to find out target URI in order to generate correct AR contents. On the other hand, in order to recognize target landmarks, images of selected building or landmark are captured from different angle and distance. This procedure looks like a similar processing of building a connection between the real building and the virtual information existed in the Linked Open Data. In our experiments, search range in the database is reduced by clustering images into groups according to their coordinates. A Grid-base clustering method and user location information are used to restrict the retrieval range. Comparing the existed research using cluster and GPS information the retrieval time is around 70~80ms. Experiment results show our approach the retrieval time reduces to around 18~20ms in average. Therefore the totally processing time is reduced from 490~540ms to 438~480ms. The performance improvement will be more obvious when the database growing. It demonstrates the proposed system is efficient and robust in many cases.

Object Recognition using SIFT and Tree Structure (SIFT와 트리구조를 이용한 내용기반 물체인식)

  • Joo, Jung-Kyoung;Lee, Hyun-Chang
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2008.06a
    • /
    • pp.33-38
    • /
    • 2008
  • 최근 컴퓨터비전이나 로봇 공학 분야에서 가격이 저렴한 웹캠을 이용한 영상, 즉 2차원 영상으로부터 물체를 인식하는 연구가 활발히 이루어지고 있다. 이러한 로봇이나 비전에서 물체를 찾아내는 여러 가지 방향들이 제시되고 있으며, 지속적으로 로봇은 사람과 유사해져가고 있다. 이를 실현하기 위해서는 사람이 사과를 보고 사과라고 알기 때문에 사과라고 인식하듯이 로봇 또한 미리 알고 있어야 한다는 가정 하에 내용기반의 물체인식이 필요하다. 그러나 엄청난 양의 내용의 데이터베이스가 필요하다. 그래서 용량은 하드웨어기술로 커버가 가능하지만 화면상에 있는 물체들을 빠르게 데이터베이스상의 자료와 매칭이 되어야한다. 본 논문에서는 이미지를 SIFT(Scale Invariant Feature Transform)알고리즘으로 BTS(Binary Search Tree)로 트리구조의 데이터베이스를 구축하여 많은 양의 데이터베이스 중 빠르게 검색하여 화면에 있는 물체를 인식하는 방법을 제안하였다.

  • PDF

A Study on Metadata for Images and Videos (이미지 및 비디오 메타데이터에 관한 연구)

  • Im, Eun-Ju;Kim, Seong-Hee
    • Proceedings of the Korean Society for Information Management Conference
    • /
    • 2005.08a
    • /
    • pp.117-124
    • /
    • 2005
  • 이미지와 비디오가 효율적으로 검색되고 이용되기 위해서는 메타데이터가 구조적으로 기술될 필요가 있다. 본 논문에서는 메타데이터 구축 체계의 기반으로 MPEG-7과 Plateau, ViMod에 대해 분석하였다. MPEG-7에서는 저작 및 생산, 이용, 컨텐츠의 구조적, 개념적 측면 등과 같은 다양한 관점에서 시청각 정보가 기술된다. Plateau는 다양한 응용 분야에 적합하도록 모델링 하였기 때문에 모델이 간단명료하고 응용분야에서는 전문성을 띠고 있으나 범용성이 부족하다는 단점이 있다. ViMod는 질의의 특성을 질의의 확실성, 질의의 차원성, 데이터 의존성 관점에서 정지영상과 비디오 영상으로 분류하고 그 하위 레벨을 내용 의존적인 영상과 내용 독립적인 영상으로 분류하였다.

  • PDF

Methods for Video Caption Extraction and Extracted Caption Image Enhancement (영화 비디오 자막 추출 및 추출된 자막 이미지 향상 방법)

  • Kim, So-Myung;Kwak, Sang-Shin;Choi, Yeong-Woo;Chung, Kyu-Sik
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.4
    • /
    • pp.235-247
    • /
    • 2002
  • For an efficient indexing and retrieval of digital video data, research on video caption extraction and recognition is required. This paper proposes methods for extracting artificial captions from video data and enhancing their image quality for an accurate Hangul and English character recognition. In the proposed methods, we first find locations of beginning and ending frames of the same caption contents and combine those multiple frames in each group by logical operation to remove background noises. During this process an evaluation is performed for detecting the integrated results with different caption images. After the multiple video frames are integrated, four different image enhancement techniques are applied to the image: resolution enhancement, contrast enhancement, stroke-based binarization, and morphological smoothing operations. By applying these operations to the video frames we can even improve the image quality of phonemes with complex strokes. Finding the beginning and ending locations of the frames with the same caption contents can be effectively used for the digital video indexing and browsing. We have tested the proposed methods with the video caption images containing both Hangul and English characters from cinema, and obtained the improved results of the character recognition.

The Extraction of Effective Index Database from Voice Database and Information Retrieval (음성 데이터베이스로부터의 효율적인 색인데이터베이스 구축과 정보검색)

  • Park Mi-Sung
    • Journal of Korean Library and Information Science Society
    • /
    • v.35 no.3
    • /
    • pp.271-291
    • /
    • 2004
  • Such information services source like digital library has been asked information services of atypical multimedia database like image, voice, VOD/AOD. Examined in this study are suggestions such as word-phrase generator, syllable recoverer, morphological analyzer, corrector for voice processing. Suggested voice processing technique transform voice database into tort database, then extract index database from text database. On top of this, the study suggest a information retrieval model to use in extracted index database, voice full-text information retrieval.

  • PDF

Image retrieval using multiresolution image partition (다해상도 이미지 분할을 이용한 영상 검색)

  • Ahn, Ill-Whan;Kim, Woo-Sung;Lee, In-Sue
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2000.04a
    • /
    • pp.874-878
    • /
    • 2000
  • 본 논문에서는 내용 기반 영상 검색 방법 중 "외각선 영역의 색상 분포에 의한 영상 검색"을 제안한다. 영상의 변화가 큰 곳은 해상도를 높게, 낮은 곳은 해상도를 낮게 데이터를 샘플링하여 비교할 데이터의 양을 줄이고, 외각선 영역의 색상을 검출하는데 사용한다. 이때 에지 트리(Edge Tree)를 이용하여 에지(Edge)의 위치 정보와 색상 정보를 유지하며, 검색을 가능하게 하였다. 이와 같은 방법을 사용해서 원하는 영상을 찾을 수 있음을 확인하였다.

  • PDF

Implementation of a Video Retrieval System Using Annotation and Comparison Area Learning of Key-Frames (키 프레임의 주석과 비교 영역 학습을 이용한 비디오 검색 시스템의 구현)

  • Lee Keun-Wang;Kim Hee-Sook;Lee Jong-Hee
    • Journal of Korea Multimedia Society
    • /
    • v.8 no.2
    • /
    • pp.269-278
    • /
    • 2005
  • In order to process video data effectively, it is required that the content information of video data is loaded in database and semantics-based retrieval method can be available for various queries of users. In this paper, we propose a video retrieval system which support semantics retrieval of various users for massive video data by user's keywords and comparison area learning based on automatic agent. By user's fundamental query and selection of image for key frame that extracted from query, the agent gives the detail shape for annotation of extracted key frame. Also, key frame selected by user becomes a query image and searches the most similar key frame through color histogram comparison and comparison area learning method that proposed. From experiment, the designed and implemented system showed high precision ratio in performance assessment more than 93 percents.

  • PDF