• Title/Summary/Keyword: 질감정보

Search Result 316, Processing Time 0.025 seconds

A Study on Increasing the Efficiency of Image Search Using Image Attribute in the area of content-Based Image Retrieval (내용기반 이미지 검색에 있어 이미지 속성정보를 활용한 검색 효율성 향상)

  • Mo, Yeong-Il;Lee, Cheol-Gyu
    • Journal of the Korea Society for Simulation
    • /
    • v.18 no.2
    • /
    • pp.39-48
    • /
    • 2009
  • This study reviews the limit of image search by considering on the image search methods related to content-based image retrieval and suggests a user interface for more efficient content-based image retrieval and the ways to utilize image properties. For now, most studies on image search are being performed focusing on content-based image retrieval; they try to search based on the image's colors, texture, shapes, and the overall form of the image. However, the results are not satisfactory because there are various technological limits. Accordingly, this study suggests a new retrieval system which adapts content-based image retrieval and the conventional keyword search method. This is about a way to attribute properties to images using texts and a fast way to search images by expressing the attribute of images as keywords and utilizing them to search images. Also, the study focuses on a simulation for a user interface to make query language on the Internet and a search for clothes in an online shopping mall as an application of the retrieval system based on image attribute. This study will contribute to adding a new purchase pattern in online shopping malls and to the development of the area of similar image search.

Comparative Analysis of Self-supervised Deephashing Models for Efficient Image Retrieval System (효율적인 이미지 검색 시스템을 위한 자기 감독 딥해싱 모델의 비교 분석)

  • Kim Soo In;Jeon Young Jin;Lee Sang Bum;Kim Won Gyum
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.12
    • /
    • pp.519-524
    • /
    • 2023
  • In hashing-based image retrieval, the hash code of a manipulated image is different from the original image, making it difficult to search for the same image. This paper proposes and evaluates a self-supervised deephashing model that generates perceptual hash codes from feature information such as texture, shape, and color of images. The comparison models are autoencoder-based variational inference models, but the encoder is designed with a fully connected layer, convolutional neural network, and transformer modules. The proposed model is a variational inference model that includes a SimAM module of extracting geometric patterns and positional relationships within images. The SimAM module can learn latent vectors highlighting objects or local regions through an energy function using the activation values of neurons and surrounding neurons. The proposed method is a representation learning model that can generate low-dimensional latent vectors from high-dimensional input images, and the latent vectors are binarized into distinguishable hash code. From the experimental results on public datasets such as CIFAR-10, ImageNet, and NUS-WIDE, the proposed model is superior to the comparative model and analyzed to have equivalent performance to the supervised learning-based deephashing model. The proposed model can be used in application systems that require low-dimensional representation of images, such as image search or copyright image determination.

Story-based Information Retrieval (스토리 기반의 정보 검색 연구)

  • You, Eun-Soon;Park, Seung-Bo
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.81-96
    • /
    • 2013
  • Video information retrieval has become a very important issue because of the explosive increase in video data from Web content development. Meanwhile, content-based video analysis using visual features has been the main source for video information retrieval and browsing. Content in video can be represented with content-based analysis techniques, which can extract various features from audio-visual data such as frames, shots, colors, texture, or shape. Moreover, similarity between videos can be measured through content-based analysis. However, a movie that is one of typical types of video data is organized by story as well as audio-visual data. This causes a semantic gap between significant information recognized by people and information resulting from content-based analysis, when content-based video analysis using only audio-visual data of low level is applied to information retrieval of movie. The reason for this semantic gap is that the story line for a movie is high level information, with relationships in the content that changes as the movie progresses. Information retrieval related to the story line of a movie cannot be executed by only content-based analysis techniques. A formal model is needed, which can determine relationships among movie contents, or track meaning changes, in order to accurately retrieve the story information. Recently, story-based video analysis techniques have emerged using a social network concept for story information retrieval. These approaches represent a story by using the relationships between characters in a movie, but these approaches have problems. First, they do not express dynamic changes in relationships between characters according to story development. Second, they miss profound information, such as emotions indicating the identities and psychological states of the characters. Emotion is essential to understanding a character's motivation, conflict, and resolution. Third, they do not take account of events and background that contribute to the story. As a result, this paper reviews the importance and weaknesses of previous video analysis methods ranging from content-based approaches to story analysis based on social network. Also, we suggest necessary elements, such as character, background, and events, based on narrative structures introduced in the literature. We extract characters' emotional words from the script of the movie Pretty Woman by using the hierarchical attribute of WordNet, which is an extensive English thesaurus. WordNet offers relationships between words (e.g., synonyms, hypernyms, hyponyms, antonyms). We present a method to visualize the emotional pattern of a character over time. Second, a character's inner nature must be predetermined in order to model a character arc that can depict the character's growth and development. To this end, we analyze the amount of the character's dialogue in the script and track the character's inner nature using social network concepts, such as in-degree (incoming links) and out-degree (outgoing links). Additionally, we propose a method that can track a character's inner nature by tracing indices such as degree, in-degree, and out-degree of the character network in a movie through its progression. Finally, the spatial background where characters meet and where events take place is an important element in the story. We take advantage of the movie script to extracting significant spatial background and suggest a scene map describing spatial arrangements and distances in the movie. Important places where main characters first meet or where they stay during long periods of time can be extracted through this scene map. In view of the aforementioned three elements (character, event, background), we extract a variety of information related to the story and evaluate the performance of the proposed method. We can track story information extracted over time and detect a change in the character's emotion or inner nature, spatial movement, and conflicts and resolutions in the story.

An Efficient Block Segmentation and Classification Method for Document Image Analysis Using SGLDM and BP (공간의존행렬과 신경망을 이용한 문서영상의 효과적인 블록분할과 유형분류)

  • Kim, Jung-Su;Lee, Jeong-Hwan;Choe, Heung-Mun
    • The Transactions of the Korea Information Processing Society
    • /
    • v.2 no.6
    • /
    • pp.937-946
    • /
    • 1995
  • We proposed and efficient block segmentation and classification method for the document analysis using SGLDM(spatial gray level dependence matrix) and BP (back Propagation) neural network. Seven texture features are extracted directly from the SGLDM of each gray-level block image, and by using the nonlinear classifier of neural network BP, we can classify document blocks into 9 categories. The proposed method classifies the equation block, the table block and the flow chart block, which are mostly composed of the characters, out of the blocks that are conventionally classified as non-character blocks. By applying Sobel operator on the gray-level document image beforebinarization, we can reduce the effect of the background noises, and by using the additional horizontal-vertical smoothing as well as the vertical-horizontal smoothing of images, we can obtain an effective block segmentation that does not lead to the segmentation into small pieces. The result of experiment shows that a document can be segmented and classified into the character blocks of large fonts, small fonts, the character recognigible candidates of tables, flow charts, equations, and the non-character blocks of photos, figures, and graphs.

  • PDF

The Method of Wet Road Surface Condition Detection With Image Processing at Night (영상처리기반 야간 젖은 노면 판별을 위한 방법론)

  • KIM, Youngmin;BAIK, Namcheol
    • Journal of Korean Society of Transportation
    • /
    • v.33 no.3
    • /
    • pp.284-293
    • /
    • 2015
  • The objective of this paper is to determine the conditions of road surface by utilizing the images collected from closed-circuit television (CCTV) cameras installed on roadside. First, a technique was examined to detect wet surfaces at nighttime. From the literature reviews, it was revealed that image processing using polarization is one of the preferred options. However, it is hard to use the polarization characteristics of road surface images at nighttime because of irregular or no light situations. In this study, we proposes a new discriminant for detecting wet and dry road surfaces using CCTV image data at night. To detect the road surface conditions with night vision, we applied the wavelet packet transform for analyzing road surface textures. Additionally, to apply the luminance feature of night CCTV images, we set the intensity histogram based on HSI(Hue Saturation Intensity) color model. With a set of 200 images taken from the field, we constructed a detection criteria hyperplane with SVM (Support Vector Machine). We conducted field tests to verify the detection ability of the wet road surfaces and obtained reliable results. The outcome of this study is also expected to be used for monitoring road surfaces to improve safety.

Computer-Aided Diagnosis for Liver Cirrhosis using Texture features Information Analysis in Computed Tomography (컴퓨터단층영상에서 TIA를 이용한 간경화의 컴퓨터보조진단)

  • Kim, Chang-Soo;Ko, Seong-Jin;Kang, Se-Sik;Kim, Jung-Hoon;Kim, Dong-Hyun;Choi, Seok-Yoon
    • The Journal of the Korea Contents Association
    • /
    • v.12 no.4
    • /
    • pp.358-366
    • /
    • 2012
  • Cirrhosis is a consequence of chronic liver disease characterized by replacement of liver tissue by fibrosis, scar tissue and regenerative nodules leading to loss of liver function. Liver Cirrhosis is most commonly caused by alcoholism, hepatitis B and C, and fatty liver disease, but has many other possible causes. Some cases are idiopathic disease from unknown cause. Abdomen of liver Computed tomography(CT) is one of the primary imaging procedures for evaluating liver disease such as liver cirrhosis, Alcoholic liver disease(ALD), cancer, and interval changes because it is economical and easy to use. The purpose of this study is to detect technique for computer-aided diagnosis(CAD) to identify liver cirrhosis in abdomen CT. We experimented on the principal components analysis(PCA) algorithm in the other method and suggested texture information analysis(TIA). Forty clinical cases involving a total of 634 CT sectional images were used in this study. Liver cirrhosis was detected by PCA method(detection rate of 35%), and by TIA methods(detection rate of 100%-AGI, TM, MU, EN). Our present results show that our method can be regarded as a technique for CAD systems to detect liver cirrhosis in CT liver images.

Content Based Image Retrieval using 8AB Representation of Spatial Relations between Objects (객체 위치 관계의 8AB 표현을 이용한 내용 기반 영상 검색 기법)

  • Joo, Chan-Hye;Chung, Chin-Wan;Park, Ho-Hyun;Lee, Seok-Lyong;Kim, Sang-Hee
    • Journal of KIISE:Databases
    • /
    • v.34 no.4
    • /
    • pp.304-314
    • /
    • 2007
  • Content Based Image Retrieval (CBIR) is to store and retrieve images using the feature description of image contents. In order to support more accurate image retrieval, it has become necessary to develop features that can effectively describe image contents. The commonly used low-level features, such as color, texture, and shape features may not be directly mapped to human visual perception. In addition, such features cannot effectively describe a single image that contains multiple objects of interest. As a result, the research on feature descriptions has shifted to focus on higher-level features, which support representations more similar to human visual perception like spatial relationships between objects. Nevertheless, the prior works on the representation of spatial relations still have shortcomings, particularly with respect to supporting rotational invariance, Rotational invariance is a key requirement for a feature description to provide robust and accurate retrieval of images. This paper proposes a high-level feature named 8AB (8 Angular Bin) that effectively describes the spatial relations of objects in an image while providing rotational invariance. With this representation, a similarity calculation and a retrieval technique are also proposed. In addition, this paper proposes a search-space pruning technique, which supports efficient image retrieval using the 8AB feature. The 8AB feature is incorporated into a CBIR system, and the experiments over both real and synthetic image sets show the effectiveness of 8AB as a high-level feature and the efficiency of the pruning technique.

Quantitative assessment of Endorectal Ultrasonography by using GLCM Algorithm (GLCM알고리즘을 이용한 경직장 초음파 영상의 정량적 평가)

  • Nho, Da-Jung;Kang, Min-Ji;Kim, Yoo-Kyeong;Seo, Ah-Reum;Lee, In-Ho;Jeong, Hee-Seong;Jo, Jin-Yeong;Ko, Seong-Jin
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2015.05a
    • /
    • pp.383-387
    • /
    • 2015
  • Bowel and rectal diseases are on the increase by irregular life and westernized eating habits of modern people, especially rectal cancer, which accounts for 50% of the entire colon cancer. For the initial rectal cancer, because there is no portion projecting on the surface, if not see inside the tissue with ultrasound, you make an errors that misdiagnosis as rectal abscess. However there is a need for more accurate diagnosis, because it is sometimes difficult to distinguish abscess from rectal cancer depending on staging, in spite of the ultrasonic diagnosis. Therefore, this study was performed quantitative analysis by using a computer algorithm for rectal cancer and abscess image. Each of 20 cases about normal, abscess and cancer by setting analysis region ($50{\times}50$ pixels) applies to GLCM algorithm and Autocorrelation, Max probability, Sum average, Sum variance in each image were analyzed by comparing the 4 single parameter. Consequently, The high lesion detection efficiency was presented 100% by the 3 parameter of Autocorrelation, Max probability, Sum variance and the parameter of Sum average presents 95% in cancer, more than 90% in abscess. Those parameters are valuable in distinction standard about normal, cancer and abscess in rectum. It is sufficient availability as a computer assisted diagnosis system depended on clinical using.

  • PDF

High-resolution 3D Object Reconstruction using Multiple Cameras (다수의 카메라를 활용한 고해상도 3차원 객체 복원 시스템)

  • Hwang, Sung Soo;Yoo, Jisung;Kim, Hee-Dong;Kim, Sujung;Paeng, Kyunghyun;Kim, Seong Dae
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.10
    • /
    • pp.150-161
    • /
    • 2013
  • This paper presents a new system which produces high resolution 3D contents by capturing multiview images of an object using multiple cameras, and estimating geometric and texture information of the object from the captured images. Even though a variety of multiview image-based 3D reconstruction systems have been proposed, it was difficult to generate high resolution 3D contents because multiview image-based 3D reconstruction requires a large amount of memory and computation. In order to reduce computational complexity and memory size for 3D reconstruction, the proposed system predetermines the regions in input images where an object can exist to extract object boundaries fast. And for fast computation of a visual hull, the system represents silhouettes and 3D-2D projection/back-projection relations by chain codes and 1D homographies, respectively. The geometric data of the reconstructed object is compactly represented by a 3D segment-based data format which is called DoCube, and the 3D object is finally reconstructed after 3D mesh generation and texture mapping are performed. Experimental results show that the proposed system produces 3D object contents of $800{\times}800{\times}800$ resolution with a rate of 2.2 seconds per frame.

A Study on the Improvement of Skin Loss Area in Skin Color Extraction for Face Detection (얼굴 검출을 위한 피부색 추출 과정에서 피부색 손실 영역 개선에 관한 연구)

  • Kim, Dong In;Lee, Gang Seong;Han, Kun Hee;Lee, Sang Hun
    • Journal of the Korea Convergence Society
    • /
    • v.10 no.5
    • /
    • pp.1-8
    • /
    • 2019
  • In this paper, we propose an improved facial skin color extraction method to solve the problem that facial surface is lost due to shadow or illumination in skin color extraction process and skin color extraction is not possible. In the conventional HSV method, when facial surface is brightly illuminated by light, the skin color component is lost in the skin color extraction process, so that a loss area appears on the face surface. In order to solve these problems, we extract the skin color, determine the elements in the H channel value range of the skin color in the HSV color space among the lost skin elements, and combine the coordinates of the lost part with the coordinates of the original image, To minimize the number of In the face detection process, the face was detected using the LBP Cascade Classifier, which represents texture feature information in the extracted skin color image. Experimental results show that the proposed method improves the detection rate and accuracy by 5.8% and 9.6%, respectively, compared with conventional RGB and HSV skin color extraction and face detection using the LBP cascade classifier method.