• Title/Summary/Keyword: Automatic Extraction

Search Result 879, Processing Time 0.024 seconds

An Automatic Business Service Identification for Effective Relevant Information Retrieval of Defense Digital Archive (국방 디지털 아카이브의 효율적 연관정보 검색을 위한 자동화된 비즈니스 서비스 식별)

  • Byun, Young-Tae;Hwang, Sang-Kyu;Jung, Chan-Ki
    • Journal of the Korean Society for information Management
    • /
    • v.27 no.4
    • /
    • pp.33-47
    • /
    • 2010
  • The growth of IT technology and the popularity of network based information sharing increase the number of digital contents in military area. Thus, there arise issues of finding suitable public information with the growing number of long-term preservation of digital public information. According to the source of raw data and the time of compilation may be variable and there can be existed in many correlations about digital contents. The business service ontology makes knowledge explicit and allows for knowledge sharing among information provider and information consumer for public digital archive engaged in improving the searching ability of digital public information. The business service ontology is at the interface as a bridge between information provider and information consumer. However, according to the difficulty of semantic knowledge extraction for the business process analysis, it is hard to realize the automation of constructing business service ontology for mapping from unformed activities to a unit of business service. To solve the problem, we propose a new business service auto-acquisition method for the first step of constructing a business service ontology based on Enterprise Architecture.

Multiple Cause Model-based Topic Extraction and Semantic Kernel Construction from Text Documents (다중요인모델에 기반한 텍스트 문서에서의 토픽 추출 및 의미 커널 구축)

  • 장정호;장병탁
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.5
    • /
    • pp.595-604
    • /
    • 2004
  • Automatic analysis of concepts or semantic relations from text documents enables not only an efficient acquisition of relevant information, but also a comparison of documents in the concept level. We present a multiple cause model-based approach to text analysis, where latent topics are automatically extracted from document sets and similarity between documents is measured by semantic kernels constructed from the extracted topics. In our approach, a document is assumed to be generated by various combinations of underlying topics. A topic is defined by a set of words that are related to the same topic or cooccur frequently within a document. In a network representing a multiple-cause model, each topic is identified by a group of words having high connection weights from a latent node. In order to facilitate teaming and inferences in multiple-cause models, some approximation methods are required and we utilize an approximation by Helmholtz machines. In an experiment on TDT-2 data set, we extract sets of meaningful words where each set contains some theme-specific terms. Using semantic kernels constructed from latent topics extracted by multiple cause models, we also achieve significant improvements over the basic vector space model in terms of retrieval effectiveness.

Automatic Extraction of Initial Training Data Using National Land Cover Map and Unsupervised Classification and Updating Land Cover Map (국가토지피복도와 무감독분류를 이용한 초기 훈련자료 자동추출과 토지피복지도 갱신)

  • Soungki, Lee;Seok Keun, Choi;Sintaek, Noh;Noyeol, Lim;Juweon, Choi
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.33 no.4
    • /
    • pp.267-275
    • /
    • 2015
  • Those land cover maps have widely been used in various fields, such as environmental studies, military strategies as well as in decision-makings. This study proposes a method to extract training data, automatically and classify the cover using ingle satellite images and national land cover maps, provided by the Ministry of Environment. For this purpose, as the initial training data, those three were used; the unsupervised classification, the ISODATA, and the existing land cover maps. The class was classified and named automatically using the class information in the existing land cover maps to overcome the difficulty in selecting classification by each class and in naming class by the unsupervised classification; so as achieve difficulty in selecting the training data in supervised classification. The extracted initial training data were utilized as the training data of MLC for the land cover classification of target satellite images, which increase the accuracy of unsupervised classification. Finally, the land cover maps could be extracted from updated training data that has been applied by an iterative method. Also, in order to reduce salt and pepper occurring in the pixel classification method, the MRF was applied in each repeated phase to enhance the accuracy of classification. It was verified quantitatively and visually that the proposed method could effectively generate the land cover maps.

A Novel Method for Automated Honeycomb Segmentation in HRCT Using Pathology-specific Morphological Analysis (병리특이적 형태분석 기법을 이용한 HRCT 영상에서의 새로운 봉와양폐 자동 분할 방법)

  • Kim, Young Jae;Kim, Tae Yun;Lee, Seung Hyun;Kim, Kwang Gi;Kim, Jong Hyo
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.2
    • /
    • pp.109-114
    • /
    • 2012
  • Honeycombs are dense structures that small cysts, which generally have about 2~10 mm in diameter, are surrounded by the wall of fibrosis. When honeycomb is found in the patients, the incidence of acute exacerbation is generally very high. Thus, the observation and quantitative measurement of honeycomb are considered as a significant marker for clinical diagnosis. In this point of view, we propose an automatic segmentation method using morphological image processing and assessment of the degree of clustering techniques. Firstly, image noises were removed by the Gaussian filtering and then a morphological dilation method was applied to segment lung regions. Secondly, honeycomb cyst candidates were detected through the 8-neighborhood pixel exploration, and then non-cyst regions were removed using the region growing method and wall pattern testing. Lastly, final honeycomb regions were segmented through the extraction of dense regions which are consisted of two or more cysts using cluster analysis. The proposed method applied to 80 High resolution computed tomography (HRCT) images and achieved a sensitivity of 89.4% and PPV (Positive Predictive Value) of 72.2%.

Development of an Image Processing System for the Large Size High Resolution Satellite Images (대용량 고해상 위성영상처리 시스템 개발)

  • 김경옥;양영규;안충현
    • Korean Journal of Remote Sensing
    • /
    • v.14 no.4
    • /
    • pp.376-391
    • /
    • 1998
  • Images from satellites will have 1 to 3 meter ground resolution and will be very useful for analyzing current status of earth surface. An image processing system named GeoWatch with more intelligent image processing algorithms has been designed and implemented to support the detailed analysis of the land surface using high-resolution satellite imagery. The GeoWatch is a valuable tool for satellite image processing such as digitizing, geometric correction using ground control points, interactive enhancement, various transforms, arithmetic operations, calculating vegetation indices. It can be used for investigating various facts such as the change detection, land cover classification, capacity estimation of the industrial complex, urban information extraction, etc. using more intelligent analysis method with a variety of visual techniques. The strong points of this system are flexible algorithm-save-method for efficient handling of large size images (e.g. full scenes), automatic menu generation and powerful visual programming environment. Most of the existing image processing systems use general graphic user interfaces. In this paper we adopted visual program language for remotely sensed image processing for its powerful programmability and ease of use. This system is an integrated raster/vector analysis system and equipped with many useful functions such as vector overlay, flight simulation, 3D display, and object modeling techniques, etc. In addition to the modules for image and digital signal processing, the system provides many other utilities such as a toolbox and an interactive image editor. This paper also presents several cases of image analysis methods with AI (Artificial Intelligent) technique and design concept for visual programming environment.

Automatic Extraction of River Levee Slope Using MMS Point Cloud Data (MMS 포인트 클라우드를 활용한 하천제방 경사도 자동 추출에 관한 연구)

  • Kim, Cheolhwan;Lee, Jisang;Choi, Wonjun;Kim, Wondae;Sohn, Hong-Gyoo
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.5_3
    • /
    • pp.1425-1434
    • /
    • 2021
  • Continuous and periodic data acquisition must be preceded to maintain and manage the river facilities effectively. Adapting the existing general facilities methods, which include river surveying methods such as terrestrial laser scanners, total stations, and Global Navigation Satellite System (GNSS), has limitation in terms of its costs, manpower, and times to acquire spatial information since the river facilities are distributed across the wide and long area. On the other hand, the Mobile Mapping System (MMS) has comparative advantage in acquiring the data of river facilities since it constructs three-dimensional spatial information while moving. By using the MMS, 184,646,009 points could be attained for Anyang stream with a length of 4 kilometers only in 20 minutes. Levee points were divided at intervals of 10 meters so that about 378 levee cross sections were generated. In addition, the waterside maximum and average slope could be automatically calculated by separating slope plane form levee point cloud, and the accuracy of RMSE was confirmed by comparing with manually calculated slope. The reference slope was calculated manually by plotting point cloud of levee slope plane and selecting two points that use location information when calculating the slope. Also, as a result of comparing the water side slope with slope standard in basic river plan for Anyang stream, it is confirmed that inspecting the river facilities with the MMS point cloud is highly recommended than the existing river survey.

Design and Implementation of OpenCV-based Inventory Management System to build Small and Medium Enterprise Smart Factory (중소기업 스마트공장 구축을 위한 OpenCV 기반 재고관리 시스템의 설계 및 구현)

  • Jang, Su-Hwan;Jeong, Jopil
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.1
    • /
    • pp.161-170
    • /
    • 2019
  • Multi-product mass production small and medium enterprise factories have a wide variety of products and a large number of products, wasting manpower and expenses for inventory management. In addition, there is no way to check the status of inventory in real time, and it is suffering economic damage due to excess inventory and shortage of stock. There are many ways to build a real-time data collection environment, but most of them are difficult to afford for small and medium-sized companies. Therefore, smart factories of small and medium enterprises are faced with difficult reality and it is hard to find appropriate countermeasures. In this paper, we implemented the contents of extension of existing inventory management method through character extraction on label with barcode and QR code, which are widely adopted as current product management technology, and evaluated the effect. Technically, through preprocessing using OpenCV for automatic recognition and classification of stock labels and barcodes, which is a method for managing input and output of existing products through computer image processing, and OCR (Optical Character Recognition) function of Google vision API. And it is designed to recognize the barcode through Zbar. We propose a method to manage inventory by real-time image recognition through Raspberry Pi without using expensive equipment.

A computer vision-based approach for behavior recognition of gestating sows fed different fiber levels during high ambient temperature

  • Kasani, Payam Hosseinzadeh;Oh, Seung Min;Choi, Yo Han;Ha, Sang Hun;Jun, Hyungmin;Park, Kyu hyun;Ko, Han Seo;Kim, Jo Eun;Choi, Jung Woo;Cho, Eun Seok;Kim, Jin Soo
    • Journal of Animal Science and Technology
    • /
    • v.63 no.2
    • /
    • pp.367-379
    • /
    • 2021
  • The objectives of this study were to evaluate convolutional neural network models and computer vision techniques for the classification of swine posture with high accuracy and to use the derived result in the investigation of the effect of dietary fiber level on the behavioral characteristics of the pregnant sow under low and high ambient temperatures during the last stage of gestation. A total of 27 crossbred sows (Yorkshire × Landrace; average body weight, 192.2 ± 4.8 kg) were assigned to three treatments in a randomized complete block design during the last stage of gestation (days 90 to 114). The sows in group 1 were fed a 3% fiber diet under neutral ambient temperature; the sows in group 2 were fed a diet with 3% fiber under high ambient temperature (HT); the sows in group 3 were fed a 6% fiber diet under HT. Eight popular deep learning-based feature extraction frameworks (DenseNet121, DenseNet201, InceptionResNetV2, InceptionV3, MobileNet, VGG16, VGG19, and Xception) used for automatic swine posture classification were selected and compared using the swine posture image dataset that was constructed under real swine farm conditions. The neural network models showed excellent performance on previously unseen data (ability to generalize). The DenseNet121 feature extractor achieved the best performance with 99.83% accuracy, and both DenseNet201 and MobileNet showed an accuracy of 99.77% for the classification of the image dataset. The behavior of sows classified by the DenseNet121 feature extractor showed that the HT in our study reduced (p < 0.05) the standing behavior of sows and also has a tendency to increase (p = 0.082) lying behavior. High dietary fiber treatment tended to increase (p = 0.064) lying and decrease (p < 0.05) the standing behavior of sows, but there was no change in sitting under HT conditions.

An Automatic ROI Extraction and Its Mask Generation based on Wavelet of Low DOF Image (피사계 심도가 낮은 이미지에서 웨이블릿 기반의 자동 ROI 추출 및 마스크 생성)

  • Park, Sun-Hwa;Seo, Yeong-Geon;Lee, Bu-Kweon;Kang, Ki-Jun;Kim, Ho-Yong;Kim, Hyung-Jun;Kim, Sang-Bok
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.3
    • /
    • pp.93-101
    • /
    • 2009
  • This paper suggests a new algorithm automatically searching for Region-of-Interest(ROI) with high speed, using the edge information of high frequency subband transformed with wavelet. The proposed method executes a searching algorithm of 4-direction object boundary by the unit of block using the edge information, and detects ROIs. The whole image is splitted by $64{\times}64$ or $32{\times}32$ sized blocks and the blocks can be ROI block or background block according to taking the edges or not. The 4-directions searche the image from the outside to the center and the algorithm uses a feature that the low-DOF image has some edges as one goes to center. After searching all the edges, the method regards the inner blocks of the edges as ROI, and makes the ROI masks and sends them to server. This is one of the dynamic ROI method. The existing methods have had some problems of complicated filtering and region merge, but this method improved considerably the problems. Also, it was possible to apply to an application requiring real-time processing caused by the process of the unit of block.

Detection of Plastic Greenhouses by Using Deep Learning Model for Aerial Orthoimages (딥러닝 모델을 이용한 항공정사영상의 비닐하우스 탐지)

  • Byunghyun Yoon;Seonkyeong Seong;Jaewan Choi
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.2
    • /
    • pp.183-192
    • /
    • 2023
  • The remotely sensed data, such as satellite imagery and aerial photos, can be used to extract and detect some objects in the image through image interpretation and processing techniques. Significantly, the possibility for utilizing digital map updating and land monitoring has been increased through automatic object detection since spatial resolution of remotely sensed data has improved and technologies about deep learning have been developed. In this paper, we tried to extract plastic greenhouses into aerial orthophotos by using fully convolutional densely connected convolutional network (FC-DenseNet), one of the representative deep learning models for semantic segmentation. Then, a quantitative analysis of extraction results had performed. Using the farm map of the Ministry of Agriculture, Food and Rural Affairsin Korea, training data was generated by labeling plastic greenhouses into Damyang and Miryang areas. And then, FC-DenseNet was trained through a training dataset. To apply the deep learning model in the remotely sensed imagery, instance norm, which can maintain the spectral characteristics of bands, was used as normalization. In addition, optimal weights for each band were determined by adding attention modules in the deep learning model. In the experiments, it was found that a deep learning model can extract plastic greenhouses. These results can be applied to digital map updating of Farm-map and landcover maps.