• Title/Summary/Keyword: Automatic Extraction Algorithm

Search Result 296, Processing Time 0.025 seconds

Detection of Address Region of Standard Postal Label Images Acquired from CCD Scanner System (CCD스캐너 시스템에서 획득된 표준 택배 라벨 영상의 주소 영역 검출)

  • 원철호;송병섭;박희준;이수형;임성운;구본후
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.8 no.2
    • /
    • pp.30-37
    • /
    • 2003
  • To effectively control a vast amount of postal packages, we need the automatic system for extracting the address region from CCD scanner images. In this paper, we propose a address region extraction algorithm in the standard postal label. We used geometric characteristics of the underlying address regions and defined several criteria for fast detection of address regions. As a result, we accomplished a successful detection and classification of the postal package labels in real time.

  • PDF

Automated Feature-Based Registration for Reverse Engineering of Human Models

  • Jun, Yong-Tae;Choi, Kui-Won
    • Journal of Mechanical Science and Technology
    • /
    • v.19 no.12
    • /
    • pp.2213-2223
    • /
    • 2005
  • In order to reconstruct a full 3D human model in reverse engineering (RE), a 3D scanner needs to be placed arbitrarily around the target model to capture all part of the scanned surface. Then, acquired multiple scans must be registered and merged since each scanned data set taken from different position is just given in its own local co-ordinate system. The goal of the registration is to create a single model by aligning all individual scans. It usually consists of two sub-steps: rough and fine registration. The fine registration process can only be performed after an initial position is approximated through the rough registration. Hence an automated rough registration process is crucial to realize a completely automatic RE system. In this paper an automated rough registration method for aligning multiple scans of complex human face is presented. The proposed method automatically aligns the meshes of different scans with the information of features that are extracted from the estimated principal curvatures of triangular meshes of the human face. Then the roughly aligned scanned data sets are further precisely enhanced with a fine registration step with the recently popular Iterative Closest Point (ICP) algorithm. Some typical examples are presented and discussed to validate the proposed system.

A Study on improving the performance of License Plate Recognition (자동차 번호판 인식 성능 향상에 관한 연구)

  • Eom, Gi-Yeol
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2006.11a
    • /
    • pp.203-207
    • /
    • 2006
  • Nowadays, Cars are continuing to grow at an alarming rate but they also cause many problems such as traffic accident, pollutions and so on. One of the most effective methods that prevent traffic accidents is the use of traffic monitoring systems, which are already widely used in many countries. The monitoring system is beginning to be used in domestic recently. An intelligent monitoring system generates photo images of cars as well as identifies cars by recognizing their plates. That is, the system automatically recognizes characters of vehicle plates. An automatic vehicle plate recognition consists of two main module: a vehicle plate locating module and a vehicle plate number identification module. We study for a vehicle plate number identification module in this paper. We use image preprocessing, feature extraction, multi-layer neural networks for recognizing characters of vehicle plates and we present a feature-comparison method for improving the performance of vehicle plate number identification module. In the experiment on identifying vehicle plate number, 300 images taken from various scenes were used. Of which, 8 images have been failed to identify vehicle plate number and the overall rate of success for our vehicle plate recognition algorithm is 98%.

  • PDF

Automatic Malware Detection Rule Generation and Verification System (악성코드 침입탐지시스템 탐지규칙 자동생성 및 검증시스템)

  • Kim, Sungho;Lee, Suchul
    • Journal of Internet Computing and Services
    • /
    • v.20 no.2
    • /
    • pp.9-19
    • /
    • 2019
  • Service and users over the Internet are increasing rapidly. Cyber attacks are also increasing. As a result, information leakage and financial damage are occurring. Government, public agencies, and companies are using security systems that use signature-based detection rules to respond to known malicious codes. However, it takes a long time to generate and validate signature-based detection rules. In this paper, we propose and develop signature based detection rule generation and verification systems using the signature extraction scheme developed based on the LDA(latent Dirichlet allocation) algorithm and the traffic analysis technique. Experimental results show that detection rules are generated and verified much more quickly than before.

An Efficient Machine Learning-based Text Summarization in the Malayalam Language

  • P Haroon, Rosna;Gafur M, Abdul;Nisha U, Barakkath
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.6
    • /
    • pp.1778-1799
    • /
    • 2022
  • Automatic text summarization is a procedure that packs enormous content into a more limited book that incorporates significant data. Malayalam is one of the toughest languages utilized in certain areas of India, most normally in Kerala and in Lakshadweep. Natural language processing in the Malayalam language is relatively low due to the complexity of the language as well as the scarcity of available resources. In this paper, a way is proposed to deal with the text summarization process in Malayalam documents by training a model based on the Support Vector Machine classification algorithm. Different features of the text are taken into account for training the machine so that the system can output the most important data from the input text. The classifier can classify the most important, important, average, and least significant sentences into separate classes and based on this, the machine will be able to create a summary of the input document. The user can select a compression ratio so that the system will output that much fraction of the summary. The model performance is measured by using different genres of Malayalam documents as well as documents from the same domain. The model is evaluated by considering content evaluation measures precision, recall, F score, and relative utility. Obtained precision and recall value shows that the model is trustable and found to be more relevant compared to the other summarizers.

Automatic Extraction of Buildings using Aerial Photo and Airborne LIDAR Data (항공사진과 항공레이저 데이터를 이용한 건물 자동추출)

  • 조우석;이영진;좌윤석
    • Korean Journal of Remote Sensing
    • /
    • v.19 no.4
    • /
    • pp.307-317
    • /
    • 2003
  • This paper presents an algorithm that automatically extracts buildings among many different features on the earth surface by fusing LIDAR data with panchromatic aerial images. The proposed algorithm consists of three stages such as point level process, polygon level process, parameter space level process. At the first stage, we eliminate gross errors and apply a local maxima filter to detect building candidate points from the raw laser scanning data. After then, a grouping procedure is performed for segmenting raw LIDAR data and the segmented LIDAR data is polygonized by the encasing polygon algorithm developed in the research. At the second stage, we eliminate non-building polygons using several constraints such as area and circularity. At the last stage, all the polygons generated at the second stage are projected onto the aerial stereo images through collinearity condition equations. Finally, we fuse the projected encasing polygons with edges detected by image processing for refining the building segments. The experimental results showed that the RMSEs of building corners in X, Y and Z were 8.1cm, 24.7cm, 35.9cm, respectively.

A Study on Knowledge Entity Extraction Method for Individual Stocks Based on Neural Tensor Network (뉴럴 텐서 네트워크 기반 주식 개별종목 지식개체명 추출 방법에 관한 연구)

  • Yang, Yunseok;Lee, Hyun Jun;Oh, Kyong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.25-38
    • /
    • 2019
  • Selecting high-quality information that meets the interests and needs of users among the overflowing contents is becoming more important as the generation continues. In the flood of information, efforts to reflect the intention of the user in the search result better are being tried, rather than recognizing the information request as a simple string. Also, large IT companies such as Google and Microsoft focus on developing knowledge-based technologies including search engines which provide users with satisfaction and convenience. Especially, the finance is one of the fields expected to have the usefulness and potential of text data analysis because it's constantly generating new information, and the earlier the information is, the more valuable it is. Automatic knowledge extraction can be effective in areas where information flow is vast, such as financial sector, and new information continues to emerge. However, there are several practical difficulties faced by automatic knowledge extraction. First, there are difficulties in making corpus from different fields with same algorithm, and it is difficult to extract good quality triple. Second, it becomes more difficult to produce labeled text data by people if the extent and scope of knowledge increases and patterns are constantly updated. Third, performance evaluation is difficult due to the characteristics of unsupervised learning. Finally, problem definition for automatic knowledge extraction is not easy because of ambiguous conceptual characteristics of knowledge. So, in order to overcome limits described above and improve the semantic performance of stock-related information searching, this study attempts to extract the knowledge entity by using neural tensor network and evaluate the performance of them. Different from other references, the purpose of this study is to extract knowledge entity which is related to individual stock items. Various but relatively simple data processing methods are applied in the presented model to solve the problems of previous researches and to enhance the effectiveness of the model. From these processes, this study has the following three significances. First, A practical and simple automatic knowledge extraction method that can be applied. Second, the possibility of performance evaluation is presented through simple problem definition. Finally, the expressiveness of the knowledge increased by generating input data on a sentence basis without complex morphological analysis. The results of the empirical analysis and objective performance evaluation method are also presented. The empirical study to confirm the usefulness of the presented model, experts' reports about individual 30 stocks which are top 30 items based on frequency of publication from May 30, 2017 to May 21, 2018 are used. the total number of reports are 5,600, and 3,074 reports, which accounts about 55% of the total, is designated as a training set, and other 45% of reports are designated as a testing set. Before constructing the model, all reports of a training set are classified by stocks, and their entities are extracted using named entity recognition tool which is the KKMA. for each stocks, top 100 entities based on appearance frequency are selected, and become vectorized using one-hot encoding. After that, by using neural tensor network, the same number of score functions as stocks are trained. Thus, if a new entity from a testing set appears, we can try to calculate the score by putting it into every single score function, and the stock of the function with the highest score is predicted as the related item with the entity. To evaluate presented models, we confirm prediction power and determining whether the score functions are well constructed by calculating hit ratio for all reports of testing set. As a result of the empirical study, the presented model shows 69.3% hit accuracy for testing set which consists of 2,526 reports. this hit ratio is meaningfully high despite of some constraints for conducting research. Looking at the prediction performance of the model for each stocks, only 3 stocks, which are LG ELECTRONICS, KiaMtr, and Mando, show extremely low performance than average. this result maybe due to the interference effect with other similar items and generation of new knowledge. In this paper, we propose a methodology to find out key entities or their combinations which are necessary to search related information in accordance with the user's investment intention. Graph data is generated by using only the named entity recognition tool and applied to the neural tensor network without learning corpus or word vectors for the field. From the empirical test, we confirm the effectiveness of the presented model as described above. However, there also exist some limits and things to complement. Representatively, the phenomenon that the model performance is especially bad for only some stocks shows the need for further researches. Finally, through the empirical study, we confirmed that the learning method presented in this study can be used for the purpose of matching the new text information semantically with the related stocks.

Performance evaluation of vessel extraction algorithm applied to Aortic root segmentation in CT Angiography (CT Angiography 영상에서 대동맥 추출을 위한 혈관 분할 알고리즘 성능 평가)

  • Kim, Tae-Hyong;Hwang, Young-sang;Shin, Ki-Young
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.9 no.2
    • /
    • pp.196-204
    • /
    • 2016
  • World Health Organization reported that heart-related diseases such as coronary artery stenoses show the highest occurrence rate which may cause heart attack. Using Computed Tomography angiography images will allow radiologists to detect and have intervention by creating 3D roadmapping of the vessels. However, it is often complex and difficult do reconstruct 3D vessel which causes very large amount of time and previous researches were studied to segment vessels more accurate automatically. Therefore, in this paper, Region Competition, Geodesic Active Contour (GAC), Multi-atlas based segmentation and Active Shape Model algorithms were applied to segment aortic root from CTA images and the results were analyzed by using mean Hausdorff distance, volume to volume measure, computational time, user-interaction and coronary ostium detection rate. As a result, Extracted 3D aortic model using GAC showed the highest accuracy but also showed highest user-interaction results. Therefore, it is important to improve automatic segmentation algorithm in future

Development of continuous blood pressure measurement system using ECG and PPG (ECG와 PPG를 이용한 실시간 연속 혈압 측정 시스템)

  • Kim, Jong-Hwa;Whang, Min-Cheol;Nam, Ki-Chang
    • Science of Emotion and Sensibility
    • /
    • v.11 no.2
    • /
    • pp.235-244
    • /
    • 2008
  • This study is to develop automatic extraction system of continuous blood pressure using ECG (Electrocardiogram) and PPG(Photoplethysmography) for u-health care technology. PTT (Pulse Transit Time) was determined from peak difference between ECG and PPG and its inverse made to get blood pressure. Since the peaks were vulnerable to be contaminated from noise and variation of amplitude, this study developed the adaptive algorithm for peak calculation in any noise condition. The developed method of the adaptive peak calculation was proven to make the standard deviations of PPT decrease to 28% and the detection of noise increase to 18%. Also, the correlation model such as blood pressure = -0.044 $\cdot$ PTT + 133.592 has successfully been determined for predicting the continuous pressure measured without using cuff but with using PPG and ECG, only.

  • PDF

Automatic Face Extraction with Unification of Brightness Distribution in Candidate Region and Triangle Structure among Facial Features (후보영역의 밝기 분산과 얼굴특징의 삼각형 배치구조를 결합한 얼굴의 자동 검출)

  • 이칠우;최정주
    • Journal of Korea Multimedia Society
    • /
    • v.3 no.1
    • /
    • pp.23-33
    • /
    • 2000
  • In this paper, we describe an algorithm which can extract human faces with natural pose from complex backgrounds. This method basically adopts the concept that facial region has the nearly same gray level for all pixels within appropriately scaled blocks. Based on the idea, we develop a hierarchial process that first, a block image data with pyramid structure of input image is generated, and some candidate regions for facial regions in the block image are Quickly determined, then finally the detailed facial features; organs are decided. To find the features easily, we introduce a local gray level transform which emphasizes dark and small regions, and estimate the geometrical triangle constraints among the facial features. The merit of our method is that we can be freed from the parameter assignment problem since the algorithm utilize a simple brightness computation, consequently robust systems not being depended on specific parameter values can be easily constructed.

  • PDF