• Title/Summary/Keyword: Automatic Classification System

Search Result 417, Processing Time 0.028 seconds

Effect of Difference of Land Cover Conditions on Urban Thermal Environment in Daegu Using Satellite and AWS Data (위성 및 AWS 자료를 이용한 지표면 피복 조건의 차이가 대구의 도시 열환경에 미치는 영향)

  • Ahn, Ji-Suk;Kim, Hae-Dong;Kim, Sang-Woo
    • Journal of Environmental Science International
    • /
    • v.19 no.3
    • /
    • pp.281-293
    • /
    • 2010
  • The present study explores time and spatial thermal environment for Daegu, which is a city built on a basin area, according to varying land cover conditions of the earth's surface by analyzing data derived from meteorological observation and satellite images. The study has classified land use by utilizing MODIS satellite images and analyzed land surface temperature. Also, by using data acquired from automatic weather system, the study has evaluated the effects of atmospheric heating caused by city pavements by analyzing the sensible heat flux between the city's land surface and the atmosphere. The results are as follows. 1) Classification of land use in the Daegu area shows 46.64% of urban and built-up area, 1.39% of watersides, 35.19% of forest, 11.43% of crops, and 5.37% grasslands. 2) During the weekdays throughout the year, the land surface temperature was high for Dalseogu, Bukgu, and Seogu regions where industrial complexes could be found. Comparatively, lower temperature could be observed in the woodlands. 3) While the land surface temperature displayed the effects of pushing air upwards during the weekdays in urban areas, the reverse was true for forest regions. During the night, the temperature did not exert any significant influence on air movement.

Automatic Identification of Database Workloads by using SVM Workload Classifier (SVM 워크로드 분류기를 통한 자동화된 데이터베이스 워크로드 식별)

  • Kim, So-Yeon;Roh, Hong-Chan;Park, Sang-Hyun
    • The Journal of the Korea Contents Association
    • /
    • v.10 no.4
    • /
    • pp.84-90
    • /
    • 2010
  • DBMS is used for a range of applications from data warehousing through on-line transaction processing. As a result of this demand, DBMS has continued to grow in terms of its size. This growth invokes the most important issue of manually tuning the performance of DBMS. The DBMS tuning should be adaptive to the type of the workload put upon it. But, identifying workloads in mixed database applications might be quite difficult. Therefore, a method is necessary for identifying workloads in the mixed database environment. In this paper, we propose a SVM workload classifier to automatically identify a DBMS workload. Database workloads are collected in TPC-C and TPC-W benchmark while changing the resource parameters. Parameters for SVM workload classifier, C and kernel parameter, were chosen experimentally. The experiments revealed that the accuracy of the proposed SVM workload classifier is about 9% higher than that of Decision tree, Naive Bayes, Multilayer perceptron and K-NN classifier.

Study on the Development of Auto-classification Algorithm for Ginseng Seedling using SVM (Support Vector Machine) (SVM(Support Vector Machine)을 이용한 묘삼 자동등급 판정 알고리즘 개발에 관한 연구)

  • Oh, Hyun-Keun;Lee, Hoon-Soo;Chung, Sun-Ok;Cho, Byoung-Kwan
    • Journal of Biosystems Engineering
    • /
    • v.36 no.1
    • /
    • pp.40-47
    • /
    • 2011
  • Image analysis algorithm for the quality evaluation of ginseng seedling was investigated. The images of ginseng seedling were acquired with a color CCD camera and processed with the image analysis methods, such as binary conversion, labeling, and thinning. The processed images were used to calculate the length and weight of ginseng seedlings. The length and weight of the samples could be predicted with standard errors of 0.343 mm, and 0.0214 g respectively, $R^2$ values of 0.8738 and 0.9835 respectively. For the evaluation of the three quality grades of Gab, Eul, and abnormal ginseng seedlings, features from the processed images were extracted. The features combined with the ratio of the lengths and areas of the ginseng seedlings efficiently differentiate the abnormal shapes from the normal ones of the samples. The grade levels were evaluated with an efficient pattern recognition method of support vector machine analysis. The quality grade of ginseng seedling could be evaluated with an accuracy of 95% and 97% for training and validation, respectively. The result indicates that color image analysis with support vector machine algorithm has good potential to be used for the development of an automatic sorting system for ginseng seedling.

An Automatic Pattern Recognition Algorithm for Identifying the Spatio-temporal Congestion Evolution Patterns in Freeway Historic Data (고속도로 이력데이터에 포함된 정체 시공간 전개 패턴 자동인식 알고리즘 개발)

  • Park, Eun Mi;Oh, Hyun Sun
    • Journal of Korean Society of Transportation
    • /
    • v.32 no.5
    • /
    • pp.522-530
    • /
    • 2014
  • Spatio-temporal congestion evolution pattern can be reproduced using the VDS(Vehicle Detection System) historic speed dataset in the TMC(Traffic Management Center)s. Such dataset provides a pool of spatio-temporally experienced traffic conditions. Traffic flow pattern is known as spatio-temporally recurred, and even non-recurrent congestion caused by incidents has patterns according to the incident conditions. These imply that the information should be useful for traffic prediction and traffic management. Traffic flow predictions are generally performed using black-box approaches such as neural network, genetic algorithm, and etc. Black-box approaches are not designed to provide an explanation of their modeling and reasoning process and not to estimate the benefits and the risks of the implementation of such a solution. TMCs are reluctant to employ the black-box approaches even though there are numerous valuable articles. This research proposes a more readily understandable and intuitively appealing data-driven approach and developes an algorithm for identifying congestion patterns for recurrent and non-recurrent congestion management and information provision.

Automatic Title Detection by Spatial Feature and Projection Profile for Document Images (공간 정보와 투영 프로파일을 이용한 문서 영상에서의 타이틀 영역 추출)

  • Park, Hyo-Jin;Kim, Bo-Ram;Kim, Wook-Hyun
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.11 no.3
    • /
    • pp.209-214
    • /
    • 2010
  • This paper proposes an algorithm of segmentation and title detection for document image. The automated title detection method that we have developed is composed of two phases, segmentation and title area detection. In the first phase, we extract and segment the document image. To perform this operation, the binary map is segmented by combination of morphological operation and CCA(connected component algorithm). The first phase provides segmented regions that would be detected as title area for the second stage. Candidate title areas are detected using geometric information, then we can extract the title region that is performed by removing non-title regions. After classification step that removes non-text regions, projection is performed to detect a title region. From the fact that usually the largest font is used for the title in the document, horizontal projection is performed within text areas. In this paper, we proposed a method of segmentation and title detection for various forms of document images using geometric features and projection profile analysis. The proposed system is expected to have various applications, such as document title recognition, multimedia data searching, real-time image processing and so on.

A Study on Varieties of Subject Access and Usabilities of the National Library of Korea Subject Headings (주제 접근의 다양성과 국립중앙도서관 주제명 표목의 활용가능성에 관한 연구)

  • Chung, Yeon Kyoung
    • Journal of the Korean BIBLIA Society for library and Information Science
    • /
    • v.25 no.4
    • /
    • pp.171-185
    • /
    • 2014
  • The purposes of this study are to examine the various methods of subject access in the rapidly changing environment and to suggest the future of subject access in National Library of Korea (NLK). First of all, current status and problems of Library of Congress Subject Headings List as an representative subject headings in the world and the ways of improving effectiveness of subject retrieval were dealt with. As the ways of improving subject access, social bookmarking, folksonomy, tagging, facet applications, automatic assignment of keyword, thesauri, classification system, and auto-assigned search box were suggested. Finally, current status of NLK subject headings and the ways of improving for utilization of the subject headings as subject access were provided.

Online Game Identity Theft Detection Model based on Hacker's Behavior Analysis (온라인게임 계정도용 탐지모델에 관한 연구)

  • Choi, Hwa-Jae;Woo, Ji-Young;Kim, Huy-Kang
    • Journal of Korea Game Society
    • /
    • v.11 no.6
    • /
    • pp.81-93
    • /
    • 2011
  • Identity theft happens frequently in popular MMORPG(Massively Multi-player Online Role Playing Games) where profits can be gained easily. In spite of the importance of security about identity theft in MMORPG, few methods to prevent and detect identity theft in online games have been proposed. In this study, we investigate real identity theft cases of an online game and define the representative patterns of identity theft as the speedy type, cautious type, and bold type. We then propose the automatic identity theft detection model based on the multi-class classification. We verify the system with one of the leading online games in Korea. The multi-class detection model outperforms the existing binary-class one(hacked or not).

Effects of herbal Cp soap on acne skin (한약 저온숙성비누가 여드름 피부에 미치는 영향)

  • Choi, Sang Rak;Seo, Bu Il;Koo, Jin Suk
    • The Korea Journal of Herbology
    • /
    • v.34 no.3
    • /
    • pp.37-44
    • /
    • 2019
  • Objectives : Acne is a common disease that affects more than 70% of adolescents. Acne patients have a poor quality of life compared to patients with other skin diseases. We tried to demonstrate the effectiveness of acne treatment using cleansing soap which is easily used in daily life. Methods : We selected 20 students with acne symptoms on their facial skin. We made herbal Cp (Cold process) soaps using Seosiokyongsan, Kyungohkgo, Hwangryunhaedoktang and Baeksoooh and distributed them to experiment participants. We let them wash their face in the morning and evening for 6 weeks using herbal Cp soap. Prior to the experiment, their skin condition was checked and assessed using A-ONE Smart One-Click Automatic Facial Diagnosis System three times at 3-week intervals. Acne status was classified into 6 stages according to KAGS and acne status was also measured 3 times in total. After the experiment, the changes of skin were analyzed through facial analysis test. Results : Based on the KAGS classification, the condition of acne has improved as a whole. The state of moisture was gradually increased and the state of skin oil was significantly decreased after 6 weeks of using soap compared to before using soap. Conclusions : Cp soaps made from four kinds of herbal medicine are believed to improve the condition of acne by increasing the moisture of the facial skin and decreasing the skin oil content.

Novel Intent based Dimension Reduction and Visual Features Semi-Supervised Learning for Automatic Visual Media Retrieval

  • kunisetti, Subramanyam;Ravichandran, Suban
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.230-240
    • /
    • 2022
  • Sharing of online videos via internet is an emerging and important concept in different types of applications like surveillance and video mobile search in different web related applications. So there is need to manage personalized web video retrieval system necessary to explore relevant videos and it helps to peoples who are searching for efficient video relates to specific big data content. To evaluate this process, attributes/features with reduction of dimensionality are computed from videos to explore discriminative aspects of scene in video based on shape, histogram, and texture, annotation of object, co-ordination, color and contour data. Dimensionality reduction is mainly depends on extraction of feature and selection of feature in multi labeled data retrieval from multimedia related data. Many of the researchers are implemented different techniques/approaches to reduce dimensionality based on visual features of video data. But all the techniques have disadvantages and advantages in reduction of dimensionality with advanced features in video retrieval. In this research, we present a Novel Intent based Dimension Reduction Semi-Supervised Learning Approach (NIDRSLA) that examine the reduction of dimensionality with explore exact and fast video retrieval based on different visual features. For dimensionality reduction, NIDRSLA learns the matrix of projection by increasing the dependence between enlarged data and projected space features. Proposed approach also addressed the aforementioned issue (i.e. Segmentation of video with frame selection using low level features and high level features) with efficient object annotation for video representation. Experiments performed on synthetic data set, it demonstrate the efficiency of proposed approach with traditional state-of-the-art video retrieval methodologies.

A New Approach to Automatic Keyword Generation Using Inverse Vector Space Model (키워드 자동 생성에 대한 새로운 접근법: 역 벡터공간모델을 이용한 키워드 할당 방법)

  • Cho, Won-Chin;Rho, Sang-Kyu;Yun, Ji-Young Agnes;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.21 no.1
    • /
    • pp.103-122
    • /
    • 2011
  • Recently, numerous documents have been made available electronically. Internet search engines and digital libraries commonly return query results containing hundreds or even thousands of documents. In this situation, it is virtually impossible for users to examine complete documents to determine whether they might be useful for them. For this reason, some on-line documents are accompanied by a list of keywords specified by the authors in an effort to guide the users by facilitating the filtering process. In this way, a set of keywords is often considered a condensed version of the whole document and therefore plays an important role for document retrieval, Web page retrieval, document clustering, summarization, text mining, and so on. Since many academic journals ask the authors to provide a list of five or six keywords on the first page of an article, keywords are most familiar in the context of journal articles. However, many other types of documents could not benefit from the use of keywords, including Web pages, email messages, news reports, magazine articles, and business papers. Although the potential benefit is large, the implementation itself is the obstacle; manually assigning keywords to all documents is a daunting task, or even impractical in that it is extremely tedious and time-consuming requiring a certain level of domain knowledge. Therefore, it is highly desirable to automate the keyword generation process. There are mainly two approaches to achieving this aim: keyword assignment approach and keyword extraction approach. Both approaches use machine learning methods and require, for training purposes, a set of documents with keywords already attached. In the former approach, there is a given set of vocabulary, and the aim is to match them to the texts. In other words, the keywords assignment approach seeks to select the words from a controlled vocabulary that best describes a document. Although this approach is domain dependent and is not easy to transfer and expand, it can generate implicit keywords that do not appear in a document. On the other hand, in the latter approach, the aim is to extract keywords with respect to their relevance in the text without prior vocabulary. In this approach, automatic keyword generation is treated as a classification task, and keywords are commonly extracted based on supervised learning techniques. Thus, keyword extraction algorithms classify candidate keywords in a document into positive or negative examples. Several systems such as Extractor and Kea were developed using keyword extraction approach. Most indicative words in a document are selected as keywords for that document and as a result, keywords extraction is limited to terms that appear in the document. Therefore, keywords extraction cannot generate implicit keywords that are not included in a document. According to the experiment results of Turney, about 64% to 90% of keywords assigned by the authors can be found in the full text of an article. Inversely, it also means that 10% to 36% of the keywords assigned by the authors do not appear in the article, which cannot be generated through keyword extraction algorithms. Our preliminary experiment result also shows that 37% of keywords assigned by the authors are not included in the full text. This is the reason why we have decided to adopt the keyword assignment approach. In this paper, we propose a new approach for automatic keyword assignment namely IVSM(Inverse Vector Space Model). The model is based on a vector space model. which is a conventional information retrieval model that represents documents and queries by vectors in a multidimensional space. IVSM generates an appropriate keyword set for a specific document by measuring the distance between the document and the keyword sets. The keyword assignment process of IVSM is as follows: (1) calculating the vector length of each keyword set based on each keyword weight; (2) preprocessing and parsing a target document that does not have keywords; (3) calculating the vector length of the target document based on the term frequency; (4) measuring the cosine similarity between each keyword set and the target document; and (5) generating keywords that have high similarity scores. Two keyword generation systems were implemented applying IVSM: IVSM system for Web-based community service and stand-alone IVSM system. Firstly, the IVSM system is implemented in a community service for sharing knowledge and opinions on current trends such as fashion, movies, social problems, and health information. The stand-alone IVSM system is dedicated to generating keywords for academic papers, and, indeed, it has been tested through a number of academic papers including those published by the Korean Association of Shipping and Logistics, the Korea Research Academy of Distribution Information, the Korea Logistics Society, the Korea Logistics Research Association, and the Korea Port Economic Association. We measured the performance of IVSM by the number of matches between the IVSM-generated keywords and the author-assigned keywords. According to our experiment, the precisions of IVSM applied to Web-based community service and academic journals were 0.75 and 0.71, respectively. The performance of both systems is much better than that of baseline systems that generate keywords based on simple probability. Also, IVSM shows comparable performance to Extractor that is a representative system of keyword extraction approach developed by Turney. As electronic documents increase, we expect that IVSM proposed in this paper can be applied to many electronic documents in Web-based community and digital library.