• Title/Summary/Keyword: Pattern Processing

Search Result 2,352, Processing Time 0.027 seconds

An Efficient Clustering Algorithm based on Heuristic Evolution (휴리스틱 진화에 기반한 효율적 클러스터링 알고리즘)

  • Ryu, Joung-Woo;Kang, Myung-Ku;Kim, Myung-Won
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.1_2
    • /
    • pp.80-90
    • /
    • 2002
  • Clustering is a useful technique for grouping data points such that points within a single group/cluster have similar characteristics. Many clustering algorithms have been developed and used in engineering applications including pattern recognition and image processing etc. Recently, it has drawn increasing attention as one of important techniques in data mining. However, clustering algorithms such as K-means and Fuzzy C-means suffer from difficulties. Those are the needs to determine the number of clusters apriori and the clustering results depending on the initial set of clusters which fails to gain desirable results. In this paper, we propose a new clustering algorithm, which solves mentioned problems. In our method we use evolutionary algorithm to solve the local optima problem that clustering converges to an undesirable state starting with an inappropriate set of clusters. We also adopt a new measure that represents how well data are clustered. The measure is determined in terms of both intra-cluster dispersion and inter-cluster separability. Using the measure, in our method the number of clusters is automatically determined as the result of optimization process. And also, we combine heuristic that is problem-specific knowledge with a evolutionary algorithm to speed evolutionary algorithm search. We have experimented our algorithm with several sets of multi-dimensional data and it has been shown that one algorithm outperforms the existing algorithms.

The Integration System for International Procurement Information Processing (국제입찰정보 통합시스템의 설계 및 구현)

  • Yoon, Jong-Wan;Lee, Jong-Woo;Park, Chan-Young
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.1
    • /
    • pp.71-81
    • /
    • 2002
  • The lack of specialties of the existing commercial web search systems stems from the fact that they have no capabilities to extract and gather the meaningful information from each information domain they cover. We are sure, however, that the necessity for the information integration system, not just search system, will be likely to become larger in the future. In this paper, we propose a design and implementation of an information integration system called TIC(target information collector). TIC is able to extract meaningful information from a specific information area in the internet and integrate them for the commercial service. We also show the evaluation results of our implementation. For the experiments we applied our TIC to the international procurement information area. The international procurement information is publicly and freely announced by each government to the world. To automatically extract common properties from the related source sites, we adopt information pointing technique using inter-HTML tag pattern parsing. And through the information integration framework design, we can easily implement a site-specific information integration engine. By running our TIC for about 8 months, we find out it can remove considerable amount of the duplicated information, and as a result, we can obtain high quality international procurement information. The main contribution of this paper is to present a framework design and it's implementation for extracting the information of a specific area and then integrating them into a meaningful one.

Analysis of Network Traffic with Urban Area Characteristics for Mobile Network Traffic Model (이동통신 네트워크 트래픽 모델을 위한 도시 지역 이동통신 트래픽 특성 분석)

  • Yoon, Young-Hyun
    • The KIPS Transactions:PartC
    • /
    • v.10C no.4
    • /
    • pp.471-478
    • /
    • 2003
  • Traditionally,, analysis, simulation and measurement have all been used to evaluate the performance of network protocols and functional entities that support mobile wireless service. Simulation methods are useful for testing the complex systems which have the very complicate interactions between components. To develop a mobile call simulator which is used to examine, validate, and predict the performance of mobile wireless call procedures must have the teletraffic model, which is to describe the mobile communication environments. Mobile teletraffic model is consists of 2 sub-models, traffic source and network traffic model. In this paper, we analyzed the network traffic data which are gathered from selected Base Stations (BSs) to define the mobile teletraffic model. We defined 4 types of cell location-Residential, Commercial, Industrial, and Afforest zone. We selected some Base Stations (BSs) which are represented cell location types in Seoul city, and gathered real data from them And then, we present the call rate per hour, cail distribution pattern per day, busy hours, loose hours, the maximum number of call, and the minimum number of calls based on defined cell location types. Those parameters are very important to test the mobile communication system´s performance and reliability and are very useful for defining the mobile network traffic model or for working the existed mobile simulation programs as input parameters.

The Construction of Multiform User Profiles Based on Transaction for Effective Recommendation and Segmentation (효과적인 추천과 세분화를 위한 트랜잭션 기반 여러 형태 사용자 프로파일의 구축)

  • Koh, Jae-Jin;An, Hyoung-Keun
    • The KIPS Transactions:PartD
    • /
    • v.13D no.5 s.108
    • /
    • pp.661-670
    • /
    • 2006
  • With the development of e-Commerce and the proliferation of easily accessible information, information filtering systems such as recommender and SDI systems have become popular to prune large information spaces so that users are directed toward those items that best meet their needs and preferences. Until now, many information filtering methods have been proposed to support filtering systems. XML is emerging as a new standard for information. Recently, filtering systems need new approaches in dealing with XML documents. So, in this paper our system suggests a method to create multiform user profiles with XML's ability to represent structure. This system consists of two parts; one is an administrator profile definition part that an administrator defines to analyze users purchase pattern before a transaction such as purchase happens directly. an other is a user profile creation part module which is applied by the defined profile. Administrator profiles are made from DTD information and it is supposed to point the specific part of a document conforming to the DTD. Proposed system builds user's profile more accurately to get adaptability for user's behavior of buying and provide useful product information without inefficient searching based on such user's profile.

A Study on the Most Frequent Diseases of Health Insurance Program and the Primary Care Physicians in Korea (의료보험 다빈도 상병과 1차진료 의사에 관한 연구)

  • 김철환;문옥륜
    • Health Policy and Management
    • /
    • v.3 no.1
    • /
    • pp.124-145
    • /
    • 1993
  • General practitioners, internists, pediatricians, and family physicians are classified as so-called primary care physicians in the United States. We carried out this study for the purpose of answering the following question; "Who are the primary care physicians in Korea\ulcorner" We analyzed the 663, 154 claims which were drawn from the health insurance processing file made during the period of one month, April 1992 on the basis of systemic random sampling technique. The 663, 154 cases were matched with the doctor's file registered at the National Federation Medical Insurance by using the indivisual physician code number and analyzed according to the kind of specialty. If we follow the Geyman's definition of primary care physician in the United States, this study shows that they can take care of 43.2% of the total private clinic's claims in Korea. Provided that general practitioners and family physicians are considered the same way as in the United Kingdom, they could with only 8.3% of the total claims in Korea. The most frequent diseases are those which rank first to 46th in the total private clinic's claims. The proportion of the most frequent diseases was highest for pediatricians(90.4%) and followed by internists(81.4%), otolaryngologists(78.7%) and family physicians(76.5%). The proportion of the most frequent diseases in the most common 46 diseases was highest for radiologists(80.4%) and the next was as follows : general practitioners(78.3%), family physicians(67.4%), and internists(67.4%). We classified the most common 20 diseases of each specialty into 17 categories of ICD-9 and compared it with those of general practitioners. The specialists who had managed a similar disease pattern to those of general practitioners were identified as anesthesiologists, family physicians, general surgeons, and internists. Some specialists practicing at private clinics managed the diseases which were not quite appropriate for their specialties. After we evaluated each specialty by the most common diseases, the most frequent diseases, and the most frequent 20 diseases of each specialty in terms of the 17 categories of ICD-9, a tentative assumption is made that the primary physicians in the Republic of Korea are general practitioners, anesthesiologists, family physicians, internists, and general surgeons. This study has concluded that the categories of the primary care physicians are so diverse that their roles and distributions are distorted accordingly. Vigorous health policy efforts in correcting the malcomposition need to be made for the better provision of primary health care in Korea. in Korea.

  • PDF

A MapReduce-Based Workflow BIG-Log Clustering Technique (맵리듀스기반 워크플로우 빅-로그 클러스터링 기법)

  • Jin, Min-Hyuck;Kim, Kwanghoon Pio
    • Journal of Internet Computing and Services
    • /
    • v.20 no.1
    • /
    • pp.87-96
    • /
    • 2019
  • In this paper, we propose a MapReduce-supported clustering technique for collecting and classifying distributed workflow enactment event logs as a preprocessing tool. Especially, we would call the distributed workflow enactment event logs as Workflow BIG-Logs, because they are satisfied with as well as well-fitted to the 5V properties of BIG-Data like Volume, Velocity, Variety, Veracity and Value. The clustering technique we develop in this paper is intentionally devised for the preprocessing phase of a specific workflow process mining and analysis algorithm based upon the workflow BIG-Logs. In other words, It uses the Map-Reduce framework as a Workflow BIG-Logs processing platform, it supports the IEEE XES standard data format, and it is eventually dedicated for the preprocessing phase of the ${\rho}$-Algorithm that is a typical workflow process mining algorithm based on the structured information control nets. More precisely, The Workflow BIG-Logs can be classified into two types: of activity-based clustering patterns and performer-based clustering patterns, and we try to implement an activity-based clustering pattern algorithm based upon the Map-Reduce framework. Finally, we try to verify the proposed clustering technique by carrying out an experimental study on the workflow enactment event log dataset released by the BPI Challenges.

Implementation of A Security Token System using Fingerprint Verification (지문 인증을 이용한 보안 토큰 시스템 구현)

  • 문대성;길연희;안도성;반성범;정용화;정교일
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.13 no.4
    • /
    • pp.63-70
    • /
    • 2003
  • In the modern electronic world, the authentication of a person is an important task in many areas of online-transactions. Using biometrics to authenticate a person's identity has several advantages over the present practices of Personal Identification Numbers(PINs) and passwords. To gain maximum security in the verification system using biometrics, the computation of the verification as well as the store of the biometric pattern has to be taken place in the security token(smart card, USB token). However, there is an open issue of integrating biometrics into the security token because of its limited resources(memory space, processing power). In this paper, we describe our implementation of the USB security token system having 206MHz StrongARM CPU, 16MBytes flash memory, and 1MBytes RAM. Also, we evaluate the performance of a light-weighted In-gerprint verification algorithm that can be executed in the restricted environments. Based on experimental results, we confirmed that the RAM requirement of the proposed algorithm was about 6.8 KBytes and the Equal Error Rate(EER) was 1.7%.

Study of Riverline Change around Sannam Wetland in the Hangang River Estuaty using LANDSAT Image Processing (LANDSAT 위성사진을 활용한 한강하구 산남습지 인근 하안선 변화 연구)

  • Youn, Sukzun;Lee, Samhee;Jang, Changhwan
    • Journal of Wetlands Research
    • /
    • v.23 no.2
    • /
    • pp.154-162
    • /
    • 2021
  • The naturally opened Han river estuary is a place where the flows of the Han river, Imjin river, Yaesung river meet with West Sea of Korea, so the hydrodynamic mechanism(Impact-Response) structure of Han river estuary is complex. Continuous observation and measurement due to the morphological characteristics at the estuary are required to maintain the estuary environment and river management facilities. However, the Sannam wetland(the study area) is in the military operation area. Therefore, Sannam wetland has the limited access under the control from military office. In 2020, there had a natural disaster due to flooding in August and COVID-19, and it made a survey hard. The noncontact survey technique, the analysis of LANDSAT images at Sannam wetland, was applied to analyze riverbed fluctuation and morphological transformation around Sannam wetland. LANDSAT images obtained from EarthExplorer, USGS and analyzed by QGIS. The analysis was performed based on the area and the distance near Sannam wetland. As a result, an erosion was happened on the downstream of the study area, and the upstream of the study area did not have any serious sediment transport. Considering the resolution of LANDSAT images, this noncontect survey technique is applicable to manage the study area. From the analysis of LANDSAT images, it is assumed that the tidal effect is greater than the inflow from the upstream. The pattern change of tidal response causes the damage of the river facilities near the Hangang river estuary.

A Study on the Design of Prediction Model for Safety Evaluation of Partial Discharge (부분 방전의 안전도 평가를 위한 예측 모델 설계)

  • Lee, Su-Il;Ko, Dae-Sik
    • Journal of Platform Technology
    • /
    • v.8 no.3
    • /
    • pp.10-21
    • /
    • 2020
  • Partial discharge occurs a lot in high-voltage power equipment such as switchgear, transformers, and switch gears. Partial discharge shortens the life of the insulator and causes insulation breakdown, resulting in large-scale damage such as a power outage. There are several types of partial discharge that occur inside the product and the surface. In this paper, we design a predictive model that can predict the pattern and probability of occurrence of partial discharge. In order to analyze the designed model, learning data for each type of partial discharge was collected through the UHF sensor by using a simulator that generates partial discharge. The predictive model designed in this paper was designed based on CNN during deep learning, and the model was verified through learning. To learn about the designed model, 5000 training data were created, and the form of training data was used as input data for the model by pre-processing the 3D raw data input from the UHF sensor as 2D data. As a result of the experiment, it was found that the accuracy of the model designed through learning has an accuracy of 0.9972. It was found that the accuracy of the proposed model was higher in the case of learning by making the data into a two-dimensional image and learning it in the form of a grayscale image.

  • PDF

Blocking Intelligent Dos Attack with SDN (SDN과 허니팟 기반 동적 파라미터 조절을 통한 지능적 서비스 거부 공격 차단)

  • Yun, Junhyeok;Mun, Sungsik;Kim, Mihui
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.11 no.1
    • /
    • pp.23-34
    • /
    • 2022
  • With the development of network technology, the application area has also been diversified, and protocols for various purposes have been developed and the amount of traffic has exploded. Therefore, it is difficult for the network administrator to meet the stability and security standards of the network with the existing traditional switching and routing methods. Software Defined Networking (SDN) is a new networking paradigm proposed to solve this problem. SDN enables efficient network management by programming network operations. This has the advantage that network administrators can flexibly respond to various types of attacks. In this paper, we design a threat level management module, an attack detection module, a packet statistics module, and a flow rule generator that collects attack information through the controller and switch, which are components of SDN, and detects attacks based on these attributes of SDN. It proposes a method to block denial of service attacks (DoS) of advanced attackers by programming and applying honeypot. In the proposed system, the attack packet can be quickly delivered to the honeypot according to the modifiable flow rule, and the honeypot that received the attack packets analyzed the intelligent attack pattern based on this. According to the analysis results, the attack detection module and the threat level management module are adjusted to respond to intelligent attacks. The performance and feasibility of the proposed system was shown by actually implementing the proposed system, performing intelligent attacks with various attack patterns and attack levels, and checking the attack detection rate compared to the existing system.