• Title/Summary/Keyword: XML 데이터 필터링

Search Result 18, Processing Time 0.025 seconds

RFID Middleware System based on XML for Processing Large-Scale Data (대용량 데이터처리를 위한 XML기반의 RFID 미들웨어시스템)

  • Park, Byoung-Seob
    • The Journal of the Korea Contents Association
    • /
    • v.7 no.7
    • /
    • pp.31-38
    • /
    • 2007
  • We implement the RFID middleware system based on XML for large-scale data processing. The Implemented middleware system are consist of the reader interface for tag data collection, the event manager for a data filtering, and application interface for the RFID application. The implemented RFID middleware system is to support both a fixed type's reader and portable type's reader. we analyze the middleware function with four application accessing protocol, HTTP, XML, JMS, and SOAP, and demonstrate a filtering speed in terms of CPU utilization.

A Filtering Technique of Streaming XML Data based Postfix Sharing for Partial matching Path Queries (부분매칭 경로질의를 위한 포스트픽스 공유에 기반한 스트리밍 XML 데이타 필터링 기법)

  • Park Seog;Kim Young-Soo
    • Journal of KIISE:Databases
    • /
    • v.33 no.1
    • /
    • pp.138-149
    • /
    • 2006
  • As the environment with sensor network and ubiquitous computing is emerged, there are many demands of handling continuous, fast data such as streaming data. As work about streaming data has begun, work about management of streaming data in Publish-Subscribe system is started. The recent emergence of XML as a standard for information exchange on Internet has led to more interest in Publish - Subscribe system. A filtering technique of streaming XML data in the existing Publish- Subscribe system is using some schemes based on automata and YFilter, which is one of filtering techniques, is very popular. YFilter exploits commonality among path queries by sharing the common prefixes of the paths so that they are processed at most one and that is using the top-down approach. However, because partial matching path queries interrupt the common prefix sharing and don't calculate from root, throughput of YFilter decreases. So we use sharing of commonality among path queries with the common postfixes of the paths and use the bottom-up approach instead of the top-down approach. This filtering technique is called as PoSFilter. And we verify this technique through comparing with YFilter about throughput.

Data Modeling Method of NETCONF Protocol's Content Layer Applying VTD-XML (VTD-XML을 적용한 NETCONF 프로토콜 Content 계층의 데이터 모델링 기법)

  • Lee, Yang Min;Lee, Jae Kee
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.4 no.11
    • /
    • pp.383-390
    • /
    • 2015
  • It is appropriate to use the NETCONF to monitor and manage today's complex networks that are composed of variety links and heterogeneous equipment. Since the first standard of the NETCONF are launched, there have been several revisions, but disadvantages of each layer capabilities is still present and the most typical disadvantage is XML document processing efficiency of the Content layer. In this paper, we perform data modeling by constructing a sub-tree based on the dependencies between Content layer data, and suggest method of extract efficiently data from XML by applying the extended VTD-XML technique for the XPath query. We performs experiment to compare NETCONF in proposed method to NETCONF in previous studies and NETCONF standard. we validate superiority of improved NETCONF in the paper. As experimental results, we verify that improved NETCONF is better than the other two NETCONF each 4% and 10% in terms of query processing rate, and faster than each 3.9 seconds and 10.4 seconds in terms of query processing speed.

Design and Implementation of XML based Global Peer-to-Peer Engine (XML기반 전역 Peer-to-Peer 엔진 설계 및 구현)

  • Kwon Tae-suk;Lee Il-su;Lee Sung-young
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.1B
    • /
    • pp.73-85
    • /
    • 2004
  • SIn this paper, we introduce our experience for designing and implementing new concept of a global XML-based Peer-to-Peer (P2P) engine to support various P2P applications, and interconnection among PC, Web and mobile computing environments. The proposed P2P engine can support to heterogeneous data exchanges and web interconnection by facilitating with the text-base XML while message exchange are necessary. It is also to provide multi-level security functions as well as to apply different types of security algorithms. The system consist of four modules; a message dispatcher to scheduling and filtering the message, a SecureNet to providing security services and data transmission, a Discovery Manager to constructing peer-to-peer networking, and a Repository Manager to processing data management including XML documents. As a feasibility test, we implement various P2P services such as chatting as a communication service, white-board as an authoring tool set during collaborative working, and a file system as a file sharing service. We also compared the proposed system to a Gnutella in order to measure performance of the systems.

An Effective Query Rewriting Method in Secure XML Access Control (안전한 XML 접근제어에서 효율적인 질의 재작성 기법)

  • An, Dong-Chan;Byun, Chang-Woo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.5 s.43
    • /
    • pp.127-137
    • /
    • 2006
  • We propose two phase filtering scheme to develop an efficient mechanism for XML databases to control query-based access. An access control environment for XML documents and some techniques to deal with fine-grained authorization priorities and conflict resolution issues are proposed. Despite this, relatively little work has been done to enforce access controls particularly for XML databases in the case of query-based access. The basic idea utilized is that a user query interaction with only necessary access control rules is modified to an alternative form through a query optimization technique, which is guaranteed to have no access violations using tree-aware metadata of XML schemas. The scheme can be applied to any XML database management system and has several advantages such as small execution time overhead, fine-grained controls, and safe and correct query modification. The experimental results clearly demonstrate the efficiency of the approach.

  • PDF

Development of geo-coding module prototype on water hazard information (수재해 정보 지오코딩 모듈 프로토타입 개발)

  • BAECK, Seung Hyub;PARK, Gwang-Ha;HWANG, Eui-Ho;CHAE, Hyo-Sok
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2017.05a
    • /
    • pp.476-476
    • /
    • 2017
  • 최근 갑작스런 폭우로 인한 제방 붕괴, 침수 및 지진 등과 같은 재해 발생 시 추가 피해를 방지하고 주민들의 긴급대피를 도운 건 SNS를 통한 현장 정보와 경보 메시지의 지속적인 전파이다. 최근의 SNS는 재난정보에서도 활용할 수 있을 정도로 진화하였다. 국가재난정보 중 수재해 관련 정보를 추출하여 다양한 주제도위에 중첩으로 공간정보를 제공할 수 있는 재난정보 제공을 위한 웹서비스를 개발하고자 하였다. 수재해 정보를 필터링하기 위하여 우선 관련된 키워드 선정이 필요하며, 기본적인 키워드는 하천일람표를 참고하여 6개 권역 및 하천이름을 선정하였다. 또한, 한강 홍수 통제소의 수자원 용어사전과 (사)한국물학술단체연합회에서 발간한 물용어집을 참고하여 수재해 관련 용어들 약 300여개를 추가하였다. 선정된 용어들은 1차적으로 적재된 데이터베이스에서 수재해 정보 관련 필터링을 하는데 사용되며, 비정형 데이터들을 필터링하고 주소 정보 검색 및 추출을 통하여 정형화 하게 된다. 추출된 주소정보에 대하여 개발한 지오코딩 모듈을 적용하여 수재해 항목에 대해 좌표정보를 업데이트 하게 된다. 가뭄, 집중호우, 홍수 등의 수재해 정보별, 또한 일자별 그룹화 및 구조화를 진행하고 해당되는 정보를 공간정보 오픈플랫폼 API를 활용하여 지도상에 가시화할 수 있다. 개발한 지오코딩 모듈을 이용하여 실제 테이블 정보를 구성하여 데이터베이스에 수재해 정보 지오코딩 테이블을 구성하여 테스트 모의하였다. 재난정보 중 홍수, 가뭄에 대한 선택정보와 시간정보를 매개변수로 받는 XML 웹서비스 테스트로 검증을 하였다. 본 연구를 통하여 재난정보 가시화에 있어서 사용자가 조회하고자 하는 유형별, 날짜별 선택이 가능한 공간적 정보를 검색 및 확인할 수 있게 되었다. 개발한 수재해 정보 지오코딩 모듈 프로토 타입은 수재해 정보 플랫폼 융합기술 연구단에서 개발하는 핵심 목표시스템 내 재난정보 제공시스템에 적용 가능하며, 수재해 정보에 대하여 대국민 서비스가 가능할 것으로 사료된다.

  • PDF

Decision Method of Importance of E-Mail based on User Profiles (사용자 프로파일에 기반한 전자 메일의 중요도 결정)

  • Lee, Samuel Sang-Kon
    • The KIPS Transactions:PartB
    • /
    • v.15B no.5
    • /
    • pp.493-500
    • /
    • 2008
  • Although modern day people gather many data from the network, the users want only the information needed. Using this technology, the users can extract on the data that satisfy the query. As the previous studies use the single data in the document, frequency of the data for example, it cannot be considered as the effective data clustering method. What is needed is the effective clustering technology that can process the electronic network documents such as the e-mail or XML that contain the tags of various formats. This paper describes the study of extracting the information from the user query based on the multi-attributes. It proposes a method of extracting the data such as the sender, text type, time limit syntax in the text, and title from the e-mail and using such data for filtering. It also describes the experiment to verify that the multi-attribute based clustering method is more accurate than the existing clustering methods using only the word frequency.

Development of Intelligent Job Classification System based on Job Posting on Job Sites (구인구직사이트의 구인정보 기반 지능형 직무분류체계의 구축)

  • Lee, Jung Seung
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.123-139
    • /
    • 2019
  • The job classification system of major job sites differs from site to site and is different from the job classification system of the 'SQF(Sectoral Qualifications Framework)' proposed by the SW field. Therefore, a new job classification system is needed for SW companies, SW job seekers, and job sites to understand. The purpose of this study is to establish a standard job classification system that reflects market demand by analyzing SQF based on job offer information of major job sites and the NCS(National Competency Standards). For this purpose, the association analysis between occupations of major job sites is conducted and the association rule between SQF and occupation is conducted to derive the association rule between occupations. Using this association rule, we proposed an intelligent job classification system based on data mapping the job classification system of major job sites and SQF and job classification system. First, major job sites are selected to obtain information on the job classification system of the SW market. Then We identify ways to collect job information from each site and collect data through open API. Focusing on the relationship between the data, filtering only the job information posted on each job site at the same time, other job information is deleted. Next, we will map the job classification system between job sites using the association rules derived from the association analysis. We will complete the mapping between these market segments, discuss with the experts, further map the SQF, and finally propose a new job classification system. As a result, more than 30,000 job listings were collected in XML format using open API in 'WORKNET,' 'JOBKOREA,' and 'saramin', which are the main job sites in Korea. After filtering out about 900 job postings simultaneously posted on multiple job sites, 800 association rules were derived by applying the Apriori algorithm, which is a frequent pattern mining. Based on 800 related rules, the job classification system of WORKNET, JOBKOREA, and saramin and the SQF job classification system were mapped and classified into 1st and 4th stages. In the new job taxonomy, the first primary class, IT consulting, computer system, network, and security related job system, consisted of three secondary classifications, five tertiary classifications, and five fourth classifications. The second primary classification, the database and the job system related to system operation, consisted of three secondary classifications, three tertiary classifications, and four fourth classifications. The third primary category, Web Planning, Web Programming, Web Design, and Game, was composed of four secondary classifications, nine tertiary classifications, and two fourth classifications. The last primary classification, job systems related to ICT management, computer and communication engineering technology, consisted of three secondary classifications and six tertiary classifications. In particular, the new job classification system has a relatively flexible stage of classification, unlike other existing classification systems. WORKNET divides jobs into third categories, JOBKOREA divides jobs into second categories, and the subdivided jobs into keywords. saramin divided the job into the second classification, and the subdivided the job into keyword form. The newly proposed standard job classification system accepts some keyword-based jobs, and treats some product names as jobs. In the classification system, not only are jobs suspended in the second classification, but there are also jobs that are subdivided into the fourth classification. This reflected the idea that not all jobs could be broken down into the same steps. We also proposed a combination of rules and experts' opinions from market data collected and conducted associative analysis. Therefore, the newly proposed job classification system can be regarded as a data-based intelligent job classification system that reflects the market demand, unlike the existing job classification system. This study is meaningful in that it suggests a new job classification system that reflects market demand by attempting mapping between occupations based on data through the association analysis between occupations rather than intuition of some experts. However, this study has a limitation in that it cannot fully reflect the market demand that changes over time because the data collection point is temporary. As market demands change over time, including seasonal factors and major corporate public recruitment timings, continuous data monitoring and repeated experiments are needed to achieve more accurate matching. The results of this study can be used to suggest the direction of improvement of SQF in the SW industry in the future, and it is expected to be transferred to other industries with the experience of success in the SW industry.