• Title/Summary/Keyword: 메타정보

Search Result 2,838, Processing Time 0.024 seconds

Cancellation of MRI Artifact due to Rotational Motion (회전운동에 기인한 MRI 아티팩트의 제거)

  • 김응규
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.4
    • /
    • pp.411-419
    • /
    • 2004
  • When the imaging object rotates in image plane during MRI scan, its rotation causes phase error and non-uniform sampling to MRI signal. The model of the problem including phase error non-uniform sampling of MRI signal showed that the MRI signals corrupted by rotations about an arbitrary center and the origin in image plane are different in their phases. Therefore the following methods are presented to improve the quality of the MR image which includes the artifact. The first, assuming that the angle of 2-D rotational motion is already known and the position of 2-D rotational center is unknown, an algorithm to correct the artifact which is based on the phase correction is presented. The second, in case of 2-D rotational motion with unknown rotational center and unknown rotational angle, an algorithm is presented to correct the MRI artifact. At this case, the energy of an ideal MR image is minimum outside the boundary of the imaging object to estimate unknown motion parameters and the measured energy increases when the imaging object has an rotation. By using this property, an evaluation function is defined to estimate unknown values of rotational angle at each phase encoding step. Finally, the effectiveness of this presented techniques is shown by using a phantom image with simulated motion and a real image with 2-D translational shift and rotation.

A Effective Ant Colony Algorithm applied to the Graph Coloring Problem (그래프 착색 문제에 적용된 효과적인 Ant Colony Algorithm에 관한 연구)

  • Ahn, Sang-Huck;Lee, Seung-Gwan;Chung, Tae-Choong
    • The KIPS Transactions:PartB
    • /
    • v.11B no.2
    • /
    • pp.221-226
    • /
    • 2004
  • Ant Colony System(ACS) Algorithm is new meta-heuristic for hard combinational optimization problem. It is a population-based approach that uses exploitation of positive feedback as well as greedy search. Recently, various methods and solutions are proposed to solve optimal solution of graph coloring problem that assign to color for adjacency node($v_i, v_j$) that they has not same color. In this paper introducing ANTCOL Algorithm that is method to solve solution by Ant Colony System algorithm that is not method that it is known well as solution of existent graph coloring problem. After introducing ACS algorithm and Assignment Type Problem, show the wav how to apply ACS to solve ATP And compare graph coloring result and execution time when use existent generating functions(ANT_Random, ANT_LF, ANT_SL, ANT_DSATUR, ANT_RLF method) with ANT_XRLF method that use XRLF that apply Randomize to RLF to solve ANTCOL. Also compare graph coloring result and execution time when use method to add re-search to ANT_XRLF(ANT_XRLF_R) with existent generating functions.

The Design and Implementation of the System for Processing Well-Formed XML Document on the Client-side (클라이언트 상의 Well-Formed XML 문서 처리 시스템의 설계 및 구현)

  • Song, Jong-Chul;Moon, Byung-Joo;Hong, Gi-Chai;Cheong, Hyun-Soo;Kim, Gyu-Tae;Lee, Soo-Youn
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.10
    • /
    • pp.3236-3246
    • /
    • 2000
  • XML is a meta-language as SGML and also can be xonsructed as an Internet versionof simplified SGML being used in confunction with XLL. Xpointer and XSL. Also W3C established DTDless Well-Formed XML document to use XML document on the Web. But it isnt offered system that consists of browsing, link and DTD generating facihty, and efficiently processes DTDless Well-Formed XML document. This paper studies on an implementation and design of system to process DTDless Well-Formed XML document on the client-side. This system consists of Well-Formed XML viewer displaying Well-Formed XML documet, XLL Processor processing Xll and Auto DTD generator constructing automatically DTDs based on multiple documents of the same class. This study focuses on automatic DTD generation during hyperlink navigation and an implementation of extended links based on XLL and Xpointer. ID and Xpointer location address are used as the address mode in the links. As a result of implement of this system, it conforms to validationof extended link facihties, extracts DTD from Well-Fromed XML Documents including same root element at the same class and constructs generalized DTD.

  • PDF

Development of Collaborative Environment for Community-driven Scientific Data Curation (커뮤니티 주도적 과학 데이터 큐레이션 협업 환경의 개발)

  • Choi, Dong-Hoon;Park, Jae-Won;Kim, ByungKyu;Shin, Jin-Sup
    • The Journal of the Korea Contents Association
    • /
    • v.17 no.9
    • /
    • pp.1-11
    • /
    • 2017
  • The importance of data curation is increasingly recognized as the need of data reuse drastically grows. Due to recent data explosion, scientists invest almost 90% of their efforts in the retrieval and collection of data needed to their study. In this paper, we deal with the development and application of a collaborative environment for community-driven data curation which is essential to enhance scientific data reusability and citability. The collaborative scientific data curation environment focuses on the cross-linking between data (or data collections) and their associated literatures to capture and organize inter-relations among research results in a specific domain. Also, plenty of contextual information is provided as metadata in order to support users in understanding data. The cross-linking has been realized by using DOI system to guarantee global accessibility to data and their relationships to literatures. The curation environment has been adopted to build a community-driven curated DB by a globally well-known intrinsically-disorderd protein research group. The curated DB will drastically reduce researchers' efforts to retrieve and collect the data required for scientific discovery.

A Study on Costs of Digital Preservation (디지털 보존의 비용요소에 관한 연구)

  • Chung, Hye-Kyung
    • Journal of the Korean Society for information Management
    • /
    • v.22 no.1 s.55
    • /
    • pp.47-64
    • /
    • 2005
  • To guarantee the long-term access to digital material, digital preservation needs to be systemized, and detailed investigation on cost elements of digital preservation should be done for the continued support of budget. To meet the needs in this area, this paper categorized the digital preservation cost into direct and indirect cost through deriving common elements used in prior research on this issue. For case analysis, two institutions, representing domestic University Library and National Library of Korea under large-scale digitization currently, are selected to analyze the current status of digital preservation and estimate the preservation cost. The case analysis shows the systematic preservation function should be performed to guarantee the long-term access digital material, even though a basic digital preservation is currently conducted. It was projected that the digital preservation cost for the two libraries, accounting for $11.8\%$ and $8.6\%$ of digitization cost, respectively, should be injected every year. However, the estimated figures are very conservative, because the cost for estimating the preservation function, such as installing digital repository and producing meta data, was excluded in the estimation. This proves that digital preservation is a synthetic activity linked directly and indirectly to various activities from production to access of digital object and an essential costs that should be considered from the beginning stage of digitization project.

Service Level Agreement Specification Model of Software and Its Mediation Mechanism for Cloud Service Broker (클라우드 서비스 브로커를 위한 소프트웨어의 서비스 수준 합의 명세 모델과 중개 방법)

  • Nam, Taewoo;Yeom, Keunhyuk
    • Journal of KIISE
    • /
    • v.42 no.5
    • /
    • pp.591-600
    • /
    • 2015
  • SLA (Service Level Agreement) is an essential factor that must be guaranteed to provide a reliable and consistent service to user in cloud computing environment. Especially, a contract between user and service provider with SLA is important in an environment using a cloud service brokerage. The cloud computing is classified into IaaS, PaaS, and SaaS according to IT resources of the various cloud service. The existing SLA is difficult to reflect the quality factors of service, because it only considers factors about the physical Network environment and have no methodological approach. In this paper, we suggested a method to specify the quality characteristics of software and proposed a mechanism and structure that can exchange SLA specification between the service provider and consumer. We defined a meta-model for the SLA specification in the SaaS level, and quality requirements of the SaaS were described by the proposed specification language. Through case studies, we verified proposed specification language that can present a variety of software quality factors. By using the UDDI-based mediation process and architecture to interchange this specification, it is stored in the repository of quality specifications and exchanged during service binding time.

Real-time and Parallel Semantic Translation Technique for Large-Scale Streaming Sensor Data in an IoT Environment (사물인터넷 환경에서 대용량 스트리밍 센서데이터의 실시간·병렬 시맨틱 변환 기법)

  • Kwon, SoonHyun;Park, Dongwan;Bang, Hyochan;Park, Youngtack
    • Journal of KIISE
    • /
    • v.42 no.1
    • /
    • pp.54-67
    • /
    • 2015
  • Nowadays, studies on the fusion of Semantic Web technologies are being carried out to promote the interoperability and value of sensor data in an IoT environment. To accomplish this, the semantic translation of sensor data is essential for convergence with service domain knowledge. The existing semantic translation technique, however, involves translating from static metadata into semantic data(RDF), and cannot properly process real-time and large-scale features in an IoT environment. Therefore, in this paper, we propose a technique for translating large-scale streaming sensor data generated in an IoT environment into semantic data, using real-time and parallel processing. In this technique, we define rules for semantic translation and store them in the semantic repository. The sensor data is translated in real-time with parallel processing using these pre-defined rules and an ontology-based semantic model. To improve the performance, we use the Apache Storm, a real-time big data analysis framework for parallel processing. The proposed technique was subjected to performance testing with the AWS observation data of the Meteorological Administration, which are large-scale streaming sensor data for demonstration purposes.

Optimizing a Multimedia File System for Streaming Severs (스트리밍 서버를 위한 멀티미디어 파일 시스템 최적화)

  • 박진연;김두한;원유집;류연승
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.5_6
    • /
    • pp.268-278
    • /
    • 2004
  • In this paper, we describe our experience in the design and implementation of the SMART file system to handle multimedia workload. Our work has three design objectives: (ⅰ) efficient support for sequential workload, (ⅱ) avoiding disk fragmentation, (ⅲ) logical unit based file access. To achieve these three objectives, we develop a file system where a file consists of linked list of Data Unit Group. Instead of tree like structure of the legacy Unix file system, we use single level file structure. Our file system can also access the file based upon the logical unit which can be video frame or audio samples. Data Unit Group is a group of logical data units which is allocated continuous disk blocks. At the beginning of each Data Unit Group, there exists an index array. Each index points to the beginning of logical data units, e.g. frames in the Data Unit Group. This index array enables the random access and sequencial access of semantic data units. SMART file system is elaborately tailored to effectively support multimedia workload. We perform physical experiments and compare the performance of SMART file system with EXT2 file system and SGI XFS file system. In this experiment, SMART file system exhibits superior performance under streaming workload.

Designing the Record Management Functions for Record Content Using Advantages of Cloud Storage (클라우드 저장소 장점을 활용한 기록 콘텐츠 관리기능 설계)

  • Yim, Jin-Hee
    • Journal of Korean Society of Archives and Records Management
    • /
    • v.19 no.3
    • /
    • pp.271-292
    • /
    • 2019
  • Recently, the central administrative agency changed its business management system to cloud-based On-nara 2.0. To transfer and manage the records of the cloud business management system, the National Archives Service has developed and distributed a cloud-based records management system. It serves as an opportunity to maximize the benefits of cloud computing and redesign the records management to be more effective and efficient. The process and method of electronic record management can be transformed through digital technologies. First, we can change the transfer method for electronic records. When the business and the records management systems share the same cloud storage, it is possible to transfer the content files between the two systems without moving the contents files physically, thus copying only the metadata and reducing the cost and the risk of integrity damage. Second, the strategy for allocating storage space for contents can be conceived. Assuming that the cloud storage is shared by the business and the record management systems, it is advantageous to distinguish the storage location based on the retention period of the content files. Third, systems that access content files, such as records creation, records management, and information disclosure systems, can share the cloud storage and minimize the duplication of content files.

A Packet Processing of Handling Large-capacity Traffic over 20Gbps Method Using Multi Core and Huge Page Memory Approache

  • Kwon, Young-Sun;Park, Byeong-Chan;Chang, Hoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.6
    • /
    • pp.73-80
    • /
    • 2021
  • In this paper, we propose a packet processing method capable of handling large-capacity traffic over 20Gbps using multi-core and huge page memory approaches. As ICT technology advances, the global average monthly traffic is expected to reach 396 exabytes by 2022. With the increase in network traffic, cyber threats are also increasing, increasing the importance of traffic analysis. Traffic analyzed as an existing high-cost foreign product simply stores statistical data and visually shows it. Network administrators introduce and analyze many traffic analysis systems to analyze traffic in various sections, but they cannot check the aggregated traffic of the entire network. In addition, since most of the existing equipment is of the 10Gbps class, it cannot handle the increasing traffic every year at a fast speed. In this paper, as a method of processing large-capacity traffic over 20Gbps, the process of processing raw packets without copying from single-core and basic SMA memory approaches to high-performance packet reception, packet detection, and statistics using multi-core and NUMA memory approaches suggest When using the proposed method, it was confirmed that more than 50% of the traffic was processed compared to the existing equipment.