• Title/Summary/Keyword: meta information

Search Result 1,241, Processing Time 0.026 seconds

Monitoring Ground-level SO2 Concentrations Based on a Stacking Ensemble Approach Using Satellite Data and Numerical Models (위성 자료와 수치모델 자료를 활용한 스태킹 앙상블 기반 SO2 지상농도 추정)

  • Choi, Hyunyoung;Kang, Yoojin;Im, Jungho;Shin, Minso;Park, Seohui;Kim, Sang-Min
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1053-1066
    • /
    • 2020
  • Sulfur dioxide (SO2) is primarily released through industrial, residential, and transportation activities, and creates secondary air pollutants through chemical reactions in the atmosphere. Long-term exposure to SO2 can result in a negative effect on the human body causing respiratory or cardiovascular disease, which makes the effective and continuous monitoring of SO2 crucial. In South Korea, SO2 monitoring at ground stations has been performed, but this does not provide spatially continuous information of SO2 concentrations. Thus, this research estimated spatially continuous ground-level SO2 concentrations at 1 km resolution over South Korea through the synergistic use of satellite data and numerical models. A stacking ensemble approach, fusing multiple machine learning algorithms at two levels (i.e., base and meta), was adopted for ground-level SO2 estimation using data from January 2015 to April 2019. Random forest and extreme gradient boosting were used as based models and multiple linear regression was adopted for the meta-model. The cross-validation results showed that the meta-model produced the improved performance by 25% compared to the base models, resulting in the correlation coefficient of 0.48 and root-mean-square-error of 0.0032 ppm. In addition, the temporal transferability of the approach was evaluated for one-year data which were not used in the model development. The spatial distribution of ground-level SO2 concentrations based on the proposed model agreed with the general seasonality of SO2 and the temporal patterns of emission sources.

Recommending Core and Connecting Keywords of Research Area Using Social Network and Data Mining Techniques (소셜 네트워크와 데이터 마이닝 기법을 활용한 학문 분야 중심 및 융합 키워드 추천 서비스)

  • Cho, In-Dong;Kim, Nam-Gyu
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.1
    • /
    • pp.127-138
    • /
    • 2011
  • The core service of most research portal sites is providing relevant research papers to various researchers that match their research interests. This kind of service may only be effective and easy to use when a user can provide correct and concrete information about a paper such as the title, authors, and keywords. However, unfortunately, most users of this service are not acquainted with concrete bibliographic information. It implies that most users inevitably experience repeated trial and error attempts of keyword-based search. Especially, retrieving a relevant research paper is more difficult when a user is novice in the research domain and does not know appropriate keywords. In this case, a user should perform iterative searches as follows : i) perform an initial search with an arbitrary keyword, ii) acquire related keywords from the retrieved papers, and iii) perform another search again with the acquired keywords. This usage pattern implies that the level of service quality and user satisfaction of a portal site are strongly affected by the level of keyword management and searching mechanism. To overcome this kind of inefficiency, some leading research portal sites adopt the association rule mining-based keyword recommendation service that is similar to the product recommendation of online shopping malls. However, keyword recommendation only based on association analysis has limitation that it can show only a simple and direct relationship between two keywords. In other words, the association analysis itself is unable to present the complex relationships among many keywords in some adjacent research areas. To overcome this limitation, we propose the hybrid approach for establishing association network among keywords used in research papers. The keyword association network can be established by the following phases : i) a set of keywords specified in a certain paper are regarded as co-purchased items, ii) perform association analysis for the keywords and extract frequent patterns of keywords that satisfy predefined thresholds of confidence, support, and lift, and iii) schematize the frequent keyword patterns as a network to show the core keywords of each research area and connecting keywords among two or more research areas. To estimate the practical application of our approach, we performed a simple experiment with 600 keywords. The keywords are extracted from 131 research papers published in five prominent Korean journals in 2009. In the experiment, we used the SAS Enterprise Miner for association analysis and the R software for social network analysis. As the final outcome, we presented a network diagram and a cluster dendrogram for the keyword association network. We summarized the results in Section 4 of this paper. The main contribution of our proposed approach can be found in the following aspects : i) the keyword network can provide an initial roadmap of a research area to researchers who are novice in the domain, ii) a researcher can grasp the distribution of many keywords neighboring to a certain keyword, and iii) researchers can get some idea for converging different research areas by observing connecting keywords in the keyword association network. Further studies should include the following. First, the current version of our approach does not implement a standard meta-dictionary. For practical use, homonyms, synonyms, and multilingual problems should be resolved with a standard meta-dictionary. Additionally, more clear guidelines for clustering research areas and defining core and connecting keywords should be provided. Finally, intensive experiments not only on Korean research papers but also on international papers should be performed in further studies.

A Content Analysis of the Trends in Vision Research With Focus on Visual Search, Eye Movement, and Eye Track

  • Rhie, Ye Lim;Lim, Ji Hyoun;Yun, Myung Hwan
    • Journal of the Ergonomics Society of Korea
    • /
    • v.33 no.1
    • /
    • pp.69-76
    • /
    • 2014
  • Objective: This study aims to present literature providing researchers with insights on specific fields of research and highlighting the major issues in the research topics. A systematic review is suggested using content analysis on literatures regarding "visual search", "eye movement", and "eye track". Background: Literature review can be classified as "narrative" or "systematic" depending on its approach in structuring the content of the research. Narrative review is a traditional approach that describes the current state of a study field and discusses relevant topics. However, since literatures on specific area cover a broad range, reviewers inherently give subjective weight on specific issues. On the contrary, systematic review applies explicit structured methodology to observe the study trends quantitatively. Method: We collected meta-data of journal papers using three search keywords: visual search, eye movement, and eye track. The collected information contains an unstructured data set including many natural languages which compose titles and abstracts, while the keyword of the journal paper is the only structured one. Based on the collected terms, seven categories were evaluated by inductive categorization and quantitative analysis from the chronological trend of the research area. Results: Unstructured information contains heavier content on "stimuli" and "condition" categories as compared with structured information. Studies on visual search cover a wide range of cognitive area whereas studies on eye movement and eye track are closely related to the physiological aspect. In addition, experimental studies show an increasing trend as opposed to the theoretical studies. Conclusion: By systematic review, we could quantitatively identify the characteristic of the research keyword which presented specific research topics. We also found out that the structured information was more suitable to observe the aim of the research. Chronological analysis on the structured keyword data showed that studies on "physical eye movement" and "cognitive process" were jointly studied in increasing fashion. Application: While conventional narrative literature reviews were largely dependent on authors' instinct, quantitative approach enabled more objective and macroscopic views. Moreover, the characteristics of information type were specified by comparing unstructured and structured information. Systematic literature review also could be used to support the authors' instinct in narrative literature reviews.

A Method to Find Feature Set for Detecting Various Denial Service Attacks in Power Grid (전력망에서의 다양한 서비스 거부 공격 탐지 위한 특징 선택 방법)

  • Lee, DongHwi;Kim, Young-Dae;Park, Woo-Bin;Kim, Joon-Seok;Kang, Seung-Ho
    • KEPCO Journal on Electric Power and Energy
    • /
    • v.2 no.2
    • /
    • pp.311-316
    • /
    • 2016
  • Network intrusion detection system based on machine learning method such as artificial neural network is quite dependent on the selected features in terms of accuracy and efficiency. Nevertheless, choosing the optimal combination of features, which guarantees accuracy and efficienty, from generally used many features to detect network intrusion requires extensive computing resources. In this paper, we deal with a optimal feature selection problem to determine 6 denial service attacks and normal usage provided by NSL-KDD data. We propose a optimal feature selection algorithm. Proposed algorithm is based on the multi-start local search algorithm, one of representative meta-heuristic algorithm for solving optimization problem. In order to evaluate the performance of our proposed algorithm, comparison with a case of all 41 features used against NSL-KDD data is conducted. In addtion, comparisons between 3 well-known machine learning methods (multi-layer perceptron., Bayes classifier, and Support vector machine) are performed to find a machine learning method which shows the best performance combined with the proposed feature selection method.

Brain MRI Template-Driven Medical Images Mapping Method Based on Semantic Features for Ischemic Stroke (허혈성 뇌졸중을 위한 뇌 자기공명영상의 의미적 특징 기반 템플릿 중심 의료 영상 매핑 기법)

  • Park, Ye-Seul;Lee, Meeyeon;Lee, Jung-Won
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.2
    • /
    • pp.69-78
    • /
    • 2016
  • Ischemic stroke is a disease that the brain tissues cannot function by reducing blood flow due to thrombosis or embolisms. Due to the nature of the disease, it is most important to identify the status of cerebral vessel and the medical images are necessarily used for its diagnosis. Among many indicators, brain MRI is most widely utilized because experts can effectively obtain the semantic information such as cerebral anatomy aiding the diagnosis with it. However, in case of emergency diseases like ischemic stroke, even though a intelligent system is required for supporting the prompt diagnosis and treatment, the current systems have some difficulties to provide the information of medical images intuitively. In other words, as the current systems have managed the medical images based on the basic meta-data such as image name, ID and so on, they cannot consider semantic information inherent in medical images. Therefore, in this paper, to provide core information like cerebral anatomy contained in brain MRI, we suggest a template-driven medical images mapping method. The key idea of the method is defining the mapping characteristics between anatomic feature and representative images by using template images that can be representative of the whole brain MRI image set and revealing the semantic relations that only medical experts can check between images. With our method, it will be possible to manage the medical images based on semantic.

A Study on Next-Generation Data Protection Based on Non File System for Spreading Smart Factory (스마트팩토리 확산을 위한 비파일시스템(None File System) 기반의 차세대 데이터보호에 관한 연구)

  • Kim, Seungyong;Hwang, Incheol;Kim, Dongsik
    • Journal of the Society of Disaster Information
    • /
    • v.17 no.1
    • /
    • pp.176-183
    • /
    • 2021
  • Purpose: The introduction of smart factories that reflect the 4th industrial revolution technologies such as AI, IoT, and VR, has been actively promoted in Korea. However, in order to solve various problems arising from existing file-based operating systems, this research will focus on identifying and verifying non-file system-based data protection technology. Method: The research will measure security storage that cannot be identified or controlled by the operating system. How to activate secure storage based on the input of digital key values. Establish a control unit that provides input and output information based on BIOS activation. Observe non-file-type structure so that mapping behavior using second meta-data can be performed according to the activation of the secure storage. Result: First, the creation of non-file system-based secure storage's data input/output were found to match the hash function value of the sample data with the hash function value of the normal storage and data. Second, the data protection performance experiments in secure storage were compared to the hash function value of the original file with the hash function value of the secure storage after ransomware activity to verify data protection performance against malicious ransomware. Conclusion: Smart factory technology is a nationally promoted technology that is being introduced to the public and this research implemented and experimented on a new concept of data protection technology to protect crucial data within the information system. In order to protect sensitive data, implementation of non-file-type secure storage technology that is non-dependent on file system is highly recommended. This research has proven the security and safety of such technology and verified its purpose.

A Research in Applying Big Data and Artificial Intelligence on Defense Metadata using Multi Repository Meta-Data Management (MRMM) (국방 빅데이터/인공지능 활성화를 위한 다중메타데이터 저장소 관리시스템(MRMM) 기술 연구)

  • Shin, Philip Wootaek;Lee, Jinhee;Kim, Jeongwoo;Shin, Dongsun;Lee, Youngsang;Hwang, Seung Ho
    • Journal of Internet Computing and Services
    • /
    • v.21 no.1
    • /
    • pp.169-178
    • /
    • 2020
  • The reductions of troops/human resources, and improvement in combat power have made Korean Department of Defense actively adapt 4th Industrial Revolution technology (Artificial Intelligence, Big Data). The defense information system has been developed in various ways according to the task and the uniqueness of each military. In order to take full advantage of the 4th Industrial Revolution technology, it is necessary to improve the closed defense datamanagement system.However, the establishment and usage of data standards in all information systems for the utilization of defense big data and artificial intelligence has limitations due to security issues, business characteristics of each military, anddifficulty in standardizing large-scale systems. Based on the interworking requirements of each system, data sharing is limited through direct linkage through interoperability agreement between systems. In order to implement smart defense using the 4th Industrial Revolution technology, it is urgent to prepare a system that can share defense data and make good use of it. To technically support the defense, it is critical to develop Multi Repository Meta-Data Management (MRMM) that supports systematic standard management of defense data that manages enterprise standard and standard mapping for each system and promotes data interoperability through linkage between standards which obeys the Defense Interoperability Management Development Guidelines. We introduced MRMM, and implemented by using vocabulary similarity using machine learning and statistical approach. Based on MRMM, We expect to simplify the standardization integration of all military databases using artificial intelligence and bigdata. This will lead to huge reduction of defense budget while increasing combat power for implementing smart defense.

Semantic Process Retrieval with Similarity Algorithms (유사도 알고리즘을 활용한 시맨틱 프로세스 검색방안)

  • Lee, Hong-Joo;Klein, Mark
    • Asia pacific journal of information systems
    • /
    • v.18 no.1
    • /
    • pp.79-96
    • /
    • 2008
  • One of the roles of the Semantic Web services is to execute dynamic intra-organizational services including the integration and interoperation of business processes. Since different organizations design their processes differently, the retrieval of similar semantic business processes is necessary in order to support inter-organizational collaborations. Most approaches for finding services that have certain features and support certain business processes have relied on some type of logical reasoning and exact matching. This paper presents our approach of using imprecise matching for expanding results from an exact matching engine to query the OWL(Web Ontology Language) MIT Process Handbook. MIT Process Handbook is an electronic repository of best-practice business processes. The Handbook is intended to help people: (1) redesigning organizational processes, (2) inventing new processes, and (3) sharing ideas about organizational practices. In order to use the MIT Process Handbook for process retrieval experiments, we had to export it into an OWL-based format. We model the Process Handbook meta-model in OWL and export the processes in the Handbook as instances of the meta-model. Next, we need to find a sizable number of queries and their corresponding correct answers in the Process Handbook. Many previous studies devised artificial dataset composed of randomly generated numbers without real meaning and used subjective ratings for correct answers and similarity values between processes. To generate a semantic-preserving test data set, we create 20 variants for each target process that are syntactically different but semantically equivalent using mutation operators. These variants represent the correct answers of the target process. We devise diverse similarity algorithms based on values of process attributes and structures of business processes. We use simple similarity algorithms for text retrieval such as TF-IDF and Levenshtein edit distance to devise our approaches, and utilize tree edit distance measure because semantic processes are appeared to have a graph structure. Also, we design similarity algorithms considering similarity of process structure such as part process, goal, and exception. Since we can identify relationships between semantic process and its subcomponents, this information can be utilized for calculating similarities between processes. Dice's coefficient and Jaccard similarity measures are utilized to calculate portion of overlaps between processes in diverse ways. We perform retrieval experiments to compare the performance of the devised similarity algorithms. We measure the retrieval performance in terms of precision, recall and F measure? the harmonic mean of precision and recall. The tree edit distance shows the poorest performance in terms of all measures. TF-IDF and the method incorporating TF-IDF measure and Levenshtein edit distance show better performances than other devised methods. These two measures are focused on similarity between name and descriptions of process. In addition, we calculate rank correlation coefficient, Kendall's tau b, between the number of process mutations and ranking of similarity values among the mutation sets. In this experiment, similarity measures based on process structure, such as Dice's, Jaccard, and derivatives of these measures, show greater coefficient than measures based on values of process attributes. However, the Lev-TFIDF-JaccardAll measure considering process structure and attributes' values together shows reasonably better performances in these two experiments. For retrieving semantic process, we can think that it's better to consider diverse aspects of process similarity such as process structure and values of process attributes. We generate semantic process data and its dataset for retrieval experiment from MIT Process Handbook repository. We suggest imprecise query algorithms that expand retrieval results from exact matching engine such as SPARQL, and compare the retrieval performances of the similarity algorithms. For the limitations and future work, we need to perform experiments with other dataset from other domain. And, since there are many similarity values from diverse measures, we may find better ways to identify relevant processes by applying these values simultaneously.

Assessment of Rocks and Alteration Information Extraction using ASTER data for Övörkhangaii Province, Mongolia (ASTER 영상자료를 활용한 몽골 오보르항가이(Övörkhangai) 일대 암상 빛 변질 정보추출의 활용가능성 평가)

  • Jeong, Yongsik;Yu, Jaehyung;Koh, Sang-Mo;Heo, Chul-Ho
    • Economic and Environmental Geology
    • /
    • v.48 no.4
    • /
    • pp.325-335
    • /
    • 2015
  • This study examined the possibility to extract potential alteration zones and lithologic information based on ASTER band ratio techniques for mineralized area located in ${\ddot{O}}v{\ddot{o}}rkhangai$ province, Mongolia, and the effectiveness of remote sensing as a preliminary exploration tool for mineral exploration was tested. The results of ABRLO, PBRLO, and PrBRLO models indicated that the detection of argillic zone requires the verification of the samples to verify hydrothermal alteration minerals as clay minerals can formed by weathering process, whereas phyllic-propylitic zones were considerably related to the spatial distribution of the intrusive bodies, geological structures, and ore distribution. QI and MI results showed that QI is more useful for sedimentary rocks such as conglomerate and sandstone than meta-sedimentary like quartzite, and MI faced relatively uncertain in detection of felsic or mafic silicate rocks. QI and MI may require additional geologic information such as the characteristics of samples and geological survey data to improve extraction of lithologic information, and, if so, it is expected that remote sensing technique would contribute significantly as a preliminary geological survey method.

Segmenting Korean Millennial Consumers of Sharing Economy Services on Social Networking: A Psychographic-based Approach (소셜 네트워크 기반 공유경제 서비스에 관한 밀레니얼스 소비자 세분화 연구: 사이코그래픽 관점에서)

  • Lee, Jae Heon;Choi, Jae Won;Kim, Ki Youn
    • Journal of Internet Computing and Services
    • /
    • v.16 no.6
    • /
    • pp.109-121
    • /
    • 2015
  • The purpose of this qualitative study is to explore consumer behavioral trends, psychological characteristics and various cognitive types of Millennial Generation consumers, primarily in their 20s, who are familiar with sharing economy services based on the emerging social networking technology. Using Q methodology, this paper theoretically defines four and interprets via a social science perspective four different types of these young consumers who are skilled at state-of-the-art ICT equipment, devices or online networking services. Sharing economy services in Korea's academic and industrial services are influenced by government policy, and related research is relatively new. This study is focused on discovering unique psychographic characteristics called 'schemata' that include personal interest, preference, attitude, and opinion. On the basis of 40 Q-sorted data samples, the analysis examined 180 collected statements from meta-studies and interviews with 35 individuals born between 1997 and 1992. As a result, four consumer groups were identifies: Type 1 'Early majority', Type 2 'Laggard', Type 3 'Opinion leader', and Type 4 'Late majority'. The results of this research can be used to explore to study in greater detail the behavior and psychological aspects of Millennial General consumers'.