• Title/Summary/Keyword: Complex Query

Search Result 128, Processing Time 0.024 seconds

An Ambient Service Model for Providing Web's Stores Information on Map Interface Hierarchically through User-Context-Based Search (사용자 상황기반 검색을 통해 웹상의 상점정보를 지도상에 계층적으로 제공하는 엠비언트 서비스 모델)

  • Seo, Kyung-Seok;Lee, Ryong;Jang, Yong-Hee;Kwon, Yang-Jin
    • Spatial Information Research
    • /
    • v.18 no.2
    • /
    • pp.57-65
    • /
    • 2010
  • Users often visit many stores while comparing the products for purchasing products or products related to it. Given a service providing location information of these stores, users can make their purchase efficiently because of reducing the time and effort they spent for wandering around and obtaining new purchase opportunities by knowing a kind of relevant stores near there. In this paper, for the purpose of providing relevant stores information efficiently, we suggest an Ambient Service Model that consists of three layers: "structured(purchase-related) information space", "real space", and "ambient information space". In the model, stores information collected from the web is grouped and structured automatically by relationships in terms of purchase. And users search relevant stores information by using an Ambient Query that is created by their context in real space. Finally, users obtain relevant stores information that is in the form of hierarchy structure on map interface. Then, users can search other kinds of relevant stores information additionally by using hierarchy structure. Consequently, It is possible to develope a service that users can obtain relevant stores information intuitively without complex search processes through the model. Also, we expect that the model can be used for developing services that provide objects information related to various objects besides stores.

Design and Implementation of Real-Time Support System for Purchasing Activities Based on Ambient Service Model (엠비언트 서비스 모델 기반의 실시간 구매활동 지원 시스템 설계 및 구현)

  • Seo, Kyung-Seok;Lee, Ryong;Jang, Yong-Hee;Kwon, Yong-Jin
    • Spatial Information Research
    • /
    • v.18 no.2
    • /
    • pp.67-75
    • /
    • 2010
  • When people are shopping at a large shopping mall, they usually become to go around many stores for looking for better products and comparing them. In this paper, we design and implement a purchasing activity support system based on an Ambient Service Model that provides relevant stores information on map interface hierarchically through user contexts based search, to support such user's purchasing activities with the help of relevant information. In this system, users can search for relevant stores information through the system by Ambient Query which is created by their location and stores information with a mobile device. Then, users obtain relevant stores information provided in the form of hierarchy of keywords as a highly condensed summary and easily figure out the locations of the stores on a map interface. Moreover, users search additional other kinds of relevant stores information over the hierarchy of keywords. Eventually, users can obtain relevant stores information intuitively and conveniently without complex search processes. We implemented this system by integrating the subordinate technologies such as RFID, map-based, location-based and ontology technology. We also performed experiments on a well-known shopping region (Ilsan Lapesta shopping mall, Goyang-city Gyeonggi-do, Korea). Finally, we also confirmed that users' shopping activities were significantly improved by utility the present system.

WPS-based Satellite Image Processing onWeb Framework and Cloud Computing Environment (클라우드 컴퓨팅과 웹 프레임워크 환경에서 WPS 기반 위성영상 정보처리)

  • Yoon, Gooseon;Lee, Kiwon
    • Korean Journal of Remote Sensing
    • /
    • v.31 no.6
    • /
    • pp.561-570
    • /
    • 2015
  • Till now, applications of many kinds of satellite images have been accentuated in the datacentric scientific studies, researches regarding system development and concerned technologies for them are on the un-matured stage. Especially, satellite image processing requires large volume data handling and specific analysis functionalities, so that practical necessity of base study for system development is emphasized on. In the view of information system, various edged trends such as web standards, cloud computing, or web framework are utilized owing to their application benefits proven and business needs. Considered these aspects, a testing implementation was carried out using OpenStack cloud computing environment and e-government framework. As for the processing functions, WPS in GeoServer, as one of OGC web standards, was applied to perform interoperable data processing scheme between two or more remote servers. Working with the server implemented, client-side was also developed using several open sources such as HTML 5, jQuery, and OpenLayers. If it is that completed further experiments onsite applications with actual multi-data sets and extension of on-demand functionalities with the result of this study, it will be referred as an example case model for complicated and complex system design and implementation which needs cloud computing, geo-spatial web standards and web framework.

A Study on Spatial Data Integration using Graph Database: Focusing on Real Estate (그래프 데이터베이스를 활용한 공간 데이터 통합 방안 연구: 부동산 분야를 중심으로)

  • Ju-Young KIM;Seula PARK;Ki-Yun YU
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.26 no.3
    • /
    • pp.12-36
    • /
    • 2023
  • Graph databases, which store different types of data and their relationships modeled as a graph, can be effective in managing and analyzing real estate spatial data linked by complex relationships. However, they are not widely used due to the limited spatial functionalities of graph databases. In this study, we propose a uniform grid-based real estate spatial data management approach using a graph database to respond to various real estate-related spatial questions. By analyzing the real estate community to identify relevant data and utilizing national point numbers as unit grids, we construct a graph schema that linking diverse real estate data, and create a test database. After building a test database, we tested basic topological relationships and spatial functions using the Jackpine benchmark, and further conducted query tests based on various scenarios to verify the appropriateness of the proposed method. The results show that the proposed method successfully executed 25 out of 29 spatial topological relationships and spatial functions, and achieved about 97% accuracy for the 25 functions and 15 scenarios. The significance of this study lies in proposing an efficient data integration method that can respond to real estate-related spatial questions, considering the limited spatial operation capabilities of graph databases. However, there are limitations such as the creation of incorrect spatial topological relationships due to the use of grid-based indexes and inefficiency of queries due to list comparisons, which need to be improved in follow-up studies.

Species Identification and Monitoring of Labeling Compliance for Commercial Pufferfish Products Sold in Korean On-line Markets (국내 온라인 유통 복어 제품의 종판별 및 표시사항 모니터링 연구)

  • Ji Young Lee;Kun Hee Kim;Tae Sun Kang
    • Journal of Food Hygiene and Safety
    • /
    • v.38 no.6
    • /
    • pp.464-475
    • /
    • 2023
  • In this study, based on an analysis of two DNA barcode markers (cytochrome c oxidase subunit I and cytochrome b genes), we performed species identification and monitored labeling compliance for 50 commercial pufferfish products sold in on-line markets in Korea. Using these barcode sequences as a query for species identification and phylogenetic analysis, we screened the GenBank database. A total of seven pufferfish species (Takifugu chinensis, T. pseudommus, T. xanthopterus, T. alboplumbeus, T. porphyreus, T. vermicularis, and Lagocephalus cheesemanii) were identified and we detected 35 products (70%) that were non-compliant with the corresponding label information. Moreover, the labels on 12 commercial products contained only the general common name (i.e., pufferfish), although not the scientific or Korean names for the 21 edible pufferfish species. Furthermore, the proportion of mislabeled highly processed products (n = 9, 81.8%) was higher than that of simply processed products (n = 26, 66.7%). With respect to the country of origin, the percentage of mislabeled Chinese products (n = 8, 80%) was higher than that of Korean products (n = 26, 66.7%). In addition, the market and dialect names of different pufferfish species were labeled only as Jolbok or Milbok, whereas two non-edible pufferfish species (T. vermicularis and T. pseudommus) were used in six commercial pufferfish products described as JolboK and Gumbok on their labels, which could be attributable to the complex classification system used for pufferfish. These monitoring results highlight the necessity to develop genetic methods that can be used to identify the 21 edible pufferfish species, as well as the need for regulatory monitoring of commercial pufferfish products.

Ontology-Based Process-Oriented Knowledge Map Enabling Referential Navigation between Knowledge (지식 간 상호참조적 네비게이션이 가능한 온톨로지 기반 프로세스 중심 지식지도)

  • Yoo, Kee-Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.61-83
    • /
    • 2012
  • A knowledge map describes the network of related knowledge into the form of a diagram, and therefore underpins the structure of knowledge categorizing and archiving by defining the relationship of the referential navigation between knowledge. The referential navigation between knowledge means the relationship of cross-referencing exhibited when a piece of knowledge is utilized by a user. To understand the contents of the knowledge, a user usually requires additionally information or knowledge related with each other in the relation of cause and effect. This relation can be expanded as the effective connection between knowledge increases, and finally forms the network of knowledge. A network display of knowledge using nodes and links to arrange and to represent the relationship between concepts can provide a more complex knowledge structure than a hierarchical display. Moreover, it can facilitate a user to infer through the links shown on the network. For this reason, building a knowledge map based on the ontology technology has been emphasized to formally as well as objectively describe the knowledge and its relationships. As the necessity to build a knowledge map based on the structure of the ontology has been emphasized, not a few researches have been proposed to fulfill the needs. However, most of those researches to apply the ontology to build the knowledge map just focused on formally expressing knowledge and its relationships with other knowledge to promote the possibility of knowledge reuse. Although many types of knowledge maps based on the structure of the ontology were proposed, no researches have tried to design and implement the referential navigation-enabled knowledge map. This paper addresses a methodology to build the ontology-based knowledge map enabling the referential navigation between knowledge. The ontology-based knowledge map resulted from the proposed methodology can not only express the referential navigation between knowledge but also infer additional relationships among knowledge based on the referential relationships. The most highlighted benefits that can be delivered by applying the ontology technology to the knowledge map include; formal expression about knowledge and its relationships with others, automatic identification of the knowledge network based on the function of self-inference on the referential relationships, and automatic expansion of the knowledge-base designed to categorize and store knowledge according to the network between knowledge. To enable the referential navigation between knowledge included in the knowledge map, and therefore to form the knowledge map in the format of a network, the ontology must describe knowledge according to the relation with the process and task. A process is composed of component tasks, while a task is activated after any required knowledge is inputted. Since the relation of cause and effect between knowledge can be inherently determined by the sequence of tasks, the referential relationship between knowledge can be circuitously implemented if the knowledge is modeled to be one of input or output of each task. To describe the knowledge with respect to related process and task, the Protege-OWL, an editor that enables users to build ontologies for the Semantic Web, is used. An OWL ontology-based knowledge map includes descriptions of classes (process, task, and knowledge), properties (relationships between process and task, task and knowledge), and their instances. Given such an ontology, the OWL formal semantics specifies how to derive its logical consequences, i.e. facts not literally present in the ontology, but entailed by the semantics. Therefore a knowledge network can be automatically formulated based on the defined relationships, and the referential navigation between knowledge is enabled. To verify the validity of the proposed concepts, two real business process-oriented knowledge maps are exemplified: the knowledge map of the process of 'Business Trip Application' and 'Purchase Management'. By applying the 'DL-Query' provided by the Protege-OWL as a plug-in module, the performance of the implemented ontology-based knowledge map has been examined. Two kinds of queries to check whether the knowledge is networked with respect to the referential relations as well as the ontology-based knowledge network can infer further facts that are not literally described were tested. The test results show that not only the referential navigation between knowledge has been correctly realized, but also the additional inference has been accurately performed.

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.

A Ranking Algorithm for Semantic Web Resources: A Class-oriented Approach (시맨틱 웹 자원의 랭킹을 위한 알고리즘: 클래스중심 접근방법)

  • Rho, Sang-Kyu;Park, Hyun-Jung;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.17 no.4
    • /
    • pp.31-59
    • /
    • 2007
  • We frequently use search engines to find relevant information in the Web but still end up with too much information. In order to solve this problem of information overload, ranking algorithms have been applied to various domains. As more information will be available in the future, effectively and efficiently ranking search results will become more critical. In this paper, we propose a ranking algorithm for the Semantic Web resources, specifically RDF resources. Traditionally, the importance of a particular Web page is estimated based on the number of key words found in the page, which is subject to manipulation. In contrast, link analysis methods such as Google's PageRank capitalize on the information which is inherent in the link structure of the Web graph. PageRank considers a certain page highly important if it is referred to by many other pages. The degree of the importance also increases if the importance of the referring pages is high. Kleinberg's algorithm is another link-structure based ranking algorithm for Web pages. Unlike PageRank, Kleinberg's algorithm utilizes two kinds of scores: the authority score and the hub score. If a page has a high authority score, it is an authority on a given topic and many pages refer to it. A page with a high hub score links to many authoritative pages. As mentioned above, the link-structure based ranking method has been playing an essential role in World Wide Web(WWW), and nowadays, many people recognize the effectiveness and efficiency of it. On the other hand, as Resource Description Framework(RDF) data model forms the foundation of the Semantic Web, any information in the Semantic Web can be expressed with RDF graph, making the ranking algorithm for RDF knowledge bases greatly important. The RDF graph consists of nodes and directional links similar to the Web graph. As a result, the link-structure based ranking method seems to be highly applicable to ranking the Semantic Web resources. However, the information space of the Semantic Web is more complex than that of WWW. For instance, WWW can be considered as one huge class, i.e., a collection of Web pages, which has only a recursive property, i.e., a 'refers to' property corresponding to the hyperlinks. However, the Semantic Web encompasses various kinds of classes and properties, and consequently, ranking methods used in WWW should be modified to reflect the complexity of the information space in the Semantic Web. Previous research addressed the ranking problem of query results retrieved from RDF knowledge bases. Mukherjea and Bamba modified Kleinberg's algorithm in order to apply their algorithm to rank the Semantic Web resources. They defined the objectivity score and the subjectivity score of a resource, which correspond to the authority score and the hub score of Kleinberg's, respectively. They concentrated on the diversity of properties and introduced property weights to control the influence of a resource on another resource depending on the characteristic of the property linking the two resources. A node with a high objectivity score becomes the object of many RDF triples, and a node with a high subjectivity score becomes the subject of many RDF triples. They developed several kinds of Semantic Web systems in order to validate their technique and showed some experimental results verifying the applicability of their method to the Semantic Web. Despite their efforts, however, there remained some limitations which they reported in their paper. First, their algorithm is useful only when a Semantic Web system represents most of the knowledge pertaining to a certain domain. In other words, the ratio of links to nodes should be high, or overall resources should be described in detail, to a certain degree for their algorithm to properly work. Second, a Tightly-Knit Community(TKC) effect, the phenomenon that pages which are less important but yet densely connected have higher scores than the ones that are more important but sparsely connected, remains as problematic. Third, a resource may have a high score, not because it is actually important, but simply because it is very common and as a consequence it has many links pointing to it. In this paper, we examine such ranking problems from a novel perspective and propose a new algorithm which can solve the problems under the previous studies. Our proposed method is based on a class-oriented approach. In contrast to the predicate-oriented approach entertained by the previous research, a user, under our approach, determines the weights of a property by comparing its relative significance to the other properties when evaluating the importance of resources in a specific class. This approach stems from the idea that most queries are supposed to find resources belonging to the same class in the Semantic Web, which consists of many heterogeneous classes in RDF Schema. This approach closely reflects the way that people, in the real world, evaluate something, and will turn out to be superior to the predicate-oriented approach for the Semantic Web. Our proposed algorithm can resolve the TKC(Tightly Knit Community) effect, and further can shed lights on other limitations posed by the previous research. In addition, we propose two ways to incorporate data-type properties which have not been employed even in the case when they have some significance on the resource importance. We designed an experiment to show the effectiveness of our proposed algorithm and the validity of ranking results, which was not tried ever in previous research. We also conducted a comprehensive mathematical analysis, which was overlooked in previous research. The mathematical analysis enabled us to simplify the calculation procedure. Finally, we summarize our experimental results and discuss further research issues.