• Title/Summary/Keyword: Information Attributes

Search Result 2,155, Processing Time 0.03 seconds

Cognition and Satisfaction of Customer in Home-delivered Meal (가정배달급식에 대한 고객의 인식 및 만족도 조사)

  • 김혜영;류시현
    • Korean journal of food and cookery science
    • /
    • v.19 no.4
    • /
    • pp.529-538
    • /
    • 2003
  • The objectives of this study were to measure customers' cognition and overall satisfaction, and to identify relatively important attributes for the overall satisfaction, of home-delivered meals. Questionnaires were distributed to 243 customers. The statistical data analyses were completed by x$^2$-tests, ANOV A, factor analysis, reliability analysis and regression analysis using SPSS version 10. 56.6% of customers get obtained information from the internet, with 31.3% of these using this method at least once a week, but 72.9% of customers used this method less than once per years. The major reasons for ordering home-delivered meals were tired of cooking, more economical and no time to cook. The results were significantly different in relation to age, occupation and monthly income. The major reasons for hesitation about ordering home-delivered meals were meals should be prepared in households, not sanitary and the use of too many artificial flavors. The results for this factor were significantly different in relation to gender, age and monthly income(p<0.01). The most preferred kinds of home-delivery meals were Korean soup (guk), stew, soup (tang), speciality dishes and party dishes. The customer's cognition of kindness of the delivery staff was highest, with food temperature being the lowest among the options. The food and service level factors were derived from a factor based analysis of customer's cognition towards home-delivered meals. The customer's cognition of food taste, food quantity, kindness of delivery staff and packaging container shape were significantly different according to the use frequency and use period. The packaging method, sanitation, kindness of delivery staff, price and taste were the most relatively important attributes for overall satisfaction with home-delivered meals.

Prefetching based on the Type-Level Access Pattern in Object-Relational DBMSs (객체관계형 DBMS에서 타입수준 액세스 패턴을 이용한 선인출 전략)

  • Han, Wook-Shin;Moon, Yang-Sae;Whang, Kyu-Young
    • Journal of KIISE:Databases
    • /
    • v.28 no.4
    • /
    • pp.529-544
    • /
    • 2001
  • Prefetching is an effective method to minimize the number of roundtrips between the client and the server in database management systems. In this paper we propose new notions of the type-level access pattern and the type-level access locality and developed an efficient prefetchin policy based on the notions. The type-level access patterns is a sequence of attributes that are referenced in accessing the objects: the type-level access locality a phenomenon that regular and repetitive type-level access patterns exist. Existing prefetching methods are based on object-level or page-level access patterns, which consist of object0ids of page-ids of the objects accessed. However, the drawback of these methods is that they work only when exactly the same objects or pages are accessed repeatedly. In contrast, even though the same objects are not accessed repeatedly, our technique effectively prefetches objects if the same attributes are referenced repeatedly, i,e of there is type-level access locality. Many navigational applications in Object-Relational Database Management System(ORDBMs) have type-level access locality. Therefore our technique can be employed in ORDBMs to effectively reduce the number of roundtrips thereby significantly enhancing the performance. We have conducted extensive experiments in a prototype ORDBMS to show the effectiveness of our algorithm. Experimental results using the 007 benchmark and a real GIS application show that our technique provides orders of magnitude improvements in the roundtrips and several factors of improvements in overall performance over on-demand fetching and context-based prefetching, which a state-of the art prefetching method. These results indicate that our approach significantly and is a practical method that can be implemented in commercial ORDMSs.

  • PDF

Classification, Analysis on Attributes and Sustainable Management Plan of Biotop Established in Pohang City (포항시 비오톱의 유형 구분, 속성 분석 및 복원 방안)

  • Jung, Song Hie;Kim, Dong Uk;Lim, Bong Soon;Kim, A Reum;Seol, Jaewon;Lee, Chang Seok
    • Korean Journal of Ecology and Environment
    • /
    • v.52 no.3
    • /
    • pp.245-265
    • /
    • 2019
  • Biotope, which represents the characteristic habitats of living organisms, need to be identified as essential for the efficient creation and sustainable management of urban ecosystems. This study was carried out to provide the basic information for ecological urban planning by analyzing types and attributes of the biotop established throughout the whole area of the Pohang city, a representative industrial city in Korea. The biotop established in Pohang city is composed of 12 types including forests (coniferous, deciduous, and mixed forests), agricultural fields (rice paddy and upland field), green facilities, river, reservoir, bare ground, residential area, public facilities, commercial area, industrial area, roads, and schools. As a result of analyzing the properties according to biotop types, industrial, commercial and residential areas, which represent urban areas, was dominated by introduced vegetation. Moreover the introduced vegetation is usually composed of exotic plants or modified forms for landscape architecture and horticulture rather than native plants, which reflects ecological property of both region and site. As the distance from the urban center increases, the agricultural field showed a form of typical farmland, whereas the closer it is, the more form of greenhouse farming. Natural green spaces were divided into riparian vegetation established along the stream and forest vegetation. Forest vegetation is consisted of secondary forests (seven communities) and plantations (three communities). The urban landscape of Pohang city is dominated by the industrial area. Among them, the steel industry, which occurs large amounts of heat pollution and carbon dioxide, occupies a large proportion. On the other hand, green space is very insufficient in quantity and inferior in quality. This study proposed several restoration plans and further, a green network, which ties the existing green spaces and the green space to be restored as a strategy to improve the environmental quality in this area.

A Study on the Acceptance of Convergence System of Broadcasting, and Telecommunication, and Their Relative Efficiency Focusing on IPFV (방송과 통신 융합시스템의 수용 및 상대적 효능에 관한 연구: IPTV를 중심으로)

  • Um, Myoung-Yong;Lee, Sang-Ho;Kim, Jai-Beam
    • Asia pacific journal of information systems
    • /
    • v.19 no.3
    • /
    • pp.25-49
    • /
    • 2009
  • Advances in technology have resulted in the emergence of new information systems. The convergence of IT and manufacturing sectors has blurred the boundaries among industries. Also, such convergence has become established as a paradigm to build a new area. Especially the convergence of broadcasting and telecommunication, notably in the case of IPTV (Internet Protocol Television), is among the most salient examples of its kind in recent years as a major case of disruptive technology innovation. Despite its much fanfare, such convergence, however, has not fulfilled the expectation; it has not produced positive economic effects while negatively affecting the growth of IPIV. Stakeholders in and around IPIV including telecommunication companies, broadcasting corporations, and government bodies wish to gain control of IPTV under their wings. IPTV has drifted in the midst of conflicts among the stakeholders in and around IPTV, particularly telecommunication and broadcasting organizations in a broad sense. Our empirical research intends to deal with how audiences accept IPTV and how firms provide IPTV services to utilize their resources. Three research questions in this paper include, first, whether Technology Acceptance Model (TAM) can sufficiently explain the acceptance of IPTV as an information system. The second question concerns with empirically testing the playful aspect of IPTV to increase its audience acceptance. Last, but not least, this paper deals with how firms can efficiently and effectively allocate their limited resources to increase IPTV viewers. To answer those three main questions of our study, we collect data from 197 current subscribers of high speed internet service and/or cable/satellite television. Empirical results show that 'perceived usefulness (PU) $\rightarrow$ Intention to use' and 'perceived ease of use (PEU) $\rightarrow$ Intention to use' are significant. Also, 'perceived ease of use' is significantly related to 'perceived usefulness.' Perceived ease of handling IPTV without much effort can positively influence the perceived value of IPTV. In this regard, engineers and designers of IPTV should pay more attention to the user-friendly interface of IPTV. In addition, 'perceived playfulness (PP)' of IPTV is positively related to 'intention to use'. Flow, fun and entertainment have recently gained greater attention in the research concerned with information systems. Such attention is due to the changing features of information systems in recent years that combine the functional and leisure attributes. These results give practical implications to the design of IPTV that reflects not just leisure but also functional elements. This paper also investigates the relationship between 'perceived ease of use (PEU)' and 'perceived playfulness (PP).' PEU is positively related to pp. Audiences without fear can be attracted more easily to the user-friendly IPTV, thereby perceiving the fun and entertainment with ease. Practical implications from this finding are that, to attract more interest and involvement from the audience, IPTV needs to be designed with similar or even more user friendly interface. Of the factors related to 'intention to use', 'perceived usefulness (PU)' and 'perceived ease of use (PEU)' have greater impacts than 'perceived playfulness (PP).' Between PU and PEU, their impacts on 'intention to use' are not significantly different statistically. Managerial implications of this finding are that firms in preparation for the launch of IPTV service should prioritize the functions and interface of IPTV. This empirical paper also provides further insight into the ways in which firms can strategically allocate their limited resources so as to appeal to viewers, both current and potential, of IPTV.

A Semantic Classification Model for e-Catalogs (전자 카탈로그를 위한 의미적 분류 모형)

  • Kim Dongkyu;Lee Sang-goo;Chun Jonghoon;Choi Dong-Hoon
    • Journal of KIISE:Databases
    • /
    • v.33 no.1
    • /
    • pp.102-116
    • /
    • 2006
  • Electronic catalogs (or e-catalogs) hold information about the goods and services offered or requested by the participants, and consequently, form the basis of an e-commerce transaction. Catalog management is complicated by a number of factors and product classification is at the core of these issues. Classification hierarchy is used for spend analysis, custom3 regulation, and product identification. Classification is the foundation on which product databases are designed, and plays a central role in almost all aspects of management and use of product information. However, product classification has received little formal treatment in terms of underlying model, operations, and semantics. We believe that the lack of a logical model for classification Introduces a number of problems not only for the classification itself but also for the product database in general. It needs to meet diverse user views to support efficient and convenient use of product information. It needs to be changed and evolved very often without breaking consistency in the cases of introduction of new products, extinction of existing products, class reorganization, and class specialization. It also needs to be merged and mapped with other classification schemes without information loss when B2B transactions occur. For these requirements, a classification scheme should be so dynamic that it takes in them within right time and cost. The existing classification schemes widely used today such as UNSPSC and eClass, however, have a lot of limitations to meet these requirements for dynamic features of classification. In this paper, we try to understand what it means to classify products and present how best to represent classification schemes so as to capture the semantics behind the classifications and facilitate mappings between them. Product information implies a plenty of semantics such as class attributes like material, time, place, etc., and integrity constraints. In this paper, we analyze the dynamic features of product databases and the limitation of existing code based classification schemes. And describe the semantic classification model, which satisfies the requirements for dynamic features oi product databases. It provides a means to explicitly and formally express more semantics for product classes and organizes class relationships into a graph. We believe the model proposed in this paper satisfies the requirements and challenges that have been raised by previous works.

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.

KANO-TOPSIS Model for AI Based New Product Development: Focusing on the Case of Developing Voice Assistant System for Vehicles (KANO-TOPSIS 모델을 이용한 지능형 신제품 개발: 차량용 음성비서 시스템 개발 사례)

  • Yang, Sungmin;Tak, Junhyuk;Kwon, Donghwan;Chung, Doohee
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.287-310
    • /
    • 2022
  • Companies' interest in developing AI-based intelligent new products is increasing. Recently, the main concern of companies is to innovate customer experience and create new values by developing new products through the effective use of Artificial intelligence technology. However, due to the nature of products based on radical technologies such as artificial intelligence, intelligent products differ from existing products and development methods, so it is clear that there is a limitation to applying the existing development methodology as it is. This study proposes a new research method based on KANO-TOPSIS for the successful development of AI-based intelligent new products by using car voice assistants as an example. Using the KANO model, select and evaluate functions that customers think are necessary for new products, and use the TOPSIS method to derives priorities by finding the importance of functions that customers need. For the analysis, major categories such as vehicle condition check and function control elements, driving-related elements, characteristics of voice assistant itself, infotainment elements, and daily life support elements were selected and customer demand attributes were subdivided. As a result of the analysis, high recognition accuracy should be considered as a top priority in the development of car voice assistants. Infotainment elements that provide customized content based on driver's biometric information and usage habits showed lower priorities than expected, while functions related to driver safety such as vehicle condition notification, driving assistance, and security, also showed as the functions that should be developed preferentially. This study is meaningful in that it presented a new product development methodology suitable for the characteristics of AI-based intelligent new products with innovative characteristics through an excellent model combining KANO and TOPSIS.

Development of Plant BIM Library according to Object Geometry and Attribute Information Guidelines (객체 형상 및 속성정보 지침에 따른 수목 BIM 라이브러리 개발)

  • Kim, Bok-Young
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.52 no.2
    • /
    • pp.51-63
    • /
    • 2024
  • While the government policy to fully adopt BIM in the construction sector is being implemented, the construction and utilization of landscape BIM models are facing challenges due to problems such as limitations in BIM authoring tools, difficulties in modeling natural materials, and a shortage in BIM content including libraries. In particular, plants, fundamental design elements in the field of landscape architecture, must be included in BIM models, yet they are often omitted during the modeling process, or necessary information is not included, which further compromises the quality of the BIM data. This study aimed to contribute to the construction and utilization of landscape BIM models by developing a plant library that complies with BIM standards and is applicable to the landscape industry. The plant library of trees and shrubs was developed in Revit by modeling 3D shapes and collecting attribute items. The geometric information is simplified to express the unique characteristics of each plant species at LOD200, LOD300, and LOD350 levels. The attribute information includes properties on plant species identification, such as species name, specifications, and quantity estimation, as well as ecological attributes and environmental performance information, totaling 24 items. The names of the files were given so that the hierarchy of an object in the landscape field could be revealed and the object name could classify the plant itself. Its usability was examined by building a landscape BIM model of an apartment complex. The result showed that the plant library facilitated the construction process of the landscape BIM model. It was also confirmed that the library was properly operated in the basic utilization of the BIM model, such as 2D documentation, quantity takeoff, and design review. However, the library lacked ground cover, and had limitations in those variables such as the environmental performance of plants because various databases for some materials have not yet been established. Further efforts are needed to develop BIM modeling tools, techniques, and various databases for natural materials. Moreover, entities and systems responsible for creating, managing, distributing, and disseminating BIM libraries must be established.

Mapping Categories of Heterogeneous Sources Using Text Analytics (텍스트 분석을 통한 이종 매체 카테고리 다중 매핑 방법론)

  • Kim, Dasom;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.4
    • /
    • pp.193-215
    • /
    • 2016
  • In recent years, the proliferation of diverse social networking services has led users to use many mediums simultaneously depending on their individual purpose and taste. Besides, while collecting information about particular themes, they usually employ various mediums such as social networking services, Internet news, and blogs. However, in terms of management, each document circulated through diverse mediums is placed in different categories on the basis of each source's policy and standards, hindering any attempt to conduct research on a specific category across different kinds of sources. For example, documents containing content on "Application for a foreign travel" can be classified into "Information Technology," "Travel," or "Life and Culture" according to the peculiar standard of each source. Likewise, with different viewpoints of definition and levels of specification for each source, similar categories can be named and structured differently in accordance with each source. To overcome these limitations, this study proposes a plan for conducting category mapping between different sources with various mediums while maintaining the existing category system of the medium as it is. Specifically, by re-classifying individual documents from the viewpoint of diverse sources and storing the result of such a classification as extra attributes, this study proposes a logical layer by which users can search for a specific document from multiple heterogeneous sources with different category names as if they belong to the same source. Besides, by collecting 6,000 articles of news from two Internet news portals, experiments were conducted to compare accuracy among sources, supervised learning and semi-supervised learning, and homogeneous and heterogeneous learning data. It is particularly interesting that in some categories, classifying accuracy of semi-supervised learning using heterogeneous learning data proved to be higher than that of supervised learning and semi-supervised learning, which used homogeneous learning data. This study has the following significances. First, it proposes a logical plan for establishing a system to integrate and manage all the heterogeneous mediums in different classifying systems while maintaining the existing physical classifying system as it is. This study's results particularly exhibit very different classifying accuracies in accordance with the heterogeneity of learning data; this is expected to spur further studies for enhancing the performance of the proposed methodology through the analysis of characteristics by category. In addition, with an increasing demand for search, collection, and analysis of documents from diverse mediums, the scope of the Internet search is not restricted to one medium. However, since each medium has a different categorical structure and name, it is actually very difficult to search for a specific category insofar as encompassing heterogeneous mediums. The proposed methodology is also significant for presenting a plan that enquires into all the documents regarding the standards of the relevant sites' categorical classification when the users select the desired site, while maintaining the existing site's characteristics and structure as it is. This study's proposed methodology needs to be further complemented in the following aspects. First, though only an indirect comparison and evaluation was made on the performance of this proposed methodology, future studies would need to conduct more direct tests on its accuracy. That is, after re-classifying documents of the object source on the basis of the categorical system of the existing source, the extent to which the classification was accurate needs to be verified through evaluation by actual users. In addition, the accuracy in classification needs to be increased by making the methodology more sophisticated. Furthermore, an understanding is required that the characteristics of some categories that showed a rather higher classifying accuracy of heterogeneous semi-supervised learning than that of supervised learning might assist in obtaining heterogeneous documents from diverse mediums and seeking plans that enhance the accuracy of document classification through its usage.

Applying Meta-model Formalization of Part-Whole Relationship to UML: Experiment on Classification of Aggregation and Composition (UML의 부분-전체 관계에 대한 메타모델 형식화 이론의 적용: 집합연관 및 복합연관 판별 실험)

  • Kim, Taekyung
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.99-118
    • /
    • 2015
  • Object-oriented programming languages have been widely selected for developing modern information systems. The use of concepts relating to object-oriented (OO, in short) programming has reduced efforts of reusing pre-existing codes, and the OO concepts have been proved to be a useful in interpreting system requirements. In line with this, we have witnessed that a modern conceptual modeling approach supports features of object-oriented programming. Unified Modeling Language or UML becomes one of de-facto standards for information system designers since the language provides a set of visual diagrams, comprehensive frameworks and flexible expressions. In a modeling process, UML users need to consider relationships between classes. Based on an explicit and clear representation of classes, the conceptual model from UML garners necessarily attributes and methods for guiding software engineers. Especially, identifying an association between a class of part and a class of whole is included in the standard grammar of UML. The representation of part-whole relationship is natural in a real world domain since many physical objects are perceived as part-whole relationship. In addition, even abstract concepts such as roles are easily identified by part-whole perception. It seems that a representation of part-whole in UML is reasonable and useful. However, it should be admitted that the use of UML is limited due to the lack of practical guidelines on how to identify a part-whole relationship and how to classify it into an aggregate- or a composite-association. Research efforts on developing the procedure knowledge is meaningful and timely in that misleading perception to part-whole relationship is hard to be filtered out in an initial conceptual modeling thus resulting in deterioration of system usability. The current method on identifying and classifying part-whole relationships is mainly counting on linguistic expression. This simple approach is rooted in the idea that a phrase of representing has-a constructs a par-whole perception between objects. If the relationship is strong, the association is classified as a composite association of part-whole relationship. In other cases, the relationship is an aggregate association. Admittedly, linguistic expressions contain clues for part-whole relationships; therefore, the approach is reasonable and cost-effective in general. Nevertheless, it does not cover concerns on accuracy and theoretical legitimacy. Research efforts on developing guidelines for part-whole identification and classification has not been accumulated sufficient achievements to solve this issue. The purpose of this study is to provide step-by-step guidelines for identifying and classifying part-whole relationships in the context of UML use. Based on the theoretical work on Meta-model Formalization, self-check forms that help conceptual modelers work on part-whole classes are developed. To evaluate the performance of suggested idea, an experiment approach was adopted. The findings show that UML users obtain better results with the guidelines based on Meta-model Formalization compared to a natural language classification scheme conventionally recommended by UML theorists. This study contributed to the stream of research effort about part-whole relationships by extending applicability of Meta-model Formalization. Compared to traditional approaches that target to establish criterion for evaluating a result of conceptual modeling, this study expands the scope to a process of modeling. Traditional theories on evaluation of part-whole relationship in the context of conceptual modeling aim to rule out incomplete or wrong representations. It is posed that qualification is still important; but, the lack of consideration on providing a practical alternative may reduce appropriateness of posterior inspection for modelers who want to reduce errors or misperceptions about part-whole identification and classification. The findings of this study can be further developed by introducing more comprehensive variables and real-world settings. In addition, it is highly recommended to replicate and extend the suggested idea of utilizing Meta-model formalization by creating different alternative forms of guidelines including plugins for integrated development environments.